arch/arm/include/asm/assembler.h
540
.type \name , #object
arch/parisc/math-emu/cnv_float.h
20
#define Dintp1(object) (object)
arch/parisc/math-emu/cnv_float.h
21
#define Dintp2(object) (object)
arch/parisc/math-emu/cnv_float.h
23
#define Duintp1(object) (object)
arch/parisc/math-emu/cnv_float.h
24
#define Duintp2(object) (object)
arch/parisc/math-emu/cnv_float.h
26
#define Qintp0(object) (object)
arch/parisc/math-emu/cnv_float.h
27
#define Qintp1(object) (object)
arch/parisc/math-emu/cnv_float.h
28
#define Qintp2(object) (object)
arch/parisc/math-emu/cnv_float.h
29
#define Qintp3(object) (object)
arch/parisc/math-emu/dbl_float.h
18
#define Dbl_sign(object) Dsign(object)
arch/parisc/math-emu/dbl_float.h
19
#define Dbl_exponent(object) Dexponent(object)
arch/parisc/math-emu/dbl_float.h
20
#define Dbl_signexponent(object) Dsignexponent(object)
arch/parisc/math-emu/dbl_float.h
21
#define Dbl_mantissap1(object) Dmantissap1(object)
arch/parisc/math-emu/dbl_float.h
22
#define Dbl_mantissap2(object) Dmantissap2(object)
arch/parisc/math-emu/dbl_float.h
23
#define Dbl_exponentmantissap1(object) Dexponentmantissap1(object)
arch/parisc/math-emu/dbl_float.h
24
#define Dbl_allp1(object) Dallp1(object)
arch/parisc/math-emu/dbl_float.h
25
#define Dbl_allp2(object) Dallp2(object)
arch/parisc/math-emu/float.h
105
#define Dallp1(object) (object)
arch/parisc/math-emu/float.h
106
#define Dsign(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
107
#define Dsignedsign(object) Bitfield_signed_extract( 0, 1,object)
arch/parisc/math-emu/float.h
108
#define Dexponent(object) Bitfield_extract( 1, 11,object)
arch/parisc/math-emu/float.h
109
#define Dmantissap1(object) Bitfield_mask( 12, 20,object)
arch/parisc/math-emu/float.h
110
#define Dsignaling(object) Bitfield_extract( 12, 1,object)
arch/parisc/math-emu/float.h
111
#define Dsignalingnan(object) Bitfield_extract( 1, 12,object)
arch/parisc/math-emu/float.h
112
#define Dhigh2mantissa(object) Bitfield_extract( 12, 2,object)
arch/parisc/math-emu/float.h
113
#define Dexponentmantissap1(object) Bitfield_mask( 1, 31,object)
arch/parisc/math-emu/float.h
114
#define Dsignexponent(object) Bitfield_extract( 0, 12,object)
arch/parisc/math-emu/float.h
115
#define Dhidden(object) Bitfield_extract( 11, 1,object)
arch/parisc/math-emu/float.h
116
#define Dhiddenoverflow(object) Bitfield_extract( 10, 1,object)
arch/parisc/math-emu/float.h
117
#define Dhiddenhigh7mantissa(object) Bitfield_extract( 11, 8,object)
arch/parisc/math-emu/float.h
118
#define Dhiddenhigh3mantissa(object) Bitfield_extract( 11, 4,object)
arch/parisc/math-emu/float.h
119
#define Dlowp1(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
120
#define Dlow31p1(object) Bitfield_mask( 1, 31,object)
arch/parisc/math-emu/float.h
121
#define Dhighp1(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
122
#define Dhigh4p1(object) Bitfield_extract( 0, 4,object)
arch/parisc/math-emu/float.h
123
#define Dhigh31p1(object) Bitfield_extract( 0, 31,object)
arch/parisc/math-emu/float.h
124
#define Dsignedhigh31p1(object) Bitfield_signed_extract( 0, 31,object)
arch/parisc/math-emu/float.h
125
#define Dbit3p1(object) Bitfield_extract( 3, 1,object)
arch/parisc/math-emu/float.h
127
#define Deposit_dsign(object,value) Bitfield_deposit(value,0,1,object)
arch/parisc/math-emu/float.h
128
#define Deposit_dexponent(object,value) Bitfield_deposit(value,1,11,object)
arch/parisc/math-emu/float.h
129
#define Deposit_dmantissap1(object,value) Bitfield_deposit(value,12,20,object)
arch/parisc/math-emu/float.h
130
#define Deposit_dhigh2mantissa(object,value) Bitfield_deposit(value,12,2,object)
arch/parisc/math-emu/float.h
131
#define Deposit_dexponentmantissap1(object,value) \
arch/parisc/math-emu/float.h
132
Bitfield_deposit(value,1,31,object)
arch/parisc/math-emu/float.h
133
#define Deposit_dsignexponent(object,value) Bitfield_deposit(value,0,12,object)
arch/parisc/math-emu/float.h
134
#define Deposit_dlowp1(object,value) Bitfield_deposit(value,31,1,object)
arch/parisc/math-emu/float.h
135
#define Deposit_dhigh4p1(object,value) Bitfield_deposit(value,0,4,object)
arch/parisc/math-emu/float.h
137
#define Is_dsign(object) Bitfield_mask( 0, 1,object)
arch/parisc/math-emu/float.h
138
#define Is_dsignaling(object) Bitfield_mask( 12, 1,object)
arch/parisc/math-emu/float.h
139
#define Is_dhidden(object) Bitfield_mask( 11, 1,object)
arch/parisc/math-emu/float.h
140
#define Is_dhiddenoverflow(object) Bitfield_mask( 10, 1,object)
arch/parisc/math-emu/float.h
141
#define Is_dlowp1(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
142
#define Is_dhighp1(object) Bitfield_mask( 0, 1,object)
arch/parisc/math-emu/float.h
143
#define Is_dbit3p1(object) Bitfield_mask( 3, 1,object)
arch/parisc/math-emu/float.h
145
#define Dallp2(object) (object)
arch/parisc/math-emu/float.h
146
#define Dmantissap2(object) (object)
arch/parisc/math-emu/float.h
147
#define Dlowp2(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
148
#define Dlow4p2(object) Bitfield_mask( 28, 4,object)
arch/parisc/math-emu/float.h
149
#define Dlow31p2(object) Bitfield_mask( 1, 31,object)
arch/parisc/math-emu/float.h
150
#define Dhighp2(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
151
#define Dhigh31p2(object) Bitfield_extract( 0, 31,object)
arch/parisc/math-emu/float.h
152
#define Dbit2p2(object) Bitfield_extract( 2, 1,object)
arch/parisc/math-emu/float.h
153
#define Dbit3p2(object) Bitfield_extract( 3, 1,object)
arch/parisc/math-emu/float.h
154
#define Dbit21p2(object) Bitfield_extract( 21, 1,object)
arch/parisc/math-emu/float.h
155
#define Dbit28p2(object) Bitfield_extract( 28, 1,object)
arch/parisc/math-emu/float.h
156
#define Dbit29p2(object) Bitfield_extract( 29, 1,object)
arch/parisc/math-emu/float.h
157
#define Dbit30p2(object) Bitfield_extract( 30, 1,object)
arch/parisc/math-emu/float.h
158
#define Dbit31p2(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
160
#define Deposit_dlowp2(object,value) Bitfield_deposit(value,31,1,object)
arch/parisc/math-emu/float.h
162
#define Is_dlowp2(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
163
#define Is_dhighp2(object) Bitfield_mask( 0, 1,object)
arch/parisc/math-emu/float.h
164
#define Is_dbit2p2(object) Bitfield_mask( 2, 1,object)
arch/parisc/math-emu/float.h
165
#define Is_dbit3p2(object) Bitfield_mask( 3, 1,object)
arch/parisc/math-emu/float.h
166
#define Is_dbit21p2(object) Bitfield_mask( 21, 1,object)
arch/parisc/math-emu/float.h
167
#define Is_dbit28p2(object) Bitfield_mask( 28, 1,object)
arch/parisc/math-emu/float.h
168
#define Is_dbit29p2(object) Bitfield_mask( 29, 1,object)
arch/parisc/math-emu/float.h
169
#define Is_dbit30p2(object) Bitfield_mask( 30, 1,object)
arch/parisc/math-emu/float.h
170
#define Is_dbit31p2(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
251
#define Extall(object) (object)
arch/parisc/math-emu/float.h
252
#define Extsign(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
253
#define Exthigh31(object) Bitfield_extract( 0, 31,object)
arch/parisc/math-emu/float.h
254
#define Extlow31(object) Bitfield_extract( 1, 31,object)
arch/parisc/math-emu/float.h
255
#define Extlow(object) Bitfield_extract( 31, 1,object)
arch/parisc/math-emu/float.h
261
#define Sextallp1(object) (object)
arch/parisc/math-emu/float.h
262
#define Sextallp2(object) (object)
arch/parisc/math-emu/float.h
263
#define Sextlowp1(object) Bitfield_extract( 31, 1,object)
arch/parisc/math-emu/float.h
264
#define Sexthighp2(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
265
#define Sextlow31p2(object) Bitfield_extract( 1, 31,object)
arch/parisc/math-emu/float.h
266
#define Sexthiddenoverflow(object) Bitfield_extract( 4, 1,object)
arch/parisc/math-emu/float.h
267
#define Is_sexthiddenoverflow(object) Bitfield_mask( 4, 1,object)
arch/parisc/math-emu/float.h
273
#define Dextallp1(object) (object)
arch/parisc/math-emu/float.h
274
#define Dextallp2(object) (object)
arch/parisc/math-emu/float.h
275
#define Dextallp3(object) (object)
arch/parisc/math-emu/float.h
276
#define Dextallp4(object) (object)
arch/parisc/math-emu/float.h
277
#define Dextlowp2(object) Bitfield_extract( 31, 1,object)
arch/parisc/math-emu/float.h
278
#define Dexthighp3(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
279
#define Dextlow31p3(object) Bitfield_extract( 1, 31,object)
arch/parisc/math-emu/float.h
280
#define Dexthiddenoverflow(object) Bitfield_extract( 10, 1,object)
arch/parisc/math-emu/float.h
281
#define Is_dexthiddenoverflow(object) Bitfield_mask( 10, 1,object)
arch/parisc/math-emu/float.h
282
#define Deposit_dextlowp4(object,value) Bitfield_deposit(value,31,1,object)
arch/parisc/math-emu/float.h
410
#define Cbit(object) Bitfield_extract( 5, 1,object)
arch/parisc/math-emu/float.h
411
#define Tbit(object) Bitfield_extract( 25, 1,object)
arch/parisc/math-emu/float.h
412
#define Roundingmode(object) Bitfield_extract( 21, 2,object)
arch/parisc/math-emu/float.h
413
#define Invalidtrap(object) Bitfield_extract( 27, 1,object)
arch/parisc/math-emu/float.h
414
#define Divisionbyzerotrap(object) Bitfield_extract( 28, 1,object)
arch/parisc/math-emu/float.h
415
#define Overflowtrap(object) Bitfield_extract( 29, 1,object)
arch/parisc/math-emu/float.h
416
#define Underflowtrap(object) Bitfield_extract( 30, 1,object)
arch/parisc/math-emu/float.h
417
#define Inexacttrap(object) Bitfield_extract( 31, 1,object)
arch/parisc/math-emu/float.h
418
#define Invalidflag(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
419
#define Divisionbyzeroflag(object) Bitfield_extract( 1, 1,object)
arch/parisc/math-emu/float.h
420
#define Overflowflag(object) Bitfield_extract( 2, 1,object)
arch/parisc/math-emu/float.h
421
#define Underflowflag(object) Bitfield_extract( 3, 1,object)
arch/parisc/math-emu/float.h
422
#define Inexactflag(object) Bitfield_extract( 4, 1,object)
arch/parisc/math-emu/float.h
423
#define Allflags(object) Bitfield_extract( 0, 5,object)
arch/parisc/math-emu/float.h
455
#define Allexception(object) (object)
arch/parisc/math-emu/float.h
456
#define Exceptiontype(object) Bitfield_extract( 0, 6,object)
arch/parisc/math-emu/float.h
457
#define Instructionfield(object) Bitfield_mask( 6,26,object)
arch/parisc/math-emu/float.h
458
#define Parmfield(object) Bitfield_extract( 23, 3,object)
arch/parisc/math-emu/float.h
459
#define Rabit(object) Bitfield_extract( 24, 1,object)
arch/parisc/math-emu/float.h
460
#define Ibit(object) Bitfield_extract( 25, 1,object)
arch/parisc/math-emu/float.h
462
#define Set_exceptiontype(object,value) Bitfield_deposit(value, 0, 6,object)
arch/parisc/math-emu/float.h
463
#define Set_parmfield(object,value) Bitfield_deposit(value, 23, 3,object)
arch/parisc/math-emu/float.h
464
#define Set_exceptiontype_and_instr_field(exception,instruction,object) \
arch/parisc/math-emu/float.h
465
object = exception << 26 | instruction
arch/parisc/math-emu/float.h
474
#define Greaterthanbit(object) Bitfield_extract( 27, 1,object)
arch/parisc/math-emu/float.h
475
#define Lessthanbit(object) Bitfield_extract( 28, 1,object)
arch/parisc/math-emu/float.h
476
#define Equalbit(object) Bitfield_extract( 29, 1,object)
arch/parisc/math-emu/float.h
477
#define Unorderedbit(object) Bitfield_extract( 30, 1,object)
arch/parisc/math-emu/float.h
478
#define Exceptionbit(object) Bitfield_extract( 31, 1,object)
arch/parisc/math-emu/float.h
48
#define Sall(object) (object)
arch/parisc/math-emu/float.h
49
#define Ssign(object) Bitfield_extract( 0, 1,object)
arch/parisc/math-emu/float.h
50
#define Ssignedsign(object) Bitfield_signed_extract( 0, 1,object)
arch/parisc/math-emu/float.h
51
#define Sexponent(object) Bitfield_extract( 1, 8,object)
arch/parisc/math-emu/float.h
52
#define Smantissa(object) Bitfield_mask( 9, 23,object)
arch/parisc/math-emu/float.h
53
#define Ssignaling(object) Bitfield_extract( 9, 1,object)
arch/parisc/math-emu/float.h
54
#define Ssignalingnan(object) Bitfield_extract( 1, 9,object)
arch/parisc/math-emu/float.h
55
#define Shigh2mantissa(object) Bitfield_extract( 9, 2,object)
arch/parisc/math-emu/float.h
56
#define Sexponentmantissa(object) Bitfield_mask( 1, 31,object)
arch/parisc/math-emu/float.h
57
#define Ssignexponent(object) Bitfield_extract( 0, 9,object)
arch/parisc/math-emu/float.h
58
#define Shidden(object) Bitfield_extract( 8, 1,object)
arch/parisc/math-emu/float.h
59
#define Shiddenoverflow(object) Bitfield_extract( 7, 1,object)
arch/parisc/math-emu/float.h
60
#define Shiddenhigh7mantissa(object) Bitfield_extract( 8, 8,object)
arch/parisc/math-emu/float.h
61
#define Shiddenhigh3mantissa(object) Bitfield_extract( 8, 4,object)
arch/parisc/math-emu/float.h
62
#define Slow(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
63
#define Slow4(object) Bitfield_mask( 28, 4,object)
arch/parisc/math-emu/float.h
64
#define Slow31(object) Bitfield_mask( 1, 31,object)
arch/parisc/math-emu/float.h
65
#define Shigh31(object) Bitfield_extract( 0, 31,object)
arch/parisc/math-emu/float.h
66
#define Ssignedhigh31(object) Bitfield_signed_extract( 0, 31,object)
arch/parisc/math-emu/float.h
67
#define Shigh4(object) Bitfield_extract( 0, 4,object)
arch/parisc/math-emu/float.h
68
#define Sbit24(object) Bitfield_extract( 24, 1,object)
arch/parisc/math-emu/float.h
69
#define Sbit28(object) Bitfield_extract( 28, 1,object)
arch/parisc/math-emu/float.h
70
#define Sbit29(object) Bitfield_extract( 29, 1,object)
arch/parisc/math-emu/float.h
71
#define Sbit30(object) Bitfield_extract( 30, 1,object)
arch/parisc/math-emu/float.h
72
#define Sbit31(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
74
#define Deposit_ssign(object,value) Bitfield_deposit(value,0,1,object)
arch/parisc/math-emu/float.h
75
#define Deposit_sexponent(object,value) Bitfield_deposit(value,1,8,object)
arch/parisc/math-emu/float.h
76
#define Deposit_smantissa(object,value) Bitfield_deposit(value,9,23,object)
arch/parisc/math-emu/float.h
77
#define Deposit_shigh2mantissa(object,value) Bitfield_deposit(value,9,2,object)
arch/parisc/math-emu/float.h
78
#define Deposit_sexponentmantissa(object,value) \
arch/parisc/math-emu/float.h
79
Bitfield_deposit(value,1,31,object)
arch/parisc/math-emu/float.h
80
#define Deposit_ssignexponent(object,value) Bitfield_deposit(value,0,9,object)
arch/parisc/math-emu/float.h
81
#define Deposit_slow(object,value) Bitfield_deposit(value,31,1,object)
arch/parisc/math-emu/float.h
82
#define Deposit_shigh4(object,value) Bitfield_deposit(value,0,4,object)
arch/parisc/math-emu/float.h
84
#define Is_ssign(object) Bitfield_mask( 0, 1,object)
arch/parisc/math-emu/float.h
85
#define Is_ssignaling(object) Bitfield_mask( 9, 1,object)
arch/parisc/math-emu/float.h
86
#define Is_shidden(object) Bitfield_mask( 8, 1,object)
arch/parisc/math-emu/float.h
87
#define Is_shiddenoverflow(object) Bitfield_mask( 7, 1,object)
arch/parisc/math-emu/float.h
88
#define Is_slow(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/float.h
89
#define Is_sbit24(object) Bitfield_mask( 24, 1,object)
arch/parisc/math-emu/float.h
90
#define Is_sbit28(object) Bitfield_mask( 28, 1,object)
arch/parisc/math-emu/float.h
91
#define Is_sbit29(object) Bitfield_mask( 29, 1,object)
arch/parisc/math-emu/float.h
92
#define Is_sbit30(object) Bitfield_mask( 30, 1,object)
arch/parisc/math-emu/float.h
93
#define Is_sbit31(object) Bitfield_mask( 31, 1,object)
arch/parisc/math-emu/fpbits.h
40
#define Bitfield_extract(start, length, object) \
arch/parisc/math-emu/fpbits.h
41
((object) >> (HOSTWDSZ - (start) - (length)) & \
arch/parisc/math-emu/fpbits.h
44
#define Bitfield_signed_extract(start, length, object) \
arch/parisc/math-emu/fpbits.h
45
((int)((object) << start) >> (HOSTWDSZ - (length)))
arch/parisc/math-emu/fpbits.h
47
#define Bitfield_mask(start, len, object) \
arch/parisc/math-emu/fpbits.h
48
((object) & (((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len)))
arch/parisc/math-emu/fpbits.h
50
#define Bitfield_deposit(value,start,len,object) object = \
arch/parisc/math-emu/fpbits.h
51
((object) & ~(((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len))) | \
arch/parisc/math-emu/sgl_float.h
19
#define Sgl_sign(object) Ssign(object)
arch/parisc/math-emu/sgl_float.h
20
#define Sgl_exponent(object) Sexponent(object)
arch/parisc/math-emu/sgl_float.h
21
#define Sgl_signexponent(object) Ssignexponent(object)
arch/parisc/math-emu/sgl_float.h
22
#define Sgl_mantissa(object) Smantissa(object)
arch/parisc/math-emu/sgl_float.h
23
#define Sgl_exponentmantissa(object) Sexponentmantissa(object)
arch/parisc/math-emu/sgl_float.h
24
#define Sgl_all(object) Sall(object)
drivers/acpi/acpi_processor.c
281
union acpi_object object = { .processor = { 0 } };
drivers/acpi/acpi_processor.c
282
struct acpi_buffer buffer = { sizeof(union acpi_object), &object };
drivers/acpi/acpi_processor.c
312
pr->acpi_id = object.processor.proc_id;
drivers/acpi/acpi_processor.c
391
if (!object.processor.pblk_address)
drivers/acpi/acpi_processor.c
393
else if (object.processor.pblk_length != 6)
drivers/acpi/acpi_processor.c
395
object.processor.pblk_length);
drivers/acpi/acpi_processor.c
397
pr->throttling.address = object.processor.pblk_address;
drivers/acpi/acpi_processor.c
401
pr->pblk = object.processor.pblk_address;
drivers/acpi/acpi_processor.c
533
union acpi_object object = {};
drivers/acpi/acpi_processor.c
534
struct acpi_buffer buffer = { sizeof(union acpi_object), &object };
drivers/acpi/acpi_processor.c
545
acpi_id = object.processor.proc_id;
drivers/acpi/acpi_processor.c
731
union acpi_object object = { 0 };
drivers/acpi/acpi_processor.c
732
struct acpi_buffer buffer = { sizeof(union acpi_object), &object };
drivers/acpi/acpi_processor.c
743
uid = object.processor.proc_id;
drivers/acpi/acpica/acdebug.h
101
acpi_status acpi_db_convert_to_package(char *string, union acpi_object *object);
drivers/acpi/acpica/acdebug.h
105
char *string, union acpi_object *object);
drivers/acpi/acpica/acdispat.h
289
acpi_ds_obj_stack_push(void *object, struct acpi_walk_state *walk_state);
drivers/acpi/acpica/acdispat.h
329
acpi_ds_result_pop(union acpi_operand_object **object,
drivers/acpi/acpica/acdispat.h
333
acpi_ds_result_push(union acpi_operand_object *object,
drivers/acpi/acpica/acinterp.h
394
acpi_ex_dump_object_descriptor(union acpi_operand_object *object, u32 flags);
drivers/acpi/acpica/aclocal.h
134
union acpi_operand_object *object; /* Interpreter object */
drivers/acpi/acpica/aclocal.h
584
union acpi_operand_object *object;
drivers/acpi/acpica/acnamesp.h
296
union acpi_operand_object *object, acpi_object_type type);
drivers/acpi/acpica/acobject.h
345
void *object; /* name_op=>HANDLE to obj, index_op=>union acpi_operand_object */
drivers/acpi/acpica/acobject.h
471
union acpi_operand_object object;
drivers/acpi/acpica/acutils.h
227
const char *acpi_ut_get_node_name(void *object);
drivers/acpi/acpica/acutils.h
229
const char *acpi_ut_get_descriptor_name(void *object);
drivers/acpi/acpica/acutils.h
231
const char *acpi_ut_get_reference_name(union acpi_operand_object *object);
drivers/acpi/acpica/acutils.h
289
acpi_ut_update_object_reference(union acpi_operand_object *object, u16 action);
drivers/acpi/acpica/acutils.h
367
void acpi_ut_add_reference(union acpi_operand_object *object);
drivers/acpi/acpica/acutils.h
369
void acpi_ut_remove_reference(union acpi_operand_object *object);
drivers/acpi/acpica/acutils.h
371
void acpi_ut_delete_internal_package_object(union acpi_operand_object *object);
drivers/acpi/acpica/acutils.h
373
void acpi_ut_delete_internal_simple_object(union acpi_operand_object *object);
drivers/acpi/acpica/acutils.h
449
void acpi_ut_delete_object_desc(union acpi_operand_object *object);
drivers/acpi/acpica/acutils.h
451
u8 acpi_ut_valid_internal_object(void *object);
drivers/acpi/acpica/acutils.h
518
*object, u16 action);
drivers/acpi/acpica/acutils.h
525
acpi_ut_create_update_state_and_push(union acpi_operand_object *object,
drivers/acpi/acpica/dbcmds.c
519
status = acpi_rs_create_resource_list(node->object, &return_buffer);
drivers/acpi/acpica/dbcmds.c
538
acpi_ut_debug_dump_buffer((u8 *)node->object->buffer.pointer,
drivers/acpi/acpica/dbcmds.c
539
node->object->buffer.length,
drivers/acpi/acpica/dbconvert.c
101
acpi_db_convert_to_buffer(char *string, union acpi_object *object)
drivers/acpi/acpica/dbconvert.c
145
object->type = ACPI_TYPE_BUFFER;
drivers/acpi/acpica/dbconvert.c
146
object->buffer.pointer = buffer;
drivers/acpi/acpica/dbconvert.c
147
object->buffer.length = length;
drivers/acpi/acpica/dbconvert.c
165
acpi_status acpi_db_convert_to_package(char *string, union acpi_object *object)
drivers/acpi/acpica/dbconvert.c
199
object->type = ACPI_TYPE_PACKAGE;
drivers/acpi/acpica/dbconvert.c
200
object->package.count = i;
drivers/acpi/acpica/dbconvert.c
201
object->package.elements = elements;
drivers/acpi/acpica/dbconvert.c
225
char *string, union acpi_object *object)
drivers/acpi/acpica/dbconvert.c
232
object->type = ACPI_TYPE_STRING;
drivers/acpi/acpica/dbconvert.c
233
object->string.pointer = string;
drivers/acpi/acpica/dbconvert.c
234
object->string.length = (u32)strlen(string);
drivers/acpi/acpica/dbconvert.c
239
status = acpi_db_convert_to_buffer(string, object);
drivers/acpi/acpica/dbconvert.c
244
status = acpi_db_convert_to_package(string, object);
drivers/acpi/acpica/dbconvert.c
249
object->type = ACPI_TYPE_INTEGER;
drivers/acpi/acpica/dbconvert.c
250
status = acpi_ut_strtoul64(string, &object->integer.value);
drivers/acpi/acpica/dbmethod.c
140
obj_desc = node->object;
drivers/acpi/acpica/dbmethod.c
183
obj_desc = walk_state->arguments[index].object;
drivers/acpi/acpica/dbmethod.c
206
obj_desc = walk_state->local_variables[index].object;
drivers/acpi/acpica/dbmethod.c
284
obj_desc = method->object;
drivers/acpi/acpica/dbnames.c
532
if (node->object->field.region_obj->region.space_id !=
drivers/acpi/acpica/dbnames.c
772
union acpi_operand_object *object;
drivers/acpi/acpica/dbnames.c
792
node = (struct acpi_namespace_node *)node->object;
drivers/acpi/acpica/dbnames.c
809
object = acpi_ns_get_attached_object(node);
drivers/acpi/acpica/dbnames.c
810
if (object) {
drivers/acpi/acpica/dbnames.c
812
if (ACPI_GET_DESCRIPTOR_TYPE(object) != ACPI_DESC_TYPE_OPERAND) {
drivers/acpi/acpica/dbnames.c
815
object, acpi_ut_get_descriptor_name(object));
drivers/acpi/acpica/dbobject.c
255
[obj_desc->reference.value].object;
drivers/acpi/acpica/dbobject.c
268
[obj_desc->reference.value].object;
drivers/acpi/acpica/dbobject.c
282
object);
drivers/acpi/acpica/dbobject.c
284
(obj_desc->reference.object);
drivers/acpi/acpica/dbobject.c
313
if (!obj_desc->reference.object) {
drivers/acpi/acpica/dbobject.c
322
(obj_desc->reference.object)) {
drivers/acpi/acpica/dbobject.c
326
object);
drivers/acpi/acpica/dbobject.c
332
(obj_desc->reference.object);
drivers/acpi/acpica/dbobject.c
418
obj_desc = walk_state->local_variables[i].object;
drivers/acpi/acpica/dbobject.c
433
obj_desc = walk_state->local_variables[i].object;
drivers/acpi/acpica/dbobject.c
488
obj_desc = walk_state->arguments[i].object;
drivers/acpi/acpica/dbobject.c
501
node->object->method.param_count);
drivers/acpi/acpica/dbobject.c
504
obj_desc = walk_state->arguments[i].object;
drivers/acpi/acpica/dbtest.c
251
obj_desc = node->object;
drivers/acpi/acpica/dsfield.c
524
if (info.region_node->object->region.space_id ==
drivers/acpi/acpica/dsfield.c
526
region_node->object->field.internal_pcc_buffer =
drivers/acpi/acpica/dsfield.c
527
ACPI_ALLOCATE_ZEROED(info.region_node->object->region.
drivers/acpi/acpica/dsfield.c
529
if (!region_node->object->field.internal_pcc_buffer) {
drivers/acpi/acpica/dsmthdat.c
111
if (walk_state->local_variables[index].object) {
drivers/acpi/acpica/dsmthdat.c
115
object));
drivers/acpi/acpica/dsmthdat.c
127
if (walk_state->arguments[index].object) {
drivers/acpi/acpica/dsmthdat.c
130
walk_state->arguments[index].object));
drivers/acpi/acpica/dsmthdat.c
25
union acpi_operand_object *object,
drivers/acpi/acpica/dsmthdat.c
282
union acpi_operand_object *object,
drivers/acpi/acpica/dsmthdat.c
291
"NewObj %p Type %2.2X, Refs=%u [%s]\n", object,
drivers/acpi/acpica/dsmthdat.c
292
type, object->common.reference_count,
drivers/acpi/acpica/dsmthdat.c
293
acpi_ut_get_type_name(object->common.type)));
drivers/acpi/acpica/dsmthdat.c
308
acpi_ut_add_reference(object);
drivers/acpi/acpica/dsmthdat.c
312
node->object = object;
drivers/acpi/acpica/dsmthdat.c
341
union acpi_operand_object *object;
drivers/acpi/acpica/dsmthdat.c
361
object = node->object;
drivers/acpi/acpica/dsmthdat.c
365
if (!object) {
drivers/acpi/acpica/dsmthdat.c
377
object = acpi_ut_create_integer_object((u64) 0);
drivers/acpi/acpica/dsmthdat.c
378
if (!object) {
drivers/acpi/acpica/dsmthdat.c
382
node->object = object;
drivers/acpi/acpica/dsmthdat.c
417
*dest_desc = object;
drivers/acpi/acpica/dsmthdat.c
418
acpi_ut_add_reference(object);
drivers/acpi/acpica/dsmthdat.c
445
union acpi_operand_object *object;
drivers/acpi/acpica/dsmthdat.c
458
object = acpi_ns_get_attached_object(node);
drivers/acpi/acpica/dsmthdat.c
465
node->object = NULL;
drivers/acpi/acpica/dsmthdat.c
467
if ((object) &&
drivers/acpi/acpica/dsmthdat.c
468
(ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_OPERAND)) {
drivers/acpi/acpica/dsmthdat.c
474
acpi_ut_remove_reference(object);
drivers/acpi/acpica/dsmthdat.c
598
object,
drivers/acpi/acpica/dsmthdat.c
657
union acpi_operand_object *object;
drivers/acpi/acpica/dsmthdat.c
670
object = acpi_ns_get_attached_object(node);
drivers/acpi/acpica/dsmthdat.c
671
if (!object) {
drivers/acpi/acpica/dsmthdat.c
680
return_VALUE(object->type);
drivers/acpi/acpica/dsobject.c
481
object));
drivers/acpi/acpica/dsobject.c
501
object));
drivers/acpi/acpica/dsobject.c
514
obj_desc->reference.object =
drivers/acpi/acpica/dsobject.c
515
op->common.node->object;
drivers/acpi/acpica/dspkginit.c
468
resolved_node->object);
drivers/acpi/acpica/dspkginit.c
516
acpi_ut_remove_reference(original_node->object);
drivers/acpi/acpica/dswexec.c
483
node->object);
drivers/acpi/acpica/dswexec.c
564
node->object,
drivers/acpi/acpica/dswload2.c
633
if (node->object) {
drivers/acpi/acpica/dswstate.c
112
acpi_ds_result_push(union acpi_operand_object *object,
drivers/acpi/acpica/dswstate.c
147
if (!object) {
drivers/acpi/acpica/dswstate.c
157
state->results.obj_desc[index] = object;
drivers/acpi/acpica/dswstate.c
161
object,
drivers/acpi/acpica/dswstate.c
164
object), walk_state,
drivers/acpi/acpica/dswstate.c
276
acpi_ds_obj_stack_push(void *object, struct acpi_walk_state *walk_state)
drivers/acpi/acpica/dswstate.c
285
object, walk_state, walk_state->num_operands));
drivers/acpi/acpica/dswstate.c
291
walk_state->operands[walk_state->operand_index] = object;
drivers/acpi/acpica/dswstate.c
299
object,
drivers/acpi/acpica/dswstate.c
302
object), walk_state,
drivers/acpi/acpica/dswstate.c
38
acpi_ds_result_pop(union acpi_operand_object **object,
drivers/acpi/acpica/dswstate.c
74
*object = state->results.obj_desc[index];
drivers/acpi/acpica/dswstate.c
75
if (!*object) {
drivers/acpi/acpica/dswstate.c
91
"Obj=%p [%s] Index=%X State=%p Num=%X\n", *object,
drivers/acpi/acpica/dswstate.c
92
acpi_ut_get_object_type_name(*object),
drivers/acpi/acpica/evregion.c
862
(next_node->object) &&
drivers/acpi/acpica/evregion.c
863
(next_node->object->region.space_id == space_id)) {
drivers/acpi/acpica/evxfgpe.c
886
if (node->object) {
drivers/acpi/acpica/excreate.c
52
target_node->object);
drivers/acpi/acpica/excreate.c
80
alias_node->object =
drivers/acpi/acpica/excreate.c
87
alias_node->object =
drivers/acpi/acpica/exdebug.c
234
node)->object,
drivers/acpi/acpica/exdebug.c
239
} else if (source_desc->reference.object) {
drivers/acpi/acpica/exdebug.c
241
(source_desc->reference.object) ==
drivers/acpi/acpica/exdebug.c
250
object), level + 4, 0);
drivers/acpi/acpica/exdebug.c
252
object_desc = source_desc->reference.object;
drivers/acpi/acpica/exdump.c
1006
object)->
drivers/acpi/acpica/exdump.c
1012
obj_desc->reference.object);
drivers/acpi/acpica/exdump.c
1145
obj_desc = ((struct acpi_namespace_node *)obj_desc)->object;
drivers/acpi/acpica/exdump.c
198
{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(reference.object), "Object Desc"},
drivers/acpi/acpica/exdump.c
274
{ACPI_EXD_LIST, ACPI_EXD_NSOFFSET(object), "Object List"},
drivers/acpi/acpica/exdump.c
645
acpi_os_printf("%p\n", obj_desc->reference.object);
drivers/acpi/acpica/exdump.c
656
acpi_os_printf("%p [%s]\n", obj_desc->reference.object,
drivers/acpi/acpica/exdump.c
662
object)->common.
drivers/acpi/acpica/exdump.c
991
} else if (obj_desc->reference.object) {
drivers/acpi/acpica/exdump.c
995
obj_desc->reference.object);
drivers/acpi/acpica/exmisc.c
61
referenced_obj = obj_desc->reference.object;
drivers/acpi/acpica/exmisc.c
95
reference_obj->reference.object = referenced_obj;
drivers/acpi/acpica/exoparg1.c
1003
return_desc = operand[0]->reference.object;
drivers/acpi/acpica/exoparg1.c
845
operand[0]->reference.object;
drivers/acpi/acpica/exoparg1.c
946
operand[0]->reference.object;
drivers/acpi/acpica/exoparg2.c
404
return_desc->reference.object = operand[0];
drivers/acpi/acpica/exprep.c
439
second_desc = info->connection_node->object;
drivers/acpi/acpica/exresnte.c
73
node = ACPI_CAST_PTR(struct acpi_namespace_node, node->object);
drivers/acpi/acpica/exresolv.c
231
(stack_desc->reference.node)->object;
drivers/acpi/acpica/exresolv.c
382
node = obj_desc->reference.object;
drivers/acpi/acpica/exresop.c
23
acpi_object_type this_type, void *object);
drivers/acpi/acpica/exresop.c
41
acpi_object_type this_type, void *object)
drivers/acpi/acpica/exresop.c
59
(((union acpi_operand_object *)object)->common.flags &
drivers/acpi/acpica/exresop.c
69
acpi_ut_get_type_name(this_type), object));
drivers/acpi/acpica/exstore.c
128
object, walk_state,
drivers/acpi/acpica/exstore.c
246
index_desc->reference.object)->common.
drivers/acpi/acpica/exstore.c
257
index_desc->reference.object)->common.
drivers/acpi/acpica/exstore.c
279
obj_desc = index_desc->reference.object;
drivers/acpi/acpica/nsaccess.c
657
if (!this_node->object) {
drivers/acpi/acpica/nsaccess.c
663
this_node->object)->type)) {
drivers/acpi/acpica/nsaccess.c
666
this_node->object;
drivers/acpi/acpica/nsalloc.c
91
obj_desc = node->object;
drivers/acpi/acpica/nsarguments.c
151
aml_param_count = node->object->method.param_count;
drivers/acpi/acpica/nsarguments.c
229
aml_param_count = node->object->method.param_count;
drivers/acpi/acpica/nsconvert.c
477
new_object->reference.object = node->object;
drivers/acpi/acpica/nsconvert.c
484
acpi_ut_add_reference(node->object);
drivers/acpi/acpica/nseval.c
76
info->node->object);
drivers/acpi/acpica/nsobject.c
102
obj_desc = ((struct acpi_namespace_node *)object)->object;
drivers/acpi/acpica/nsobject.c
103
object_type = ((struct acpi_namespace_node *)object)->type;
drivers/acpi/acpica/nsobject.c
111
obj_desc = (union acpi_operand_object *)object;
drivers/acpi/acpica/nsobject.c
123
if (node->object) {
drivers/acpi/acpica/nsobject.c
145
last_obj_desc->common.next_object = node->object;
drivers/acpi/acpica/nsobject.c
149
node->object = obj_desc;
drivers/acpi/acpica/nsobject.c
174
obj_desc = node->object;
drivers/acpi/acpica/nsobject.c
195
node->object = NULL;
drivers/acpi/acpica/nsobject.c
200
node->object = obj_desc->common.next_object;
drivers/acpi/acpica/nsobject.c
204
if (node->object &&
drivers/acpi/acpica/nsobject.c
205
(node->object->common.type != ACPI_TYPE_LOCAL_DATA)) {
drivers/acpi/acpica/nsobject.c
206
node->object = node->object->common.next_object;
drivers/acpi/acpica/nsobject.c
257
if (!node->object ||
drivers/acpi/acpica/nsobject.c
258
((ACPI_GET_DESCRIPTOR_TYPE(node->object) != ACPI_DESC_TYPE_OPERAND)
drivers/acpi/acpica/nsobject.c
259
&& (ACPI_GET_DESCRIPTOR_TYPE(node->object) !=
drivers/acpi/acpica/nsobject.c
261
|| ((node->object)->common.type == ACPI_TYPE_LOCAL_DATA)) {
drivers/acpi/acpica/nsobject.c
265
return_PTR(node->object);
drivers/acpi/acpica/nsobject.c
323
obj_desc = node->object;
drivers/acpi/acpica/nsobject.c
349
node->object = data_desc;
drivers/acpi/acpica/nsobject.c
377
obj_desc = node->object;
drivers/acpi/acpica/nsobject.c
38
union acpi_operand_object *object, acpi_object_type type)
drivers/acpi/acpica/nsobject.c
385
node->object = obj_desc->common.next_object;
drivers/acpi/acpica/nsobject.c
420
obj_desc = node->object;
drivers/acpi/acpica/nsobject.c
57
if (!object && (ACPI_TYPE_ANY != type)) {
drivers/acpi/acpica/nsobject.c
77
if (node->object == object) {
drivers/acpi/acpica/nsobject.c
80
object, node));
drivers/acpi/acpica/nsobject.c
87
if (!object) {
drivers/acpi/acpica/nsobject.c
96
else if ((ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_NAMED) &&
drivers/acpi/acpica/nsobject.c
97
((struct acpi_namespace_node *)object)->object) {
drivers/acpi/acpica/nssearch.c
102
node->object);
drivers/acpi/acpica/nssearch.c
312
acpi_ut_remove_reference((*return_node)->object);
drivers/acpi/acpica/nssearch.c
313
(*return_node)->object = NULL;
drivers/acpi/acpica/nsxfeval.c
498
node = info->return_object->reference.object;
drivers/acpi/acpica/nsxfeval.c
500
obj_desc = node->object;
drivers/acpi/acpica/nsxfname.c
268
param_count = node->object->method.param_count;
drivers/acpi/acpica/tbxfload.c
357
acpi_status acpi_unload_parent_table(acpi_handle object)
drivers/acpi/acpica/tbxfload.c
360
ACPI_CAST_PTR(struct acpi_namespace_node, object);
drivers/acpi/acpica/tbxfload.c
369
if (!object) {
drivers/acpi/acpica/utcache.c
153
acpi_status acpi_os_release_object(struct acpi_memory_list *cache, void *object)
drivers/acpi/acpica/utcache.c
159
if (!cache || !object) {
drivers/acpi/acpica/utcache.c
166
ACPI_FREE(object);
drivers/acpi/acpica/utcache.c
180
memset(object, 0xCA, cache->object_size);
drivers/acpi/acpica/utcache.c
181
ACPI_SET_DESCRIPTOR_TYPE(object, ACPI_DESC_TYPE_CACHED);
drivers/acpi/acpica/utcache.c
185
ACPI_SET_DESCRIPTOR_PTR(object, cache->list_head);
drivers/acpi/acpica/utcache.c
186
cache->list_head = object;
drivers/acpi/acpica/utcache.c
211
void *object;
drivers/acpi/acpica/utcache.c
232
object = cache->list_head;
drivers/acpi/acpica/utcache.c
233
cache->list_head = ACPI_GET_DESCRIPTOR_PTR(object);
drivers/acpi/acpica/utcache.c
240
ACPI_GET_FUNCTION_NAME, object,
drivers/acpi/acpica/utcache.c
250
memset(object, 0, cache->object_size);
drivers/acpi/acpica/utcache.c
271
object = ACPI_ALLOCATE_ZEROED(cache->object_size);
drivers/acpi/acpica/utcache.c
272
if (!object) {
drivers/acpi/acpica/utcache.c
277
return_PTR(object);
drivers/acpi/acpica/utcopy.c
507
internal_object->reference.object =
drivers/acpi/acpica/utcopy.c
734
acpi_ut_add_reference(source_desc->reference.object);
drivers/acpi/acpica/utdecode.c
239
const char *acpi_ut_get_node_name(void *object)
drivers/acpi/acpica/utdecode.c
241
struct acpi_namespace_node *node = (struct acpi_namespace_node *)object;
drivers/acpi/acpica/utdecode.c
245
if (!object) {
drivers/acpi/acpica/utdecode.c
251
if ((object == ACPI_ROOT_OBJECT) || (object == acpi_gbl_root_node)) {
drivers/acpi/acpica/utdecode.c
305
const char *acpi_ut_get_descriptor_name(void *object)
drivers/acpi/acpica/utdecode.c
308
if (!object) {
drivers/acpi/acpica/utdecode.c
312
if (ACPI_GET_DESCRIPTOR_TYPE(object) > ACPI_DESC_TYPE_MAX) {
drivers/acpi/acpica/utdecode.c
316
return (acpi_gbl_desc_type_names[ACPI_GET_DESCRIPTOR_TYPE(object)]);
drivers/acpi/acpica/utdecode.c
343
const char *acpi_ut_get_reference_name(union acpi_operand_object *object)
drivers/acpi/acpica/utdecode.c
346
if (!object) {
drivers/acpi/acpica/utdecode.c
350
if (ACPI_GET_DESCRIPTOR_TYPE(object) != ACPI_DESC_TYPE_OPERAND) {
drivers/acpi/acpica/utdecode.c
354
if (object->common.type != ACPI_TYPE_LOCAL_REFERENCE) {
drivers/acpi/acpica/utdecode.c
358
if (object->reference.class > ACPI_REFCLASS_MAX) {
drivers/acpi/acpica/utdecode.c
362
return (acpi_gbl_ref_class_names[object->reference.class]);
drivers/acpi/acpica/utdelete.c
101
obj_pointer = object->package.elements;
drivers/acpi/acpica/utdelete.c
110
if (object->device.gpe_block) {
drivers/acpi/acpica/utdelete.c
111
(void)acpi_ev_delete_gpe_block(object->device.
drivers/acpi/acpica/utdelete.c
122
handler_desc = object->common_notify.handler;
drivers/acpi/acpica/utdelete.c
134
object, object->mutex.os_mutex));
drivers/acpi/acpica/utdelete.c
136
if (object == acpi_gbl_global_lock_mutex) {
drivers/acpi/acpica/utdelete.c
145
acpi_os_delete_mutex(object->mutex.os_mutex);
drivers/acpi/acpica/utdelete.c
148
acpi_ex_unlink_mutex(object);
drivers/acpi/acpica/utdelete.c
149
acpi_os_delete_mutex(object->mutex.os_mutex);
drivers/acpi/acpica/utdelete.c
157
object, object->event.os_semaphore));
drivers/acpi/acpica/utdelete.c
159
(void)acpi_os_delete_semaphore(object->event.os_semaphore);
drivers/acpi/acpica/utdelete.c
160
object->event.os_semaphore = ACPI_SEMAPHORE_NULL;
drivers/acpi/acpica/utdelete.c
166
"***** Method %p\n", object));
drivers/acpi/acpica/utdelete.c
170
if (object->method.mutex) {
drivers/acpi/acpica/utdelete.c
171
acpi_os_delete_mutex(object->method.mutex->mutex.
drivers/acpi/acpica/utdelete.c
173
acpi_ut_delete_object_desc(object->method.mutex);
drivers/acpi/acpica/utdelete.c
174
object->method.mutex = NULL;
drivers/acpi/acpica/utdelete.c
177
if (object->method.node) {
drivers/acpi/acpica/utdelete.c
178
object->method.node = NULL;
drivers/acpi/acpica/utdelete.c
18
static void acpi_ut_delete_internal_obj(union acpi_operand_object *object);
drivers/acpi/acpica/utdelete.c
185
"***** Region %p\n", object));
drivers/acpi/acpica/utdelete.c
191
if (!(object->region.node->flags & ANOBJ_TEMPORARY)) {
drivers/acpi/acpica/utdelete.c
192
acpi_ut_remove_address_range(object->region.space_id,
drivers/acpi/acpica/utdelete.c
193
object->region.node);
drivers/acpi/acpica/utdelete.c
196
second_desc = acpi_ns_get_secondary_object(object);
drivers/acpi/acpica/utdelete.c
203
handler_desc = object->region.handler;
drivers/acpi/acpica/utdelete.c
21
acpi_ut_update_ref_count(union acpi_operand_object *object, u32 action);
drivers/acpi/acpica/utdelete.c
214
if (next_desc == object) {
drivers/acpi/acpica/utdelete.c
242
address_space.setup(object,
drivers/acpi/acpica/utdelete.c
260
if (object->field.internal_pcc_buffer) {
drivers/acpi/acpica/utdelete.c
261
ACPI_FREE(object->field.internal_pcc_buffer);
drivers/acpi/acpica/utdelete.c
269
"***** Buffer Field %p\n", object));
drivers/acpi/acpica/utdelete.c
271
second_desc = acpi_ns_get_secondary_object(object);
drivers/acpi/acpica/utdelete.c
280
"***** Bank Field %p\n", object));
drivers/acpi/acpica/utdelete.c
282
second_desc = acpi_ns_get_secondary_object(object);
drivers/acpi/acpica/utdelete.c
291
"***** Address handler %p\n", object));
drivers/acpi/acpica/utdelete.c
293
acpi_os_delete_mutex(object->address_space.context_mutex);
drivers/acpi/acpica/utdelete.c
313
ACPI_GET_FUNCTION_NAME, object,
drivers/acpi/acpica/utdelete.c
314
acpi_ut_get_object_type_name(object)));
drivers/acpi/acpica/utdelete.c
316
acpi_ut_delete_object_desc(object);
drivers/acpi/acpica/utdelete.c
36
static void acpi_ut_delete_internal_obj(union acpi_operand_object *object)
drivers/acpi/acpica/utdelete.c
365
acpi_ut_update_ref_count(union acpi_operand_object *object, u32 action)
drivers/acpi/acpica/utdelete.c
374
if (!object) {
drivers/acpi/acpica/utdelete.c
383
original_count = object->common.reference_count;
drivers/acpi/acpica/utdelete.c
391
object->common.reference_count = new_count;
drivers/acpi/acpica/utdelete.c
399
object));
drivers/acpi/acpica/utdelete.c
404
object, object->common.type,
drivers/acpi/acpica/utdelete.c
405
acpi_ut_get_object_type_name(object),
drivers/acpi/acpica/utdelete.c
416
object->common.reference_count = new_count;
drivers/acpi/acpica/utdelete.c
424
object));
drivers/acpi/acpica/utdelete.c
430
ACPI_GET_FUNCTION_NAME, object,
drivers/acpi/acpica/utdelete.c
431
object->common.type, new_count));
drivers/acpi/acpica/utdelete.c
436
acpi_ut_delete_internal_obj(object);
drivers/acpi/acpica/utdelete.c
45
ACPI_FUNCTION_TRACE_PTR(ut_delete_internal_obj, object);
drivers/acpi/acpica/utdelete.c
456
new_count, object, object->common.type, message));
drivers/acpi/acpica/utdelete.c
47
if (!object) {
drivers/acpi/acpica/utdelete.c
482
acpi_ut_update_object_reference(union acpi_operand_object *object, u16 action)
drivers/acpi/acpica/utdelete.c
493
while (object) {
drivers/acpi/acpica/utdelete.c
497
if (ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_NAMED) {
drivers/acpi/acpica/utdelete.c
499
"Object %p is NS handle\n", object));
drivers/acpi/acpica/utdelete.c
507
switch (object->common.type) {
drivers/acpi/acpica/utdelete.c
518
object->common_notify.notify_list[i];
drivers/acpi/acpica/utdelete.c
534
for (i = 0; i < object->package.count; i++) {
drivers/acpi/acpica/utdelete.c
539
next_object = object->package.elements[i];
drivers/acpi/acpica/utdelete.c
55
switch (object->common.type) {
drivers/acpi/acpica/utdelete.c
577
next_object = object->buffer_field.buffer_obj;
drivers/acpi/acpica/utdelete.c
582
next_object = object->bank_field.bank_obj;
drivers/acpi/acpica/utdelete.c
584
acpi_ut_create_update_state_and_push(object->
drivers/acpi/acpica/utdelete.c
59
"**** String %p, ptr %p\n", object,
drivers/acpi/acpica/utdelete.c
596
next_object = object->index_field.index_obj;
drivers/acpi/acpica/utdelete.c
598
acpi_ut_create_update_state_and_push(object->
drivers/acpi/acpica/utdelete.c
60
object->string.pointer));
drivers/acpi/acpica/utdelete.c
614
if ((object->reference.class == ACPI_REFCLASS_INDEX) ||
drivers/acpi/acpica/utdelete.c
615
(object->reference.class == ACPI_REFCLASS_NAME)) {
drivers/acpi/acpica/utdelete.c
616
next_object = object->reference.object;
drivers/acpi/acpica/utdelete.c
632
acpi_ut_update_ref_count(object, action);
drivers/acpi/acpica/utdelete.c
633
object = NULL;
drivers/acpi/acpica/utdelete.c
638
object = next_object;
drivers/acpi/acpica/utdelete.c
64
if (!(object->common.flags & AOPOBJ_STATIC_POINTER)) {
drivers/acpi/acpica/utdelete.c
642
object = state->update.object;
drivers/acpi/acpica/utdelete.c
677
void acpi_ut_add_reference(union acpi_operand_object *object)
drivers/acpi/acpica/utdelete.c
68
obj_pointer = object->string.pointer;
drivers/acpi/acpica/utdelete.c
684
if (!acpi_ut_valid_internal_object(object)) {
drivers/acpi/acpica/utdelete.c
690
object, object->common.reference_count));
drivers/acpi/acpica/utdelete.c
694
(void)acpi_ut_update_object_reference(object, REF_INCREMENT);
drivers/acpi/acpica/utdelete.c
710
void acpi_ut_remove_reference(union acpi_operand_object *object)
drivers/acpi/acpica/utdelete.c
719
if (!object ||
drivers/acpi/acpica/utdelete.c
720
(ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_NAMED)) {
drivers/acpi/acpica/utdelete.c
726
if (!acpi_ut_valid_internal_object(object)) {
drivers/acpi/acpica/utdelete.c
732
ACPI_GET_FUNCTION_NAME, object,
drivers/acpi/acpica/utdelete.c
733
object->common.reference_count));
drivers/acpi/acpica/utdelete.c
740
(void)acpi_ut_update_object_reference(object, REF_DECREMENT);
drivers/acpi/acpica/utdelete.c
75
"**** Buffer %p, ptr %p\n", object,
drivers/acpi/acpica/utdelete.c
76
object->buffer.pointer));
drivers/acpi/acpica/utdelete.c
80
if (!(object->common.flags & AOPOBJ_STATIC_POINTER)) {
drivers/acpi/acpica/utdelete.c
84
obj_pointer = object->buffer.pointer;
drivers/acpi/acpica/utdelete.c
92
object->package.count));
drivers/acpi/acpica/utinit.c
190
acpi_gbl_root_node_struct.object = NULL;
drivers/acpi/acpica/utmisc.c
158
acpi_ut_create_update_state_and_push(union acpi_operand_object *object,
drivers/acpi/acpica/utmisc.c
168
if (!object) {
drivers/acpi/acpica/utmisc.c
172
state = acpi_ut_create_update_state(object, action);
drivers/acpi/acpica/utobject.c
107
object->common.type = (u8) type;
drivers/acpi/acpica/utobject.c
111
object->common.reference_count = 1;
drivers/acpi/acpica/utobject.c
115
return_PTR(object);
drivers/acpi/acpica/utobject.c
305
u8 acpi_ut_valid_internal_object(void *object)
drivers/acpi/acpica/utobject.c
312
if (!object) {
drivers/acpi/acpica/utobject.c
319
switch (ACPI_GET_DESCRIPTOR_TYPE(object)) {
drivers/acpi/acpica/utobject.c
330
object, acpi_ut_get_descriptor_name(object)));
drivers/acpi/acpica/utobject.c
355
union acpi_operand_object *object;
drivers/acpi/acpica/utobject.c
359
object = acpi_os_acquire_object(acpi_gbl_operand_cache);
drivers/acpi/acpica/utobject.c
360
if (!object) {
drivers/acpi/acpica/utobject.c
369
ACPI_SET_DESCRIPTOR_TYPE(object, ACPI_DESC_TYPE_OPERAND);
drivers/acpi/acpica/utobject.c
372
object, (u32) sizeof(union acpi_operand_object)));
drivers/acpi/acpica/utobject.c
374
return_PTR(object);
drivers/acpi/acpica/utobject.c
389
void acpi_ut_delete_object_desc(union acpi_operand_object *object)
drivers/acpi/acpica/utobject.c
391
ACPI_FUNCTION_TRACE_PTR(ut_delete_object_desc, object);
drivers/acpi/acpica/utobject.c
395
if (ACPI_GET_DESCRIPTOR_TYPE(object) != ACPI_DESC_TYPE_OPERAND) {
drivers/acpi/acpica/utobject.c
397
"%p is not an ACPI Operand object [%s]", object,
drivers/acpi/acpica/utobject.c
398
acpi_ut_get_descriptor_name(object)));
drivers/acpi/acpica/utobject.c
402
(void)acpi_os_release_object(acpi_gbl_operand_cache, object);
drivers/acpi/acpica/utobject.c
60
union acpi_operand_object *object;
drivers/acpi/acpica/utobject.c
68
object =
drivers/acpi/acpica/utobject.c
71
if (!object) {
drivers/acpi/acpica/utobject.c
74
kmemleak_not_leak(object);
drivers/acpi/acpica/utobject.c
87
acpi_ut_delete_object_desc(object);
drivers/acpi/acpica/utobject.c
96
object->common.next_object = second_object;
drivers/acpi/acpica/utosi.c
403
string_desc = walk_state->arguments[0].object;
drivers/acpi/acpica/utstate.c
157
*object, u16 action)
drivers/acpi/acpica/utstate.c
173
state->update.object = object;
drivers/acpi/acpica/uttrack.c
654
(descriptor->object.common.
drivers/acpi/acpica/uttrack.c
656
descriptor->object.common.
drivers/acpi/dptf/dptf_power.c
25
#define DPTF_POWER_SHOW(name, object) \
drivers/acpi/dptf/dptf_power.c
34
status = acpi_evaluate_integer(acpi_dev->handle, #object,\
drivers/acpi/osl.c
1642
acpi_status acpi_os_release_object(acpi_cache_t *cache, void *object)
drivers/acpi/osl.c
1644
kmem_cache_free(cache, object);
drivers/android/binder.c
1784
struct binder_object *object)
drivers/android/binder.c
1790
read_size = min_t(size_t, sizeof(*object), buffer->data_size - offset);
drivers/android/binder.c
1796
if (copy_from_user(object, u + offset, read_size))
drivers/android/binder.c
1799
if (binder_alloc_copy_from_buffer(&proc->alloc, object, buffer,
drivers/android/binder.c
1805
hdr = &object->hdr;
drivers/android/binder.c
1857
struct binder_object *object,
drivers/android/binder.c
1875
object_size = binder_get_object(proc, NULL, b, object_offset, object);
drivers/android/binder.c
1876
if (!object_size || object->hdr.type != BINDER_TYPE_PTR)
drivers/android/binder.c
1881
return &object->bbo;
drivers/android/binder.c
2051
struct binder_object object;
drivers/android/binder.c
2058
object_offset, &object);
drivers/android/binder.c
2064
hdr = &object.hdr;
drivers/android/binder.c
2734
struct binder_object object;
drivers/android/binder.c
2741
parent = binder_validate_ptr(target_proc, b, &object, bp->parent,
drivers/android/binder.c
3463
struct binder_object object;
drivers/android/binder.c
3500
t->buffer, object_offset, &object);
drivers/android/binder.c
3518
hdr = &object.hdr;
drivers/base/cacheinfo.c
648
#define show_one(file_name, object) \
drivers/base/cacheinfo.c
653
return sysfs_emit(buf, "%u\n", this_leaf->object); \
drivers/cpufreq/cpufreq.c
708
#define show_one(file_name, object) \
drivers/cpufreq/cpufreq.c
712
return sysfs_emit(buf, "%u\n", policy->object); \
drivers/cpufreq/cpufreq.c
752
#define store_one(file_name, object) \
drivers/cpufreq/cpufreq.c
763
ret = freq_qos_update_request(policy->object##_freq_req, val);\
drivers/cpufreq/intel_pstate.c
1515
#define show_one(file_name, object) \
drivers/cpufreq/intel_pstate.c
1519
return sprintf(buf, "%u\n", global.object); \
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1558
ATOM_OBJECT *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1564
object = get_bios_object(bp, object_id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1566
if (!object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1569
record = get_encoder_cap_record(bp, object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1589
ATOM_OBJECT *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1594
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1599
offset = le16_to_cpu(object->usRecordOffset)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
193
ATOM_OBJECT *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
199
object = get_bios_object(bp, object_id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
201
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
206
number = get_src_obj_list(bp, object, &id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
2095
static uint32_t get_src_obj_list(struct bios_parser *bp, ATOM_OBJECT *object,
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
2101
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
2106
offset = le16_to_cpu(object->usSrcDstTableOffset)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
221
ATOM_OBJECT *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
229
object = get_bios_object(bp, id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
231
if (!object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
234
offset = le16_to_cpu(object->usRecordOffset)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
267
ATOM_OBJECT *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
273
object = get_bios_object(bp, id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
275
if (!object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
278
record = get_hpd_record(bp, object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
2795
ATOM_OBJECT *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
2800
object = NULL;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
2821
object = &object_table->asObjects[i];
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
2822
record_offset = object->usRecordOffset +
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
291
ATOM_OBJECT *object,
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
297
offset = le16_to_cpu(object->usRecordOffset)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
333
ATOM_OBJECT *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
341
object = get_bios_object(bp, connector_object_id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
343
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
348
if (bios_parser_get_device_tag_record(bp, object, &record)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
59
static uint32_t get_src_obj_list(struct bios_parser *bp, ATOM_OBJECT *object,
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
67
ATOM_OBJECT *object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
75
ATOM_OBJECT *object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
947
ATOM_OBJECT *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
952
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
957
offset = le16_to_cpu(object->usRecordOffset)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2128
struct atom_display_object_path_v2 *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2140
object = get_bios_object(bp, object_id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2142
if (!object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2145
record = get_encoder_cap_record(bp, object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2176
struct atom_display_object_path_v2 *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2181
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2186
offset = object->encoder_recordoffset + bp->object_info_tbl_offset;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2213
struct atom_display_object_path_v2 *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2218
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2223
offset = object->disp_recordoffset + bp->object_info_tbl_offset;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2249
struct atom_display_object_path_v3 *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2254
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2259
offset = object->disp_recordoffset + bp->object_info_tbl_offset;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2289
struct atom_display_object_path_v2 *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2300
object = get_bios_object(bp, object_id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2302
if (!object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2305
record = get_disp_connector_caps_record(bp, object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2335
struct atom_display_object_path_v3 *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2340
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
2345
offset = object->disp_recordoffset + bp->object_info_tbl_offset;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
3232
struct atom_display_object_path_v2 *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
3248
object = &v1_4->display_path[i];
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
3250
(object->disp_recordoffset) +
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
3347
struct atom_display_object_path_v3 *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
3364
object = &v1_5->display_path[i];
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
3366
(object->disp_recordoffset) +
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
3403
connector_id = object_id_from_bios_object_id(object->display_objid);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
390
struct atom_display_object_path_v2 *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
414
object = get_bios_object(bp, id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
416
if (!object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
419
offset = object->disp_recordoffset + bp->object_info_tbl_offset;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
528
struct atom_display_object_path_v3 *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
533
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
538
offset = object->disp_recordoffset + bp->object_info_tbl_offset;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
567
struct atom_display_object_path_v2 *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
577
object = get_bios_object(bp, id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
579
if (!object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
582
record = get_hpd_record(bp, object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
605
struct atom_display_object_path_v2 *object)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
610
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
615
offset = le16_to_cpu(object->disp_recordoffset)
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
769
struct atom_display_object_path_v2 *object;
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
781
object = get_bios_object(bp, connector_object_id);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
783
if (!object) {
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
789
info->dev_id = device_type_from_device_id(object->device_tag);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
83
struct atom_display_object_path_v2 *object);
drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c
87
struct atom_display_object_path_v2 *object);
drivers/gpu/drm/drm_crtc_internal.h
164
struct drm_mode_object *object);
drivers/gpu/drm/drm_lease.c
212
int object;
drivers/gpu/drm/drm_lease.c
226
idr_for_each_entry(leases, entry, object) {
drivers/gpu/drm/drm_lease.c
228
if (!idr_find(&dev->mode_config.object_idr, object))
drivers/gpu/drm/drm_lease.c
230
else if (_drm_has_leased(lessor, object))
drivers/gpu/drm/drm_lease.c
234
drm_dbg_lease(dev, "object %d failed %d\n", object, error);
drivers/gpu/drm/drm_lease.c
302
int object;
drivers/gpu/drm/drm_lease.c
317
idr_for_each_entry(&master->leases, entry, object)
drivers/gpu/drm/drm_lease.c
318
idr_remove(&master->leases, object);
drivers/gpu/drm/drm_lease.c
646
int object;
drivers/gpu/drm/drm_lease.c
669
idr_for_each_entry(object_idr, entry, object) {
drivers/gpu/drm/drm_lease.c
671
drm_dbg_lease(dev, "adding object %d\n", object);
drivers/gpu/drm/drm_lease.c
672
ret = put_user(object, object_ids + count);
drivers/gpu/drm/drm_mode_object.c
107
struct drm_mode_object *object)
drivers/gpu/drm/drm_mode_object.c
109
WARN_ON(!dev->driver->load && dev->registered && !object->free_cb);
drivers/gpu/drm/drm_mode_object.c
112
if (object->id) {
drivers/gpu/drm/drm_mode_object.c
113
idr_remove(&dev->mode_config.object_idr, object->id);
drivers/gpu/drm/drm_mode_object.c
114
object->id = 0;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
231
struct etnaviv_gem_object *etnaviv_obj = mapping->object;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
298
mapping->object = etnaviv_obj;
drivers/gpu/drm/etnaviv/etnaviv_gem.h
27
struct etnaviv_gem_object *object;
drivers/gpu/drm/etnaviv/etnaviv_mmu.c
117
struct etnaviv_gem_object *etnaviv_obj = mapping->object;
drivers/gpu/drm/nouveau/dispnv04/arb.c
197
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/dac.c
136
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/dac.c
239
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/dac.c
69
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
285
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/dfp.c
491
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.c
168
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.c
669
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.c
779
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.c
798
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
122
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
130
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
167
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
183
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
196
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
203
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
211
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
225
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
261
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
62
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
73
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
82
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/hw.h
93
struct nvif_object *device = &nouveau_drm(dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/overlay.c
121
struct nvif_object *dev = &drm->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/overlay.c
195
struct nvif_object *dev = &nouveau_drm(plane->dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/overlay.c
219
struct nvif_object *dev = &nouveau_drm(plane->base.dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/overlay.c
371
struct nvif_object *dev = &nouveau_drm(plane->dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/overlay.c
446
struct nvif_object *dev = &nouveau_drm(plane->dev)->client.device.object;
drivers/gpu/drm/nouveau/dispnv04/tvnv17.h
134
nvif_wr32(&device->object, reg, val);
drivers/gpu/drm/nouveau/dispnv04/tvnv17.h
140
return nvif_rd32(&device->object, reg);
drivers/gpu/drm/nouveau/dispnv50/base.c
46
cid = nvif_mclass(&disp->disp->object, bases);
drivers/gpu/drm/nouveau/dispnv50/core.c
67
cid = nvif_mclass(&disp->disp->object, cores);
drivers/gpu/drm/nouveau/dispnv50/corec37d.c
107
ret = nvif_object_ctor(&disp->disp->object, "dispCaps", 0,
drivers/gpu/drm/nouveau/dispnv50/crc.c
125
memset_io(ctx->mem.object.map.ptr, 0, ctx->mem.object.map.size);
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
101
struct crc907d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
92
struct crc907d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
70
struct crcc37d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
85
struct crcc37d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/curs.c
48
cid = nvif_mclass(&disp->disp->object, curses);
drivers/gpu/drm/nouveau/dispnv50/curs507a.c
187
ret = nvif_object_ctor(&disp->disp->object, "kmsCurs", 0, oclass,
drivers/gpu/drm/nouveau/dispnv50/disp.c
138
dmac->cur = push->cur - (u32 __iomem *)dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
145
nvif_wr32(&device->object, 0x070000, 0x00000001);
drivers/gpu/drm/nouveau/dispnv50/disp.c
147
if (!(nvif_rd32(&device->object, 0x070000) & 0x00000002))
drivers/gpu/drm/nouveau/dispnv50/disp.c
1768
if ((disp->disp->object.oclass == GT214_DISP ||
drivers/gpu/drm/nouveau/dispnv50/disp.c
1769
disp->disp->object.oclass >= GF110_DISP) &&
drivers/gpu/drm/nouveau/dispnv50/disp.c
1779
if (disp->disp->object.oclass != NV50_DISP &&
drivers/gpu/drm/nouveau/dispnv50/disp.c
1920
if (disp->disp->object.oclass < GF110_DISP) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
201
dmac->cur = push->cur - (u32 __iomem *)dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
207
push->cur = dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
220
push->bgn = dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
237
struct nvif_object *disp = &drm->display->disp.object;
drivers/gpu/drm/nouveau/dispnv50/disp.c
261
dmac->push.bgn = dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
272
args->pushbuf = nvif_handle(&dmac->push.mem.object);
drivers/gpu/drm/nouveau/dispnv50/disp.c
2870
if (disp->disp->object.oclass >= GB202_DISP)
drivers/gpu/drm/nouveau/dispnv50/disp.c
2872
else if (disp->disp->object.oclass >= TU102_DISP)
drivers/gpu/drm/nouveau/dispnv50/disp.c
2888
if (disp->disp->object.oclass >= GM107_DISP) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
2891
} else if (disp->disp->object.oclass >= GK104_DISP) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
712
if (disp->disp->object.oclass <= GT200_DISP ||
drivers/gpu/drm/nouveau/dispnv50/disp.c
713
disp->disp->object.oclass == GT206_DISP)
drivers/gpu/drm/nouveau/dispnv50/head.c
588
if (disp->disp->object.oclass < GF110_DISP)
drivers/gpu/drm/nouveau/dispnv50/head.c
593
if (disp->disp->object.oclass < GV100_DISP) {
drivers/gpu/drm/nouveau/dispnv50/head.c
617
disp->disp->object.oclass >= GF110_DISP,
drivers/gpu/drm/nouveau/dispnv50/lut.c
36
void __iomem *mem = lut->mem[buffer].object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/lut.c
70
const u32 size = disp->disp->object.oclass < GF110_DISP ? 257 : 1025;
drivers/gpu/drm/nouveau/dispnv50/oimm.c
44
cid = nvif_mclass(&disp->disp->object, oimms);
drivers/gpu/drm/nouveau/dispnv50/oimm507b.c
36
ret = nvif_object_ctor(&disp->disp->object, "kmsOvim", 0, oclass,
drivers/gpu/drm/nouveau/dispnv50/ovly.c
46
cid = nvif_mclass(&disp->disp->object, ovlys);
drivers/gpu/drm/nouveau/dispnv50/wimm.c
43
cid = nvif_mclass(&disp->disp->object, wimms);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
111
NV_DMA_IN_MEMORY, &args, argc, &ctxdma->object);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
51
nvif_object_dtor(&ctxdma->object);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
567
asyw->image.handle[0] = ctxdma->object.handle;
drivers/gpu/drm/nouveau/dispnv50/wndw.c
79
if (ctxdma->object.handle == handle)
drivers/gpu/drm/nouveau/dispnv50/wndw.c
947
cid = nvif_mclass(&disp->disp->object, wndws);
drivers/gpu/drm/nouveau/dispnv50/wndw.h
10
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/client.h
8
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/conn.h
33
return conn->object.handle;
drivers/gpu/drm/nouveau/include/nvif/conn.h
9
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/device.h
10
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/disp.h
7
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/event.h
13
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/event.h
20
return nvif_object_constructed(&event->object);
drivers/gpu/drm/nouveau/include/nvif/head.h
18
return head->object.handle;
drivers/gpu/drm/nouveau/include/nvif/head.h
9
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/ioctl.h
24
__u64 object;
drivers/gpu/drm/nouveau/include/nvif/ioctl.h
49
__u64 object;
drivers/gpu/drm/nouveau/include/nvif/mem.h
6
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/mmu.h
6
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/object.h
26
nvif_object_constructed(struct nvif_object *object)
drivers/gpu/drm/nouveau/include/nvif/object.h
28
return object->client != NULL;
drivers/gpu/drm/nouveau/include/nvif/object.h
45
#define nvif_object(a) (a)->object
drivers/gpu/drm/nouveau/include/nvif/object.h
75
struct nvif_object *object = (o); \
drivers/gpu/drm/nouveau/include/nvif/object.h
81
cnt = nvif_object_sclass_get(object, &sclass); \
drivers/gpu/drm/nouveau/include/nvif/outp.h
10
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/printf.h
10
_p->func->l(_o, "[%s/%08x:%s] "f"\n", _o->client->object.name, _o->handle, _o->name, ##a); \
drivers/gpu/drm/nouveau/include/nvif/push.h
75
u32 __o = _ppp->cur - (u32 *)_ppp->mem.object.map.ptr; \
drivers/gpu/drm/nouveau/include/nvif/push.h
76
NVIF_DEBUG(&_ppp->mem.object, "%08x: "f, __o * 4, ##a); \
drivers/gpu/drm/nouveau/include/nvif/user.h
8
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvif/vmm.h
25
struct nvif_object object;
drivers/gpu/drm/nouveau/include/nvkm/core/client.h
4
#define nvkm_client(p) container_of((p), struct nvkm_client, object)
drivers/gpu/drm/nouveau/include/nvkm/core/client.h
8
struct nvkm_object object;
drivers/gpu/drm/nouveau/include/nvkm/core/object.h
19
u64 object;
drivers/gpu/drm/nouveau/include/nvkm/core/object.h
64
struct nvkm_object *nvkm_object_search(struct nvkm_client *, u64 object,
drivers/gpu/drm/nouveau/include/nvkm/core/oclass.h
24
u64 object;
drivers/gpu/drm/nouveau/include/nvkm/core/oproxy.h
10
struct nvkm_object *object;
drivers/gpu/drm/nouveau/include/nvkm/engine/disp.h
19
struct nvkm_gsp_object object;
drivers/gpu/drm/nouveau/include/nvkm/engine/disp.h
68
struct nvkm_object object;
drivers/gpu/drm/nouveau/include/nvkm/engine/dma.h
12
struct nvkm_object object;
drivers/gpu/drm/nouveau/include/nvkm/engine/dma.h
24
struct nvkm_dmaobj *nvkm_dmaobj_search(struct nvkm_client *, u64 object);
drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h
40
struct nvkm_gsp_object object;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
205
} object;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
213
struct nvkm_gsp_object object;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
321
nvkm_gsp_rm_ctrl_get(struct nvkm_gsp_object *object, u32 cmd, u32 argc)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
323
return object->client->gsp->rm->api->ctrl->get(object, cmd, argc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
327
nvkm_gsp_rm_ctrl_push(struct nvkm_gsp_object *object, void *argv, u32 repc)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
329
return object->client->gsp->rm->api->ctrl->push(object, argv, repc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
333
nvkm_gsp_rm_ctrl_rd(struct nvkm_gsp_object *object, u32 cmd, u32 repc)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
335
void *argv = nvkm_gsp_rm_ctrl_get(object, cmd, repc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
341
ret = nvkm_gsp_rm_ctrl_push(object, &argv, repc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
348
nvkm_gsp_rm_ctrl_wr(struct nvkm_gsp_object *object, void *argv)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
350
int ret = nvkm_gsp_rm_ctrl_push(object, &argv, 0);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
358
nvkm_gsp_rm_ctrl_done(struct nvkm_gsp_object *object, void *repv)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
360
object->client->gsp->rm->api->ctrl->done(object, repv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
365
struct nvkm_gsp_object *object)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
371
object->client = parent->client;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
372
object->parent = parent;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
373
object->handle = handle;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
375
argv = gsp->rm->api->alloc->get(object, oclass, argc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
377
object->client = NULL;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
385
nvkm_gsp_rm_alloc_push(struct nvkm_gsp_object *object, void *argv)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
387
void *repv = object->client->gsp->rm->api->alloc->push(object, argv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
390
object->client = NULL;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
396
nvkm_gsp_rm_alloc_wr(struct nvkm_gsp_object *object, void *argv)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
398
void *repv = nvkm_gsp_rm_alloc_push(object, argv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
407
nvkm_gsp_rm_alloc_done(struct nvkm_gsp_object *object, void *repv)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
409
object->client->gsp->rm->api->alloc->done(object, repv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
414
struct nvkm_gsp_object *object)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
416
void *argv = nvkm_gsp_rm_alloc_get(parent, handle, oclass, argc, object);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
421
return nvkm_gsp_rm_alloc_wr(object, argv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
425
nvkm_gsp_rm_free(struct nvkm_gsp_object *object)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
427
if (object->client) {
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
428
int ret = object->client->gsp->rm->api->alloc->free(object);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
429
object->client = NULL;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
448
if (device->object.client)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
449
device->object.client->gsp->rm->api->device->dtor(device);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
472
struct nvkm_gsp_object object;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
481
struct nvkm_rm *rm = device->object.client->gsp->rm;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
492
device->object.client->gsp->rm->api->device->event.dtor(event);
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
73
struct nvkm_gsp_object object;
drivers/gpu/drm/nouveau/nouveau_abi16.c
104
if (obj->object == object)
drivers/gpu/drm/nouveau/nouveau_abi16.c
119
nouveau_abi16_obj_new(struct nouveau_abi16 *abi16, enum nouveau_abi16_obj_type type, u64 object)
drivers/gpu/drm/nouveau/nouveau_abi16.c
123
obj = nouveau_abi16_obj_find(abi16, object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
132
obj->object = object;
drivers/gpu/drm/nouveau/nouveau_abi16.c
165
nvif_object_dtor(&ntfy->object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
243
return nvif_mclass(&device->object, dmas) < 0 ? NV50_DMA_IB_MAX : 0;
drivers/gpu/drm/nouveau/nouveau_abi16.c
607
oclass, NULL, 0, &ntfy->object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
671
&ntfy->object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
702
if (ntfy->object.handle == fini->handle) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
725
obj = nouveau_abi16_obj_find(abi16, ioctl->object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
750
obj = nouveau_abi16_obj_find(abi16, ioctl->object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
777
if (ioctl->object || args->oclass != NV_DEVICE)
drivers/gpu/drm/nouveau/nouveau_abi16.c
780
obj = nouveau_abi16_obj_new(abi16, DEVICE, args->object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
791
obj = nouveau_abi16_obj_new(abi16, ENGOBJ, args->object);
drivers/gpu/drm/nouveau/nouveau_abi16.c
91
u64 object;
drivers/gpu/drm/nouveau/nouveau_abi16.c
99
nouveau_abi16_obj_find(struct nouveau_abi16 *abi16, u64 object)
drivers/gpu/drm/nouveau/nouveau_abi16.h
16
struct nvif_object object;
drivers/gpu/drm/nouveau/nouveau_backlight.c
102
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_backlight.c
68
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_backlight.c
80
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_bios.c
1924
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_bios.c
217
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_bios.c
321
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_bios.c
637
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_bo.c
1242
nvif_object_unmap_handle(&mem->mem.object);
drivers/gpu/drm/nouveau/nouveau_bo.c
1245
nvif_object_unmap_handle(&mem->mem.object);
drivers/gpu/drm/nouveau/nouveau_bo.c
1305
switch (mem->mem.object.oclass) {
drivers/gpu/drm/nouveau/nouveau_bo.c
1324
ret = nvif_object_map_handle(&mem->mem.object,
drivers/gpu/drm/nouveau/nouveau_bo.c
304
if (mmu->object.oclass >= NVIF_CLASS_MMU_GF100)
drivers/gpu/drm/nouveau/nouveau_chan.c
133
chan->chan.push.bgn = chan->chan.push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/nouveau_chan.c
171
chan->chan.push.mem.object.parent = cli->base.object.parent;
drivers/gpu/drm/nouveau/nouveau_chan.c
172
chan->chan.push.mem.object.client = &cli->base;
drivers/gpu/drm/nouveau/nouveau_chan.c
173
chan->chan.push.mem.object.name = "chanPush";
drivers/gpu/drm/nouveau/nouveau_chan.c
174
chan->chan.push.mem.object.map.ptr = chan->push.buffer->kmap.virtual;
drivers/gpu/drm/nouveau/nouveau_chan.c
236
ret = nvif_object_ctor(&device->object, "abi16PushCtxDma", 0,
drivers/gpu/drm/nouveau/nouveau_chan.c
281
cid = nvif_mclass(&device->object, hosts);
drivers/gpu/drm/nouveau/nouveau_chan.c
309
args->vmm = nvif_handle(&chan->vmm->vmm.object);
drivers/gpu/drm/nouveau/nouveau_chan.c
328
args->huserd = nvif_handle(&chan->mem_userd.object);
drivers/gpu/drm/nouveau/nouveau_chan.c
331
chan->userd = &chan->mem_userd.object;
drivers/gpu/drm/nouveau/nouveau_chan.c
339
ret = nvif_object_ctor(&device->object, "abi16ChanUser", 0, hosts[cid].oclass,
drivers/gpu/drm/nouveau/nouveau_chan.c
531
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_connector.c
1407
if (disp->disp.object.oclass >= GB202_DISP)
drivers/gpu/drm/nouveau/nouveau_connector.c
1416
if ((disp->disp.object.oclass >= G82_DISP)
drivers/gpu/drm/nouveau/nouveau_connector.c
1429
if (nvif_object_constructed(&nv_connector->conn.object)) {
drivers/gpu/drm/nouveau/nouveau_connector.c
1458
if (disp->disp.object.oclass < NV50_DISP) {
drivers/gpu/drm/nouveau/nouveau_connector.c
153
if (disp->disp.object.oclass < NV50_DISP)
drivers/gpu/drm/nouveau/nouveau_connector.c
271
if (nouveau_display(connector->dev)->disp.object.oclass < NV50_DISP) {
drivers/gpu/drm/nouveau/nouveau_connector.c
331
if (disp->disp.object.oclass < NV50_DISP)
drivers/gpu/drm/nouveau/nouveau_connector.c
423
if (nvif_object_constructed(&nv_encoder->outp.object)) {
drivers/gpu/drm/nouveau/nouveau_debugfs.c
302
return nvif_object_ctor(&drm->client.device.object, "debugfsCtrl", 0,
drivers/gpu/drm/nouveau/nouveau_debugfs.c
63
nvif_rd32(&drm->client.device.object, 0x101000));
drivers/gpu/drm/nouveau/nouveau_display.c
612
if (disp->disp.object.oclass < NV50_DISP)
drivers/gpu/drm/nouveau/nouveau_display.c
615
if (disp->disp.object.oclass < GF110_DISP)
drivers/gpu/drm/nouveau/nouveau_display.c
699
if (disp->disp.object.oclass < NV50_DISP) {
drivers/gpu/drm/nouveau/nouveau_display.c
720
if (disp->disp.object.oclass >= NV50_DISP)
drivers/gpu/drm/nouveau/nouveau_display.c
97
ret = nvif_mthd(&head->object, NVIF_HEAD_V0_SCANOUTPOS, &args, sizeof(args));
drivers/gpu/drm/nouveau/nouveau_drm.c
1163
nvif_mask(&device->object, 0x088488, (1 << 25), (1 << 25));
drivers/gpu/drm/nouveau/nouveau_drm.c
226
cli->device.object.map.ptr = NULL;
drivers/gpu/drm/nouveau/nouveau_drm.c
277
cli->device.object.map.ptr = drm->device.object.map.ptr;
drivers/gpu/drm/nouveau/nouveau_drm.c
279
ret = nvif_mmu_ctor(&cli->device.object, "drmMmu", drm->mmu.object.oclass,
drivers/gpu/drm/nouveau/nouveau_drm.c
286
ret = nvif_mclass(&cli->mmu.object, vmms);
drivers/gpu/drm/nouveau/nouveau_drm.c
298
ret = nvif_mclass(&cli->mmu.object, mems);
drivers/gpu/drm/nouveau/nouveau_drm.c
479
ret = n = nvif_object_sclass_get(&device->object, &sclass);
drivers/gpu/drm/nouveau/nouveau_drm.c
545
nouveau_drm_errorf(struct nvif_object *object, const char *fmt, ...)
drivers/gpu/drm/nouveau/nouveau_drm.c
547
struct nouveau_drm *drm = container_of(object->parent, typeof(*drm), parent);
drivers/gpu/drm/nouveau/nouveau_drm.c
559
nouveau_drm_debugf(struct nvif_object *object, const char *fmt, ...)
drivers/gpu/drm/nouveau/nouveau_drm.c
561
struct nouveau_drm *drm = container_of(object->parent, typeof(*drm), parent);
drivers/gpu/drm/nouveau/nouveau_drm.c
652
nvif_mask(&drm->client.device.object, 0x00088080, 0x00000800, 0x00000000);
drivers/gpu/drm/nouveau/nouveau_drm.c
759
drm->_client.object.parent = &drm->parent;
drivers/gpu/drm/nouveau/nouveau_drm.c
778
ret = nvif_mclass(&drm->device.object, mmus);
drivers/gpu/drm/nouveau/nouveau_drm.c
784
ret = nvif_mmu_ctor(&drm->device.object, "drmMmu", mmus[ret].oclass, &drm->mmu);
drivers/gpu/drm/nouveau/nouveau_gem.c
108
if (vmm->vmm.object.oclass < NVIF_CLASS_VMM_NV50)
drivers/gpu/drm/nouveau/nouveau_gem.c
193
if (vmm->vmm.object.oclass < NVIF_CLASS_VMM_NV50)
drivers/gpu/drm/nouveau/nouveau_gem.c
314
if (vmm->vmm.object.oclass >= NVIF_CLASS_VMM_NV50 &&
drivers/gpu/drm/nouveau/nouveau_gem.c
425
if (chan->vmm->vmm.object.oclass >= NVIF_CLASS_VMM_NV50) {
drivers/gpu/drm/nouveau/nouveau_gem.c
520
if (chan->vmm->vmm.object.oclass >= NVIF_CLASS_VMM_NV50) {
drivers/gpu/drm/nouveau/nouveau_led.c
41
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_led.c
58
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_mem.c
106
if (mmu->object.oclass >= NVIF_CLASS_MMU_GF100)
drivers/gpu/drm/nouveau/nouveau_mem.c
46
switch (vmm->object.oclass) {
drivers/gpu/drm/nouveau/nouveau_nvif.c
61
return nvkm_object_init(&client->object);
drivers/gpu/drm/nouveau/nouveau_nvif.c
74
return nvkm_object_fini(&client->object, state);
drivers/gpu/drm/nouveau/nouveau_nvif.c
80
struct nvif_object *object = (void *)(unsigned long)token;
drivers/gpu/drm/nouveau/nouveau_nvif.c
81
struct nvif_event *event = container_of(object, typeof(*event), object);
drivers/gpu/drm/nouveau/nouveau_svm.c
1002
nvif_object_map(&buffer->object, NULL, 0);
drivers/gpu/drm/nouveau/nouveau_svm.c
1008
ret = nvif_event_ctor(&buffer->object, "svmFault", id, nouveau_svm_event, true, NULL, 0,
drivers/gpu/drm/nouveau/nouveau_svm.c
1073
ret = nvif_mclass(&drm->client.device.object, buffers);
drivers/gpu/drm/nouveau/nouveau_svm.c
243
nvif_object_mthd(&svmm->vmm->vmm.object, NVIF_VMM_V0_PFNCLR,
drivers/gpu/drm/nouveau/nouveau_svm.c
350
cli->vmm.vmm.object.oclass, MANAGED,
drivers/gpu/drm/nouveau/nouveau_svm.c
384
WARN_ON(nvif_object_mthd(&svm->drm->client.vmm.vmm.object,
drivers/gpu/drm/nouveau/nouveau_svm.c
400
WARN_ON(nvif_object_mthd(&svm->drm->client.vmm.vmm.object,
drivers/gpu/drm/nouveau/nouveau_svm.c
456
struct nvif_object *memory = &buffer->object;
drivers/gpu/drm/nouveau/nouveau_svm.c
48
struct nvif_object object;
drivers/gpu/drm/nouveau/nouveau_svm.c
641
ret = nvif_object_ioctl(&svmm->vmm->vmm.object, args, size, NULL);
drivers/gpu/drm/nouveau/nouveau_svm.c
707
ret = nvif_object_ioctl(&svmm->vmm->vmm.object, args, size, NULL);
drivers/gpu/drm/nouveau/nouveau_svm.c
721
struct nvif_object *device = &svm->drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_svm.c
935
nvif_object_ioctl(&svmm->vmm->vmm.object, args,
drivers/gpu/drm/nouveau/nouveau_svm.c
954
struct nvif_object *device = &svm->drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_svm.c
969
if (!nvif_object_constructed(&buffer->object))
drivers/gpu/drm/nouveau/nouveau_svm.c
981
nvif_object_dtor(&buffer->object);
drivers/gpu/drm/nouveau/nouveau_svm.c
989
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nouveau_svm.c
996
sizeof(args), &buffer->object);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
147
switch (vmm->object.oclass) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1941
cli->vmm.vmm.object.oclass, RAW,
drivers/gpu/drm/nouveau/nouveau_vga.c
15
struct nvif_object *device = &drm->client.device.object;
drivers/gpu/drm/nouveau/nvif/chan.c
107
chan->push.mem.object.map.ptr = push;
drivers/gpu/drm/nouveau/nvif/chan.c
11
u32 put = push->bgn - (u32 *)chan->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/nvif/chan.c
119
u32 cur = push->cur - (u32 *)push->mem.object.map.ptr;
drivers/gpu/drm/nouveau/nvif/chan.c
155
push->bgn = (u32 *)push->mem.object.map.ptr + cur;
drivers/gpu/drm/nouveau/nvif/chan.c
38
const u32 *map = chan->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/nvif/client.c
35
return client->driver->suspend(client->object.priv, runtime);
drivers/gpu/drm/nouveau/nvif/client.c
41
return client->driver->resume(client->object.priv);
drivers/gpu/drm/nouveau/nvif/client.c
47
nvif_object_dtor(&client->object);
drivers/gpu/drm/nouveau/nvif/client.c
58
ret = nvif_object_ctor(parent != client ? &parent->object : NULL,
drivers/gpu/drm/nouveau/nvif/client.c
61
&client->object);
drivers/gpu/drm/nouveau/nvif/client.c
65
client->object.client = client;
drivers/gpu/drm/nouveau/nvif/client.c
66
client->object.handle = ~0;
drivers/gpu/drm/nouveau/nvif/conn.c
42
ret = nvif_event_ctor_(&conn->object, name ?: "nvifConnHpd", nvif_conn_id(conn),
drivers/gpu/drm/nouveau/nvif/conn.c
44
NVIF_DEBUG(&conn->object, "[NEW EVENT:HPD types:%02x]", types);
drivers/gpu/drm/nouveau/nvif/conn.c
51
nvif_object_dtor(&conn->object);
drivers/gpu/drm/nouveau/nvif/conn.c
63
ret = nvif_object_ctor(&disp->object, name ?: "nvifConn", id, NVIF_CLASS_CONN,
drivers/gpu/drm/nouveau/nvif/conn.c
64
&args, sizeof(args), &conn->object);
drivers/gpu/drm/nouveau/nvif/conn.c
65
NVIF_ERRON(ret, &disp->object, "[NEW conn id:%d]", id);
drivers/gpu/drm/nouveau/nvif/device.c
32
int ret = nvif_object_mthd(&device->object, NV_DEVICE_V0_TIME,
drivers/gpu/drm/nouveau/nvif/device.c
44
return nvif_object_map(&device->object, NULL, 0);
drivers/gpu/drm/nouveau/nvif/device.c
53
nvif_object_dtor(&device->object);
drivers/gpu/drm/nouveau/nvif/device.c
59
int ret = nvif_object_ctor(&client->object, name ? name : "nvifDevice", 0,
drivers/gpu/drm/nouveau/nvif/device.c
60
0x0080, NULL, 0, &device->object);
drivers/gpu/drm/nouveau/nvif/device.c
65
ret = nvif_object_mthd(&device->object, NV_DEVICE_V0_INFO,
drivers/gpu/drm/nouveau/nvif/disp.c
32
nvif_object_dtor(&disp->object);
drivers/gpu/drm/nouveau/nvif/disp.c
62
cid = nvif_sclass(&device->object, disps, oclass);
drivers/gpu/drm/nouveau/nvif/disp.c
63
disp->object.client = NULL;
drivers/gpu/drm/nouveau/nvif/disp.c
65
NVIF_DEBUG(&device->object, "[NEW disp%04x] not supported", oclass);
drivers/gpu/drm/nouveau/nvif/disp.c
71
ret = nvif_object_ctor(&device->object, name ?: "nvifDisp", 0,
drivers/gpu/drm/nouveau/nvif/disp.c
72
disps[cid].oclass, &args, sizeof(args), &disp->object);
drivers/gpu/drm/nouveau/nvif/disp.c
73
NVIF_ERRON(ret, &device->object, "[NEW disp%04x]", disps[cid].oclass);
drivers/gpu/drm/nouveau/nvif/disp.c
77
NVIF_DEBUG(&disp->object, "[NEW] conn_mask:%08x outp_mask:%08x head_mask:%08x",
drivers/gpu/drm/nouveau/nvif/driver.c
35
ret = client->driver->init(name, device, cfg, dbg, &client->object.priv);
drivers/gpu/drm/nouveau/nvif/event.c
32
int ret = nvif_mthd(&event->object, NVIF_EVENT_V0_BLOCK, NULL, 0);
drivers/gpu/drm/nouveau/nvif/event.c
33
NVIF_ERRON(ret, &event->object, "[BLOCK]");
drivers/gpu/drm/nouveau/nvif/event.c
43
int ret = nvif_mthd(&event->object, NVIF_EVENT_V0_ALLOW, NULL, 0);
drivers/gpu/drm/nouveau/nvif/event.c
44
NVIF_ERRON(ret, &event->object, "[ALLOW]");
drivers/gpu/drm/nouveau/nvif/event.c
53
nvif_object_dtor(&event->object);
drivers/gpu/drm/nouveau/nvif/event.c
73
NVIF_CLASS_EVENT, args, argc, &event->object);
drivers/gpu/drm/nouveau/nvif/fifo.c
27
struct nvif_object *object = &device->object;
drivers/gpu/drm/nouveau/nvif/fifo.c
49
ret = nvif_object_mthd(object, NV_DEVICE_V0_INFO, a, sizeof(*a));
drivers/gpu/drm/nouveau/nvif/head.c
33
int ret = nvif_event_ctor(&head->object, name ?: "nvifHeadVBlank", nvif_head_id(head),
drivers/gpu/drm/nouveau/nvif/head.c
35
NVIF_ERRON(ret, &head->object, "[NEW EVENT:VBLANK]");
drivers/gpu/drm/nouveau/nvif/head.c
42
nvif_object_dtor(&head->object);
drivers/gpu/drm/nouveau/nvif/head.c
54
ret = nvif_object_ctor(&disp->object, name ? name : "nvifHead", id, NVIF_CLASS_HEAD,
drivers/gpu/drm/nouveau/nvif/head.c
55
&args, sizeof(args), &head->object);
drivers/gpu/drm/nouveau/nvif/head.c
56
NVIF_ERRON(ret, &disp->object, "[NEW head id:%d]", args.id);
drivers/gpu/drm/nouveau/nvif/mem.c
34
ret = nvif_object_map(&mem->object, NULL, 0);
drivers/gpu/drm/nouveau/nvif/mem.c
44
nvif_object_dtor(&mem->object);
drivers/gpu/drm/nouveau/nvif/mem.c
56
mem->object.client = NULL;
drivers/gpu/drm/nouveau/nvif/mem.c
72
ret = nvif_object_ctor(&mmu->object, name ? name : "nvifMem", 0, oclass,
drivers/gpu/drm/nouveau/nvif/mem.c
73
args, sizeof(*args) + argc, &mem->object);
drivers/gpu/drm/nouveau/nvif/mem.c
93
mem->object.client = NULL;
drivers/gpu/drm/nouveau/nvif/mmu.c
121
ret = nvif_object_mthd(&mmu->object, NVIF_MMU_V0_KIND,
drivers/gpu/drm/nouveau/nvif/mmu.c
30
if (!nvif_object_constructed(&mmu->object))
drivers/gpu/drm/nouveau/nvif/mmu.c
36
nvif_object_dtor(&mmu->object);
drivers/gpu/drm/nouveau/nvif/mmu.c
58
&args, sizeof(args), &mmu->object);
drivers/gpu/drm/nouveau/nvif/mmu.c
67
ret = nvif_mclass(&mmu->object, mems);
drivers/gpu/drm/nouveau/nvif/mmu.c
84
ret = nvif_object_mthd(&mmu->object, NVIF_MMU_V0_HEAP,
drivers/gpu/drm/nouveau/nvif/mmu.c
95
ret = nvif_object_mthd(&mmu->object, NVIF_MMU_V0_TYPE,
drivers/gpu/drm/nouveau/nvif/object.c
101
nvif_object_mthd(struct nvif_object *object, u32 mthd, void *data, u32 size)
drivers/gpu/drm/nouveau/nvif/object.c
127
ret = nvif_object_ioctl(object, args, args_size, NULL);
drivers/gpu/drm/nouveau/nvif/object.c
135
nvif_object_unmap_handle(struct nvif_object *object)
drivers/gpu/drm/nouveau/nvif/object.c
144
nvif_object_ioctl(object, &args, sizeof(args), NULL);
drivers/gpu/drm/nouveau/nvif/object.c
148
nvif_object_map_handle(struct nvif_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvif/object.c
163
ret = nvif_object_ioctl(object, args, argn, NULL);
drivers/gpu/drm/nouveau/nvif/object.c
172
nvif_object_unmap(struct nvif_object *object)
drivers/gpu/drm/nouveau/nvif/object.c
174
struct nvif_client *client = object->client;
drivers/gpu/drm/nouveau/nvif/object.c
175
if (object->map.ptr) {
drivers/gpu/drm/nouveau/nvif/object.c
176
if (object->map.size) {
drivers/gpu/drm/nouveau/nvif/object.c
177
client->driver->unmap(client, object->map.ptr,
drivers/gpu/drm/nouveau/nvif/object.c
178
object->map.size);
drivers/gpu/drm/nouveau/nvif/object.c
179
object->map.size = 0;
drivers/gpu/drm/nouveau/nvif/object.c
181
object->map.ptr = NULL;
drivers/gpu/drm/nouveau/nvif/object.c
182
nvif_object_unmap_handle(object);
drivers/gpu/drm/nouveau/nvif/object.c
187
nvif_object_map(struct nvif_object *object, void *argv, u32 argc)
drivers/gpu/drm/nouveau/nvif/object.c
189
struct nvif_client *client = object->client;
drivers/gpu/drm/nouveau/nvif/object.c
191
int ret = nvif_object_map_handle(object, argv, argc, &handle, &length);
drivers/gpu/drm/nouveau/nvif/object.c
194
object->map.ptr = client->driver->map(client,
drivers/gpu/drm/nouveau/nvif/object.c
197
if (ret = -ENOMEM, object->map.ptr) {
drivers/gpu/drm/nouveau/nvif/object.c
198
object->map.size = length;
drivers/gpu/drm/nouveau/nvif/object.c
202
object->map.ptr = (void *)(unsigned long)handle;
drivers/gpu/drm/nouveau/nvif/object.c
205
nvif_object_unmap_handle(object);
drivers/gpu/drm/nouveau/nvif/object.c
211
nvif_object_dtor(struct nvif_object *object)
drivers/gpu/drm/nouveau/nvif/object.c
220
if (!nvif_object_constructed(object))
drivers/gpu/drm/nouveau/nvif/object.c
223
nvif_object_unmap(object);
drivers/gpu/drm/nouveau/nvif/object.c
224
nvif_object_ioctl(object, &args, sizeof(args), NULL);
drivers/gpu/drm/nouveau/nvif/object.c
225
object->client = NULL;
drivers/gpu/drm/nouveau/nvif/object.c
230
s32 oclass, void *data, u32 size, struct nvif_object *object)
drivers/gpu/drm/nouveau/nvif/object.c
238
object->client = NULL;
drivers/gpu/drm/nouveau/nvif/object.c
239
object->name = name ? name : "nvifObject";
drivers/gpu/drm/nouveau/nvif/object.c
240
object->handle = handle;
drivers/gpu/drm/nouveau/nvif/object.c
241
object->oclass = oclass;
drivers/gpu/drm/nouveau/nvif/object.c
242
object->map.ptr = NULL;
drivers/gpu/drm/nouveau/nvif/object.c
243
object->map.size = 0;
drivers/gpu/drm/nouveau/nvif/object.c
249
nvif_object_dtor(object);
drivers/gpu/drm/nouveau/nvif/object.c
255
nvif_object_dtor(object);
drivers/gpu/drm/nouveau/nvif/object.c
259
object->parent = parent->parent;
drivers/gpu/drm/nouveau/nvif/object.c
264
args->new.object = nvif_handle(object);
drivers/gpu/drm/nouveau/nvif/object.c
269
ret = nvif_object_ioctl(parent, args, args_size, &object->priv);
drivers/gpu/drm/nouveau/nvif/object.c
273
object->client = parent->client;
drivers/gpu/drm/nouveau/nvif/object.c
277
nvif_object_dtor(object);
drivers/gpu/drm/nouveau/nvif/object.c
31
nvif_object_ioctl(struct nvif_object *object, void *data, u32 size, void **hack)
drivers/gpu/drm/nouveau/nvif/object.c
33
struct nvif_client *client = object->client;
drivers/gpu/drm/nouveau/nvif/object.c
39
if (object != &client->object)
drivers/gpu/drm/nouveau/nvif/object.c
40
args->v0.object = nvif_handle(object);
drivers/gpu/drm/nouveau/nvif/object.c
42
args->v0.object = 0;
drivers/gpu/drm/nouveau/nvif/object.c
46
return client->driver->ioctl(client->object.priv, data, size, hack);
drivers/gpu/drm/nouveau/nvif/object.c
57
nvif_object_sclass_get(struct nvif_object *object, struct nvif_sclass **psclass)
drivers/gpu/drm/nouveau/nvif/object.c
75
ret = nvif_object_ioctl(object, args, size, NULL);
drivers/gpu/drm/nouveau/nvif/outp.c
107
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_DRIVE, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
108
NVIF_ERRON(ret, &outp->object, "[DP_DRIVE lanes:%d]", args.lanes);
drivers/gpu/drm/nouveau/nvif/outp.c
128
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_TRAIN, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
129
NVIF_ERRON(ret, &outp->object,
drivers/gpu/drm/nouveau/nvif/outp.c
152
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_RATES, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
153
NVIF_ERRON(ret, &outp->object, "[DP_RATES rates:%d]", args.rates);
drivers/gpu/drm/nouveau/nvif/outp.c
169
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_AUX_XFER, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
170
NVIF_DEBUG(&outp->object, "[DP_AUX_XFER type:%d size:%d addr:%05x] %d size:%d (ret: %d)",
drivers/gpu/drm/nouveau/nvif/outp.c
190
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_AUX_PWR, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
191
NVIF_ERRON(ret, &outp->object, "[DP_AUX_PWR state:%d]", args.state);
drivers/gpu/drm/nouveau/nvif/outp.c
208
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_HDA_ELD, mthd, sizeof(*mthd) + size);
drivers/gpu/drm/nouveau/nvif/outp.c
209
NVIF_ERRON(ret, &outp->object, "[HDA_ELD head:%d size:%d]", head, size);
drivers/gpu/drm/nouveau/nvif/outp.c
220
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_INFOFRAME, args, sizeof(*args) + size);
drivers/gpu/drm/nouveau/nvif/outp.c
221
NVIF_ERRON(ret, &outp->object, "[INFOFRAME type:%d size:%d]", type, size);
drivers/gpu/drm/nouveau/nvif/outp.c
242
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_HDMI, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
243
NVIF_ERRON(ret, &outp->object,
drivers/gpu/drm/nouveau/nvif/outp.c
261
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_LVDS, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
262
NVIF_ERRON(ret, &outp->object, "[LVDS dual:%d 8bpc:%d]", args.dual, args.bpc8);
drivers/gpu/drm/nouveau/nvif/outp.c
275
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_BL_SET, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
276
NVIF_ERRON(ret, &outp->object, "[BL_SET level:%d]", args.level);
drivers/gpu/drm/nouveau/nvif/outp.c
288
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_BL_GET, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
289
NVIF_ERRON(ret, &outp->object, "[BL_GET level:%d]", args.level);
drivers/gpu/drm/nouveau/nvif/outp.c
296
int ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_RELEASE, NULL, 0);
drivers/gpu/drm/nouveau/nvif/outp.c
297
NVIF_ERRON(ret, &outp->object, "[RELEASE]");
drivers/gpu/drm/nouveau/nvif/outp.c
309
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_ACQUIRE, args, sizeof(*args));
drivers/gpu/drm/nouveau/nvif/outp.c
325
NVIF_ERRON(ret, &outp->object, "[ACQUIRE PIOR] or:%d", args.or);
drivers/gpu/drm/nouveau/nvif/outp.c
338
NVIF_ERRON(ret, &outp->object, "[ACQUIRE SOR] or:%d link:%d", args.or, args.link);
drivers/gpu/drm/nouveau/nvif/outp.c
349
NVIF_ERRON(ret, &outp->object, "[ACQUIRE DAC] or:%d", args.or);
drivers/gpu/drm/nouveau/nvif/outp.c
364
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_INHERIT, args, sizeof(*args));
drivers/gpu/drm/nouveau/nvif/outp.c
381
NVIF_ERRON(ret && ret != -ENODEV, &outp->object, "[INHERIT proto:LVDS] ret:%d", ret);
drivers/gpu/drm/nouveau/nvif/outp.c
392
NVIF_ERRON(ret && ret != -ENODEV, &outp->object, "[INHERIT proto:TMDS] ret:%d", ret);
drivers/gpu/drm/nouveau/nvif/outp.c
403
NVIF_ERRON(ret && ret != -ENODEV, &outp->object, "[INHERIT proto:DP] ret:%d", ret);
drivers/gpu/drm/nouveau/nvif/outp.c
417
NVIF_ERRON(ret && ret != -ENODEV, &outp->object, "[INHERIT proto:RGB_CRT] ret:%d", ret);
drivers/gpu/drm/nouveau/nvif/outp.c
42
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_MST_VCPI, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
43
NVIF_ERRON(ret, &outp->object,
drivers/gpu/drm/nouveau/nvif/outp.c
430
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_LOAD_DETECT, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
431
NVIF_ERRON(ret, &outp->object, "[LOAD_DETECT data:%08x] load:%02x", args.data, args.load);
drivers/gpu/drm/nouveau/nvif/outp.c
447
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_EDID_GET, args, sizeof(*args));
drivers/gpu/drm/nouveau/nvif/outp.c
448
NVIF_ERRON(ret, &outp->object, "[EDID_GET] size:%d", args->size);
drivers/gpu/drm/nouveau/nvif/outp.c
472
ret = nvif_mthd(&outp->object, NVIF_OUTP_V0_DETECT, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
473
NVIF_ERRON(ret, &outp->object, "[DETECT] status:%02x", args.status);
drivers/gpu/drm/nouveau/nvif/outp.c
492
nvif_object_dtor(&outp->object);
drivers/gpu/drm/nouveau/nvif/outp.c
504
ret = nvif_object_ctor(&disp->object, name ?: "nvifOutp", id, NVIF_CLASS_OUTP,
drivers/gpu/drm/nouveau/nvif/outp.c
505
&args, sizeof(args), &outp->object);
drivers/gpu/drm/nouveau/nvif/outp.c
506
NVIF_ERRON(ret, &disp->object, "[NEW outp id:%d]", id);
drivers/gpu/drm/nouveau/nvif/outp.c
57
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_MST_ID_PUT, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
58
NVIF_ERRON(ret, &outp->object, "[DP_MST_ID_PUT id:%08x]", args.id);
drivers/gpu/drm/nouveau/nvif/outp.c
69
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_MST_ID_GET, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
70
NVIF_ERRON(ret, &outp->object, "[DP_MST_ID_GET] id:%08x", args.id);
drivers/gpu/drm/nouveau/nvif/outp.c
89
ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_SST, &args, sizeof(args));
drivers/gpu/drm/nouveau/nvif/outp.c
90
NVIF_ERRON(ret, &outp->object,
drivers/gpu/drm/nouveau/nvif/user.c
31
nvif_object_dtor(&device->user.object);
drivers/gpu/drm/nouveau/nvif/user.c
56
cid = nvif_mclass(&device->object, users);
drivers/gpu/drm/nouveau/nvif/user.c
60
ret = nvif_object_ctor(&device->object, name ? name : "nvifUsermode",
drivers/gpu/drm/nouveau/nvif/user.c
62
&device->user.object);
drivers/gpu/drm/nouveau/nvif/user.c
66
nvif_object_map(&device->user.object, NULL, 0);
drivers/gpu/drm/nouveau/nvif/userc361.c
30
hi = nvif_rd32(&user->object, 0x084);
drivers/gpu/drm/nouveau/nvif/userc361.c
31
lo = nvif_rd32(&user->object, 0x080);
drivers/gpu/drm/nouveau/nvif/userc361.c
32
} while (hi != nvif_rd32(&user->object, 0x084));
drivers/gpu/drm/nouveau/nvif/userc361.c
40
nvif_wr32(&user->object, 0x90, token);
drivers/gpu/drm/nouveau/nvif/vmm.c
119
return nvif_object_mthd(&vmm->object, NVIF_VMM_V0_RAW,
drivers/gpu/drm/nouveau/nvif/vmm.c
134
return nvif_object_mthd(&vmm->object, NVIF_VMM_V0_RAW,
drivers/gpu/drm/nouveau/nvif/vmm.c
148
.memory = nvif_handle(&mem->object),
drivers/gpu/drm/nouveau/nvif/vmm.c
155
return nvif_object_mthd(&vmm->object, NVIF_VMM_V0_RAW,
drivers/gpu/drm/nouveau/nvif/vmm.c
172
return nvif_object_mthd(&vmm->object, NVIF_VMM_V0_RAW,
drivers/gpu/drm/nouveau/nvif/vmm.c
187
return nvif_object_mthd(&vmm->object, NVIF_VMM_V0_RAW,
drivers/gpu/drm/nouveau/nvif/vmm.c
195
nvif_object_dtor(&vmm->object);
drivers/gpu/drm/nouveau/nvif/vmm.c
207
vmm->object.client = NULL;
drivers/gpu/drm/nouveau/nvif/vmm.c
228
ret = nvif_object_ctor(&mmu->object, name ? name : "nvifVmm", 0,
drivers/gpu/drm/nouveau/nvif/vmm.c
229
oclass, args, argn, &vmm->object);
drivers/gpu/drm/nouveau/nvif/vmm.c
246
ret = nvif_object_mthd(&vmm->object, NVIF_VMM_V0_PAGE,
drivers/gpu/drm/nouveau/nvif/vmm.c
30
return nvif_object_mthd(&vmm->object, NVIF_VMM_V0_UNMAP,
drivers/gpu/drm/nouveau/nvif/vmm.c
53
args->memory = nvif_handle(&mem->object);
drivers/gpu/drm/nouveau/nvif/vmm.c
57
ret = nvif_object_mthd(&vmm->object, NVIF_VMM_V0_MAP,
drivers/gpu/drm/nouveau/nvif/vmm.c
68
WARN_ON(nvif_object_mthd(&vmm->object, NVIF_VMM_V0_PUT,
drivers/gpu/drm/nouveau/nvif/vmm.c
98
ret = nvif_object_mthd(&vmm->object, NVIF_VMM_V0_GET,
drivers/gpu/drm/nouveau/nvkm/core/client.c
116
nvkm_object_ctor(&nvkm_client, &oclass, &client->object);
drivers/gpu/drm/nouveau/nvkm/core/client.c
52
client->object.client = oclass->client;
drivers/gpu/drm/nouveau/nvkm/core/client.c
53
client->object.handle = oclass->handle;
drivers/gpu/drm/nouveau/nvkm/core/client.c
54
client->object.object = oclass->object;
drivers/gpu/drm/nouveau/nvkm/core/client.c
56
*pobject = &client->object;
drivers/gpu/drm/nouveau/nvkm/core/client.c
76
nvkm_client_child_get(struct nvkm_object *object, int index,
drivers/gpu/drm/nouveau/nvkm/core/client.c
94
nvkm_client_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/client.c
96
return nvkm_client(object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
106
args->v0.object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
118
oclass.object = args->v0.object;
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
132
ret = oclass.ctor(&oclass, data, size, &object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
135
ret = nvkm_object_init(object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
137
list_add_tail(&object->head, &parent->tree);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
138
if (nvkm_object_insert(object)) {
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
139
client->data = object;
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
144
nvkm_object_fini(object, NVKM_POWEROFF);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
147
nvkm_object_del(&object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
153
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
160
nvif_ioctl(object, "delete size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
162
nvif_ioctl(object, "delete\n");
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
163
nvkm_object_fini(object, NVKM_POWEROFF);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
164
nvkm_object_del(&object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
172
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
179
nvif_ioctl(object, "mthd size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
181
nvif_ioctl(object, "mthd vers %d mthd %02x\n",
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
183
ret = nvkm_object_mthd(object, args->v0.method, data, size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
192
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
199
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
206
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
214
nvif_ioctl(object, "map size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
216
nvif_ioctl(object, "map vers %d\n", args->v0.version);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
217
ret = nvkm_object_map(object, data, size, &type,
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
231
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
238
nvif_ioctl(object, "unmap size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
240
nvif_ioctl(object, "unmap\n");
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
241
ret = nvkm_object_unmap(object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
267
struct nvkm_object *object;
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
270
object = nvkm_object_search(client, handle, NULL);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
271
if (IS_ERR(object)) {
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
272
nvif_ioctl(&client->object, "object not found\n");
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
273
return PTR_ERR(object);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
278
ret = nvkm_ioctl_v0[type].func(client, object, data, size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
287
struct nvkm_object *object = &client->object;
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
293
nvif_ioctl(object, "size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
296
nvif_ioctl(object,
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
298
args->v0.version, args->v0.type, args->v0.object,
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
300
ret = nvkm_ioctl_path(client, args->v0.object, args->v0.type,
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
305
nvif_ioctl(object, "return %d\n", ret);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
34
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
42
nvkm_ioctl_sclass_(struct nvkm_object *object, int index, struct nvkm_oclass *oclass)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
44
if ( object->func->uevent &&
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
45
!object->func->uevent(object, NULL, 0, NULL) && index-- == 0) {
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
53
if (object->func->sclass)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
54
return object->func->sclass(object, index, oclass);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
61
struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
69
nvif_ioctl(object, "sclass size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
71
nvif_ioctl(object, "sclass vers %d count %d\n",
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
76
while (nvkm_ioctl_sclass_(object, i, &oclass) >= 0) {
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
98
struct nvkm_object *object = NULL;
drivers/gpu/drm/nouveau/nvkm/core/object.c
102
nvkm_object_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/object.c
104
if (likely(object->func->mthd))
drivers/gpu/drm/nouveau/nvkm/core/object.c
105
return object->func->mthd(object, mthd, data, size);
drivers/gpu/drm/nouveau/nvkm/core/object.c
110
nvkm_object_ntfy(struct nvkm_object *object, u32 mthd,
drivers/gpu/drm/nouveau/nvkm/core/object.c
113
if (likely(object->func->ntfy))
drivers/gpu/drm/nouveau/nvkm/core/object.c
114
return object->func->ntfy(object, mthd, pevent);
drivers/gpu/drm/nouveau/nvkm/core/object.c
119
nvkm_object_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/core/object.c
122
if (likely(object->func->map))
drivers/gpu/drm/nouveau/nvkm/core/object.c
123
return object->func->map(object, argv, argc, type, addr, size);
drivers/gpu/drm/nouveau/nvkm/core/object.c
128
nvkm_object_unmap(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
130
if (likely(object->func->unmap))
drivers/gpu/drm/nouveau/nvkm/core/object.c
131
return object->func->unmap(object);
drivers/gpu/drm/nouveau/nvkm/core/object.c
136
nvkm_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *gpuobj,
drivers/gpu/drm/nouveau/nvkm/core/object.c
139
if (object->func->bind)
drivers/gpu/drm/nouveau/nvkm/core/object.c
140
return object->func->bind(object, gpuobj, align, pgpuobj);
drivers/gpu/drm/nouveau/nvkm/core/object.c
145
nvkm_object_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/core/object.c
164
nvif_debug(object, "%s children...\n", action);
drivers/gpu/drm/nouveau/nvkm/core/object.c
166
list_for_each_entry_reverse(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
172
nvif_debug(object, "%s running...\n", action);
drivers/gpu/drm/nouveau/nvkm/core/object.c
173
if (object->func->fini) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
174
ret = object->func->fini(object, suspend);
drivers/gpu/drm/nouveau/nvkm/core/object.c
176
nvif_error(object, "%s failed with %d\n", action, ret);
drivers/gpu/drm/nouveau/nvkm/core/object.c
183
nvif_debug(object, "%s completed in %lldus\n", action, time);
drivers/gpu/drm/nouveau/nvkm/core/object.c
187
if (object->func->init) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
188
int rret = object->func->init(object);
drivers/gpu/drm/nouveau/nvkm/core/object.c
190
nvif_fatal(object, "failed to restart, %d\n", rret);
drivers/gpu/drm/nouveau/nvkm/core/object.c
193
list_for_each_entry_continue_reverse(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
200
nvkm_object_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
206
nvif_debug(object, "init running...\n");
drivers/gpu/drm/nouveau/nvkm/core/object.c
208
if (object->func->init) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
209
ret = object->func->init(object);
drivers/gpu/drm/nouveau/nvkm/core/object.c
214
nvif_debug(object, "init children...\n");
drivers/gpu/drm/nouveau/nvkm/core/object.c
215
list_for_each_entry(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
222
nvif_debug(object, "init completed in %lldus\n", time);
drivers/gpu/drm/nouveau/nvkm/core/object.c
226
list_for_each_entry_continue_reverse(child, &object->tree, head)
drivers/gpu/drm/nouveau/nvkm/core/object.c
229
nvif_error(object, "init failed with %d\n", ret);
drivers/gpu/drm/nouveau/nvkm/core/object.c
230
if (object->func->fini)
drivers/gpu/drm/nouveau/nvkm/core/object.c
231
object->func->fini(object, NVKM_POWEROFF);
drivers/gpu/drm/nouveau/nvkm/core/object.c
236
nvkm_object_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
239
void *data = object;
drivers/gpu/drm/nouveau/nvkm/core/object.c
242
nvif_debug(object, "destroy children...\n");
drivers/gpu/drm/nouveau/nvkm/core/object.c
244
list_for_each_entry_safe(child, ctemp, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
248
nvif_debug(object, "destroy running...\n");
drivers/gpu/drm/nouveau/nvkm/core/object.c
249
nvkm_object_unmap(object);
drivers/gpu/drm/nouveau/nvkm/core/object.c
250
if (object->func->dtor)
drivers/gpu/drm/nouveau/nvkm/core/object.c
251
data = object->func->dtor(object);
drivers/gpu/drm/nouveau/nvkm/core/object.c
252
nvkm_engine_unref(&object->engine);
drivers/gpu/drm/nouveau/nvkm/core/object.c
254
nvif_debug(object, "destroy completed in %lldus...\n", time);
drivers/gpu/drm/nouveau/nvkm/core/object.c
261
struct nvkm_object *object = *pobject;
drivers/gpu/drm/nouveau/nvkm/core/object.c
262
if (object && !WARN_ON(!object->func)) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
263
*pobject = nvkm_object_dtor(object);
drivers/gpu/drm/nouveau/nvkm/core/object.c
264
nvkm_object_remove(object);
drivers/gpu/drm/nouveau/nvkm/core/object.c
265
list_del(&object->head);
drivers/gpu/drm/nouveau/nvkm/core/object.c
273
const struct nvkm_oclass *oclass, struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
275
object->func = func;
drivers/gpu/drm/nouveau/nvkm/core/object.c
276
object->client = oclass->client;
drivers/gpu/drm/nouveau/nvkm/core/object.c
277
object->engine = nvkm_engine_ref(oclass->engine);
drivers/gpu/drm/nouveau/nvkm/core/object.c
278
object->oclass = oclass->base.oclass;
drivers/gpu/drm/nouveau/nvkm/core/object.c
279
object->handle = oclass->handle;
drivers/gpu/drm/nouveau/nvkm/core/object.c
280
object->object = oclass->object;
drivers/gpu/drm/nouveau/nvkm/core/object.c
281
INIT_LIST_HEAD(&object->head);
drivers/gpu/drm/nouveau/nvkm/core/object.c
282
INIT_LIST_HEAD(&object->tree);
drivers/gpu/drm/nouveau/nvkm/core/object.c
283
RB_CLEAR_NODE(&object->node);
drivers/gpu/drm/nouveau/nvkm/core/object.c
284
WARN_ON(IS_ERR(object->engine));
drivers/gpu/drm/nouveau/nvkm/core/object.c
32
struct nvkm_object *object;
drivers/gpu/drm/nouveau/nvkm/core/object.c
39
object = rb_entry(node, typeof(*object), node);
drivers/gpu/drm/nouveau/nvkm/core/object.c
40
if (handle < object->object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
43
if (handle > object->object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
53
object = &client->object;
drivers/gpu/drm/nouveau/nvkm/core/object.c
57
if (unlikely(func && object->func != func))
drivers/gpu/drm/nouveau/nvkm/core/object.c
59
return object;
drivers/gpu/drm/nouveau/nvkm/core/object.c
63
nvkm_object_remove(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
67
spin_lock_irqsave(&object->client->obj_lock, flags);
drivers/gpu/drm/nouveau/nvkm/core/object.c
68
if (!RB_EMPTY_NODE(&object->node))
drivers/gpu/drm/nouveau/nvkm/core/object.c
69
rb_erase(&object->node, &object->client->objroot);
drivers/gpu/drm/nouveau/nvkm/core/object.c
70
spin_unlock_irqrestore(&object->client->obj_lock, flags);
drivers/gpu/drm/nouveau/nvkm/core/object.c
74
nvkm_object_insert(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/object.c
80
spin_lock_irqsave(&object->client->obj_lock, flags);
drivers/gpu/drm/nouveau/nvkm/core/object.c
81
ptr = &object->client->objroot.rb_node;
drivers/gpu/drm/nouveau/nvkm/core/object.c
85
if (object->object < this->object) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
87
} else if (object->object > this->object) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
90
spin_unlock_irqrestore(&object->client->obj_lock, flags);
drivers/gpu/drm/nouveau/nvkm/core/object.c
95
rb_link_node(&object->node, parent, ptr);
drivers/gpu/drm/nouveau/nvkm/core/object.c
96
rb_insert_color(&object->node, &object->client->objroot);
drivers/gpu/drm/nouveau/nvkm/core/object.c
97
spin_unlock_irqrestore(&object->client->obj_lock, flags);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
101
if (oproxy->object->func->fini) {
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
102
ret = oproxy->object->func->fini(oproxy->object, suspend);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
117
nvkm_oproxy_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
119
struct nvkm_oproxy *oproxy = nvkm_oproxy(object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
128
if (oproxy->object->func->init) {
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
129
ret = oproxy->object->func->init(oproxy->object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
144
nvkm_oproxy_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
146
struct nvkm_oproxy *oproxy = nvkm_oproxy(object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
149
nvkm_object_del(&oproxy->object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
27
nvkm_oproxy_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
29
return nvkm_object_mthd(nvkm_oproxy(object)->object, mthd, data, size);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
33
nvkm_oproxy_ntfy(struct nvkm_object *object, u32 mthd,
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
36
return nvkm_object_ntfy(nvkm_oproxy(object)->object, mthd, pevent);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
40
nvkm_oproxy_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
43
struct nvkm_oproxy *oproxy = nvkm_oproxy(object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
44
return nvkm_object_map(oproxy->object, argv, argc, type, addr, size);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
48
nvkm_oproxy_unmap(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
50
struct nvkm_oproxy *oproxy = nvkm_oproxy(object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
52
if (unlikely(!oproxy->object))
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
55
return nvkm_object_unmap(oproxy->object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
59
nvkm_oproxy_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
62
return nvkm_object_bind(nvkm_oproxy(object)->object,
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
67
nvkm_oproxy_sclass(struct nvkm_object *object, int index,
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
70
struct nvkm_oproxy *oproxy = nvkm_oproxy(object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
71
oclass->parent = oproxy->object;
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
72
if (!oproxy->object->func->sclass)
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
74
return oproxy->object->func->sclass(oproxy->object, index, oclass);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
78
nvkm_oproxy_uevent(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
81
struct nvkm_oproxy *oproxy = nvkm_oproxy(object);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
83
if (!oproxy->object->func->uevent)
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
86
return oproxy->object->func->uevent(oproxy->object, argv, argc, uevent);
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
90
nvkm_oproxy_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/core/oproxy.c
92
struct nvkm_oproxy *oproxy = nvkm_oproxy(object);
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
107
nvkm_ramht_insert(struct nvkm_ramht *ramht, struct nvkm_object *object,
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
118
return nvkm_ramht_update(ramht, co, object, chid,
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
60
nvkm_ramht_update(struct nvkm_ramht *ramht, int co, struct nvkm_object *object,
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
71
if (object) {
drivers/gpu/drm/nouveau/nvkm/core/ramht.c
72
ret = nvkm_object_bind(object, ramht->parent, 16, &data->inst);
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
116
struct nvkm_client *client = uevent->object.client;
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
119
return uevent->func(uevent->parent, uevent->object.object, bits);
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
121
return client->event(uevent->object.object, NULL, 0);
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
149
*pobject = &uevent->object;
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
151
nvkm_object_ctor(&nvkm_uevent, oclass, &uevent->object);
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
22
#define nvkm_uevent(p) container_of((p), struct nvkm_uevent, object)
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
29
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
61
nvkm_uevent_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc)
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
63
struct nvkm_uevent *uevent = nvkm_uevent(object);
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
76
nvkm_uevent_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
78
struct nvkm_uevent *uevent = nvkm_uevent(object);
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
85
nvkm_uevent_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
87
struct nvkm_uevent *uevent = nvkm_uevent(object);
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
96
nvkm_uevent_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/core/uevent.c
98
struct nvkm_uevent *uevent = nvkm_uevent(object);
drivers/gpu/drm/nouveau/nvkm/engine/ce/gv100.c
30
gv100_ce_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, int align,
drivers/gpu/drm/nouveau/nvkm/engine/ce/gv100.c
33
struct nvkm_device *device = object->engine->subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c
34
g84_cipher_oclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c
37
int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16,
drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c
41
nvkm_wo32(*pgpuobj, 0x00, object->oclass);
drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c
56
g84_cipher_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c
59
return nvkm_gpuobj_new(object->engine->subdev.device, 256,
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
149
nvif_ioctl(&ctrl->object, "control pstate user size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
151
nvif_ioctl(&ctrl->object,
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
170
nvkm_control_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
172
struct nvkm_control *ctrl = nvkm_control(object);
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
199
*pobject = &ctrl->object;
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
202
nvkm_object_ctor(&nvkm_control, oclass, &ctrl->object);
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
43
nvif_ioctl(&ctrl->object, "control pstate info size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
45
nvif_ioctl(&ctrl->object, "control pstate info vers %d\n",
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
81
nvif_ioctl(&ctrl->object, "control pstate attr size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c
83
nvif_ioctl(&ctrl->object,
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.h
4
#define nvkm_control(p) container_of((p), struct nvkm_control, object)
drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.h
8
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
176
struct nvkm_object *object = &udev->object;
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
183
nvif_ioctl(object, "device time size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
185
nvif_ioctl(object, "device time vers %d\n", args->v0.version);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
193
nvkm_udevice_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
195
struct nvkm_udevice *udev = nvkm_udevice(object);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
196
nvif_ioctl(object, "device mthd %08x\n", mthd);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
209
nvkm_udevice_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
212
struct nvkm_udevice *udev = nvkm_udevice(object);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
221
nvkm_udevice_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
223
struct nvkm_udevice *udev = nvkm_udevice(object);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
24
#define nvkm_udevice(p) container_of((p), struct nvkm_udevice, object)
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
242
nvkm_udevice_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
244
struct nvkm_udevice *udev = nvkm_udevice(object);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
272
nvkm_udevice_child_get(struct nvkm_object *object, int index,
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
275
struct nvkm_udevice *udev = nvkm_udevice(object);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
332
nvkm_object_ctor(&nvkm_udevice, oclass, &udev->object);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
333
*pobject = &udev->object;
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
38
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
75
struct nvkm_object *object = &udev->object;
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
85
nvif_ioctl(object, "device info size %d\n", size);
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
87
nvif_ioctl(object, "device info vers %d count %d\n",
drivers/gpu/drm/nouveau/nvkm/engine/device/user.c
97
nvif_ioctl(object, "device info vers %d\n", args->v0.version);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
100
if (object->hash < 0)
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
101
return object->hash;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
107
nvkm_disp_chan_child_get(struct nvkm_object *object, int index, struct nvkm_oclass *sclass)
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
109
struct nvkm_disp_chan *chan = nvkm_disp_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
131
nvkm_disp_chan_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
133
struct nvkm_disp_chan *chan = nvkm_disp_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
141
nvkm_disp_chan_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
143
struct nvkm_disp_chan *chan = nvkm_disp_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
150
nvkm_disp_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
152
struct nvkm_disp_chan *chan = nvkm_disp_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
200
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
202
nvkm_object_ctor(&nvkm_disp_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
30
nvkm_disp_chan_ntfy(struct nvkm_object *object, u32 type, struct nvkm_event **pevent)
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
32
struct nvkm_disp_chan *chan = nvkm_disp_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
47
nvkm_disp_chan_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
50
struct nvkm_disp_chan *chan = nvkm_disp_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
68
struct nvkm_disp_chan_object *object = container_of(base, typeof(*object), oproxy);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
70
nvkm_ramht_remove(object->disp->ramht, object->hash);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
86
struct nvkm_disp_chan_object *object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
89
if (!(object = kzalloc_obj(*object)))
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
91
nvkm_oproxy_ctor(&nvkm_disp_chan_child_func_, oclass, &object->oproxy);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
92
object->disp = disp;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
93
*pobject = &object->oproxy.base;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
95
ret = sclass->ctor(device, oclass, argv, argc, &object->oproxy.object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.c
99
object->hash = chan->func->bind(chan, object->oproxy.object, oclass->handle);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.h
19
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.h
27
struct nvkm_gsp_object object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.h
36
int (*push)(struct nvkm_disp_chan *, u64 object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.h
4
#define nvkm_disp_chan(p) container_of((p), struct nvkm_disp_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/conn.h
16
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
568
gf119_disp_dmac_bind(struct nvkm_disp_chan *chan, struct nvkm_object *object, u32 handle)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gf119.c
570
return nvkm_ramht_insert(chan->disp->ramht, object, chan->chid.user, -9, handle,
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
354
struct nvkm_object *object, u32 handle)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
356
return nvkm_ramht_insert(chan->disp->ramht, object, chan->chid.user, -9, handle,
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
794
#define gv100_disp_caps(p) container_of((p), struct gv100_disp_caps, object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
797
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
802
gv100_disp_caps_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
805
struct gv100_disp_caps *caps = gv100_disp_caps(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
827
*pobject = &caps->object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
829
nvkm_object_ctor(&gv100_disp_caps, oclass, &caps->object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/head.h
31
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
626
nv50_disp_dmac_bind(struct nvkm_disp_chan *chan, struct nvkm_object *object, u32 handle)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
628
return nvkm_ramht_insert(chan->disp->ramht, object, chan->chid.user, -10, handle,
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
684
nv50_disp_dmac_push(struct nvkm_disp_chan *chan, u64 object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/nv50.c
686
chan->memory = nvkm_umem_search(chan->object.client, object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.h
66
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/priv.h
4
#define nvkm_udisp(p) container_of((p), struct nvkm_disp, client.object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
152
nvkm_uconn_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
154
struct nvkm_conn *conn = nvkm_uconn(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
158
conn->object.func = NULL;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
192
if (!conn->object.func) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
22
#define nvkm_uconn(p) container_of((p), struct nvkm_conn, object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
252
nvkm_object_ctor(&nvkm_uconn, oclass, &conn->object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
253
*pobject = &conn->object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
34
nvkm_uconn_uevent_gsp(struct nvkm_object *object, u64 token, u32 bits)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
47
return object->client->event(token, &args, sizeof(args.v0));
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
51
nvkm_uconn_uevent_aux(struct nvkm_object *object, u64 token, u32 bits)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
64
return object->client->event(token, &args, sizeof(args.v0));
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
68
nvkm_uconn_uevent_gpio(struct nvkm_object *object, u64 token, u32 bits)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
79
return object->client->event(token, &args, sizeof(args.v0));
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
95
nvkm_uconn_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
97
struct nvkm_conn *conn = nvkm_uconn(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
31
nvkm_udisp_sclass(struct nvkm_object *object, int index, struct nvkm_oclass *sclass)
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
33
struct nvkm_disp *disp = nvkm_udisp(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
63
nvkm_udisp_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
65
struct nvkm_disp *disp = nvkm_udisp(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
68
if (object == &disp->client.object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
69
disp->client.object.func = NULL;
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
93
if (disp->client.object.func) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
97
nvkm_object_ctor(&nvkm_udisp, oclass, &disp->client.object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/udisp.c
98
*pobject = &disp->client.object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
120
if (!head->object.func) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
121
nvkm_object_ctor(&nvkm_uhead, oclass, &head->object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
122
*pobject = &head->object;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
22
#define nvkm_uhead(p) container_of((p), struct nvkm_head, object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
31
nvkm_uhead_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
33
struct nvkm_head *head = nvkm_uhead(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
75
nvkm_uhead_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
77
struct nvkm_head *head = nvkm_uhead(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
87
nvkm_uhead_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
89
struct nvkm_head *head = nvkm_uhead(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uhead.c
93
head->object.func = NULL;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
22
#define nvkm_uoutp(p) container_of((p), struct nvkm_outp, object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
544
nvkm_uoutp_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
546
struct nvkm_outp *outp = nvkm_uoutp(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
568
nvkm_uoutp_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
570
struct nvkm_outp *outp = nvkm_uoutp(object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
574
outp->object.func = NULL;
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
608
if (!outp->object.func) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
657
nvkm_object_ctor(&nvkm_uoutp, oclass, &outp->object);
drivers/gpu/drm/nouveau/nvkm/engine/disp/uoutp.c
658
*pobject = &outp->object;
drivers/gpu/drm/nouveau/nvkm/engine/dma/base.c
42
*pobject = &dmaobj->object;
drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c
37
struct nvkm_object *object;
drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c
39
object = nvkm_object_search(client, handle, &nvkm_dmaobj_func);
drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c
40
if (IS_ERR(object))
drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c
41
return (void *)object;
drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c
43
return nvkm_dmaobj(object);
drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c
79
nvkm_object_ctor(&nvkm_dmaobj_func, oclass, &dmaobj->object);
drivers/gpu/drm/nouveau/nvkm/engine/dma/user.h
4
#define nvkm_dmaobj(p) container_of((p), struct nvkm_dmaobj, object)
drivers/gpu/drm/nouveau/nvkm/engine/falcon.c
46
nvkm_falcon_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/falcon.c
49
return nvkm_gpuobj_new(object->engine->subdev.device, 256,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
163
ret = nvkm_object_bind(vctx->ectx->object, NULL, 0, &vctx->inst);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
41
nvkm_object_del(&ectx->object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
82
ret = engine->func->fifo.cclass(chan, &cclass, &ectx->object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.c
84
ret = nvkm_object_new_(engine->func->cclass, &cclass, NULL, 0, &ectx->object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/cgrp.h
24
struct nvkm_object *object;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
432
ret = nvkm_object_bind(&dmaobj->object, chan->inst, -16, &chan->push);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
112
nvkm_object_ctor(&nvkm_ucgrp, oclass, &ucgrp->object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
113
*pobject = &ucgrp->object;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
22
#define nvkm_ucgrp(p) container_of((p), struct nvkm_ucgrp, object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
32
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
46
nvkm_ucgrp_sclass(struct nvkm_object *object, int index, struct nvkm_oclass *oclass)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
48
struct nvkm_cgrp *cgrp = nvkm_ucgrp(object)->cgrp;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
66
nvkm_ucgrp_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ucgrp.c
68
struct nvkm_ucgrp *ucgrp = nvkm_ucgrp(object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
106
if (!ectx->object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
114
ret = nvkm_object_init(ectx->object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
186
.object = oclass->object,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
188
.parent = uobj->cctx->vctx->ectx->object ?: oclass->parent,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
190
}, argv, argc, &uobj->oproxy.object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
195
uobj->hash = engn->func->ramht_add(engn, uobj->oproxy.object, uobj->chan);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
204
nvkm_uchan_sclass(struct nvkm_object *object, int index, struct nvkm_oclass *oclass)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
206
struct nvkm_chan *chan = nvkm_uchan(object)->chan;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
22
#define nvkm_uchan(p) container_of((p), struct nvkm_uchan, object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
255
nvkm_uchan_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
258
struct nvkm_chan *chan = nvkm_uchan(object)->chan;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
272
nvkm_uchan_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
274
struct nvkm_chan *chan = nvkm_uchan(object)->chan;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
286
nvkm_uchan_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
288
struct nvkm_chan *chan = nvkm_uchan(object)->chan;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
302
nvkm_uchan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
304
struct nvkm_uchan *uchan = nvkm_uchan(object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
321
nvkm_uchan_chan(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
323
if (WARN_ON(object->func != &nvkm_uchan))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
326
return nvkm_uchan(object)->chan;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
37
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
383
nvkm_object_ctor(&nvkm_uchan, oclass, &uchan->object);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
384
*pobject = &uchan->object;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
42
nvkm_uchan_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
44
struct nvkm_chan *chan = nvkm_uchan(object)->chan;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
82
if (!ectx->object)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/uchan.c
90
nvkm_object_fini(ectx->object, NVKM_POWEROFF);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
153
#define gf100_gr_object(p) container_of((p), struct gf100_gr_object, object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
156
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
161
gf100_fermi_mthd_zbc_color(struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
163
struct gf100_gr *gr = gf100_gr(nvkm_gr(object->engine));
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
207
gf100_fermi_mthd_zbc_depth(struct nvkm_object *object, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
209
struct gf100_gr *gr = gf100_gr(nvkm_gr(object->engine));
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
231
gf100_fermi_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
233
nvif_ioctl(object, "fermi mthd %08x\n", mthd);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
236
return gf100_fermi_mthd_zbc_color(object, data, size);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
238
return gf100_fermi_mthd_zbc_depth(object, data, size);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
286
struct gf100_gr_object *object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
288
if (!(object = kzalloc_obj(*object)))
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
290
*pobject = &object->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
293
&gf100_gr_object_func, oclass, &object->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
294
object->chan = chan;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
320
gf100_gr_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
323
struct gf100_gr_chan *chan = gf100_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
355
gf100_gr_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
357
struct gf100_gr_chan *chan = gf100_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
389
nvkm_object_ctor(&gf100_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
392
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h
275
#define gf100_gr_chan(p) container_of((p), struct gf100_gr_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h
279
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1043
nv04_gr_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1046
int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, align,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1050
nvkm_wo32(*pgpuobj, 0x00, object->oclass);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1148
nv04_gr_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1150
struct nv04_gr_chan *chan = nv04_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1161
nv04_gr_chan_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1163
struct nv04_gr_chan *chan = nv04_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1193
nvkm_object_ctor(&nv04_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1196
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
1320
chan ? chan->object.client->name : "unknown",
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
357
#define nv04_gr_chan(p) container_of((p), struct nv04_gr_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
360
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
1012
nvkm_object_ctor(&nv10_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
1015
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
1128
chan ? chan->object.client->name : "unknown",
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
397
#define nv10_gr_chan(p) container_of((p), struct nv10_gr_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
400
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
433
struct nvkm_device *device = chan->object.engine->subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
506
struct nvkm_device *device = chan->object.engine->subdev.device;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
954
nv10_gr_chan_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
956
struct nv10_gr_chan *chan = nv10_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
971
nv10_gr_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c
973
struct nv10_gr_chan *chan = nv10_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
17
nv20_gr_chan_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
19
struct nv20_gr_chan *chan = nv20_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
30
nv20_gr_chan_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
32
struct nv20_gr_chan *chan = nv20_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
60
nv20_gr_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
62
struct nv20_gr_chan *chan = nv20_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
84
nvkm_object_ctor(&nv20_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c
87
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.h
22
#define nv20_gr_chan(p) container_of((p), struct nv20_gr_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.h
26
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv25.c
30
nvkm_object_ctor(&nv25_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv25.c
33
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv2a.c
30
nvkm_object_ctor(&nv2a_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv2a.c
33
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv30.c
31
nvkm_object_ctor(&nv30_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv30.c
34
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv34.c
30
nvkm_object_ctor(&nv34_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv34.c
33
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv35.c
30
nvkm_object_ctor(&nv35_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv35.c
33
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
130
nv40_gr_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
132
struct nv40_gr_chan *chan = nv40_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
157
nvkm_object_ctor(&nv40_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
160
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
44
nv40_gr_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
47
int ret = nvkm_gpuobj_new(object->engine->subdev.device, 20, align,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
51
nvkm_wo32(*pgpuobj, 0x00, object->oclass);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
74
nv40_gr_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
77
struct nv40_gr_chan *chan = nv40_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
92
nv40_gr_chan_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c
94
struct nv40_gr_chan *chan = nv40_gr_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.h
19
#define nv40_gr_chan(p) container_of((p), struct nv40_gr_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.h
23
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c
43
nv50_gr_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c
46
int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c
50
nvkm_wo32(*pgpuobj, 0x00, object->oclass);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c
69
nv50_gr_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c
72
struct nv50_gr *gr = nv50_gr_chan(object)->gr;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c
97
nvkm_object_ctor(&nv50_gr_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c
99
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.h
22
#define nv50_gr_chan(p) container_of((p), struct nv50_gr_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.h
26
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c
39
nv31_mpeg_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c
42
int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, align,
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c
46
nvkm_wo32(*pgpuobj, 0x00, object->oclass);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c
65
nv31_mpeg_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c
67
struct nv31_mpeg_chan *chan = nv31_mpeg_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c
94
nvkm_object_ctor(&nv31_mpeg_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c
97
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.h
21
#define nv31_mpeg_chan(p) container_of((p), struct nv31_mpeg_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.h
25
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
112
nvkm_object_ctor(&nv44_mpeg_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
115
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
41
#define nv44_mpeg_chan(p) container_of((p), struct nv44_mpeg_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
44
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
52
nv44_mpeg_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
55
struct nv44_mpeg_chan *chan = nv44_mpeg_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
56
int ret = nvkm_gpuobj_new(chan->object.engine->subdev.device, 264 * 4,
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
68
nv44_mpeg_chan_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
71
struct nv44_mpeg_chan *chan = nv44_mpeg_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
84
nv44_mpeg_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c
86
struct nv44_mpeg_chan *chan = nv44_mpeg_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv50.c
37
nv50_mpeg_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv50.c
40
int ret = nvkm_gpuobj_new(object->engine->subdev.device, 128 * 4,
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c
53
nvkm_sw_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c
55
struct nvkm_sw_chan *chan = nvkm_sw_chan(object);
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c
82
nvkm_object_ctor(&nvkm_sw_chan, oclass, &chan->object);
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.h
12
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.h
4
#define nvkm_sw_chan(p) container_of((p), struct nvkm_sw_chan, object)
drivers/gpu/drm/nouveau/nvkm/engine/sw/gf100.c
115
*pobject = &chan->base.object;
drivers/gpu/drm/nouveau/nvkm/engine/sw/gf100.c
60
struct nvkm_engine *engine = chan->base.object.engine;
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c
117
*pobject = &chan->base.object;
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv10.c
46
*pobject = &chan->object;
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv50.c
111
*pobject = &chan->base.object;
drivers/gpu/drm/nouveau/nvkm/engine/sw/nv50.c
65
struct nvkm_engine *engine = chan->base.object.engine;
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c
30
nvkm_nvsw_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent)
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c
39
return nvkm_uevent_add(uevent, &nvkm_nvsw(object)->chan->event, 0,
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c
44
nvkm_nvsw_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size)
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c
46
struct nvkm_nvsw *nvsw = nvkm_nvsw(object);
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c
69
*pobject = &nvsw->object;
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c
71
nvkm_object_ctor(&nvkm_nvsw_, oclass, &nvsw->object);
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.h
4
#define nvkm_nvsw(p) container_of((p), struct nvkm_nvsw, object)
drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.h
8
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/engine/xtensa.c
44
nvkm_xtensa_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent,
drivers/gpu/drm/nouveau/nvkm/engine/xtensa.c
47
return nvkm_gpuobj_new(object->engine->subdev.device, 0x10000, align,
drivers/gpu/drm/nouveau/nvkm/subdev/fault/priv.h
11
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/priv.h
3
#define nvkm_fault_buffer(p) container_of((p), struct nvkm_fault_buffer, object)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
107
nvkm_object_ctor(&nvkm_ufault, oclass, &buffer->object);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
108
*pobject = &buffer->object;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
32
nvkm_ufault_uevent(struct nvkm_object *object, void *argv, u32 argc, struct nvkm_uevent *uevent)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
34
struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
47
nvkm_ufault_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
50
struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
59
nvkm_ufault_fini(struct nvkm_object *object, enum nvkm_suspend_state suspend)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
61
struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
67
nvkm_ufault_init(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
69
struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
75
nvkm_ufault_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/client.c
10
const unsigned int id = client->object.handle - NVKM_RM_CLIENT(0);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/client.c
16
if (client->object.client)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/client.c
17
nvkm_gsp_rm_free(&client->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/client.c
41
client->object.client = client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
18
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
23
nvkm_rm_engine_obj_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
25
struct nvkm_rm_engine_obj *obj = container_of(object, typeof(*obj), object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
81
nvkm_object_ctor(&nvkm_rm_engine_obj, oclass, &obj->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
82
*pobject = &obj->object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
92
return nvkm_rm_engine_obj_new(&chan->rm.object, chan->id, oclass, pobject);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
14
struct r535_gr_chan *chan = container_of(oclass->parent, typeof(*chan), object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
16
return nvkm_rm_engine_obj_new(&chan->chan->rm.object, chan->chan->id, oclass, pobject);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.h
12
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
28
r535_gsp_rpc_rm_free(struct nvkm_gsp_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
30
struct nvkm_gsp_client *client = object->client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
35
client->object.handle, object->handle);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
41
rpc->params.hRoot = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
43
rpc->params.hObjectOld = object->handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
48
r535_gsp_rpc_rm_alloc_done(struct nvkm_gsp_object *object, void *params)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
52
nvkm_gsp_rpc_done(object->client->gsp, rpc);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
56
r535_gsp_rpc_rm_alloc_push(struct nvkm_gsp_object *object, void *params)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
59
struct nvkm_gsp *gsp = object->client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
78
r535_gsp_rpc_rm_alloc_get(struct nvkm_gsp_object *object, u32 oclass,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
81
struct nvkm_gsp_client *client = object->client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
86
client->object.handle, object->parent->handle,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
87
object->handle);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
97
rpc->hClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
98
rpc->hParent = object->parent->handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/alloc.c
99
rpc->hObject = object->handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/client.c
31
args = nvkm_gsp_rm_alloc_get(&client->object, handle, NV01_ROOT, sizeof(*args),
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/client.c
32
&client->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/client.c
36
args->hClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/client.c
39
return nvkm_gsp_rm_alloc_wr(&client->object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
28
r535_gsp_rpc_rm_ctrl_done(struct nvkm_gsp_object *object, void *params)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
34
nvkm_gsp_rpc_done(object->client->gsp, rpc);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
38
r535_gsp_rpc_rm_ctrl_push(struct nvkm_gsp_object *object, void **params, u32 repc)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
41
struct nvkm_gsp *gsp = object->client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
54
object->client->object.handle, object->handle, rpc->cmd, rpc->status);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
66
r535_gsp_rpc_rm_ctrl_get(struct nvkm_gsp_object *object, u32 cmd, u32 params_size)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
68
struct nvkm_gsp_client *client = object->client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
73
client->object.handle, object->handle, cmd, params_size);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
80
rpc->hClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ctrl.c
81
rpc->hObject = object->handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
106
nvkm_gsp_rm_free(&device->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
114
return nvkm_gsp_rm_alloc(&device->object, NVKM_RM_SUBDEVICE, NV20_SUBDEVICE_0,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
124
args = nvkm_gsp_rm_alloc_get(&client->object, NVKM_RM_DEVICE, NV01_DEVICE_0, sizeof(*args),
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
125
&device->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
129
args->hClientShare = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
131
ret = nvkm_gsp_rm_alloc_wr(&device->object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
137
nvkm_gsp_rm_free(&device->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
31
struct nvkm_gsp_client *client = device->object.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
41
nvkm_gsp_rm_free(&event->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
65
struct nvkm_gsp_client *client = device->object.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
72
&event->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
76
args->hParentClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/device.c
82
ret = nvkm_gsp_rm_alloc_wr(&event->object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
120
ret = rmapi->disp->chan.set_pushbuf(chan->disp, chan->object.oclass, chan->head, NULL);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
124
args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
125
(chan->object.oclass << 16) | chan->head,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
126
chan->object.oclass, sizeof(*args), &chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
132
return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1446
nvkm_gsp_rm_free(&disp->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1464
ret = nvkm_gsp_rm_alloc(&disp->rm.device.object, disp->func->root.oclass << 16,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1465
disp->func->root.oclass, 0, &disp->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
150
r535_dmac_bind(struct nvkm_disp_chan *chan, struct nvkm_object *object, u32 handle)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
152
return nvkm_ramht_insert(chan->disp->ramht, object, chan->chid.user, -9, handle,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
154
(chan->disp->rm.client.object.handle & 0x3fff));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1544
ret = nvkm_gsp_rm_alloc(&disp->rm.device.object, NVKM_RM_DISP, NV04_DISPLAY_COMMON, 0,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
173
args = nvkm_gsp_rm_alloc_get(&disp->rm.object, (oclass << 16) | inst, oclass,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
190
ret = rmapi->disp->chan.set_pushbuf(chan->disp, chan->object.oclass, chan->head, chan->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
194
return rmapi->disp->chan.dmac_alloc(chan->disp, chan->object.oclass, chan->head,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
195
chan->suspend_put, &chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
201
chan->memory = nvkm_umem_search(chan->object.client, memory);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
48
switch (chan->object.oclass & 0xff) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
69
nvkm_gsp_rm_free(&chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
106
object->client = device->object.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
107
object->parent = &device->object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
108
object->handle = handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
134
ctrl->hClient = fbsr->client.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
170
ctrl->hClient = fbsr->client.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
53
u64 phys, u64 size, struct sg_table *sgt, struct nvkm_gsp_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
55
struct nvkm_gsp_client *client = device->object.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
66
rpc->hClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
67
rpc->hDevice = device->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
120
args->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
187
&chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
197
ctrl = nvkm_gsp_rm_ctrl_get(&chan->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
204
ret = nvkm_gsp_rm_ctrl_wr(&chan->rm.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
209
ctrl = nvkm_gsp_rm_ctrl_get(&chan->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
215
ret = nvkm_gsp_rm_ctrl_wr(&chan->rm.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
287
ctrl->hClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
288
ctrl->hObject = chan->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
289
ctrl->hChanClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
44
struct nvkm_gsp *gsp = chan->rm.object.client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
64
nvkm_gsp_rm_free(&chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
80
struct nvkm_gsp *gsp = device->object.client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
86
args = nvkm_gsp_rm_alloc_get(&device->object, handle,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
153
nvkm_object_ctor(&r535_gr_chan, oclass, &grc->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
157
*pobject = &grc->object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
159
ret = r535_gr_promote_ctx(gr, false, grc->vmm, grc->mem, grc->vma, &chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
40
r535_gr_chan_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
42
struct r535_gr_chan *grc = container_of(object, typeof(*grc), object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
74
ctrl->hChanClient = vmm->rm.client.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
221
gsp->internal.client.object.client = &gsp->internal.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
222
gsp->internal.client.object.parent = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
223
gsp->internal.client.object.handle = rpc->hInternalClient;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
226
gsp->internal.device.object.client = &gsp->internal.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
227
gsp->internal.device.object.parent = &gsp->internal.client.object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
228
gsp->internal.device.object.handle = rpc->hInternalDevice;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
231
gsp->internal.device.subdevice.parent = &gsp->internal.device.object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
977
if (event->object.handle == msg->hEvent) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
104
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
131
ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
135
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
144
ctrl->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
146
ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
33
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
37
ctrl->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
39
WARN_ON(nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
45
nvkm_gsp_rm_free(&vmm->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
63
args = nvkm_gsp_rm_alloc_get(&vmm->rm.device.object, handle, FERMI_VASPACE_A,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
64
sizeof(*args), &vmm->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
72
ret = nvkm_gsp_rm_alloc_wr(&vmm->rm.object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/client.c
14
args = nvkm_gsp_rm_alloc_get(&client->object, handle, NV01_ROOT, sizeof(*args),
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/client.c
15
&client->object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/client.c
19
args->hClient = client->object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/client.c
22
return nvkm_gsp_rm_alloc_wr(&client->object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
20
args = nvkm_gsp_rm_alloc_get(&disp->rm.object, (oclass << 16) | inst, oclass,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
81
ctrl->hClient = gsp->internal.client.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
25
struct nvkm_gsp *gsp = device->object.client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
31
args = nvkm_gsp_rm_alloc_get(&device->object, handle,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
65
args->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
88
gsp->internal.client.object.client = &gsp->internal.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
89
gsp->internal.client.object.parent = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
90
gsp->internal.client.object.handle = rpc->hInternalClient;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
94
gsp->internal.device.object.client = &gsp->internal.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
95
gsp->internal.device.object.parent = &gsp->internal.client.object;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
96
gsp->internal.device.object.handle = rpc->hInternalDevice;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
99
gsp->internal.device.subdevice.parent = &gsp->internal.device.object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
124
nvkm_umem_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
126
struct nvkm_umem *umem = nvkm_umem(object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
127
spin_lock(&umem->object.client->lock);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
129
spin_unlock(&umem->object.client->lock);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
166
nvkm_object_ctor(&nvkm_umem, oclass, &umem->object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
170
*pobject = &umem->object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
182
spin_lock(&umem->object.client->lock);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
183
list_add(&umem->head, &umem->object.client->umem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
184
spin_unlock(&umem->object.client->lock);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
37
struct nvkm_client *master = client->object.client;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
39
struct nvkm_object *object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
42
object = nvkm_object_search(client, handle, &nvkm_umem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
43
if (IS_ERR(object)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
47
if (umem->object.object == handle) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
55
umem = nvkm_umem(object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
63
nvkm_umem_unmap(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
65
struct nvkm_umem *umem = nvkm_umem(object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
86
nvkm_umem_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
89
struct nvkm_umem *umem = nvkm_umem(object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.h
3
#define nvkm_umem(p) container_of((p), struct nvkm_umem, object)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.h
8
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
133
nvkm_ummu_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
135
struct nvkm_ummu *ummu = nvkm_ummu(object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
177
nvkm_object_ctor(&nvkm_ummu, oclass, &ummu->object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
179
*pobject = &ummu->object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
32
nvkm_ummu_sclass(struct nvkm_object *object, int index,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
35
struct nvkm_mmu *mmu = nvkm_ummu(object)->mmu;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.h
3
#define nvkm_ummu(p) container_of((p), struct nvkm_ummu, object)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.h
8
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
153
struct nvkm_client *client = uvmm->object.client;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
36
struct nvkm_object *object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
38
object = nvkm_object_search(client, handle, &nvkm_uvmm);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
39
if (IS_ERR(object))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
392
struct nvkm_client *client = uvmm->object.client;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
40
return (void *)object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
42
return nvkm_vmm_ref(nvkm_uvmm(object)->vmm);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
492
nvkm_uvmm_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
494
struct nvkm_uvmm *uvmm = nvkm_uvmm(object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
507
uvmm->object.client,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
518
nvkm_uvmm_dtor(struct nvkm_object *object)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
520
struct nvkm_uvmm *uvmm = nvkm_uvmm(object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
557
nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
558
*pobject = &uvmm->object;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.h
3
#define nvkm_uvmm(p) container_of((p), struct nvkm_uvmm, object)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.h
8
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/uvfn.c
22
#define nvkm_uvfn(p) container_of((p), struct nvkm_uvfn, object)
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/uvfn.c
28
struct nvkm_object object;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/uvfn.c
33
nvkm_uvfn_map(struct nvkm_object *object, void *argv, u32 argc,
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/uvfn.c
36
struct nvkm_vfn *vfn = nvkm_uvfn(object)->vfn;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/uvfn.c
62
nvkm_object_ctor(&nvkm_uvfn, oclass, &uvfn->object);
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/uvfn.c
65
*pobject = &uvfn->object;
drivers/iio/temperature/mlx90632.c
734
static s32 mlx90632_calc_temp_object_iteration(s32 prev_object_temp, s64 object,
drivers/iio/temperature/mlx90632.c
753
ir_Alpha = div64_s64((s64)object * 10000000LL, Alpha_corr);
drivers/iio/temperature/mlx90632.c
767
static s32 mlx90632_calc_temp_object(s64 object, s64 ambient, s32 Ea, s32 Eb,
drivers/iio/temperature/mlx90632.c
782
temp = mlx90632_calc_temp_object_iteration(temp, object, TAdut, TAdut4,
drivers/iio/temperature/mlx90632.c
789
static s32 mlx90632_calc_temp_object_extended(s64 object, s64 ambient, s64 reflected,
drivers/iio/temperature/mlx90632.c
806
temp = mlx90632_calc_temp_object_iteration(temp, object, TAdut, TaTr4,
drivers/iio/temperature/mlx90632.c
819
s64 object, ambient;
drivers/iio/temperature/mlx90632.c
871
object = mlx90632_preprocess_temp_obj_extended(object_new_raw,
drivers/iio/temperature/mlx90632.c
874
*val = mlx90632_calc_temp_object_extended(object, ambient,
drivers/iio/temperature/mlx90632.c
883
object = mlx90632_preprocess_temp_obj(object_new_raw,
drivers/iio/temperature/mlx90632.c
888
*val = mlx90632_calc_temp_object(object, ambient, Ea, Eb, Fa, Fb, Ga,
drivers/iio/temperature/mlx90635.c
520
static s32 mlx90635_calc_temp_object_iteration(s32 prev_object_temp, s64 object,
drivers/iio/temperature/mlx90635.c
541
calcedFa = div64_s64((s64)object * 100000000000LL, Alpha_corr);
drivers/iio/temperature/mlx90635.c
555
static s32 mlx90635_calc_temp_object(s64 object, s64 ambient, u32 Ea, u32 Eb,
drivers/iio/temperature/mlx90635.c
570
temp = mlx90635_calc_temp_object_iteration(temp, object, TAdut, TAdut4,
drivers/iio/temperature/mlx90635.c
581
s64 object, ambient;
drivers/iio/temperature/mlx90635.c
598
object = mlx90635_preprocess_temp_obj(object_raw,
drivers/iio/temperature/mlx90635.c
602
*val = mlx90635_calc_temp_object(object, ambient, Ea, Eb, Ga, Fa, Fa_scale, Fb,
drivers/infiniband/core/rdma_core.c
138
} else if (uobj->object) {
drivers/infiniband/core/rdma_core.c
145
uobj->object = NULL;
drivers/infiniband/core/rdma_core.c
488
uobj->object = filp;
drivers/infiniband/core/rdma_core.c
564
struct file *filp = uobj->object;
drivers/infiniband/core/rdma_core.c
627
struct file *filp = uobj->object;
drivers/infiniband/core/rdma_core.c
739
struct file *filp = uobj->object;
drivers/infiniband/core/rdma_core.c
793
WARN_ON(entry->object);
drivers/infiniband/core/rdma_core.c
921
obj->object = NULL;
drivers/infiniband/core/uverbs_cmd.c
1076
obj->uevent.uobject.object = cq;
drivers/infiniband/core/uverbs_cmd.c
1366
xrcd = (struct ib_xrcd *)xrcd_uobj->object;
drivers/infiniband/core/uverbs_cmd.c
1470
obj->uevent.uobject.object = qp;
drivers/infiniband/core/uverbs_cmd.c
1601
xrcd = (struct ib_xrcd *)xrcd_uobj->object;
drivers/infiniband/core/uverbs_cmd.c
1620
obj->uevent.uobject.object = qp;
drivers/infiniband/core/uverbs_cmd.c
2476
uobj->object = ah;
drivers/infiniband/core/uverbs_cmd.c
2964
obj->uevent.uobject.object = wq;
drivers/infiniband/core/uverbs_cmd.c
3156
uobj->object = rwq_ind_tbl;
drivers/infiniband/core/uverbs_cmd.c
3418
attr.ext.xrc.xrcd = (struct ib_xrcd *)xrcd_uobj->object;
drivers/infiniband/core/uverbs_cmd.c
3458
obj->uevent.uobject.object = srq;
drivers/infiniband/core/uverbs_cmd.c
473
uobj->object = pd;
drivers/infiniband/core/uverbs_cmd.c
642
obj->uobject.object = xrcd;
drivers/infiniband/core/uverbs_cmd.c
765
uobj->object = mr;
drivers/infiniband/core/uverbs_cmd.c
812
mr = uobj->object;
drivers/infiniband/core/uverbs_cmd.c
860
new_uobj->object = new_mr;
drivers/infiniband/core/uverbs_cmd.c
966
uobj->object = mw;
drivers/infiniband/core/uverbs_std_types.c
106
struct ib_xrcd *xrcd = uobject->object;
drivers/infiniband/core/uverbs_std_types.c
125
struct ib_pd *pd = uobject->object;
drivers/infiniband/core/uverbs_std_types.c
46
return rdma_destroy_ah_user((struct ib_ah *)uobject->object,
drivers/infiniband/core/uverbs_std_types.c
55
struct ib_flow *flow = (struct ib_flow *)uobject->object;
drivers/infiniband/core/uverbs_std_types.c
75
return uverbs_dealloc_mw((struct ib_mw *)uobject->object);
drivers/infiniband/core/uverbs_std_types.c
82
struct ib_rwq_ind_table *rwq_ind_tbl = uobject->object;
drivers/infiniband/core/uverbs_std_types_counters.c
42
struct ib_counters *counters = uobject->object;
drivers/infiniband/core/uverbs_std_types_counters.c
78
uobj->object = counters;
drivers/infiniband/core/uverbs_std_types_cq.c
199
obj->uevent.uobject.object = cq;
drivers/infiniband/core/uverbs_std_types_cq.c
42
struct ib_cq *cq = uobject->object;
drivers/infiniband/core/uverbs_std_types_dm.c
41
struct ib_dm *dm = uobject->object;
drivers/infiniband/core/uverbs_std_types_dm.c
82
uobj->object = dm;
drivers/infiniband/core/uverbs_std_types_dmabuf.c
138
uobj->object = uverbs_dmabuf->dmabuf->file;
drivers/infiniband/core/uverbs_std_types_dmah.c
15
struct ib_dmah *dmah = uobject->object;
drivers/infiniband/core/uverbs_std_types_dmah.c
92
uobj->object = dmah;
drivers/infiniband/core/uverbs_std_types_flow_action.c
41
struct ib_flow_action *action = uobject->object;
drivers/infiniband/core/uverbs_std_types_mr.c
142
uobj->object = mr;
drivers/infiniband/core/uverbs_std_types_mr.c
255
uobj->object = mr;
drivers/infiniband/core/uverbs_std_types_mr.c
384
uobj->object = mr;
drivers/infiniband/core/uverbs_std_types_mr.c
43
return ib_dereg_mr_user((struct ib_mr *)uobject->object,
drivers/infiniband/core/uverbs_std_types_qp.c
130
xrcd = (struct ib_xrcd *)xrcd_uobj->object;
drivers/infiniband/core/uverbs_std_types_qp.c
15
struct ib_qp *qp = uobject->object;
drivers/infiniband/core/uverbs_std_types_qp.c
265
obj->uevent.uobject.object = qp;
drivers/infiniband/core/uverbs_std_types_srq.c
113
obj->uevent.uobject.object = srq;
drivers/infiniband/core/uverbs_std_types_srq.c
14
struct ib_srq *srq = uobject->object;
drivers/infiniband/core/uverbs_std_types_srq.c
81
attr.ext.xrc.xrcd = (struct ib_xrcd *)xrcd_uobj->object;
drivers/infiniband/core/uverbs_std_types_wq.c
14
struct ib_wq *wq = uobject->object;
drivers/infiniband/core/uverbs_std_types_wq.c
80
obj->uevent.uobject.object = wq;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4756
uobj->object = entry;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4780
struct bnxt_re_user_mmap_entry *entry = uobject->object;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4894
uobj->object = entry;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4918
struct bnxt_re_user_mmap_entry *entry = uobject->object;
drivers/infiniband/hw/mlx5/devx.c
1453
struct devx_obj *obj = uobject->object;
drivers/infiniband/hw/mlx5/devx.c
1610
uobj->object = obj;
drivers/infiniband/hw/mlx5/devx.c
2090
obj = (struct devx_obj *)devx_uobj->object;
drivers/infiniband/hw/mlx5/devx.c
2420
uobj->object = obj;
drivers/infiniband/hw/mlx5/devx.c
2439
struct devx_umem *obj = uobject->object;
drivers/infiniband/hw/mlx5/devx.c
2665
struct devx_obj *obj = devx_out->uobject->object;
drivers/infiniband/hw/mlx5/devx.c
2714
(get_dec_obj_type(uobject->object, MLX5_EVENT_TYPE_MAX) !=
drivers/infiniband/hw/mlx5/devx.c
2720
obj = uobject->object;
drivers/infiniband/hw/mlx5/devx.c
646
to_mcq(uobj->object)->mcq.cqn) ==
drivers/infiniband/hw/mlx5/devx.c
651
struct mlx5_core_srq *srq = &(to_msrq(uobj->object)->msrq);
drivers/infiniband/hw/mlx5/devx.c
669
to_msrq(uobj->object)->msrq.srqn) ==
drivers/infiniband/hw/mlx5/devx.c
675
struct mlx5_ib_qp *qp = to_mqp(uobj->object);
drivers/infiniband/hw/mlx5/devx.c
703
to_mrwq(uobj->object)->core_qp.qpn) ==
drivers/infiniband/hw/mlx5/devx.c
708
to_mrwq_ind_table(uobj->object)->rqtn) ==
drivers/infiniband/hw/mlx5/devx.c
714
struct devx_obj *devx_uobj = uobj->object;
drivers/infiniband/hw/mlx5/dm.c
210
struct ib_dm *ibdm = uobj->object;
drivers/infiniband/hw/mlx5/fs.c
2493
devx_obj = arr_flow_actions[0]->object;
drivers/infiniband/hw/mlx5/fs.c
2536
to_mflow_act(arr_flow_actions[i]->object);
drivers/infiniband/hw/mlx5/fs.c
2542
arr_flow_actions[i]->object);
drivers/infiniband/hw/mlx5/fs.c
2578
struct mlx5_ib_flow_matcher *obj = uobject->object;
drivers/infiniband/hw/mlx5/fs.c
2822
struct mlx5_ib_steering_anchor *obj = uobject->object;
drivers/infiniband/hw/mlx5/fs.c
2979
uobj->object = obj;
drivers/infiniband/hw/mlx5/fs.c
3051
uobj->object = obj;
drivers/infiniband/hw/mlx5/main.c
4126
struct mlx5_user_mmap_entry *obj = uobject->object;
drivers/infiniband/hw/mlx5/main.c
4209
uobj->object = entry;
drivers/infiniband/hw/mlx5/main.c
4336
uobj->object = entry;
drivers/infiniband/hw/mlx5/qos.c
73
uobj->object = pp_entry;
drivers/infiniband/hw/mlx5/qos.c
89
struct mlx5_ib_pp *pp_entry = uobject->object;
drivers/input/gameport/gameport.c
262
void *object;
drivers/input/gameport/gameport.c
301
if (event->object == e->object) {
drivers/input/gameport/gameport.c
336
gameport_add_port(event->object);
drivers/input/gameport/gameport.c
340
gameport_attach_driver(event->object);
drivers/input/gameport/gameport.c
353
static int gameport_queue_event(void *object, struct module *owner,
drivers/input/gameport/gameport.c
370
if (event->object == object) {
drivers/input/gameport/gameport.c
393
event->object = object;
drivers/input/gameport/gameport.c
408
static void gameport_remove_pending_events(void *object)
drivers/input/gameport/gameport.c
416
if (event->object == object) {
drivers/input/gameport/gameport.c
443
gameport = event->object;
drivers/input/serio/hil_mlc.c
589
mlc->imatch = node->object.packet;
drivers/input/serio/hil_mlc.c
593
mlc->imatch = node->object.packet;
drivers/input/serio/hil_mlc.c
597
mlc->imatch = node->object.packet;
drivers/input/serio/hil_mlc.c
641
BUG_ON(node->object.func == NULL);
drivers/input/serio/hil_mlc.c
642
rc = node->object.func(mlc, node->arg);
drivers/input/serio/hil_mlc.c
674
pack = node->object.packet;
drivers/input/serio/hil_mlc.c
680
pack = node->object.packet;
drivers/input/serio/hil_mlc.c
686
pack = node->object.packet;
drivers/input/serio/serio.c
134
void *object;
drivers/input/serio/serio.c
163
static void serio_remove_duplicate_events(void *object,
drivers/input/serio/serio.c
171
if (object == e->object) {
drivers/input/serio/serio.c
197
serio_add_port(event->object);
drivers/input/serio/serio.c
201
serio_reconnect_port(event->object);
drivers/input/serio/serio.c
205
serio_disconnect_port(event->object);
drivers/input/serio/serio.c
206
serio_find_driver(event->object);
drivers/input/serio/serio.c
210
serio_reconnect_subtree(event->object);
drivers/input/serio/serio.c
214
serio_attach_driver(event->object);
drivers/input/serio/serio.c
218
serio_remove_duplicate_events(event->object, event->type);
drivers/input/serio/serio.c
225
static int serio_queue_event(void *object, struct module *owner,
drivers/input/serio/serio.c
240
if (event->object == object) {
drivers/input/serio/serio.c
261
event->object = object;
drivers/input/serio/serio.c
274
static void serio_remove_pending_events(void *object)
drivers/input/serio/serio.c
281
if (event->object == object) {
drivers/input/serio/serio.c
303
serio = event->object;
drivers/input/touchscreen/atmel_mxt_ts.c
1403
struct mxt_object *object;
drivers/input/touchscreen/atmel_mxt_ts.c
1424
object = mxt_get_object(data, type);
drivers/input/touchscreen/atmel_mxt_ts.c
1425
if (!object) {
drivers/input/touchscreen/atmel_mxt_ts.c
1440
if (size > mxt_obj_size(object)) {
drivers/input/touchscreen/atmel_mxt_ts.c
1447
size - mxt_obj_size(object), type);
drivers/input/touchscreen/atmel_mxt_ts.c
1448
} else if (mxt_obj_size(object) > size) {
drivers/input/touchscreen/atmel_mxt_ts.c
1459
mxt_obj_size(object) - size, type);
drivers/input/touchscreen/atmel_mxt_ts.c
1462
if (instance >= mxt_obj_instances(object)) {
drivers/input/touchscreen/atmel_mxt_ts.c
1467
reg = object->start_address + mxt_obj_size(object) * instance;
drivers/input/touchscreen/atmel_mxt_ts.c
1480
if (i > mxt_obj_size(object))
drivers/input/touchscreen/atmel_mxt_ts.c
1489
reg, object->type, byte_offset);
drivers/input/touchscreen/atmel_mxt_ts.c
1750
struct mxt_object *object = object_table + i;
drivers/input/touchscreen/atmel_mxt_ts.c
1753
le16_to_cpus(&object->start_address);
drivers/input/touchscreen/atmel_mxt_ts.c
1755
if (object->num_report_ids) {
drivers/input/touchscreen/atmel_mxt_ts.c
1757
reportid += object->num_report_ids *
drivers/input/touchscreen/atmel_mxt_ts.c
1758
mxt_obj_instances(object);
drivers/input/touchscreen/atmel_mxt_ts.c
1767
object->type, object->start_address,
drivers/input/touchscreen/atmel_mxt_ts.c
1768
mxt_obj_size(object), mxt_obj_instances(object),
drivers/input/touchscreen/atmel_mxt_ts.c
1771
switch (object->type) {
drivers/input/touchscreen/atmel_mxt_ts.c
1780
data->T5_msg_size = mxt_obj_size(object);
drivers/input/touchscreen/atmel_mxt_ts.c
1783
data->T5_msg_size = mxt_obj_size(object) - 1;
drivers/input/touchscreen/atmel_mxt_ts.c
1785
data->T5_address = object->start_address;
drivers/input/touchscreen/atmel_mxt_ts.c
1789
data->T6_address = object->start_address;
drivers/input/touchscreen/atmel_mxt_ts.c
1792
data->T7_address = object->start_address;
drivers/input/touchscreen/atmel_mxt_ts.c
1795
data->T71_address = object->start_address;
drivers/input/touchscreen/atmel_mxt_ts.c
1802
object->num_report_ids - 1;
drivers/input/touchscreen/atmel_mxt_ts.c
1803
data->num_touchids = object->num_report_ids;
drivers/input/touchscreen/atmel_mxt_ts.c
1810
data->T18_address = object->start_address;
drivers/input/touchscreen/atmel_mxt_ts.c
1813
data->T44_address = object->start_address;
drivers/input/touchscreen/atmel_mxt_ts.c
1827
data->num_touchids = object->num_report_ids - 2;
drivers/input/touchscreen/atmel_mxt_ts.c
1831
end_address = object->start_address
drivers/input/touchscreen/atmel_mxt_ts.c
1832
+ mxt_obj_size(object) * mxt_obj_instances(object) - 1;
drivers/input/touchscreen/atmel_mxt_ts.c
1949
struct mxt_object *object;
drivers/input/touchscreen/atmel_mxt_ts.c
1951
object = mxt_get_object(data, MXT_TOUCH_MULTI_T9);
drivers/input/touchscreen/atmel_mxt_ts.c
1952
if (!object)
drivers/input/touchscreen/atmel_mxt_ts.c
1956
object->start_address + MXT_T9_XSIZE,
drivers/input/touchscreen/atmel_mxt_ts.c
1962
object->start_address + MXT_T9_YSIZE,
drivers/input/touchscreen/atmel_mxt_ts.c
1968
object->start_address + MXT_T9_RANGE,
drivers/input/touchscreen/atmel_mxt_ts.c
1977
object->start_address + MXT_T9_ORIENT,
drivers/input/touchscreen/atmel_mxt_ts.c
1993
struct mxt_object *object;
drivers/input/touchscreen/atmel_mxt_ts.c
1998
object = mxt_get_object(data, MXT_TOUCH_MULTITOUCHSCREEN_T100);
drivers/input/touchscreen/atmel_mxt_ts.c
1999
if (!object)
drivers/input/touchscreen/atmel_mxt_ts.c
2004
object->start_address + MXT_T100_XRANGE,
drivers/input/touchscreen/atmel_mxt_ts.c
2012
object->start_address + MXT_T100_YRANGE,
drivers/input/touchscreen/atmel_mxt_ts.c
2020
object->start_address + MXT_T100_XSIZE,
drivers/input/touchscreen/atmel_mxt_ts.c
2026
object->start_address + MXT_T100_YSIZE,
drivers/input/touchscreen/atmel_mxt_ts.c
2033
object->start_address + MXT_T100_CFG1,
drivers/input/touchscreen/atmel_mxt_ts.c
2044
object->start_address + MXT_T100_TCHAUX,
drivers/input/touchscreen/atmel_mxt_ts.c
2705
struct mxt_object *object;
drivers/input/touchscreen/atmel_mxt_ts.c
2708
object = mxt_get_object(data, MXT_GEN_COMMAND_T6);
drivers/input/touchscreen/atmel_mxt_ts.c
2709
if (!object)
drivers/input/touchscreen/atmel_mxt_ts.c
2712
dbg->diag_cmd_address = object->start_address + MXT_COMMAND_DIAGNOSTIC;
drivers/input/touchscreen/atmel_mxt_ts.c
2714
object = mxt_get_object(data, MXT_DEBUG_DIAGNOSTIC_T37);
drivers/input/touchscreen/atmel_mxt_ts.c
2715
if (!object)
drivers/input/touchscreen/atmel_mxt_ts.c
2718
if (mxt_obj_size(object) != sizeof(struct t37_debug)) {
drivers/input/touchscreen/atmel_mxt_ts.c
2723
dbg->t37_address = object->start_address;
drivers/input/touchscreen/atmel_mxt_ts.c
2836
struct mxt_object *object, int instance,
drivers/input/touchscreen/atmel_mxt_ts.c
2841
if (mxt_obj_instances(object) > 1)
drivers/input/touchscreen/atmel_mxt_ts.c
2844
for (i = 0; i < mxt_obj_size(object); i++)
drivers/input/touchscreen/atmel_mxt_ts.c
2856
struct mxt_object *object;
drivers/input/touchscreen/atmel_mxt_ts.c
2869
object = data->object_table + i;
drivers/input/touchscreen/atmel_mxt_ts.c
2871
if (!mxt_object_readable(object->type))
drivers/input/touchscreen/atmel_mxt_ts.c
2874
count += sysfs_emit_at(buf, count, "T%u:\n", object->type);
drivers/input/touchscreen/atmel_mxt_ts.c
2876
for (j = 0; j < mxt_obj_instances(object); j++) {
drivers/input/touchscreen/atmel_mxt_ts.c
2877
u16 size = mxt_obj_size(object);
drivers/input/touchscreen/atmel_mxt_ts.c
2878
u16 addr = object->start_address + j * size;
drivers/input/touchscreen/atmel_mxt_ts.c
2884
count = mxt_show_instance(buf, count, object, j, obuf);
drivers/input/touchscreen/atmel_mxt_ts.c
3062
static DEVICE_ATTR(object, S_IRUGO, mxt_object_show, NULL);
drivers/input/touchscreen/atmel_mxt_ts.c
755
struct mxt_object *object;
drivers/input/touchscreen/atmel_mxt_ts.c
759
object = data->object_table + i;
drivers/input/touchscreen/atmel_mxt_ts.c
760
if (object->type == type)
drivers/input/touchscreen/atmel_mxt_ts.c
761
return object;
drivers/input/touchscreen/atmel_mxt_ts.c
804
struct mxt_object *object;
drivers/input/touchscreen/atmel_mxt_ts.c
807
object = mxt_get_object(data, type);
drivers/input/touchscreen/atmel_mxt_ts.c
808
if (!object || offset >= mxt_obj_size(object))
drivers/input/touchscreen/atmel_mxt_ts.c
811
reg = object->start_address;
drivers/net/ethernet/huawei/hinic/hinic_debugfs.c
150
out = hinic_dbg_get_sq_info(dbg->dev, dbg->object, *desc);
drivers/net/ethernet/huawei/hinic/hinic_debugfs.c
154
out = hinic_dbg_get_rq_info(dbg->dev, dbg->object, *desc);
drivers/net/ethernet/huawei/hinic/hinic_debugfs.c
190
tmp->object = data;
drivers/net/ethernet/huawei/hinic/hinic_dev.h
73
void *object;
drivers/net/ethernet/mellanox/mlx5/core/debugfs.c
474
field = qp_read_field(d->dev, d->object, desc->i, &is_str);
drivers/net/ethernet/mellanox/mlx5/core/debugfs.c
478
field = eq_read_field(d->dev, d->object, desc->i);
drivers/net/ethernet/mellanox/mlx5/core/debugfs.c
482
field = cq_read_field(d->dev, d->object, desc->i);
drivers/net/ethernet/mellanox/mlx5/core/debugfs.c
517
d->object = data;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
480
struct mlx5_eqe_obj_change *object;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
488
object = &eqe->data.obj_change;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
489
type = be16_to_cpu(object->obj_type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
494
sa_entry = xa_load(&ipsec->sadb, be32_to_cpu(object->obj_id));
drivers/net/wwan/iosm/iosm_ipc_pcie.c
238
union acpi_object *object;
drivers/net/wwan/iosm/iosm_ipc_pcie.c
247
object = acpi_evaluate_dsm(handle_acpi, &wwan_acpi_guid, 0, 3, NULL);
drivers/net/wwan/iosm/iosm_ipc_pcie.c
248
if (!object)
drivers/net/wwan/iosm/iosm_ipc_pcie.c
251
if (object->integer.value == 3)
drivers/net/wwan/iosm/iosm_ipc_pcie.c
254
ACPI_FREE(object);
drivers/platform/x86/lenovo/thinkpad_acpi.c
588
#define TPACPI_HANDLE(object, parent, paths...) \
drivers/platform/x86/lenovo/thinkpad_acpi.c
589
static acpi_handle object##_handle; \
drivers/platform/x86/lenovo/thinkpad_acpi.c
590
static const acpi_handle * const object##_parent __initconst = \
drivers/platform/x86/lenovo/thinkpad_acpi.c
592
static char *object##_paths[] __initdata = { paths }
drivers/platform/x86/lenovo/thinkpad_acpi.c
737
#define TPACPI_ACPIHANDLE_INIT(object) \
drivers/platform/x86/lenovo/thinkpad_acpi.c
738
drv_acpi_handle_init(#object, &object##_handle, *object##_parent, \
drivers/platform/x86/lenovo/thinkpad_acpi.c
739
object##_paths, ARRAY_SIZE(object##_paths))
drivers/platform/x86/sony-laptop.c
764
union acpi_object *object = __call_snc_method(handle, name, value);
drivers/platform/x86/sony-laptop.c
766
if (!object)
drivers/platform/x86/sony-laptop.c
771
} else if (object->type == ACPI_TYPE_BUFFER) {
drivers/platform/x86/sony-laptop.c
772
len = MIN(buflen, object->buffer.length);
drivers/platform/x86/sony-laptop.c
774
memcpy(buffer, object->buffer.pointer, len);
drivers/platform/x86/sony-laptop.c
776
} else if (object->type == ACPI_TYPE_INTEGER) {
drivers/platform/x86/sony-laptop.c
777
len = MIN(buflen, sizeof(object->integer.value));
drivers/platform/x86/sony-laptop.c
779
memcpy(buffer, &object->integer.value, len);
drivers/platform/x86/sony-laptop.c
782
pr_warn("Unexpected acpi_object: 0x%x\n", object->type);
drivers/platform/x86/sony-laptop.c
786
kfree(object);
drivers/tee/qcomtee/async.c
110
struct qcomtee_object *object;
drivers/tee/qcomtee/async.c
116
object = qcomtee_idx_erase(oic, msg->object_ids[i]);
drivers/tee/qcomtee/async.c
117
qcomtee_object_put(object);
drivers/tee/qcomtee/call.c
15
static int find_qtee_object(struct qcomtee_object **object, unsigned long id,
drivers/tee/qcomtee/call.c
162
struct qcomtee_object *object = arg->o;
drivers/tee/qcomtee/call.c
164
switch (typeof_qcomtee_object(object)) {
drivers/tee/qcomtee/call.c
171
if (is_qcomtee_user_object(object))
drivers/tee/qcomtee/call.c
172
return qcomtee_user_param_from_object(param, object,
drivers/tee/qcomtee/call.c
175
else if (is_qcomtee_memobj_object(object))
drivers/tee/qcomtee/call.c
176
return qcomtee_memobj_param_from_object(param, object,
drivers/tee/qcomtee/call.c
181
return qcomtee_context_add_qtee_object(param, object, ctx);
drivers/tee/qcomtee/call.c
22
*object = idr_find(&ctxdata->qtee_objects_idr, id);
drivers/tee/qcomtee/call.c
23
if (!qcomtee_object_get(*object))
drivers/tee/qcomtee/call.c
32
struct qcomtee_object *object;
drivers/tee/qcomtee/call.c
35
object = idr_remove(&ctxdata->qtee_objects_idr, id);
drivers/tee/qcomtee/call.c
37
qcomtee_object_put(object);
drivers/tee/qcomtee/call.c
399
struct qcomtee_object *object;
drivers/tee/qcomtee/call.c
430
object = ROOT_QCOMTEE_OBJECT;
drivers/tee/qcomtee/call.c
431
} else if (find_qtee_object(&object, arg->id, ctxdata)) {
drivers/tee/qcomtee/call.c
439
ret = qcomtee_object_do_invoke(oic, object, arg->op, u, &result);
drivers/tee/qcomtee/call.c
468
qcomtee_object_put(object);
drivers/tee/qcomtee/call.c
52
struct qcomtee_object *object,
drivers/tee/qcomtee/call.c
59
ret = idr_alloc(&ctxdata->qtee_objects_idr, object, 0, 0,
drivers/tee/qcomtee/call.c
601
struct qcomtee_object *object;
drivers/tee/qcomtee/call.c
607
idr_for_each_entry(&ctxdata->qtee_objects_idr, object, id)
drivers/tee/qcomtee/call.c
608
qcomtee_object_put(object);
drivers/tee/qcomtee/call.c
72
int qcomtee_context_find_qtee_object(struct qcomtee_object **object,
drivers/tee/qcomtee/call.c
78
return find_qtee_object(object, param->u.objref.id, ctxdata);
drivers/tee/qcomtee/core.c
101
queue_work(qcomtee->wq, &object->work);
drivers/tee/qcomtee/core.c
105
qcomtee_object_name(object), ret, result);
drivers/tee/qcomtee/core.c
106
qcomtee_qtee_object_free(object);
drivers/tee/qcomtee/core.c
110
static void qcomtee_release_qtee_object(struct qcomtee_object *object)
drivers/tee/qcomtee/core.c
113
tee_get_drvdata(object->info.qcomtee_async_ctx->teedev);
drivers/tee/qcomtee/core.c
115
INIT_WORK(&object->work, qcomtee_do_release_qtee_object);
drivers/tee/qcomtee/core.c
116
queue_work(qcomtee->wq, &object->work);
drivers/tee/qcomtee/core.c
121
struct qcomtee_object *object;
drivers/tee/qcomtee/core.c
124
object = container_of(refcount, struct qcomtee_object, refcount);
drivers/tee/qcomtee/core.c
133
switch (typeof_qcomtee_object(object)) {
drivers/tee/qcomtee/core.c
135
qcomtee_release_qtee_object(object);
drivers/tee/qcomtee/core.c
139
name = object->name;
drivers/tee/qcomtee/core.c
141
if (object->ops->release)
drivers/tee/qcomtee/core.c
142
object->ops->release(object);
drivers/tee/qcomtee/core.c
160
int qcomtee_object_get(struct qcomtee_object *object)
drivers/tee/qcomtee/core.c
162
if (object != &qcomtee_primordial_object &&
drivers/tee/qcomtee/core.c
163
object != NULL_QCOMTEE_OBJECT &&
drivers/tee/qcomtee/core.c
164
object != ROOT_QCOMTEE_OBJECT)
drivers/tee/qcomtee/core.c
165
return kref_get_unless_zero(&object->refcount);
drivers/tee/qcomtee/core.c
174
void qcomtee_object_put(struct qcomtee_object *object)
drivers/tee/qcomtee/core.c
176
if (object != &qcomtee_primordial_object &&
drivers/tee/qcomtee/core.c
177
object != NULL_QCOMTEE_OBJECT &&
drivers/tee/qcomtee/core.c
178
object != ROOT_QCOMTEE_OBJECT)
drivers/tee/qcomtee/core.c
179
kref_put(&object->refcount, qcomtee_object_release);
drivers/tee/qcomtee/core.c
183
struct qcomtee_object *object)
drivers/tee/qcomtee/core.c
188
return xa_alloc_cyclic(&qcomtee->xa_local_objects, idx, object,
drivers/tee/qcomtee/core.c
217
struct qcomtee_object *object,
drivers/tee/qcomtee/core.c
222
switch (typeof_qcomtee_object(object)) {
drivers/tee/qcomtee/core.c
224
if (qcomtee_idx_alloc(oic, &idx, object) < 0)
drivers/tee/qcomtee/core.c
232
*object_id = object->info.qtee_id;
drivers/tee/qcomtee/core.c
268
struct qcomtee_object *object;
drivers/tee/qcomtee/core.c
274
object = xa_load(&qcomtee->xa_local_objects, object_id);
drivers/tee/qcomtee/core.c
276
qcomtee_object_get(object);
drivers/tee/qcomtee/core.c
278
return object;
drivers/tee/qcomtee/core.c
290
int qcomtee_object_user_init(struct qcomtee_object *object,
drivers/tee/qcomtee/core.c
298
kref_init(&object->refcount);
drivers/tee/qcomtee/core.c
299
QCOMTEE_OBJECT_SET(object, QCOMTEE_OBJECT_TYPE_NULL);
drivers/tee/qcomtee/core.c
308
object->ops = ops;
drivers/tee/qcomtee/core.c
309
if (!object->ops->dispatch)
drivers/tee/qcomtee/core.c
313
object->name = kvasprintf_const(GFP_KERNEL, fmt, ap);
drivers/tee/qcomtee/core.c
314
QCOMTEE_OBJECT_SET(object, QCOMTEE_OBJECT_TYPE_CB);
drivers/tee/qcomtee/core.c
359
struct qcomtee_object **object,
drivers/tee/qcomtee/core.c
366
*object = NULL_QCOMTEE_OBJECT;
drivers/tee/qcomtee/core.c
370
*object = qcomtee_local_object_get(oic, object_id);
drivers/tee/qcomtee/core.c
371
if (*object == NULL_QCOMTEE_OBJECT)
drivers/tee/qcomtee/core.c
377
*object = qcomtee_qtee_object_alloc(oic, object_id);
drivers/tee/qcomtee/core.c
378
if (*object == NULL_QCOMTEE_OBJECT)
drivers/tee/qcomtee/core.c
396
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/core.c
412
if (qcomtee_object_id_get(oic, object, &object_id))
drivers/tee/qcomtee/core.c
55
struct qcomtee_object *object;
drivers/tee/qcomtee/core.c
57
object = kzalloc_obj(*object);
drivers/tee/qcomtee/core.c
58
if (!object)
drivers/tee/qcomtee/core.c
62
object->name = kasprintf(GFP_KERNEL, "qcomtee-%u", object_id);
drivers/tee/qcomtee/core.c
626
struct qcomtee_object *object;
drivers/tee/qcomtee/core.c
63
QCOMTEE_OBJECT_SET(object, QCOMTEE_OBJECT_TYPE_TEE, object_id);
drivers/tee/qcomtee/core.c
635
object = qcomtee_local_object_get(oic, object_id);
drivers/tee/qcomtee/core.c
636
if (object == NULL_QCOMTEE_OBJECT) {
drivers/tee/qcomtee/core.c
64
kref_init(&object->refcount);
drivers/tee/qcomtee/core.c
641
oic->object = object;
drivers/tee/qcomtee/core.c
649
qcomtee_object_put(object);
drivers/tee/qcomtee/core.c
654
qcomtee_object_get(object);
drivers/tee/qcomtee/core.c
66
object->info.qcomtee_async_ctx = qcomtee->ctx;
drivers/tee/qcomtee/core.c
668
errno = object->ops->dispatch(oic, object, op, oic->u);
drivers/tee/qcomtee/core.c
67
teedev_ctx_get(object->info.qcomtee_async_ctx);
drivers/tee/qcomtee/core.c
69
return object;
drivers/tee/qcomtee/core.c
72
static void qcomtee_qtee_object_free(struct qcomtee_object *object)
drivers/tee/qcomtee/core.c
742
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/core.c
75
teedev_ctx_put(object->info.qcomtee_async_ctx);
drivers/tee/qcomtee/core.c
758
ret = qcomtee_prepare_msg(oic, object, op, u);
drivers/tee/qcomtee/core.c
77
kfree(object->name);
drivers/tee/qcomtee/core.c
78
kfree(object);
drivers/tee/qcomtee/core.c
780
qto = oic->object;
drivers/tee/qcomtee/core.c
792
oic->object = NULL_QCOMTEE_OBJECT;
drivers/tee/qcomtee/core.c
83
struct qcomtee_object *object;
drivers/tee/qcomtee/core.c
849
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/core.c
857
if (typeof_qcomtee_object(object) != QCOMTEE_OBJECT_TYPE_TEE &&
drivers/tee/qcomtee/core.c
858
typeof_qcomtee_object(object) != QCOMTEE_OBJECT_TYPE_ROOT)
drivers/tee/qcomtee/core.c
866
return qcomtee_object_do_invoke_internal(oic, object, op, u, result);
drivers/tee/qcomtee/core.c
90
object = container_of(work, struct qcomtee_object, work);
drivers/tee/qcomtee/core.c
91
qcomtee = tee_get_drvdata(object->info.qcomtee_async_ctx->teedev);
drivers/tee/qcomtee/core.c
93
qcomtee->oic.ctx = object->info.qcomtee_async_ctx;
drivers/tee/qcomtee/core.c
95
ret = qcomtee_object_do_invoke_internal(&qcomtee->oic, object,
drivers/tee/qcomtee/mem_obj.c
103
err = qcomtee_object_user_init(&mem_object->object,
drivers/tee/qcomtee/mem_obj.c
117
*object = &no_free_ptr(mem_object)->object;
drivers/tee/qcomtee/mem_obj.c
124
struct qcomtee_object *object,
drivers/tee/qcomtee/mem_obj.c
129
mem_object = to_qcomtee_mem_object(object);
drivers/tee/qcomtee/mem_obj.c
138
qcomtee_object_put(object);
drivers/tee/qcomtee/mem_obj.c
153
int qcomtee_mem_object_map(struct qcomtee_object *object,
drivers/tee/qcomtee/mem_obj.c
157
struct qcomtee_mem_object *mem_object = to_qcomtee_mem_object(object);
drivers/tee/qcomtee/mem_obj.c
160
qcomtee_object_get(&mem_object->object);
drivers/tee/qcomtee/mem_obj.c
162
*map_object = &mem_object->object;
drivers/tee/qcomtee/mem_obj.c
36
struct qcomtee_object object;
drivers/tee/qcomtee/mem_obj.c
44
container_of((o), struct qcomtee_mem_object, object)
drivers/tee/qcomtee/mem_obj.c
49
int is_qcomtee_memobj_object(struct qcomtee_object *object)
drivers/tee/qcomtee/mem_obj.c
51
return object != NULL_QCOMTEE_OBJECT &&
drivers/tee/qcomtee/mem_obj.c
52
typeof_qcomtee_object(object) == QCOMTEE_OBJECT_TYPE_CB &&
drivers/tee/qcomtee/mem_obj.c
53
object->ops == &qcomtee_mem_object_ops;
drivers/tee/qcomtee/mem_obj.c
57
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/mem_obj.c
63
static void qcomtee_mem_object_release(struct qcomtee_object *object)
drivers/tee/qcomtee/mem_obj.c
65
struct qcomtee_mem_object *mem_object = to_qcomtee_mem_object(object);
drivers/tee/qcomtee/mem_obj.c
87
int qcomtee_memobj_param_to_object(struct qcomtee_object **object,
drivers/tee/qcomtee/primordial_obj.c
100
if (object && err)
drivers/tee/qcomtee/primordial_obj.c
101
qcomtee_object_put(object);
drivers/tee/qcomtee/primordial_obj.c
97
struct qcomtee_object *object = oic->data;
drivers/tee/qcomtee/qcomtee.h
112
int is_qcomtee_user_object(struct qcomtee_object *object);
drivers/tee/qcomtee/qcomtee.h
113
void qcomtee_user_object_set_notify(struct qcomtee_object *object, bool notify);
drivers/tee/qcomtee/qcomtee.h
115
int qcomtee_user_param_to_object(struct qcomtee_object **object,
drivers/tee/qcomtee/qcomtee.h
119
struct qcomtee_object *object,
drivers/tee/qcomtee/qcomtee.h
150
int is_qcomtee_memobj_object(struct qcomtee_object *object);
drivers/tee/qcomtee/qcomtee.h
162
int qcomtee_memobj_param_to_object(struct qcomtee_object **object,
drivers/tee/qcomtee/qcomtee.h
168
struct qcomtee_object *object,
drivers/tee/qcomtee/qcomtee.h
181
int qcomtee_mem_object_map(struct qcomtee_object *object,
drivers/tee/qcomtee/qcomtee.h
66
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/qcomtee.h
95
struct qcomtee_object *object,
drivers/tee/qcomtee/qcomtee.h
97
int qcomtee_context_find_qtee_object(struct qcomtee_object **object,
drivers/tee/qcomtee/qcomtee_object.h
163
struct qcomtee_object *object;
drivers/tee/qcomtee/qcomtee_object.h
202
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/qcomtee_object.h
212
void (*release)(struct qcomtee_object *object);
drivers/tee/qcomtee/qcomtee_object.h
214
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/qcomtee_object.h
217
struct qcomtee_object *object, int err);
drivers/tee/qcomtee/qcomtee_object.h
252
typeof_qcomtee_object(struct qcomtee_object *object)
drivers/tee/qcomtee/qcomtee_object.h
254
if (object == NULL_QCOMTEE_OBJECT)
drivers/tee/qcomtee/qcomtee_object.h
256
return object->object_type;
drivers/tee/qcomtee/qcomtee_object.h
259
static inline const char *qcomtee_object_name(struct qcomtee_object *object)
drivers/tee/qcomtee/qcomtee_object.h
261
if (object == NULL_QCOMTEE_OBJECT)
drivers/tee/qcomtee/qcomtee_object.h
264
if (!object->name)
drivers/tee/qcomtee/qcomtee_object.h
266
return object->name;
drivers/tee/qcomtee/qcomtee_object.h
278
int qcomtee_object_user_init(struct qcomtee_object *object,
drivers/tee/qcomtee/qcomtee_object.h
284
int qcomtee_object_get(struct qcomtee_object *object);
drivers/tee/qcomtee/qcomtee_object.h
285
void qcomtee_object_put(struct qcomtee_object *object);
drivers/tee/qcomtee/user_obj.c
226
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/user_obj.c
229
struct qcomtee_user_object *uo = to_qcomtee_user_object(object);
drivers/tee/qcomtee/user_obj.c
317
static void qcomtee_user_object_release(struct qcomtee_object *object)
drivers/tee/qcomtee/user_obj.c
319
struct qcomtee_user_object *uo = to_qcomtee_user_object(object);
drivers/tee/qcomtee/user_obj.c
366
int qcomtee_user_param_to_object(struct qcomtee_object **object,
drivers/tee/qcomtee/user_obj.c
381
err = qcomtee_object_user_init(&user_object->object,
drivers/tee/qcomtee/user_obj.c
39
struct qcomtee_object object;
drivers/tee/qcomtee/user_obj.c
390
*object = &no_free_ptr(user_object)->object;
drivers/tee/qcomtee/user_obj.c
397
struct qcomtee_object *object,
drivers/tee/qcomtee/user_obj.c
402
uo = to_qcomtee_user_object(object);
drivers/tee/qcomtee/user_obj.c
411
qcomtee_object_put(object);
drivers/tee/qcomtee/user_obj.c
46
container_of((o), struct qcomtee_user_object, object)
drivers/tee/qcomtee/user_obj.c
51
int is_qcomtee_user_object(struct qcomtee_object *object)
drivers/tee/qcomtee/user_obj.c
53
return object != NULL_QCOMTEE_OBJECT &&
drivers/tee/qcomtee/user_obj.c
54
typeof_qcomtee_object(object) == QCOMTEE_OBJECT_TYPE_CB &&
drivers/tee/qcomtee/user_obj.c
55
object->ops == &qcomtee_user_object_ops;
drivers/tee/qcomtee/user_obj.c
59
void qcomtee_user_object_set_notify(struct qcomtee_object *object, bool notify)
drivers/tee/qcomtee/user_obj.c
61
if (is_qcomtee_user_object(object))
drivers/tee/qcomtee/user_obj.c
62
to_qcomtee_user_object(object)->notify = notify;
drivers/vfio/pci/mlx5/cmd.c
1110
struct mlx5_eqe_obj_change *object;
drivers/vfio/pci/mlx5/cmd.c
1131
object = &eqe->data.obj_change;
drivers/vfio/pci/mlx5/cmd.c
1132
obj_id = be32_to_cpu(object->obj_id);
drivers/xen/xen-acpi-processor.c
331
union acpi_object object = { 0 };
drivers/xen/xen-acpi-processor.c
332
struct acpi_buffer buffer = { sizeof(union acpi_object), &object };
drivers/xen/xen-acpi-processor.c
344
acpi_id = object.processor.proc_id;
drivers/xen/xen-acpi-processor.c
345
pblk = object.processor.pblk_address;
fs/cachefiles/cache.c
290
struct cachefiles_object *object;
fs/cachefiles/cache.c
298
object = list_first_entry(&cache->object_list,
fs/cachefiles/cache.c
300
cachefiles_see_object(object, cachefiles_obj_see_withdrawal);
fs/cachefiles/cache.c
301
list_del_init(&object->cache_link);
fs/cachefiles/cache.c
302
fscache_withdraw_cookie(object->cookie);
fs/cachefiles/interface.c
100
object->cookie = NULL;
fs/cachefiles/interface.c
101
kmem_cache_free(cachefiles_object_jar, object);
fs/cachefiles/interface.c
115
static int cachefiles_adjust_size(struct cachefiles_object *object)
fs/cachefiles/interface.c
118
struct file *file = object->file;
fs/cachefiles/interface.c
123
ni_size = object->cookie->object_size;
fs/cachefiles/interface.c
127
object->debug_id, (unsigned long long) ni_size);
fs/cachefiles/interface.c
167
cachefiles_io_error_obj(object, "Size set failed");
fs/cachefiles/interface.c
180
struct cachefiles_object *object;
fs/cachefiles/interface.c
185
object = cachefiles_alloc_object(cookie);
fs/cachefiles/interface.c
186
if (!object)
fs/cachefiles/interface.c
189
_enter("{OBJ%x}", object->debug_id);
fs/cachefiles/interface.c
191
if (!cachefiles_cook_key(object))
fs/cachefiles/interface.c
194
cookie->cache_priv = object;
fs/cachefiles/interface.c
198
success = cachefiles_look_up_object(object);
fs/cachefiles/interface.c
202
cachefiles_see_object(object, cachefiles_obj_see_lookup_cookie);
fs/cachefiles/interface.c
205
list_add(&object->cache_link, &cache->object_list);
fs/cachefiles/interface.c
207
cachefiles_adjust_size(object);
fs/cachefiles/interface.c
215
cachefiles_see_object(object, cachefiles_obj_see_lookup_failed);
fs/cachefiles/interface.c
217
_debug("failed c=%08x o=%08x", cookie->debug_id, object->debug_id);
fs/cachefiles/interface.c
224
cachefiles_put_object(object, cachefiles_obj_put_alloc_fail);
fs/cachefiles/interface.c
233
static bool cachefiles_shorten_object(struct cachefiles_object *object,
fs/cachefiles/interface.c
236
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/interface.c
244
trace_cachefiles_trunc(object, inode, i_size, dio_size,
fs/cachefiles/interface.c
250
trace_cachefiles_io_error(object, file_inode(file), ret,
fs/cachefiles/interface.c
252
cachefiles_io_error_obj(object, "Trunc-to-size failed %d", ret);
fs/cachefiles/interface.c
253
cachefiles_remove_object_xattr(cache, object, file->f_path.dentry);
fs/cachefiles/interface.c
258
trace_cachefiles_trunc(object, inode, dio_size, new_size,
fs/cachefiles/interface.c
265
trace_cachefiles_io_error(object, file_inode(file), ret,
fs/cachefiles/interface.c
267
cachefiles_io_error_obj(object, "Trunc-to-dio-size failed %d", ret);
fs/cachefiles/interface.c
268
cachefiles_remove_object_xattr(cache, object, file->f_path.dentry);
fs/cachefiles/interface.c
27
struct cachefiles_object *object;
fs/cachefiles/interface.c
282
struct cachefiles_object *object = cachefiles_cres_object(cres);
fs/cachefiles/interface.c
283
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/interface.c
284
struct fscache_cookie *cookie = object->cookie;
fs/cachefiles/interface.c
293
cachefiles_shorten_object(object, file, new_size);
fs/cachefiles/interface.c
295
object->cookie->object_size = new_size;
fs/cachefiles/interface.c
309
static void cachefiles_commit_object(struct cachefiles_object *object,
fs/cachefiles/interface.c
31
object = kmem_cache_zalloc(cachefiles_object_jar, GFP_KERNEL);
fs/cachefiles/interface.c
314
if (test_and_clear_bit(FSCACHE_COOKIE_LOCAL_WRITE, &object->cookie->flags))
fs/cachefiles/interface.c
316
if (test_and_clear_bit(FSCACHE_COOKIE_NEEDS_UPDATE, &object->cookie->flags))
fs/cachefiles/interface.c
319
cachefiles_set_object_xattr(object);
fs/cachefiles/interface.c
32
if (!object)
fs/cachefiles/interface.c
321
if (test_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags))
fs/cachefiles/interface.c
322
cachefiles_commit_tmpfile(cache, object);
fs/cachefiles/interface.c
328
static void cachefiles_clean_up_object(struct cachefiles_object *object,
fs/cachefiles/interface.c
333
if (test_bit(FSCACHE_COOKIE_RETIRED, &object->cookie->flags)) {
fs/cachefiles/interface.c
334
if (!test_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags)) {
fs/cachefiles/interface.c
335
cachefiles_see_object(object, cachefiles_obj_see_clean_delete);
fs/cachefiles/interface.c
336
_debug("- inval object OBJ%x", object->debug_id);
fs/cachefiles/interface.c
337
cachefiles_delete_object(object, FSCACHE_OBJECT_WAS_RETIRED);
fs/cachefiles/interface.c
339
cachefiles_see_object(object, cachefiles_obj_see_clean_drop_tmp);
fs/cachefiles/interface.c
340
_debug("- inval object OBJ%x tmpfile", object->debug_id);
fs/cachefiles/interface.c
343
cachefiles_see_object(object, cachefiles_obj_see_clean_commit);
fs/cachefiles/interface.c
344
cachefiles_commit_object(object, cache);
fs/cachefiles/interface.c
347
cachefiles_unmark_inode_in_use(object, object->file);
fs/cachefiles/interface.c
349
spin_lock(&object->lock);
fs/cachefiles/interface.c
35
if (cachefiles_ondemand_init_obj_info(object, volume)) {
fs/cachefiles/interface.c
350
file = object->file;
fs/cachefiles/interface.c
351
object->file = NULL;
fs/cachefiles/interface.c
352
spin_unlock(&object->lock);
fs/cachefiles/interface.c
36
kmem_cache_free(cachefiles_object_jar, object);
fs/cachefiles/interface.c
363
struct cachefiles_object *object = cookie->cache_priv;
fs/cachefiles/interface.c
364
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/interface.c
367
_enter("o=%x", object->debug_id);
fs/cachefiles/interface.c
368
cachefiles_see_object(object, cachefiles_obj_see_withdraw_cookie);
fs/cachefiles/interface.c
370
if (!list_empty(&object->cache_link)) {
fs/cachefiles/interface.c
372
cachefiles_see_object(object, cachefiles_obj_see_withdrawal);
fs/cachefiles/interface.c
373
list_del_init(&object->cache_link);
fs/cachefiles/interface.c
377
cachefiles_ondemand_clean_object(object);
fs/cachefiles/interface.c
379
if (object->file) {
fs/cachefiles/interface.c
381
cachefiles_clean_up_object(object, cache);
fs/cachefiles/interface.c
386
cachefiles_put_object(object, cachefiles_obj_put_detach);
fs/cachefiles/interface.c
394
struct cachefiles_object *object = cookie->cache_priv;
fs/cachefiles/interface.c
398
_enter("o=%x,[%llu]", object->debug_id, object->cookie->object_size);
fs/cachefiles/interface.c
40
refcount_set(&object->ref, 1);
fs/cachefiles/interface.c
400
old_tmpfile = test_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags);
fs/cachefiles/interface.c
402
if (!object->file) {
fs/cachefiles/interface.c
408
new_file = cachefiles_create_tmpfile(object);
fs/cachefiles/interface.c
414
spin_lock(&object->lock);
fs/cachefiles/interface.c
416
old_file = object->file;
fs/cachefiles/interface.c
417
object->file = new_file;
fs/cachefiles/interface.c
418
object->content_info = CACHEFILES_CONTENT_NO_DATA;
fs/cachefiles/interface.c
419
set_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags);
fs/cachefiles/interface.c
42
spin_lock_init(&object->lock);
fs/cachefiles/interface.c
420
set_bit(FSCACHE_COOKIE_NEEDS_UPDATE, &object->cookie->flags);
fs/cachefiles/interface.c
422
spin_unlock(&object->lock);
fs/cachefiles/interface.c
43
INIT_LIST_HEAD(&object->cache_link);
fs/cachefiles/interface.c
430
struct cachefiles_volume *volume = object->volume;
fs/cachefiles/interface.c
436
cachefiles_bury_object(volume->cache, object,
fs/cachefiles/interface.c
44
object->volume = volume;
fs/cachefiles/interface.c
45
object->debug_id = atomic_inc_return(&cachefiles_object_debug_id);
fs/cachefiles/interface.c
46
object->cookie = fscache_get_cookie(cookie, fscache_cookie_get_attach_object);
fs/cachefiles/interface.c
49
trace_cachefiles_ref(object->debug_id, cookie->debug_id, 1,
fs/cachefiles/interface.c
51
return object;
fs/cachefiles/interface.c
57
void cachefiles_see_object(struct cachefiles_object *object,
fs/cachefiles/interface.c
60
trace_cachefiles_ref(object->debug_id, object->cookie->debug_id,
fs/cachefiles/interface.c
61
refcount_read(&object->ref), why);
fs/cachefiles/interface.c
67
struct cachefiles_object *cachefiles_grab_object(struct cachefiles_object *object,
fs/cachefiles/interface.c
72
__refcount_inc(&object->ref, &r);
fs/cachefiles/interface.c
73
trace_cachefiles_ref(object->debug_id, object->cookie->debug_id, r, why);
fs/cachefiles/interface.c
74
return object;
fs/cachefiles/interface.c
80
void cachefiles_put_object(struct cachefiles_object *object,
fs/cachefiles/interface.c
83
unsigned int object_debug_id = object->debug_id;
fs/cachefiles/interface.c
84
unsigned int cookie_debug_id = object->cookie->debug_id;
fs/cachefiles/interface.c
89
done = __refcount_dec_and_test(&object->ref, &r);
fs/cachefiles/interface.c
94
ASSERTCMP(object->file, ==, NULL);
fs/cachefiles/interface.c
96
kfree(object->d_name);
fs/cachefiles/interface.c
97
cachefiles_ondemand_deinit_obj_info(object);
fs/cachefiles/interface.c
98
cache = object->volume->cache->cache;
fs/cachefiles/interface.c
99
fscache_put_cookie(object->cookie, fscache_cookie_put_object);
fs/cachefiles/internal.h
142
struct cachefiles_object *object;
fs/cachefiles/internal.h
240
extern void cachefiles_see_object(struct cachefiles_object *object,
fs/cachefiles/internal.h
242
extern struct cachefiles_object *cachefiles_grab_object(struct cachefiles_object *object,
fs/cachefiles/internal.h
244
extern void cachefiles_put_object(struct cachefiles_object *object,
fs/cachefiles/internal.h
252
extern int __cachefiles_prepare_write(struct cachefiles_object *object,
fs/cachefiles/internal.h
256
extern int __cachefiles_write(struct cachefiles_object *object,
fs/cachefiles/internal.h
266
extern bool cachefiles_cook_key(struct cachefiles_object *object);
fs/cachefiles/internal.h
276
extern void cachefiles_unmark_inode_in_use(struct cachefiles_object *object,
fs/cachefiles/internal.h
279
struct cachefiles_object *object,
fs/cachefiles/internal.h
283
extern int cachefiles_delete_object(struct cachefiles_object *object,
fs/cachefiles/internal.h
285
extern bool cachefiles_look_up_object(struct cachefiles_object *object);
fs/cachefiles/internal.h
297
extern struct file *cachefiles_create_tmpfile(struct cachefiles_object *object);
fs/cachefiles/internal.h
299
struct cachefiles_object *object);
fs/cachefiles/internal.h
314
extern int cachefiles_ondemand_init_object(struct cachefiles_object *object);
fs/cachefiles/internal.h
315
extern void cachefiles_ondemand_clean_object(struct cachefiles_object *object);
fs/cachefiles/internal.h
317
extern int cachefiles_ondemand_read(struct cachefiles_object *object,
fs/cachefiles/internal.h
326
cachefiles_ondemand_object_is_##_state(const struct cachefiles_object *object) \
fs/cachefiles/internal.h
328
return object->ondemand->state == CACHEFILES_ONDEMAND_OBJSTATE_##_STATE; \
fs/cachefiles/internal.h
332
cachefiles_ondemand_set_object_##_state(struct cachefiles_object *object) \
fs/cachefiles/internal.h
334
object->ondemand->state = CACHEFILES_ONDEMAND_OBJSTATE_##_STATE; \
fs/cachefiles/internal.h
344
return cachefiles_ondemand_object_is_reopening(req->object) &&
fs/cachefiles/internal.h
355
static inline int cachefiles_ondemand_init_object(struct cachefiles_object *object)
fs/cachefiles/internal.h
360
static inline void cachefiles_ondemand_clean_object(struct cachefiles_object *object)
fs/cachefiles/internal.h
364
static inline int cachefiles_ondemand_read(struct cachefiles_object *object,
fs/cachefiles/internal.h
415
extern int cachefiles_set_object_xattr(struct cachefiles_object *object);
fs/cachefiles/internal.h
416
extern int cachefiles_check_auxdata(struct cachefiles_object *object,
fs/cachefiles/internal.h
419
struct cachefiles_object *object,
fs/cachefiles/internal.h
437
#define cachefiles_io_error_obj(object, FMT, ...) \
fs/cachefiles/internal.h
441
___cache = (object)->volume->cache; \
fs/cachefiles/internal.h
443
(object)->debug_id); \
fs/cachefiles/internal.h
58
struct cachefiles_object *object;
fs/cachefiles/io.c
145
ki->object = object;
fs/cachefiles/io.c
155
cachefiles_grab_object(object, cachefiles_obj_get_ioreq);
fs/cachefiles/io.c
157
trace_cachefiles_read(object, file_inode(file), ki->iocb.ki_pos, len - skipped);
fs/cachefiles/io.c
203
struct cachefiles_object *object;
fs/cachefiles/io.c
213
object = cachefiles_cres_object(cres);
fs/cachefiles/io.c
215
granularity = max_t(size_t, object->volume->cache->bsize, granularity);
fs/cachefiles/io.c
259
struct cachefiles_object *object = ki->object;
fs/cachefiles/io.c
268
trace_cachefiles_io_error(object, inode, ret,
fs/cachefiles/io.c
27
struct cachefiles_object *object;
fs/cachefiles/io.c
271
atomic_long_sub(ki->b_writing, &object->volume->cache->b_writing);
fs/cachefiles/io.c
272
set_bit(FSCACHE_COOKIE_HAVE_DATA, &object->cookie->flags);
fs/cachefiles/io.c
281
int __cachefiles_write(struct cachefiles_object *object,
fs/cachefiles/io.c
295
cache = object->volume->cache;
fs/cachefiles/io.c
313
ki->object = object;
fs/cachefiles/io.c
326
cachefiles_grab_object(object, cachefiles_obj_get_ioreq);
fs/cachefiles/io.c
328
trace_cachefiles_write(object, file_inode(file), ki->iocb.ki_pos, len);
fs/cachefiles/io.c
38
cachefiles_put_object(ki->object, cachefiles_obj_put_ioreq);
fs/cachefiles/io.c
384
struct cachefiles_object *object = NULL;
fs/cachefiles/io.c
421
object = cachefiles_cres_object(cres);
fs/cachefiles/io.c
422
cache = object->volume->cache;
fs/cachefiles/io.c
433
trace_cachefiles_io_error(object, file_inode(file), off,
fs/cachefiles/io.c
456
trace_cachefiles_io_error(object, file_inode(file), to,
fs/cachefiles/io.c
478
rc = cachefiles_ondemand_read(object, start, len);
fs/cachefiles/io.c
488
trace_cachefiles_prep_read(object, start, len, *_flags, ret, why, ino, netfs_ino);
fs/cachefiles/io.c
519
int __cachefiles_prepare_write(struct cachefiles_object *object,
fs/cachefiles/io.c
524
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/io.c
55
trace_cachefiles_io_error(ki->object, inode, ret,
fs/cachefiles/io.c
555
trace_cachefiles_io_error(object, file_inode(file), pos,
fs/cachefiles/io.c
574
trace_cachefiles_io_error(object, file_inode(file), pos,
fs/cachefiles/io.c
588
trace_cachefiles_io_error(object, file_inode(file), ret,
fs/cachefiles/io.c
590
cachefiles_io_error_obj(object,
fs/cachefiles/io.c
60
if (ki->object->cookie->inval_counter == ki->inval_counter)
fs/cachefiles/io.c
606
struct cachefiles_object *object = cachefiles_cres_object(cres);
fs/cachefiles/io.c
607
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/io.c
619
ret = __cachefiles_prepare_write(object, cachefiles_cres_file(cres),
fs/cachefiles/io.c
649
struct cachefiles_object *object = cachefiles_cres_object(cres);
fs/cachefiles/io.c
650
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/io.c
700
ret = __cachefiles_prepare_write(object, cachefiles_cres_file(cres),
fs/cachefiles/io.c
744
struct cachefiles_object *object = cachefiles_cres_object(cres);
fs/cachefiles/io.c
748
if (object->file) {
fs/cachefiles/io.c
749
spin_lock(&object->lock);
fs/cachefiles/io.c
750
if (!cres->cache_priv2 && object->file)
fs/cachefiles/io.c
751
cres->cache_priv2 = get_file(object->file);
fs/cachefiles/io.c
752
spin_unlock(&object->lock);
fs/cachefiles/io.c
82
struct cachefiles_object *object;
fs/cachefiles/io.c
93
object = cachefiles_cres_object(cres);
fs/cachefiles/key.c
134
object->d_name = name;
fs/cachefiles/key.c
135
_leave(" = %s", object->d_name);
fs/cachefiles/key.c
37
bool cachefiles_cook_key(struct cachefiles_object *object)
fs/cachefiles/key.c
39
const u8 *key = fscache_get_key(object->cookie), *kend;
fs/cachefiles/key.c
41
unsigned int acc, i, n, nle, nbe, keylen = object->cookie->key_len;
fs/cachefiles/namei.c
17
static bool __cachefiles_mark_inode_in_use(struct cachefiles_object *object,
fs/cachefiles/namei.c
230
struct cachefiles_object *object,
fs/cachefiles/namei.c
24
trace_cachefiles_mark_active(object, inode);
fs/cachefiles/namei.c
240
trace_cachefiles_unlink(object, d_inode(dentry)->i_ino, why);
fs/cachefiles/namei.c
254
trace_cachefiles_vfs_error(object, d_backing_inode(dir), ret,
fs/cachefiles/namei.c
268
struct cachefiles_object *object,
fs/cachefiles/namei.c
27
trace_cachefiles_mark_failed(object, inode);
fs/cachefiles/namei.c
288
ret = cachefiles_unlink(cache, object, dir, rep, why);
fs/cachefiles/namei.c
33
static bool cachefiles_mark_inode_in_use(struct cachefiles_object *object,
fs/cachefiles/namei.c
340
trace_cachefiles_vfs_error(object, d_inode(cache->graveyard),
fs/cachefiles/namei.c
39
can_use = __cachefiles_mark_inode_in_use(object, inode);
fs/cachefiles/namei.c
392
trace_cachefiles_rename(object, d_inode(rep)->i_ino, why);
fs/cachefiles/namei.c
397
trace_cachefiles_vfs_error(object, d_inode(dir), ret,
fs/cachefiles/namei.c
404
__cachefiles_unmark_inode_in_use(object, d_inode(rep));
fs/cachefiles/namei.c
414
int cachefiles_delete_object(struct cachefiles_object *object,
fs/cachefiles/namei.c
417
struct cachefiles_volume *volume = object->volume;
fs/cachefiles/namei.c
418
struct dentry *dentry = object->file->f_path.dentry;
fs/cachefiles/namei.c
419
struct dentry *fan = volume->fanout[(u8)object->cookie->key_hash];
fs/cachefiles/namei.c
422
_enter(",OBJ%x{%pD}", object->debug_id, object->file);
fs/cachefiles/namei.c
428
ret = cachefiles_unlink(volume->cache, object, fan, dentry, why);
fs/cachefiles/namei.c
437
struct file *cachefiles_create_tmpfile(struct cachefiles_object *object)
fs/cachefiles/namei.c
439
struct cachefiles_volume *volume = object->volume;
fs/cachefiles/namei.c
442
struct dentry *fan = volume->fanout[(u8)object->cookie->key_hash];
fs/cachefiles/namei.c
460
trace_cachefiles_vfs_error(object, d_inode(fan), ret,
fs/cachefiles/namei.c
463
cachefiles_io_error_obj(object, "Failed to create tmpfile");
fs/cachefiles/namei.c
467
trace_cachefiles_tmpfile(object, file_inode(file));
fs/cachefiles/namei.c
47
static void __cachefiles_unmark_inode_in_use(struct cachefiles_object *object,
fs/cachefiles/namei.c
470
if (!cachefiles_mark_inode_in_use(object, file_inode(file)))
fs/cachefiles/namei.c
473
ret = cachefiles_ondemand_init_object(object);
fs/cachefiles/namei.c
477
ni_size = object->cookie->object_size;
fs/cachefiles/namei.c
481
trace_cachefiles_trunc(object, file_inode(file), 0, ni_size,
fs/cachefiles/namei.c
488
object, file_inode(file), ret,
fs/cachefiles/namei.c
506
cachefiles_do_unmark_inode_in_use(object, file_inode(file));
fs/cachefiles/namei.c
51
trace_cachefiles_mark_inactive(object, inode);
fs/cachefiles/namei.c
516
static bool cachefiles_create_file(struct cachefiles_object *object)
fs/cachefiles/namei.c
521
ret = cachefiles_has_space(object->volume->cache, 1, 0,
fs/cachefiles/namei.c
526
file = cachefiles_create_tmpfile(object);
fs/cachefiles/namei.c
530
set_bit(FSCACHE_COOKIE_NEEDS_UPDATE, &object->cookie->flags);
fs/cachefiles/namei.c
531
set_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags);
fs/cachefiles/namei.c
533
object->file = file;
fs/cachefiles/namei.c
54
static void cachefiles_do_unmark_inode_in_use(struct cachefiles_object *object,
fs/cachefiles/namei.c
541
static bool cachefiles_open_file(struct cachefiles_object *object,
fs/cachefiles/namei.c
544
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/namei.c
551
if (!cachefiles_mark_inode_in_use(object, d_inode(dentry))) {
fs/cachefiles/namei.c
565
trace_cachefiles_vfs_error(object, d_backing_inode(dentry),
fs/cachefiles/namei.c
578
ret = cachefiles_ondemand_init_object(object);
fs/cachefiles/namei.c
58
__cachefiles_unmark_inode_in_use(object, inode);
fs/cachefiles/namei.c
582
ret = cachefiles_check_auxdata(object, file);
fs/cachefiles/namei.c
586
clear_bit(FSCACHE_COOKIE_NO_DATA_TO_READ, &object->cookie->flags);
fs/cachefiles/namei.c
588
object->file = file;
fs/cachefiles/namei.c
598
fscache_cookie_lookup_negative(object->cookie);
fs/cachefiles/namei.c
599
cachefiles_unmark_inode_in_use(object, file);
fs/cachefiles/namei.c
602
return cachefiles_create_file(object);
fs/cachefiles/namei.c
608
cachefiles_do_unmark_inode_in_use(object, d_inode(dentry));
fs/cachefiles/namei.c
616
bool cachefiles_look_up_object(struct cachefiles_object *object)
fs/cachefiles/namei.c
618
struct cachefiles_volume *volume = object->volume;
fs/cachefiles/namei.c
619
struct dentry *dentry, *fan = volume->fanout[(u8)object->cookie->key_hash];
fs/cachefiles/namei.c
622
_enter("OBJ%x,%s,", object->debug_id, object->d_name);
fs/cachefiles/namei.c
628
&QSTR(object->d_name), fan);
fs/cachefiles/namei.c
631
trace_cachefiles_lookup(object, fan, dentry);
fs/cachefiles/namei.c
636
cachefiles_io_error_obj(object, "Lookup failed");
fs/cachefiles/namei.c
646
ret = cachefiles_bury_object(volume->cache, object,
fs/cachefiles/namei.c
655
ret = cachefiles_open_file(object, dentry);
fs/cachefiles/namei.c
66
void cachefiles_unmark_inode_in_use(struct cachefiles_object *object,
fs/cachefiles/namei.c
660
_leave(" = t [%lu]", file_inode(object->file)->i_ino);
fs/cachefiles/namei.c
664
fscache_cookie_lookup_negative(object->cookie);
fs/cachefiles/namei.c
665
return cachefiles_create_file(object);
fs/cachefiles/namei.c
672
struct cachefiles_object *object)
fs/cachefiles/namei.c
674
struct cachefiles_volume *volume = object->volume;
fs/cachefiles/namei.c
675
struct dentry *dentry, *fan = volume->fanout[(u8)object->cookie->key_hash];
fs/cachefiles/namei.c
679
_enter(",%pD", object->file);
fs/cachefiles/namei.c
683
dentry = start_creating(&nop_mnt_idmap, fan, &QSTR(object->d_name));
fs/cachefiles/namei.c
687
trace_cachefiles_vfs_error(object, d_inode(fan), PTR_ERR(dentry),
fs/cachefiles/namei.c
69
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/namei.c
698
ret = cachefiles_unlink(volume->cache, object, fan, dentry,
fs/cachefiles/namei.c
708
&QSTR(object->d_name));
fs/cachefiles/namei.c
712
trace_cachefiles_vfs_error(object, d_inode(fan), PTR_ERR(dentry),
fs/cachefiles/namei.c
72
cachefiles_do_unmark_inode_in_use(object, inode);
fs/cachefiles/namei.c
721
ret = vfs_link(object->file->f_path.dentry, &nop_mnt_idmap,
fs/cachefiles/namei.c
724
trace_cachefiles_vfs_error(object, d_inode(fan), ret,
fs/cachefiles/namei.c
728
trace_cachefiles_link(object, file_inode(object->file));
fs/cachefiles/namei.c
729
spin_lock(&object->lock);
fs/cachefiles/namei.c
731
clear_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags);
fs/cachefiles/namei.c
732
spin_unlock(&object->lock);
fs/cachefiles/namei.c
74
if (!test_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags)) {
fs/cachefiles/ondemand.c
101
spin_lock(&object->lock);
fs/cachefiles/ondemand.c
102
file = object->file;
fs/cachefiles/ondemand.c
104
spin_unlock(&object->lock);
fs/cachefiles/ondemand.c
108
spin_unlock(&object->lock);
fs/cachefiles/ondemand.c
119
struct cachefiles_object *object = filp->private_data;
fs/cachefiles/ondemand.c
120
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/ondemand.c
133
req->object != object) {
fs/cachefiles/ondemand.c
140
trace_cachefiles_ondemand_cread(object, id);
fs/cachefiles/ondemand.c
195
!req->object->ondemand->ondemand_id) {
fs/cachefiles/ondemand.c
20
struct cachefiles_object *object = file->private_data;
fs/cachefiles/ondemand.c
202
info = req->object->ondemand;
fs/cachefiles/ondemand.c
248
cookie = req->object->cookie;
fs/cachefiles/ondemand.c
254
trace_cachefiles_ondemand_copen(req->object, id, size);
fs/cachefiles/ondemand.c
256
cachefiles_ondemand_set_object_open(req->object);
fs/cachefiles/ondemand.c
264
cachefiles_ondemand_set_object_close(req->object);
fs/cachefiles/ondemand.c
27
if (!object)
fs/cachefiles/ondemand.c
296
struct cachefiles_object *object;
fs/cachefiles/ondemand.c
30
info = object->ondemand;
fs/cachefiles/ondemand.c
302
object = cachefiles_grab_object(req->object,
fs/cachefiles/ondemand.c
304
cache = object->volume->cache;
fs/cachefiles/ondemand.c
31
cache = object->volume->cache;
fs/cachefiles/ondemand.c
319
&cachefiles_ondemand_fd_fops, object,
fs/cachefiles/ondemand.c
326
spin_lock(&object->ondemand->lock);
fs/cachefiles/ondemand.c
327
if (object->ondemand->ondemand_id > 0) {
fs/cachefiles/ondemand.c
328
spin_unlock(&object->ondemand->lock);
fs/cachefiles/ondemand.c
337
object->ondemand->ondemand_id = object_id;
fs/cachefiles/ondemand.c
338
spin_unlock(&object->ondemand->lock);
fs/cachefiles/ondemand.c
341
trace_cachefiles_ondemand_open(object, &req->msg, load);
fs/cachefiles/ondemand.c
353
spin_lock(&object->ondemand->lock);
fs/cachefiles/ondemand.c
355
if (object->ondemand->ondemand_id <= 0)
fs/cachefiles/ondemand.c
356
cachefiles_ondemand_set_object_close(object);
fs/cachefiles/ondemand.c
357
spin_unlock(&object->ondemand->lock);
fs/cachefiles/ondemand.c
358
cachefiles_put_object(object, cachefiles_obj_put_ondemand_fd);
fs/cachefiles/ondemand.c
367
cachefiles_ondemand_init_object(info->object);
fs/cachefiles/ondemand.c
379
struct cachefiles_object *object;
fs/cachefiles/ondemand.c
38
cachefiles_ondemand_set_object_close(object);
fs/cachefiles/ondemand.c
385
object = req->object;
fs/cachefiles/ondemand.c
386
info = object->ondemand;
fs/cachefiles/ondemand.c
387
if (cachefiles_ondemand_object_is_close(object)) {
fs/cachefiles/ondemand.c
388
cachefiles_ondemand_set_object_reopening(object);
fs/cachefiles/ondemand.c
392
if (cachefiles_ondemand_object_is_reopening(object))
fs/cachefiles/ondemand.c
450
cachefiles_grab_object(req->object, cachefiles_obj_get_read_req);
fs/cachefiles/ondemand.c
460
msg->object_id = req->object->ondemand->ondemand_id;
fs/cachefiles/ondemand.c
474
cachefiles_put_object(req->object, cachefiles_obj_put_read_req);
fs/cachefiles/ondemand.c
484
static int cachefiles_ondemand_send_req(struct cachefiles_object *object,
fs/cachefiles/ondemand.c
490
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/ondemand.c
510
req->object = object;
fs/cachefiles/ondemand.c
52
trace_cachefiles_ondemand_fd_release(object, object_id);
fs/cachefiles/ondemand.c
53
cachefiles_put_object(object, cachefiles_obj_put_ondemand_fd);
fs/cachefiles/ondemand.c
538
cachefiles_ondemand_object_is_dropping(object)) {
fs/cachefiles/ondemand.c
548
!cachefiles_ondemand_object_is_open(object)) {
fs/cachefiles/ondemand.c
549
WARN_ON_ONCE(object->ondemand->ondemand_id == 0);
fs/cachefiles/ondemand.c
602
!cachefiles_ondemand_object_is_dropping(object))
fs/cachefiles/ondemand.c
603
cachefiles_ondemand_set_object_close(object);
fs/cachefiles/ondemand.c
61
struct cachefiles_object *object = kiocb->ki_filp->private_data;
fs/cachefiles/ondemand.c
611
struct cachefiles_object *object = req->object;
fs/cachefiles/ondemand.c
612
struct fscache_cookie *cookie = object->cookie;
fs/cachefiles/ondemand.c
613
struct fscache_volume *volume = object->volume->vcookie;
fs/cachefiles/ondemand.c
62
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/ondemand.c
629
if (!(object->cookie->advice & FSCACHE_ADV_WANT_CACHE_SIZE)) {
fs/cachefiles/ondemand.c
645
struct cachefiles_object *object = req->object;
fs/cachefiles/ondemand.c
647
if (!cachefiles_ondemand_object_is_open(object))
fs/cachefiles/ondemand.c
650
trace_cachefiles_ondemand_close(object, &req->msg);
fs/cachefiles/ondemand.c
662
struct cachefiles_object *object = req->object;
fs/cachefiles/ondemand.c
668
trace_cachefiles_ondemand_read(object, &req->msg, load);
fs/cachefiles/ondemand.c
672
int cachefiles_ondemand_init_object(struct cachefiles_object *object)
fs/cachefiles/ondemand.c
674
struct fscache_cookie *cookie = object->cookie;
fs/cachefiles/ondemand.c
675
struct fscache_volume *volume = object->volume->vcookie;
fs/cachefiles/ondemand.c
678
if (!object->ondemand)
fs/cachefiles/ondemand.c
687
if (cachefiles_ondemand_object_is_open(object))
fs/cachefiles/ondemand.c
69
spin_lock(&object->lock);
fs/cachefiles/ondemand.c
695
return cachefiles_ondemand_send_req(object, CACHEFILES_OP_OPEN,
fs/cachefiles/ondemand.c
699
void cachefiles_ondemand_clean_object(struct cachefiles_object *object)
fs/cachefiles/ondemand.c
70
file = object->file;
fs/cachefiles/ondemand.c
705
if (!object->ondemand)
fs/cachefiles/ondemand.c
708
cachefiles_ondemand_send_req(object, CACHEFILES_OP_CLOSE, 0,
fs/cachefiles/ondemand.c
711
if (!object->ondemand->ondemand_id)
fs/cachefiles/ondemand.c
715
cache = object->volume->cache;
fs/cachefiles/ondemand.c
717
cachefiles_ondemand_set_object_dropping(object);
fs/cachefiles/ondemand.c
719
if (req->object == object) {
fs/cachefiles/ondemand.c
72
spin_unlock(&object->lock);
fs/cachefiles/ondemand.c
728
cancel_work_sync(&object->ondemand->ondemand_work);
fs/cachefiles/ondemand.c
731
int cachefiles_ondemand_init_obj_info(struct cachefiles_object *object,
fs/cachefiles/ondemand.c
737
object->ondemand = kzalloc_obj(struct cachefiles_ondemand_info);
fs/cachefiles/ondemand.c
738
if (!object->ondemand)
fs/cachefiles/ondemand.c
741
object->ondemand->object = object;
fs/cachefiles/ondemand.c
742
spin_lock_init(&object->ondemand->lock);
fs/cachefiles/ondemand.c
743
INIT_WORK(&object->ondemand->ondemand_work, ondemand_object_worker);
fs/cachefiles/ondemand.c
747
void cachefiles_ondemand_deinit_obj_info(struct cachefiles_object *object)
fs/cachefiles/ondemand.c
749
kfree(object->ondemand);
fs/cachefiles/ondemand.c
750
object->ondemand = NULL;
fs/cachefiles/ondemand.c
753
int cachefiles_ondemand_read(struct cachefiles_object *object,
fs/cachefiles/ondemand.c
758
return cachefiles_ondemand_send_req(object, CACHEFILES_OP_READ,
fs/cachefiles/ondemand.c
76
spin_unlock(&object->lock);
fs/cachefiles/ondemand.c
79
ret = __cachefiles_prepare_write(object, file, &pos, &aligned_len, len, true);
fs/cachefiles/ondemand.c
84
trace_cachefiles_ondemand_fd_write(object, file_inode(file), pos, len);
fs/cachefiles/ondemand.c
85
ret = __cachefiles_write(object, file, pos, iter, NULL, NULL);
fs/cachefiles/ondemand.c
97
struct cachefiles_object *object = filp->private_data;
fs/cachefiles/xattr.c
102
int cachefiles_check_auxdata(struct cachefiles_object *object, struct file *file)
fs/cachefiles/xattr.c
106
unsigned int len = object->cookie->aux_len, tlen;
fs/cachefiles/xattr.c
107
const void *p = fscache_get_aux(object->cookie);
fs/cachefiles/xattr.c
123
trace_cachefiles_vfs_error(object, file_inode(file), xlen,
fs/cachefiles/xattr.c
128
object,
fs/cachefiles/xattr.c
138
} else if (be64_to_cpu(buf->object_size) != object->cookie->object_size) {
fs/cachefiles/xattr.c
150
trace_cachefiles_coherency(object, file_inode(file)->i_ino,
fs/cachefiles/xattr.c
161
struct cachefiles_object *object,
fs/cachefiles/xattr.c
176
trace_cachefiles_vfs_error(object, d_inode(dentry), ret,
fs/cachefiles/xattr.c
197
struct cachefiles_object *object = cookie->cache_priv;
fs/cachefiles/xattr.c
198
struct cachefiles_cache *cache = object->volume->cache;
fs/cachefiles/xattr.c
200
_enter("c=%08x", object->cookie->debug_id);
fs/cachefiles/xattr.c
202
if (!test_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags)) {
fs/cachefiles/xattr.c
204
cachefiles_set_object_xattr(object);
fs/cachefiles/xattr.c
39
int cachefiles_set_object_xattr(struct cachefiles_object *object)
fs/cachefiles/xattr.c
43
struct file *file = object->file;
fs/cachefiles/xattr.c
44
unsigned int len = object->cookie->aux_len;
fs/cachefiles/xattr.c
51
_enter("%x,#%d", object->debug_id, len);
fs/cachefiles/xattr.c
57
buf->object_size = cpu_to_be64(object->cookie->object_size);
fs/cachefiles/xattr.c
60
buf->content = object->content_info;
fs/cachefiles/xattr.c
61
if (test_bit(FSCACHE_COOKIE_LOCAL_WRITE, &object->cookie->flags))
fs/cachefiles/xattr.c
64
memcpy(buf->data, fscache_get_aux(object->cookie), len);
fs/cachefiles/xattr.c
77
trace_cachefiles_vfs_error(object, file_inode(file), ret,
fs/cachefiles/xattr.c
79
trace_cachefiles_coherency(object, file_inode(file)->i_ino,
fs/cachefiles/xattr.c
85
object,
fs/cachefiles/xattr.c
88
trace_cachefiles_coherency(object, file_inode(file)->i_ino,
fs/gfs2/glock.c
814
void glock_set_object(struct gfs2_glock *gl, void *object)
fs/gfs2/glock.c
820
gl->gl_object = object;
fs/gfs2/glock.c
831
void glock_clear_object(struct gfs2_glock *gl, void *object)
fs/gfs2/glock.c
839
if (gfs2_assert_warn(glock_sbd(gl), prev_object == object))
fs/gfs2/glock.h
280
void glock_set_object(struct gfs2_glock *gl, void *object);
fs/gfs2/glock.h
281
void glock_clear_object(struct gfs2_glock *gl, void *object);
include/acpi/acpiosxf.h
219
acpi_status acpi_os_release_object(acpi_cache_t * cache, void *object);
include/acpi/acpixf.h
476
acpi_unload_parent_table(acpi_handle object))
include/acpi/acpixf.h
531
acpi_get_name(acpi_handle object, u32 name_type,
include/acpi/acpixf.h
538
acpi_attach_data(acpi_handle object,
include/acpi/acpixf.h
542
acpi_detach_data(acpi_handle object,
include/acpi/acpixf.h
545
acpi_get_data(acpi_handle object,
include/acpi/acpixf.h
556
acpi_evaluate_object(acpi_handle object,
include/acpi/acpixf.h
563
acpi_evaluate_object_typed(acpi_handle object,
include/acpi/acpixf.h
572
acpi_get_object_info(acpi_handle object,
include/acpi/acpixf.h
584
acpi_get_type(acpi_handle object,
include/acpi/acpixf.h
588
acpi_get_parent(acpi_handle object,
include/acpi/acpixf.h
976
acpi_get_data_full(acpi_handle object,
include/acpi/actypes.h
1064
void (*acpi_object_handler) (acpi_handle object, void *data);
include/acpi/actypes.h
1067
acpi_status (*acpi_init_handler) (acpi_handle object, u32 function);
include/acpi/actypes.h
1135
acpi_status (*acpi_walk_callback) (acpi_handle object,
include/linux/assoc_array.h
35
unsigned long (*get_object_key_chunk)(const void *object, int level);
include/linux/assoc_array.h
38
bool (*compare_object)(const void *object, const void *index_key);
include/linux/assoc_array.h
43
int (*diff_objects)(const void *object, const void *index_key);
include/linux/assoc_array.h
46
void (*free_object)(void *object);
include/linux/assoc_array.h
61
int (*iterator)(const void *object,
include/linux/assoc_array.h
72
void *object);
include/linux/assoc_array.h
74
void *object);
include/linux/assoc_array.h
84
bool (*iterator)(void *object, void *iterator_data),
include/linux/bug.h
85
void mem_dump_obj(void *object);
include/linux/bug.h
87
static inline void mem_dump_obj(void *object) {}
include/linux/debugobjects.h
34
void *object;
include/linux/hil_mlc.h
97
} object;
include/linux/kasan.h
137
void __kasan_unpoison_new_object(struct kmem_cache *cache, void *object);
include/linux/kasan.h
149
void *object)
include/linux/kasan.h
152
__kasan_unpoison_new_object(cache, object);
include/linux/kasan.h
155
void __kasan_poison_new_object(struct kmem_cache *cache, void *object);
include/linux/kasan.h
166
void *object)
include/linux/kasan.h
169
__kasan_poison_new_object(cache, object);
include/linux/kasan.h
173
const void *object);
include/linux/kasan.h
175
struct kmem_cache *cache, const void *object)
include/linux/kasan.h
178
return __kasan_init_slab_obj(cache, object);
include/linux/kasan.h
179
return (void *)object;
include/linux/kasan.h
182
bool __kasan_slab_pre_free(struct kmem_cache *s, void *object,
include/linux/kasan.h
196
void *object)
include/linux/kasan.h
199
return __kasan_slab_pre_free(s, object, _RET_IP_);
include/linux/kasan.h
203
bool __kasan_slab_free(struct kmem_cache *s, void *object, bool init,
include/linux/kasan.h
230
void *object, bool init,
include/linux/kasan.h
235
return __kasan_slab_free(s, object, init, still_accessible,
include/linux/kasan.h
248
void *object, gfp_t flags, bool init);
include/linux/kasan.h
250
struct kmem_cache *s, void *object, gfp_t flags, bool init)
include/linux/kasan.h
253
return __kasan_slab_alloc(s, object, flags, init);
include/linux/kasan.h
254
return object;
include/linux/kasan.h
257
void * __must_check __kasan_kmalloc(struct kmem_cache *s, const void *object,
include/linux/kasan.h
260
const void *object, size_t size, gfp_t flags)
include/linux/kasan.h
263
return __kasan_kmalloc(s, object, size, flags);
include/linux/kasan.h
264
return (void *)object;
include/linux/kasan.h
277
void * __must_check __kasan_krealloc(const void *object,
include/linux/kasan.h
279
static __always_inline void * __must_check kasan_krealloc(const void *object,
include/linux/kasan.h
283
return __kasan_krealloc(object, new_size, flags);
include/linux/kasan.h
284
return (void *)object;
include/linux/kasan.h
418
void *object) {}
include/linux/kasan.h
420
void *object) {}
include/linux/kasan.h
422
const void *object)
include/linux/kasan.h
424
return (void *)object;
include/linux/kasan.h
427
static inline bool kasan_slab_pre_free(struct kmem_cache *s, void *object)
include/linux/kasan.h
432
static inline bool kasan_slab_free(struct kmem_cache *s, void *object,
include/linux/kasan.h
439
static inline void *kasan_slab_alloc(struct kmem_cache *s, void *object,
include/linux/kasan.h
442
return object;
include/linux/kasan.h
444
static inline void *kasan_kmalloc(struct kmem_cache *s, const void *object,
include/linux/kasan.h
447
return (void *)object;
include/linux/kasan.h
453
static inline void *kasan_krealloc(const void *object, size_t new_size,
include/linux/kasan.h
456
return (void *)object;
include/linux/kfence.h
222
bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
include/linux/kfence.h
246
static inline bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
include/linux/kmsan.h
100
void kmsan_slab_alloc(struct kmem_cache *s, void *object, gfp_t flags);
include/linux/kmsan.h
109
void kmsan_slab_free(struct kmem_cache *s, void *object);
include/linux/kmsan.h
332
static inline void kmsan_slab_alloc(struct kmem_cache *s, void *object,
include/linux/kmsan.h
337
static inline void kmsan_slab_free(struct kmem_cache *s, void *object)
include/linux/mlx5/driver.h
222
void *object;
include/linux/slab.h
523
bool kmem_dump_obj(void *object);
include/linux/slab.h
525
static inline bool kmem_dump_obj(void *object) { return false; }
include/linux/vmalloc.h
330
bool vmalloc_dump_obj(void *object);
include/linux/vmalloc.h
332
static inline bool vmalloc_dump_obj(void *object) { return false; }
include/media/media-request.h
284
int (*prepare)(struct media_request_object *object);
include/media/media-request.h
285
void (*unprepare)(struct media_request_object *object);
include/media/media-request.h
286
void (*queue)(struct media_request_object *object);
include/media/media-request.h
287
void (*unbind)(struct media_request_object *object);
include/media/media-request.h
288
void (*release)(struct media_request_object *object);
include/rdma/ib_verbs.h
1568
void *object; /* containing object */
include/rdma/uverbs_ioctl.h
701
return attr->obj_attr.uobject->object;
include/rdma/uverbs_std_types.h
123
uobj->object = action;
include/rdma/uverbs_std_types.h
152
uobj->object = ibflow;
include/rdma/uverbs_std_types.h
38
return uobj->object;
lib/assoc_array.c
1028
void assoc_array_insert_set_object(struct assoc_array_edit *edit, void *object)
lib/assoc_array.c
1030
BUG_ON(!object);
lib/assoc_array.c
1031
edit->leaf = assoc_array_leaf_to_ptr(object);
lib/assoc_array.c
134
int (*iterator)(const void *object,
lib/assoc_array.c
1453
bool (*iterator)(void *object, void *iterator_data),
lib/assoc_array.c
964
void *object)
lib/assoc_array.c
976
BUG_ON(assoc_array_ptr_is_meta(object));
lib/assoc_array.c
983
edit->leaf = assoc_array_leaf_to_ptr(object);
lib/debugobjects.c
1012
struct debug_obj o = { .object = addr, .state = ODEBUG_STATE_NOTAVAILABLE, .descr = descr };
lib/debugobjects.c
1053
struct debug_obj o = { .object = addr, .state = ODEBUG_STATE_NOTAVAILABLE, .descr = descr };
lib/debugobjects.c
1108
oaddr = (unsigned long) obj->object;
lib/debugobjects.c
440
if (obj->object == addr)
lib/debugobjects.c
480
obj->object = addr;
lib/debugobjects.c
627
descr->debug_hint(obj->object) : NULL;
lib/debugobjects.c
632
obj->object, descr->name, hint);
lib/debugobjects.c
823
struct debug_obj o = { .object = addr, .state = ODEBUG_STATE_NOTAVAILABLE, .descr = descr };
lib/debugobjects.c
880
struct debug_obj o = { .object = addr, .state = ODEBUG_STATE_NOTAVAILABLE, .descr = descr };
mm/kasan/common.c
166
void __kasan_unpoison_new_object(struct kmem_cache *cache, void *object)
mm/kasan/common.c
168
kasan_unpoison(object, cache->object_size, false);
mm/kasan/common.c
171
void __kasan_poison_new_object(struct kmem_cache *cache, void *object)
mm/kasan/common.c
173
kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE),
mm/kasan/common.c
188
const void *object, bool init)
mm/kasan/common.c
205
return init ? kasan_random_tag() : get_tag(object);
mm/kasan/common.c
209
const void *object)
mm/kasan/common.c
213
kasan_init_object_meta(cache, object);
mm/kasan/common.c
216
object = set_tag(object, assign_tag(cache, object, true));
mm/kasan/common.c
218
return (void *)object;
mm/kasan/common.c
222
static bool check_slab_allocation(struct kmem_cache *cache, void *object,
mm/kasan/common.c
225
void *tagged_object = object;
mm/kasan/common.c
227
object = kasan_reset_tag(object);
mm/kasan/common.c
229
if (unlikely(nearest_obj(cache, virt_to_slab(object), object) != object)) {
mm/kasan/common.c
242
static inline void poison_slab_object(struct kmem_cache *cache, void *object,
mm/kasan/common.c
245
void *tagged_object = object;
mm/kasan/common.c
247
object = kasan_reset_tag(object);
mm/kasan/common.c
249
kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE),
mm/kasan/common.c
256
bool __kasan_slab_pre_free(struct kmem_cache *cache, void *object,
mm/kasan/common.c
259
if (is_kfence_address(object))
mm/kasan/common.c
261
return check_slab_allocation(cache, object, ip);
mm/kasan/common.c
264
bool __kasan_slab_free(struct kmem_cache *cache, void *object, bool init,
mm/kasan/common.c
267
if (is_kfence_address(object))
mm/kasan/common.c
285
poison_slab_object(cache, object, init);
mm/kasan/common.c
295
if (kasan_quarantine_put(cache, object))
mm/kasan/common.c
329
static inline void unpoison_slab_object(struct kmem_cache *cache, void *object,
mm/kasan/common.c
336
kasan_unpoison(object, cache->object_size, init);
mm/kasan/common.c
340
kasan_save_alloc_info(cache, object, flags);
mm/kasan/common.c
344
void *object, gfp_t flags, bool init)
mm/kasan/common.c
352
if (unlikely(object == NULL))
mm/kasan/common.c
355
if (is_kfence_address(object))
mm/kasan/common.c
356
return (void *)object;
mm/kasan/common.c
362
tag = assign_tag(cache, object, false);
mm/kasan/common.c
363
tagged_object = set_tag(object, tag);
mm/kasan/common.c
372
const void *object, size_t size, gfp_t flags)
mm/kasan/common.c
383
kasan_poison_last_granule((void *)object, size);
mm/kasan/common.c
386
redzone_start = round_up((unsigned long)(object + size),
mm/kasan/common.c
388
redzone_end = round_up((unsigned long)(object + cache->object_size),
mm/kasan/common.c
398
kasan_save_alloc_info(cache, (void *)object, flags);
mm/kasan/common.c
402
void * __must_check __kasan_kmalloc(struct kmem_cache *cache, const void *object,
mm/kasan/common.c
408
if (unlikely(object == NULL))
mm/kasan/common.c
411
if (is_kfence_address(object))
mm/kasan/common.c
412
return (void *)object;
mm/kasan/common.c
415
poison_kmalloc_redzone(cache, object, size, flags);
mm/kasan/common.c
418
return (void *)object;
mm/kasan/common.c
459
void * __must_check __kasan_krealloc(const void *object, size_t size, gfp_t flags)
mm/kasan/common.c
466
if (unlikely(object == ZERO_SIZE_PTR))
mm/kasan/common.c
467
return (void *)object;
mm/kasan/common.c
469
if (is_kfence_address(object))
mm/kasan/common.c
470
return (void *)object;
mm/kasan/common.c
477
kasan_unpoison(object, size, false);
mm/kasan/common.c
479
slab = virt_to_slab(object);
mm/kasan/common.c
483
poison_kmalloc_large_redzone(object, size, flags);
mm/kasan/common.c
485
poison_kmalloc_redzone(slab->slab_cache, object, size, flags);
mm/kasan/common.c
487
return (void *)object;
mm/kasan/generic.c
468
const void *object)
mm/kasan/generic.c
472
return (void *)object + cache->kasan_info.alloc_meta_offset;
mm/kasan/generic.c
476
const void *object)
mm/kasan/generic.c
481
return (void *)object + cache->kasan_info.free_meta_offset;
mm/kasan/generic.c
484
void kasan_init_object_meta(struct kmem_cache *cache, const void *object)
mm/kasan/generic.c
488
alloc_meta = kasan_get_alloc_meta(cache, object);
mm/kasan/generic.c
507
static void release_free_meta(const void *object, struct kasan_free_meta *meta)
mm/kasan/generic.c
510
if (*(u8 *)kasan_mem_to_shadow(object) != KASAN_SLAB_FREE_META)
mm/kasan/generic.c
514
*(u8 *)kasan_mem_to_shadow(object) = KASAN_SLAB_FREE;
mm/kasan/generic.c
544
void *object;
mm/kasan/generic.c
550
object = nearest_obj(cache, slab, addr);
mm/kasan/generic.c
551
alloc_meta = kasan_get_alloc_meta(cache, object);
mm/kasan/generic.c
559
void kasan_save_alloc_info(struct kmem_cache *cache, void *object, gfp_t flags)
mm/kasan/generic.c
563
alloc_meta = kasan_get_alloc_meta(cache, object);
mm/kasan/generic.c
573
void kasan_save_free_info(struct kmem_cache *cache, void *object)
mm/kasan/generic.c
577
free_meta = kasan_get_free_meta(cache, object);
mm/kasan/generic.c
582
release_free_meta(object, free_meta);
mm/kasan/generic.c
587
*(u8 *)kasan_mem_to_shadow(object) = KASAN_SLAB_FREE_META;
mm/kasan/kasan.h
219
void *object;
mm/kasan/kasan.h
359
size_t kasan_get_alloc_size(void *object, struct kmem_cache *cache);
mm/kasan/kasan.h
376
void kasan_print_aux_stacks(struct kmem_cache *cache, const void *object);
mm/kasan/kasan.h
378
static inline void kasan_print_aux_stacks(struct kmem_cache *cache, const void *object) { }
mm/kasan/kasan.h
383
void kasan_report_invalid_free(void *object, unsigned long ip, enum kasan_report_type type);
mm/kasan/kasan.h
389
const void *object);
mm/kasan/kasan.h
391
const void *object);
mm/kasan/kasan.h
392
void kasan_init_object_meta(struct kmem_cache *cache, const void *object);
mm/kasan/kasan.h
394
static inline void kasan_init_object_meta(struct kmem_cache *cache, const void *object) { }
mm/kasan/kasan.h
400
void kasan_save_alloc_info(struct kmem_cache *cache, void *object, gfp_t flags);
mm/kasan/kasan.h
402
void kasan_save_free_info(struct kmem_cache *cache, void *object);
mm/kasan/kasan.h
405
bool kasan_quarantine_put(struct kmem_cache *cache, void *object);
mm/kasan/kasan.h
409
static inline bool kasan_quarantine_put(struct kmem_cache *cache, void *object) { return false; }
mm/kasan/quarantine.c
145
void *object = qlink_to_object(qlink, cache);
mm/kasan/quarantine.c
146
struct kasan_free_meta *free_meta = kasan_get_free_meta(cache, object);
mm/kasan/quarantine.c
163
___cache_free(cache, object, _THIS_IP_);
mm/kasan/quarantine.c
185
bool kasan_quarantine_put(struct kmem_cache *cache, void *object)
mm/kasan/quarantine.c
190
struct kasan_free_meta *meta = kasan_get_free_meta(cache, object);
mm/kasan/report.c
296
unsigned long object_addr = (unsigned long)info->object;
mm/kasan/report.c
302
info->object, info->cache->name, info->cache->object_size);
mm/kasan/report.c
347
kasan_print_aux_stacks(info->cache, info->object);
mm/kasan/report.c
381
if (info->cache && info->object) {
mm/kasan/report.c
503
info->object = nearest_obj(info->cache, slab, addr);
mm/kasan/report.c
506
info->alloc_size = kasan_get_alloc_size(info->object, info->cache);
mm/kasan/report.c
511
info->cache = info->object = NULL;
mm/kasan/report_generic.c
168
if (!info->cache || !info->object)
mm/kasan/report_generic.c
171
alloc_meta = kasan_get_alloc_meta(info->cache, info->object);
mm/kasan/report_generic.c
176
if (*(u8 *)kasan_mem_to_shadow(info->object) == KASAN_SLAB_FREE_META) {
mm/kasan/report_generic.c
178
free_meta = kasan_get_free_meta(info->cache, info->object);
mm/kasan/report_generic.c
189
void kasan_print_aux_stacks(struct kmem_cache *cache, const void *object)
mm/kasan/report_generic.c
193
alloc_meta = kasan_get_alloc_meta(cache, object);
mm/kasan/report_generic.c
46
size_t kasan_get_alloc_size(void *object, struct kmem_cache *cache)
mm/kasan/report_generic.c
60
shadow = (u8 *)kasan_mem_to_shadow(object);
mm/kasan/report_hw_tags.c
30
size_t kasan_get_alloc_size(void *object, struct kmem_cache *cache)
mm/kasan/report_hw_tags.c
46
memory_tag = hw_get_mem_tag(object + i * KASAN_GRANULE_SIZE);
mm/kasan/report_sw_tags.c
48
size_t kasan_get_alloc_size(void *object, struct kmem_cache *cache)
mm/kasan/report_sw_tags.c
62
shadow = (u8 *)kasan_mem_to_shadow(object);
mm/kasan/report_tags.c
37
if ((!info->cache || !info->object) && !info->bug_type) {
mm/kasan/report_tags.c
61
if (kasan_reset_tag(entry->ptr) != info->object ||
mm/kasan/tags.c
132
entry->ptr = object;
mm/kasan/tags.c
140
void kasan_save_alloc_info(struct kmem_cache *cache, void *object, gfp_t flags)
mm/kasan/tags.c
142
save_stack_info(cache, object, flags, false);
mm/kasan/tags.c
145
void kasan_save_free_info(struct kmem_cache *cache, void *object)
mm/kasan/tags.c
147
save_stack_info(cache, object, 0, true);
mm/kasan/tags.c
97
static void save_stack_info(struct kmem_cache *cache, void *object,
mm/kfence/report.c
299
bool __kfence_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
mm/kfence/report.c
301
struct kfence_metadata *meta = addr_to_metadata((unsigned long)object);
mm/kfence/report.c
311
kpp->kp_ptr = object;
mm/kmemleak.c
1004
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
1008
object->flags |= OBJECT_FULL_SCAN;
mm/kmemleak.c
1012
size = untagged_objp + object->size - untagged_ptr;
mm/kmemleak.c
1013
} else if (untagged_ptr + size > untagged_objp + object->size) {
mm/kmemleak.c
1015
dump_object_info(object);
mm/kmemleak.c
1024
hlist_add_head(&area->node, &object->area_list);
mm/kmemleak.c
1026
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
1027
put_object(object);
mm/kmemleak.c
1039
struct kmemleak_object *object;
mm/kmemleak.c
1041
object = find_and_get_object(ptr, 0);
mm/kmemleak.c
1042
if (!object) {
mm/kmemleak.c
1048
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
1049
object->excess_ref = excess_ref;
mm/kmemleak.c
1050
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
1051
put_object(object);
mm/kmemleak.c
1062
struct kmemleak_object *object;
mm/kmemleak.c
1064
object = find_and_get_object(ptr, 0);
mm/kmemleak.c
1065
if (!object) {
mm/kmemleak.c
1070
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
1071
object->flags |= OBJECT_NO_SCAN;
mm/kmemleak.c
1072
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
1073
put_object(object);
mm/kmemleak.c
1203
struct kmemleak_object *object;
mm/kmemleak.c
1212
object = find_and_get_object((unsigned long)ptr, 1);
mm/kmemleak.c
1213
if (!object) {
mm/kmemleak.c
1222
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
1223
object->trace_handle = trace_handle;
mm/kmemleak.c
1224
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
1226
put_object(object);
mm/kmemleak.c
1386
static bool update_checksum(struct kmemleak_object *object)
mm/kmemleak.c
1388
u32 old_csum = object->checksum;
mm/kmemleak.c
1390
if (WARN_ON_ONCE(object->flags & OBJECT_PHYS))
mm/kmemleak.c
1395
if (object->flags & OBJECT_PERCPU) {
mm/kmemleak.c
1398
object->checksum = 0;
mm/kmemleak.c
1400
void *ptr = per_cpu_ptr((void __percpu *)object->pointer, cpu);
mm/kmemleak.c
1402
object->checksum ^= crc32(0, kasan_reset_tag((void *)ptr), object->size);
mm/kmemleak.c
1405
object->checksum = crc32(0, kasan_reset_tag((void *)object->pointer), object->size);
mm/kmemleak.c
1410
return object->checksum != old_csum;
mm/kmemleak.c
1416
static void update_refs(struct kmemleak_object *object)
mm/kmemleak.c
1418
if (!color_white(object)) {
mm/kmemleak.c
1429
object->count++;
mm/kmemleak.c
1430
if (color_gray(object)) {
mm/kmemleak.c
1432
WARN_ON(!get_object(object));
mm/kmemleak.c
1433
list_add_tail(&object->gray_list, &gray_list);
mm/kmemleak.c
1440
struct kmemleak_object *object;
mm/kmemleak.c
1459
object = __lookup_object(pointer, 1, objflags);
mm/kmemleak.c
1460
if (!object)
mm/kmemleak.c
1462
if (object == scanned)
mm/kmemleak.c
1471
raw_spin_lock_nested(&object->lock, SINGLE_DEPTH_NESTING);
mm/kmemleak.c
1473
if (color_gray(object)) {
mm/kmemleak.c
1474
excess_ref = object->excess_ref;
mm/kmemleak.c
1478
update_refs(object);
mm/kmemleak.c
1480
raw_spin_unlock(&object->lock);
mm/kmemleak.c
1483
object = lookup_object(excess_ref, 0);
mm/kmemleak.c
1484
if (!object)
mm/kmemleak.c
1486
if (object == scanned)
mm/kmemleak.c
1489
raw_spin_lock_nested(&object->lock, SINGLE_DEPTH_NESTING);
mm/kmemleak.c
1490
update_refs(object);
mm/kmemleak.c
1491
raw_spin_unlock(&object->lock);
mm/kmemleak.c
1566
static void scan_object(struct kmemleak_object *object)
mm/kmemleak.c
1575
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
1576
if (object->flags & OBJECT_NO_SCAN)
mm/kmemleak.c
1578
if (!(object->flags & OBJECT_ALLOCATED))
mm/kmemleak.c
1582
if (object->flags & OBJECT_PERCPU) {
mm/kmemleak.c
1586
void *start = per_cpu_ptr((void __percpu *)object->pointer, cpu);
mm/kmemleak.c
1587
void *end = start + object->size;
mm/kmemleak.c
1589
scan_block(start, end, object);
mm/kmemleak.c
1591
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
1593
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
1594
if (!(object->flags & OBJECT_ALLOCATED))
mm/kmemleak.c
1597
} else if (hlist_empty(&object->area_list) ||
mm/kmemleak.c
1598
object->flags & OBJECT_FULL_SCAN) {
mm/kmemleak.c
1599
void *start = object->flags & OBJECT_PHYS ?
mm/kmemleak.c
1600
__va((phys_addr_t)object->pointer) :
mm/kmemleak.c
1601
(void *)object->pointer;
mm/kmemleak.c
1602
void *end = start + object->size;
mm/kmemleak.c
1607
scan_block(start, next, object);
mm/kmemleak.c
1613
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
1615
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
1616
} while (object->flags & OBJECT_ALLOCATED);
mm/kmemleak.c
1618
hlist_for_each_entry(area, &object->area_list, node)
mm/kmemleak.c
1621
object);
mm/kmemleak.c
1624
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
1633
struct kmemleak_object *object, *tmp;
mm/kmemleak.c
1640
object = list_entry(gray_list.next, typeof(*object), gray_list);
mm/kmemleak.c
1641
while (&object->gray_list != &gray_list) {
mm/kmemleak.c
1646
scan_object(object);
mm/kmemleak.c
1648
tmp = list_entry(object->gray_list.next, typeof(*object),
mm/kmemleak.c
1652
list_del(&object->gray_list);
mm/kmemleak.c
1653
put_object(object);
mm/kmemleak.c
1655
object = tmp;
mm/kmemleak.c
1665
static void kmemleak_cond_resched(struct kmemleak_object *object)
mm/kmemleak.c
1667
if (!get_object(object))
mm/kmemleak.c
1671
if (object->del_state & DELSTATE_REMOVED)
mm/kmemleak.c
1673
object->del_state |= DELSTATE_NO_DELETE;
mm/kmemleak.c
1681
if (object->del_state & DELSTATE_REMOVED)
mm/kmemleak.c
1682
list_del_rcu(&object->object_list);
mm/kmemleak.c
1683
object->del_state &= ~DELSTATE_NO_DELETE;
mm/kmemleak.c
1686
put_object(object);
mm/kmemleak.c
1696
struct kmemleak_object *object;
mm/kmemleak.c
1705
list_for_each_entry_rcu(object, &object_list, object_list) {
mm/kmemleak.c
1706
raw_spin_lock_irq(&object->lock);
mm/kmemleak.c
1712
if (atomic_read(&object->use_count) > 1) {
mm/kmemleak.c
1714
atomic_read(&object->use_count));
mm/kmemleak.c
1715
dump_object_info(object);
mm/kmemleak.c
1720
if ((object->flags & OBJECT_PHYS) &&
mm/kmemleak.c
1721
!(object->flags & OBJECT_NO_SCAN)) {
mm/kmemleak.c
1722
unsigned long phys = object->pointer;
mm/kmemleak.c
1725
PHYS_PFN(phys + object->size) > max_low_pfn)
mm/kmemleak.c
1726
__paint_it(object, KMEMLEAK_BLACK);
mm/kmemleak.c
1730
object->count = 0;
mm/kmemleak.c
1731
if (color_gray(object) && get_object(object))
mm/kmemleak.c
1732
list_add_tail(&object->gray_list, &gray_list);
mm/kmemleak.c
1734
raw_spin_unlock_irq(&object->lock);
mm/kmemleak.c
1737
kmemleak_cond_resched(object);
mm/kmemleak.c
1805
list_for_each_entry_rcu(object, &object_list, object_list) {
mm/kmemleak.c
1807
kmemleak_cond_resched(object);
mm/kmemleak.c
1814
if (!color_white(object))
mm/kmemleak.c
1816
raw_spin_lock_irq(&object->lock);
mm/kmemleak.c
1817
if (color_white(object) && (object->flags & OBJECT_ALLOCATED)
mm/kmemleak.c
1818
&& update_checksum(object) && get_object(object)) {
mm/kmemleak.c
1820
object->count = object->min_count;
mm/kmemleak.c
1821
list_add_tail(&object->gray_list, &gray_list);
mm/kmemleak.c
1823
raw_spin_unlock_irq(&object->lock);
mm/kmemleak.c
1842
list_for_each_entry_rcu(object, &object_list, object_list) {
mm/kmemleak.c
1844
kmemleak_cond_resched(object);
mm/kmemleak.c
1851
if (!color_white(object))
mm/kmemleak.c
1853
raw_spin_lock_irq(&object->lock);
mm/kmemleak.c
1854
if (unreferenced_object(object) &&
mm/kmemleak.c
1855
!(object->flags & OBJECT_REPORTED)) {
mm/kmemleak.c
1856
object->flags |= OBJECT_REPORTED;
mm/kmemleak.c
1859
print_unreferenced(NULL, object);
mm/kmemleak.c
1863
raw_spin_unlock_irq(&object->lock);
mm/kmemleak.c
1947
struct kmemleak_object *object;
mm/kmemleak.c
1956
list_for_each_entry_rcu(object, &object_list, object_list) {
mm/kmemleak.c
1959
if (get_object(object))
mm/kmemleak.c
1962
object = NULL;
mm/kmemleak.c
1964
return object;
mm/kmemleak.c
2012
struct kmemleak_object *object = v;
mm/kmemleak.c
2015
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
2016
if ((object->flags & OBJECT_REPORTED) && unreferenced_object(object))
mm/kmemleak.c
2017
print_unreferenced(seq, object);
mm/kmemleak.c
2018
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
2037
struct kmemleak_object *object;
mm/kmemleak.c
2039
object = __find_and_get_object(addr, 1, objflags);
mm/kmemleak.c
2040
if (!object)
mm/kmemleak.c
2043
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
2044
dump_object_info(object);
mm/kmemleak.c
2045
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
2047
put_object(object);
mm/kmemleak.c
2080
struct kmemleak_object *object;
mm/kmemleak.c
2083
list_for_each_entry_rcu(object, &object_list, object_list) {
mm/kmemleak.c
2084
raw_spin_lock_irq(&object->lock);
mm/kmemleak.c
2085
if ((object->flags & OBJECT_REPORTED) &&
mm/kmemleak.c
2086
unreferenced_object(object))
mm/kmemleak.c
2087
__paint_it(object, KMEMLEAK_GREY);
mm/kmemleak.c
2088
raw_spin_unlock_irq(&object->lock);
mm/kmemleak.c
2197
struct kmemleak_object *object, *tmp;
mm/kmemleak.c
2204
list_for_each_entry_safe(object, tmp, &object_list, object_list) {
mm/kmemleak.c
2205
__remove_object(object);
mm/kmemleak.c
2206
__delete_object(object);
mm/kmemleak.c
293
struct kmemleak_object *object)
mm/kmemleak.c
295
const u8 *ptr = (const u8 *)object->pointer;
mm/kmemleak.c
298
if (WARN_ON_ONCE(object->flags & OBJECT_PHYS))
mm/kmemleak.c
301
if (object->flags & OBJECT_PERCPU)
mm/kmemleak.c
302
ptr = (const u8 *)this_cpu_ptr((void __percpu *)object->pointer);
mm/kmemleak.c
305
len = min_t(size_t, object->size, HEX_MAX_LINES * HEX_ROW_SIZE);
mm/kmemleak.c
307
if (object->flags & OBJECT_PERCPU)
mm/kmemleak.c
326
static bool color_white(const struct kmemleak_object *object)
mm/kmemleak.c
328
return object->count != KMEMLEAK_BLACK &&
mm/kmemleak.c
329
object->count < object->min_count;
mm/kmemleak.c
332
static bool color_gray(const struct kmemleak_object *object)
mm/kmemleak.c
334
return object->min_count != KMEMLEAK_BLACK &&
mm/kmemleak.c
335
object->count >= object->min_count;
mm/kmemleak.c
343
static bool unreferenced_object(struct kmemleak_object *object)
mm/kmemleak.c
345
return (color_white(object) && object->flags & OBJECT_ALLOCATED) &&
mm/kmemleak.c
346
time_before_eq(object->jiffies + jiffies_min_age,
mm/kmemleak.c
350
static const char *__object_type_str(struct kmemleak_object *object)
mm/kmemleak.c
352
if (object->flags & OBJECT_PHYS)
mm/kmemleak.c
354
if (object->flags & OBJECT_PERCPU)
mm/kmemleak.c
364
struct kmemleak_object *object)
mm/kmemleak.c
370
nr_entries = stack_depot_fetch(object->trace_handle, &entries);
mm/kmemleak.c
372
__object_type_str(object),
mm/kmemleak.c
373
object->pointer, object->size);
mm/kmemleak.c
375
object->comm, object->pid, object->jiffies);
mm/kmemleak.c
376
hex_dump_object(seq, object);
mm/kmemleak.c
377
warn_or_seq_printf(seq, " backtrace (crc %x):\n", object->checksum);
mm/kmemleak.c
390
static void dump_object_info(struct kmemleak_object *object)
mm/kmemleak.c
393
__object_type_str(object), object->pointer, object->size);
mm/kmemleak.c
395
object->comm, object->pid, object->jiffies);
mm/kmemleak.c
396
pr_notice(" min_count = %d\n", object->min_count);
mm/kmemleak.c
397
pr_notice(" count = %d\n", object->count);
mm/kmemleak.c
398
pr_notice(" flags = 0x%x\n", object->flags);
mm/kmemleak.c
399
pr_notice(" checksum = %u\n", object->checksum);
mm/kmemleak.c
401
if (object->trace_handle)
mm/kmemleak.c
402
stack_depot_print(object->trace_handle);
mm/kmemleak.c
427
struct kmemleak_object *object;
mm/kmemleak.c
430
object = rb_entry(rb, struct kmemleak_object, rb_node);
mm/kmemleak.c
431
untagged_objp = (unsigned long)kasan_reset_tag((void *)object->pointer);
mm/kmemleak.c
434
rb = object->rb_node.rb_left;
mm/kmemleak.c
435
else if (untagged_objp + object->size <= untagged_ptr)
mm/kmemleak.c
436
rb = object->rb_node.rb_right;
mm/kmemleak.c
438
return object;
mm/kmemleak.c
447
dump_object_info(object);
mm/kmemleak.c
467
static int get_object(struct kmemleak_object *object)
mm/kmemleak.c
469
return atomic_inc_not_zero(&object->use_count);
mm/kmemleak.c
478
struct kmemleak_object *object;
mm/kmemleak.c
483
object = kmem_cache_alloc_noprof(object_cache,
mm/kmemleak.c
485
if (object)
mm/kmemleak.c
486
return object;
mm/kmemleak.c
491
object = list_first_entry_or_null(&mem_pool_free_list,
mm/kmemleak.c
492
typeof(*object), object_list);
mm/kmemleak.c
493
if (object)
mm/kmemleak.c
494
list_del(&object->object_list);
mm/kmemleak.c
496
object = &mem_pool[--mem_pool_free_count];
mm/kmemleak.c
503
return object;
mm/kmemleak.c
509
static void mem_pool_free(struct kmemleak_object *object)
mm/kmemleak.c
513
if (object < mem_pool || object >= ARRAY_END(mem_pool)) {
mm/kmemleak.c
514
kmem_cache_free(object_cache, object);
mm/kmemleak.c
520
list_add(&object->object_list, &mem_pool_free_list);
mm/kmemleak.c
531
struct kmemleak_object *object =
mm/kmemleak.c
538
hlist_for_each_entry_safe(area, tmp, &object->area_list, node) {
mm/kmemleak.c
542
mem_pool_free(object);
mm/kmemleak.c
552
static void put_object(struct kmemleak_object *object)
mm/kmemleak.c
554
if (!atomic_dec_and_test(&object->use_count))
mm/kmemleak.c
558
WARN_ON(object->flags & OBJECT_ALLOCATED);
mm/kmemleak.c
566
call_rcu(&object->rcu, free_object_rcu);
mm/kmemleak.c
568
free_object_rcu(&object->rcu);
mm/kmemleak.c
578
struct kmemleak_object *object;
mm/kmemleak.c
582
object = __lookup_object(ptr, alias, objflags);
mm/kmemleak.c
586
if (object && !get_object(object))
mm/kmemleak.c
587
object = NULL;
mm/kmemleak.c
590
return object;
mm/kmemleak.c
603
static void __remove_object(struct kmemleak_object *object)
mm/kmemleak.c
605
rb_erase(&object->rb_node, object_tree(object->flags));
mm/kmemleak.c
606
if (!(object->del_state & DELSTATE_NO_DELETE))
mm/kmemleak.c
607
list_del_rcu(&object->object_list);
mm/kmemleak.c
608
object->del_state |= DELSTATE_REMOVED;
mm/kmemleak.c
615
struct kmemleak_object *object;
mm/kmemleak.c
617
object = __lookup_object(ptr, alias, objflags);
mm/kmemleak.c
618
if (object)
mm/kmemleak.c
619
__remove_object(object);
mm/kmemleak.c
621
return object;
mm/kmemleak.c
633
struct kmemleak_object *object;
mm/kmemleak.c
636
object = __find_and_remove_object(ptr, alias, objflags);
mm/kmemleak.c
639
return object;
mm/kmemleak.c
663
struct kmemleak_object *object;
mm/kmemleak.c
665
object = mem_pool_alloc(gfp);
mm/kmemleak.c
666
if (!object) {
mm/kmemleak.c
672
INIT_LIST_HEAD(&object->object_list);
mm/kmemleak.c
673
INIT_LIST_HEAD(&object->gray_list);
mm/kmemleak.c
674
INIT_HLIST_HEAD(&object->area_list);
mm/kmemleak.c
675
raw_spin_lock_init(&object->lock);
mm/kmemleak.c
676
atomic_set(&object->use_count, 1);
mm/kmemleak.c
677
object->excess_ref = 0;
mm/kmemleak.c
678
object->count = 0; /* white color initially */
mm/kmemleak.c
679
object->checksum = 0;
mm/kmemleak.c
680
object->del_state = 0;
mm/kmemleak.c
684
object->pid = 0;
mm/kmemleak.c
685
strscpy(object->comm, "hardirq");
mm/kmemleak.c
687
object->pid = 0;
mm/kmemleak.c
688
strscpy(object->comm, "softirq");
mm/kmemleak.c
690
object->pid = current->pid;
mm/kmemleak.c
697
strscpy(object->comm, current->comm);
mm/kmemleak.c
701
object->trace_handle = set_track_prepare();
mm/kmemleak.c
703
return object;
mm/kmemleak.c
706
static int __link_object(struct kmemleak_object *object, unsigned long ptr,
mm/kmemleak.c
715
object->flags = OBJECT_ALLOCATED | objflags;
mm/kmemleak.c
716
object->pointer = ptr;
mm/kmemleak.c
717
object->size = kfence_ksize((void *)ptr) ?: size;
mm/kmemleak.c
718
object->min_count = min_count;
mm/kmemleak.c
719
object->jiffies = jiffies;
mm/kmemleak.c
761
rb_link_node(&object->rb_node, rb_parent, link);
mm/kmemleak.c
762
rb_insert_color(&object->rb_node, object_tree(objflags));
mm/kmemleak.c
763
list_add_tail_rcu(&object->object_list, &object_list);
mm/kmemleak.c
775
struct kmemleak_object *object;
mm/kmemleak.c
779
object = __alloc_object(gfp);
mm/kmemleak.c
780
if (!object)
mm/kmemleak.c
784
ret = __link_object(object, ptr, size, min_count, objflags);
mm/kmemleak.c
787
mem_pool_free(object);
mm/kmemleak.c
814
static void __delete_object(struct kmemleak_object *object)
mm/kmemleak.c
818
WARN_ON(!(object->flags & OBJECT_ALLOCATED));
mm/kmemleak.c
819
WARN_ON(atomic_read(&object->use_count) < 1);
mm/kmemleak.c
825
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
826
object->flags &= ~OBJECT_ALLOCATED;
mm/kmemleak.c
827
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
828
put_object(object);
mm/kmemleak.c
837
struct kmemleak_object *object;
mm/kmemleak.c
839
object = find_and_remove_object(ptr, 0, objflags);
mm/kmemleak.c
840
if (!object)
mm/kmemleak.c
846
__delete_object(object);
mm/kmemleak.c
857
struct kmemleak_object *object, *object_l, *object_r;
mm/kmemleak.c
869
object = __find_and_remove_object(ptr, 1, objflags);
mm/kmemleak.c
870
if (!object)
mm/kmemleak.c
878
start = object->pointer;
mm/kmemleak.c
879
end = object->pointer + object->size;
mm/kmemleak.c
882
object->min_count, objflags))
mm/kmemleak.c
886
object->min_count, objflags))
mm/kmemleak.c
891
if (object) {
mm/kmemleak.c
892
__delete_object(object);
mm/kmemleak.c
907
static void __paint_it(struct kmemleak_object *object, int color)
mm/kmemleak.c
909
object->min_count = color;
mm/kmemleak.c
911
object->flags |= OBJECT_NO_SCAN;
mm/kmemleak.c
914
static void paint_it(struct kmemleak_object *object, int color)
mm/kmemleak.c
918
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
919
__paint_it(object, color);
mm/kmemleak.c
920
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
925
struct kmemleak_object *object;
mm/kmemleak.c
927
object = __find_and_get_object(ptr, 0, objflags);
mm/kmemleak.c
928
if (!object)
mm/kmemleak.c
934
paint_it(object, color);
mm/kmemleak.c
935
put_object(object);
mm/kmemleak.c
963
struct kmemleak_object *object;
mm/kmemleak.c
965
object = find_and_get_object(ptr, 0);
mm/kmemleak.c
966
if (!object) {
mm/kmemleak.c
972
raw_spin_lock_irqsave(&object->lock, flags);
mm/kmemleak.c
973
object->checksum = 0;
mm/kmemleak.c
974
raw_spin_unlock_irqrestore(&object->lock, flags);
mm/kmemleak.c
975
put_object(object);
mm/kmemleak.c
985
struct kmemleak_object *object;
mm/kmemleak.c
990
object = find_and_get_object(ptr, 1);
mm/kmemleak.c
991
if (!object) {
mm/kmemleak.c
998
untagged_objp = (unsigned long)kasan_reset_tag((void *)object->pointer);
mm/kmsan/hooks.c
48
void kmsan_slab_alloc(struct kmem_cache *s, void *object, gfp_t flags)
mm/kmsan/hooks.c
50
if (unlikely(object == NULL))
mm/kmsan/hooks.c
63
kmsan_internal_unpoison_memory(object, s->object_size,
mm/kmsan/hooks.c
66
kmsan_internal_poison_memory(object, s->object_size, flags,
mm/kmsan/hooks.c
71
void kmsan_slab_free(struct kmem_cache *s, void *object)
mm/kmsan/hooks.c
87
kmsan_internal_poison_memory(object, s->object_size,
mm/slab.h
213
void (*ctor)(void *object); /* Object constructor */
mm/slab.h
282
void *object = x - (x - slab_address(slab)) % cache->size;
mm/slab.h
285
void *result = (unlikely(object > last_object)) ? last_object : object;
mm/slab.h
453
extern void print_tracking(struct kmem_cache *s, void *object);
mm/slab.h
460
static inline void print_tracking(struct kmem_cache *s, void *object)
mm/slab.h
715
void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab);
mm/slab.h
730
void skip_orig_size_check(struct kmem_cache *s, const void *object);
mm/slab_common.c
616
static void kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
mm/slab_common.c
618
if (__kfence_obj_info(kpp, object, slab))
mm/slab_common.c
620
__kmem_obj_info(kpp, object, slab);
mm/slab_common.c
638
bool kmem_dump_obj(void *object)
mm/slab_common.c
647
if (object < (void *)PAGE_SIZE || !virt_addr_valid(object))
mm/slab_common.c
649
slab = virt_to_slab(object);
mm/slab_common.c
653
kmem_obj_info(&kp, object, slab);
mm/slab_common.c
658
if (is_kfence_address(object))
mm/slab_common.c
665
ptroffset = ((char *)object - (char *)kp.kp_objp) - kp.kp_data_offset;
mm/slub.c
1000
(object - base) % s->size) {
mm/slub.c
1016
static struct track *get_track(struct kmem_cache *s, void *object,
mm/slub.c
1021
p = object + get_info_end(s);
mm/slub.c
1045
static void set_track_update(struct kmem_cache *s, void *object,
mm/slub.c
1049
struct track *p = get_track(s, object, alloc);
mm/slub.c
1060
static __always_inline void set_track(struct kmem_cache *s, void *object,
mm/slub.c
1065
set_track_update(s, object, alloc, addr, handle);
mm/slub.c
1068
static void init_tracking(struct kmem_cache *s, void *object)
mm/slub.c
1075
p = get_track(s, object, TRACK_ALLOC);
mm/slub.c
1097
void print_tracking(struct kmem_cache *s, void *object)
mm/slub.c
1103
print_track("Allocated", get_track(s, object, TRACK_ALLOC), pr_time);
mm/slub.c
1104
print_track("Freed", get_track(s, object, TRACK_FREE), pr_time);
mm/slub.c
1114
void skip_orig_size_check(struct kmem_cache *s, const void *object)
mm/slub.c
1116
set_orig_size(s, (void *)object, s->object_size);
mm/slub.c
1202
u8 *object, const char *reason)
mm/slub.c
1208
if (!object || !check_valid_pointer(s, slab, object)) {
mm/slub.c
1210
pr_err("Invalid pointer 0x%p\n", object);
mm/slub.c
1212
print_trailer(s, slab, object);
mm/slub.c
1245
static void init_object(struct kmem_cache *s, void *object, u8 val)
mm/slub.c
1247
u8 *p = kasan_reset_tag(object);
mm/slub.c
1265
poison_size = get_orig_size(s, object);
mm/slub.c
1294
u8 *object, const char *what, u8 *start, unsigned int value,
mm/slub.c
1318
object_err(s, slab, object, "Object corrupt");
mm/slub.c
1454
void *object, u8 val)
mm/slub.c
1456
u8 *p = object;
mm/slub.c
1457
u8 *endobject = object + s->object_size;
mm/slub.c
1462
if (!check_bytes_and_report(s, slab, object, "Left Redzone",
mm/slub.c
1463
object - s->red_left_pad, val, s->red_left_pad, ret))
mm/slub.c
1466
if (!check_bytes_and_report(s, slab, object, "Right Redzone",
mm/slub.c
1471
orig_size = get_orig_size(s, object);
mm/slub.c
1474
!check_bytes_and_report(s, slab, object,
mm/slub.c
1571
void *object = NULL;
mm/slub.c
1579
if (object) {
mm/slub.c
1580
object_err(s, slab, object,
mm/slub.c
1582
set_freepointer(s, object, NULL);
mm/slub.c
1592
object = fp;
mm/slub.c
1593
fp = get_freepointer(s, object);
mm/slub.c
1624
static void trace(struct kmem_cache *s, struct slab *slab, void *object,
mm/slub.c
1631
object, slab->inuse,
mm/slub.c
1635
print_section(KERN_INFO, "Object ", (void *)object,
mm/slub.c
1685
static void setup_object_debug(struct kmem_cache *s, void *object)
mm/slub.c
1690
init_object(s, object, SLUB_RED_INACTIVE);
mm/slub.c
1691
init_tracking(s, object);
mm/slub.c
1706
struct slab *slab, void *object)
mm/slub.c
1711
if (!check_valid_pointer(s, slab, object)) {
mm/slub.c
1712
object_err(s, slab, object, "Freelist Pointer check fails");
mm/slub.c
1716
if (!check_object(s, slab, object, SLUB_RED_INACTIVE))
mm/slub.c
1723
struct slab *slab, void *object, int orig_size)
mm/slub.c
1726
if (!alloc_consistency_checks(s, slab, object))
mm/slub.c
1731
trace(s, slab, object, 1);
mm/slub.c
1732
set_orig_size(s, object, orig_size);
mm/slub.c
1733
init_object(s, object, SLUB_RED_ACTIVE);
mm/slub.c
1750
struct slab *slab, void *object, unsigned long addr)
mm/slub.c
1752
if (!check_valid_pointer(s, slab, object)) {
mm/slub.c
1753
slab_err(s, slab, "Invalid object pointer 0x%p", object);
mm/slub.c
1757
if (on_freelist(s, slab, object)) {
mm/slub.c
1758
object_err(s, slab, object, "Object already free");
mm/slub.c
1762
if (!check_object(s, slab, object, SLUB_RED_ACTIVE))
mm/slub.c
1768
object);
mm/slub.c
1770
object_err(s, slab, object,
mm/slub.c
1995
static inline void setup_object_debug(struct kmem_cache *s, void *object) {}
mm/slub.c
2000
struct slab *slab, void *object, int orig_size) { return true; }
mm/slub.c
2008
void *object, u8 val) { return 1; }
mm/slub.c
2010
static inline void set_track(struct kmem_cache *s, void *object,
mm/slub.c
2353
__alloc_tagging_slab_alloc_hook(struct kmem_cache *s, void *object, gfp_t flags)
mm/slub.c
2359
if (!object)
mm/slub.c
2368
slab = virt_to_slab(object);
mm/slub.c
2369
obj_exts = prepare_slab_obj_exts_hook(s, slab, flags, object);
mm/slub.c
2376
unsigned int obj_idx = obj_to_index(s, slab, object);
mm/slub.c
2388
alloc_tagging_slab_alloc_hook(struct kmem_cache *s, void *object, gfp_t flags)
mm/slub.c
2391
__alloc_tagging_slab_alloc_hook(s, object, flags);
mm/slub.c
2430
alloc_tagging_slab_alloc_hook(struct kmem_cache *s, void *object, gfp_t flags)
mm/slub.c
2445
static void memcg_alloc_abort_single(struct kmem_cache *s, void *object);
mm/slub.c
2573
void *object;
mm/slub.c
2648
delayed_free->object = x;
mm/slub.c
2693
void *object;
mm/slub.c
2710
object = next;
mm/slub.c
2711
next = get_freepointer(s, object);
mm/slub.c
2714
if (likely(slab_free_hook(s, object, init, false))) {
mm/slub.c
2716
set_freepointer(s, object, *head);
mm/slub.c
2717
*head = object;
mm/slub.c
2719
*tail = object;
mm/slub.c
2727
} while (object != old_tail);
mm/slub.c
2732
static void *setup_object(struct kmem_cache *s, void *object)
mm/slub.c
2734
setup_object_debug(s, object);
mm/slub.c
2735
object = kasan_init_slab_obj(s, object);
mm/slub.c
2737
kasan_unpoison_new_object(s, object);
mm/slub.c
2738
s->ctor(object);
mm/slub.c
2739
kasan_poison_new_object(s, object);
mm/slub.c
2741
return object;
mm/slub.c
3652
void *object;
mm/slub.c
3665
object = slab->freelist;
mm/slub.c
3666
slab->freelist = get_freepointer(s, object);
mm/slub.c
3669
if (!alloc_debug_processing(s, slab, object, orig_size)) {
mm/slub.c
3679
return object;
mm/slub.c
3694
void *object;
mm/slub.c
3702
object = slab->freelist;
mm/slub.c
3703
slab->freelist = get_freepointer(s, object);
mm/slub.c
3706
if (!alloc_debug_processing(s, slab, object, orig_size)) {
mm/slub.c
3729
return object;
mm/slub.c
3798
void *object = NULL;
mm/slub.c
3821
object = alloc_single_from_partial(s, n, slab,
mm/slub.c
3823
if (object)
mm/slub.c
3843
object = old.freelist;
mm/slub.c
3850
return object;
mm/slub.c
3907
void *object = get_from_partial_node(s, n, pc);
mm/slub.c
3909
if (object) {
mm/slub.c
3917
return object;
mm/slub.c
3933
void *object;
mm/slub.c
3938
object = get_from_partial_node(s, get_node(s, searchnode), pc);
mm/slub.c
3939
if (object || (node != NUMA_NO_NODE && (pc->flags & __GFP_THISNODE)))
mm/slub.c
3940
return object;
mm/slub.c
4119
void *object = head;
mm/slub.c
4139
if (!free_consistency_checks(s, slab, object, addr))
mm/slub.c
4144
set_track_update(s, object, TRACK_FREE, addr, handle);
mm/slub.c
4145
trace(s, slab, object, 0);
mm/slub.c
4147
init_object(s, object, SLUB_RED_INACTIVE);
mm/slub.c
4150
if (object != tail) {
mm/slub.c
4151
object = get_freepointer(s, object);
mm/slub.c
4166
slab_fix(s, "Object at 0x%p not freed", object);
mm/slub.c
4323
void *object;
mm/slub.c
4342
object = slab->freelist;
mm/slub.c
4343
while (object && allocated < count) {
mm/slub.c
4344
p[allocated] = object;
mm/slub.c
4345
object = get_freepointer(s, object);
mm/slub.c
4351
slab->freelist = object;
mm/slub.c
4378
void *object;
mm/slub.c
4409
object = get_from_partial(s, node, &pc);
mm/slub.c
4410
if (object)
mm/slub.c
4428
object = alloc_single_from_new_slab(s, slab, orig_size, gfpflags);
mm/slub.c
4430
if (likely(object))
mm/slub.c
4433
alloc_from_new_slab(s, slab, &object, 1, allow_spin);
mm/slub.c
4436
if (likely(object))
mm/slub.c
4437
return object;
mm/slub.c
4448
set_track(s, object, TRACK_ALLOC, addr, gfpflags);
mm/slub.c
4450
return object;
mm/slub.c
4456
void *object;
mm/slub.c
4479
object = ___slab_alloc(s, gfpflags, node, addr, orig_size);
mm/slub.c
4481
return object;
mm/slub.c
4676
void *object;
mm/slub.c
4722
object = pcs->main->objects[pcs->main->size - 1];
mm/slub.c
4730
if (page_to_nid(virt_to_page(object)) != node) {
mm/slub.c
4743
return object;
mm/slub.c
4840
void *object;
mm/slub.c
4847
object = kfence_alloc(s, orig_size, gfpflags);
mm/slub.c
4848
if (unlikely(object))
mm/slub.c
4851
object = alloc_from_pcs(s, gfpflags, node);
mm/slub.c
4853
if (!object)
mm/slub.c
4854
object = __slab_alloc_node(s, gfpflags, node, addr, orig_size);
mm/slub.c
4856
maybe_wipe_obj_freeptr(s, object);
mm/slub.c
4866
slab_post_alloc_hook(s, lru, gfpflags, 1, &object, init, orig_size);
mm/slub.c
4868
return object;
mm/slub.c
526
static inline void *get_freepointer(struct kmem_cache *s, void *object)
mm/slub.c
531
object = kasan_reset_tag(object);
mm/slub.c
532
ptr_addr = (unsigned long)object + s->offset;
mm/slub.c
537
static inline void set_freepointer(struct kmem_cache *s, void *object, void *fp)
mm/slub.c
539
unsigned long freeptr_addr = (unsigned long)object + s->offset;
mm/slub.c
542
BUG_ON(object == fp); /* naive detection of double free or corruption */
mm/slub.c
5764
bool free_to_pcs(struct kmem_cache *s, void *object, bool allow_spin)
mm/slub.c
5780
pcs->main->objects[pcs->main->size++] = object;
mm/slub.c
6159
void slab_free(struct kmem_cache *s, struct slab *slab, void *object,
mm/slub.c
6162
memcg_slab_free_hook(s, slab, &object, 1);
mm/slub.c
6163
alloc_tagging_slab_free_hook(s, slab, &object, 1);
mm/slub.c
6165
if (unlikely(!slab_free_hook(s, object, slab_want_init_on_free(s), false)))
mm/slub.c
6170
if (likely(free_to_pcs(s, object, true)))
mm/slub.c
6174
__slab_free(s, slab, object, object, 1, addr);
mm/slub.c
6181
void memcg_alloc_abort_single(struct kmem_cache *s, void *object)
mm/slub.c
6183
struct slab *slab = virt_to_slab(object);
mm/slub.c
6185
alloc_tagging_slab_free_hook(s, slab, &object, 1);
mm/slub.c
6187
if (likely(slab_free_hook(s, object, slab_want_init_on_free(s), false)))
mm/slub.c
6188
__slab_free(s, slab, object, object, 1, _RET_IP_);
mm/slub.c
6213
void *object = delayed_free->object;
mm/slub.c
6214
struct slab *slab = virt_to_slab(object);
mm/slub.c
6219
if (WARN_ON(is_kfence_address(object)))
mm/slub.c
6230
if (slab_free_hook(s, object, slab_want_init_on_free(s), true)) {
mm/slub.c
6231
__slab_free(s, slab, object, object, 1, _THIS_IP_);
mm/slub.c
6328
static size_t __ksize(const void *object)
mm/slub.c
6333
if (unlikely(object == ZERO_SIZE_PTR))
mm/slub.c
6336
page = virt_to_page(object);
mm/slub.c
6347
skip_orig_size_check(slab->slab_cache, object);
mm/slub.c
6389
static void free_large_kmalloc(struct page *page, void *object)
mm/slub.c
6399
pr_warn_once("object pointer: 0x%p\n", object);
mm/slub.c
6401
kmemleak_free(object);
mm/slub.c
6402
kasan_kfree_large(object);
mm/slub.c
6403
kmsan_kfree_large(object);
mm/slub.c
6462
void kfree(const void *object)
mm/slub.c
6467
void *x = (void *)object;
mm/slub.c
6469
trace_kfree(_RET_IP_, object);
mm/slub.c
6471
if (unlikely(ZERO_OR_NULL_PTR(object)))
mm/slub.c
6474
page = virt_to_page(object);
mm/slub.c
6478
free_large_kmalloc(page, (void *)object);
mm/slub.c
6496
void kfree_nolock(const void *object)
mm/slub.c
6500
void *x = (void *)object;
mm/slub.c
6502
if (unlikely(ZERO_OR_NULL_PTR(object)))
mm/slub.c
6505
slab = virt_to_slab(object);
mm/slub.c
6904
void *object;
mm/slub.c
6909
object = p[--size];
mm/slub.c
6910
page = virt_to_page(object);
mm/slub.c
6915
free_large_kmalloc(page, object);
mm/slub.c
6928
df->tail = object;
mm/slub.c
6929
df->freelist = object;
mm/slub.c
6932
if (is_kfence_address(object))
mm/slub.c
6935
set_freepointer(df->s, object, NULL);
mm/slub.c
6939
object = p[--size];
mm/slub.c
6941
if (df->slab == virt_to_slab(object)) {
mm/slub.c
6943
set_freepointer(df->s, object, df->freelist);
mm/slub.c
6944
df->freelist = object;
mm/slub.c
7021
void *object;
mm/slub.c
7034
object = get_freelist_nofreeze(s, slab);
mm/slub.c
7036
while (object && refilled < max) {
mm/slub.c
7037
p[refilled] = object;
mm/slub.c
7038
object = get_freepointer(s, object);
mm/slub.c
7049
if (unlikely(object)) {
mm/slub.c
7050
void *head = object;
mm/slub.c
7055
tail = object;
mm/slub.c
7057
object = get_freepointer(s, object);
mm/slub.c
7058
} while (object);
mm/slub.c
731
void *object, unsigned long orig_size)
mm/slub.c
733
void *p = kasan_reset_tag(object);
mm/slub.c
744
static inline unsigned long get_orig_size(struct kmem_cache *s, void *object)
mm/slub.c
746
void *p = kasan_reset_tag(object);
mm/slub.c
748
if (is_kfence_address(object))
mm/slub.c
749
return kfence_ksize(object);
mm/slub.c
7907
void __kmem_obj_info(struct kmem_obj_info *kpp, void *object, struct slab *slab)
mm/slub.c
7917
kpp->kp_ptr = object;
mm/slub.c
7921
objp0 = kasan_reset_tag(object);
mm/slub.c
989
struct slab *slab, void *object)
mm/slub.c
993
if (!object)
mm/slub.c
997
object = kasan_reset_tag(object);
mm/slub.c
998
object = restore_red_left(s, object);
mm/slub.c
999
if (object < base || object >= base + slab->objects * s->size ||
mm/util.c
1060
void mem_dump_obj(void *object)
mm/util.c
1064
if (kmem_dump_obj(object))
mm/util.c
1067
if (vmalloc_dump_obj(object))
mm/util.c
1070
if (is_vmalloc_addr(object))
mm/util.c
1072
else if (virt_addr_valid(object))
mm/util.c
1074
else if (object == NULL)
mm/util.c
1076
else if (object == ZERO_SIZE_PTR)
mm/vmalloc.c
5140
bool vmalloc_dump_obj(void *object)
mm/vmalloc.c
5149
addr = PAGE_ALIGN((unsigned long) object);
net/ceph/messenger_v1.c
596
int end, int size, void *object)
net/ceph/messenger_v1.c
601
int ret = ceph_tcp_recvmsg(con->sock, object + have, left);
samples/bpf/fds_example.c
112
static int bpf_do_prog(const char *file, uint32_t flags, const char *object)
samples/bpf/fds_example.c
117
fd = bpf_prog_create(object);
samples/bpf/fds_example.c
143
const char *file = NULL, *object = NULL;
samples/bpf/fds_example.c
176
object = optarg;
samples/bpf/fds_example.c
190
return bpf_do_prog(file, flags, object);
samples/bpf/fds_example.c
50
static int bpf_prog_create(const char *object)
samples/bpf/fds_example.c
60
if (object) {
samples/bpf/fds_example.c
61
obj = bpf_object__open_file(object, NULL);
scripts/mod/modpost.c
1480
static void extract_crcs_for_object(const char *object, struct module *mod)
scripts/mod/modpost.c
1487
base = get_basename(object);
scripts/mod/modpost.c
1488
dirlen = base - object;
scripts/mod/modpost.c
1491
dirlen, object, base);
security/keys/internal.h
125
int (*iterator)(const void *object, void *iterator_data);
security/keys/keyring.c
1105
const void *object;
security/keys/keyring.c
1112
object = assoc_array_find(&keyring->keys, &keyring_assoc_array_ops,
security/keys/keyring.c
1115
if (object)
security/keys/keyring.c
1122
key = keyring_ptr_to_key(object);
security/keys/keyring.c
1192
static int keyring_detect_cycle_iterator(const void *object,
security/keys/keyring.c
1196
const struct key *key = keyring_ptr_to_key(object);
security/keys/keyring.c
1693
static bool keyring_gc_select_iterator(void *object, void *iterator_data)
security/keys/keyring.c
1695
struct key *key = keyring_ptr_to_key(object);
security/keys/keyring.c
1704
static int keyring_gc_check_iterator(const void *object, void *iterator_data)
security/keys/keyring.c
1706
const struct key *key = keyring_ptr_to_key(object);
security/keys/keyring.c
303
static unsigned long keyring_get_object_key_chunk(const void *object, int level)
security/keys/keyring.c
305
const struct key *key = keyring_ptr_to_key(object);
security/keys/keyring.c
309
static bool keyring_compare_object(const void *object, const void *data)
security/keys/keyring.c
312
const struct key *key = keyring_ptr_to_key(object);
security/keys/keyring.c
325
static int keyring_diff_objects(const void *object, const void *data)
security/keys/keyring.c
327
const struct key *key_a = keyring_ptr_to_key(object);
security/keys/keyring.c
387
static void keyring_free_object(void *object)
security/keys/keyring.c
389
key_put(keyring_ptr_to_key(object));
security/keys/keyring.c
42
void *object = assoc_array_ptr_to_leaf(x);
security/keys/keyring.c
43
return (struct key *)((unsigned long)object & ~KEYRING_PTR_SUBTYPE);
security/keys/keyring.c
458
static int keyring_read_iterator(const void *object, void *data)
security/keys/keyring.c
461
const struct key *key = keyring_ptr_to_key(object);
security/keys/keyring.c
574
static int keyring_search_iterator(const void *object, void *iterator_data)
security/keys/keyring.c
577
const struct key *key = keyring_ptr_to_key(object);
security/keys/keyring.c
649
const void *object;
security/keys/keyring.c
651
object = assoc_array_find(&keyring->keys,
security/keys/keyring.c
654
return object ? ctx->iterator(object, ctx) : 0;
security/landlock/fs.c
1260
WARN_ON_ONCE(inode_sec->object);
security/landlock/fs.c
1279
struct landlock_object *object;
security/landlock/fs.c
1304
object = rcu_dereference(landlock_inode(inode)->object);
security/landlock/fs.c
1305
if (!object) {
security/landlock/fs.c
1319
spin_lock(&object->lock);
security/landlock/fs.c
1320
if (object->underobj == inode) {
security/landlock/fs.c
1321
object->underobj = NULL;
security/landlock/fs.c
1322
spin_unlock(&object->lock);
security/landlock/fs.c
1332
rcu_assign_pointer(landlock_inode(inode)->object, NULL);
security/landlock/fs.c
1342
spin_unlock(&object->lock);
security/landlock/fs.c
256
struct landlock_object *object, *new_object;
security/landlock/fs.c
261
object = rcu_dereference(inode_sec->object);
security/landlock/fs.c
262
if (object) {
security/landlock/fs.c
263
if (likely(refcount_inc_not_zero(&object->usage))) {
security/landlock/fs.c
265
return object;
security/landlock/fs.c
271
spin_lock(&object->lock);
security/landlock/fs.c
272
spin_unlock(&object->lock);
security/landlock/fs.c
290
if (unlikely(rcu_access_pointer(inode_sec->object))) {
security/landlock/fs.c
305
rcu_assign_pointer(inode_sec->object, new_object);
security/landlock/fs.c
342
id.key.object = get_inode_object(d_backing_inode(path->dentry));
security/landlock/fs.c
343
if (IS_ERR(id.key.object))
security/landlock/fs.c
344
return PTR_ERR(id.key.object);
security/landlock/fs.c
352
landlock_put_object(id.key.object);
security/landlock/fs.c
379
id.key.object = rcu_dereference(landlock_inode(inode)->object);
security/landlock/fs.c
55
static void release_inode(struct landlock_object *const object)
security/landlock/fs.c
56
__releases(object->lock)
security/landlock/fs.c
58
struct inode *const inode = object->underobj;
security/landlock/fs.c
62
spin_unlock(&object->lock);
security/landlock/fs.c
70
object->underobj = NULL;
security/landlock/fs.c
77
spin_unlock(&object->lock);
security/landlock/fs.c
84
rcu_assign_pointer(landlock_inode(inode)->object, NULL);
security/landlock/fs.h
40
struct landlock_object __rcu *object;
security/landlock/object.c
42
void landlock_put_object(struct landlock_object *const object)
security/landlock/object.c
49
if (!object)
security/landlock/object.c
58
if (refcount_dec_and_lock(&object->usage, &object->lock)) {
security/landlock/object.c
59
__acquire(&object->lock);
security/landlock/object.c
64
object->underops->release(object);
security/landlock/object.c
65
kfree_rcu(object, rcu_free);
security/landlock/object.h
25
void (*release)(struct landlock_object *const object)
security/landlock/object.h
26
__releases(object->lock);
security/landlock/object.h
83
void landlock_put_object(struct landlock_object *const object);
security/landlock/object.h
85
static inline void landlock_get_object(struct landlock_object *const object)
security/landlock/object.h
87
if (object)
security/landlock/object.h
88
refcount_inc(&object->usage);
security/landlock/ruleset.c
132
WARN_ON_ONCE(!id.key.object);
security/landlock/ruleset.c
133
landlock_get_object(id.key.object);
security/landlock/ruleset.c
172
landlock_put_object(rule->key.object);
security/landlock/ruleset.c
220
if (is_object_pointer(id.type) && WARN_ON_ONCE(!id.key.object))
security/landlock/ruleset.h
47
struct landlock_object *object;
security/smack/smack.h
262
char *object;
security/smack/smack_access.c
117
int smk_access(struct smack_known *subject, struct smack_known *object,
security/smack/smack_access.c
138
if (object == &smack_known_web || subject == &smack_known_web)
security/smack/smack_access.c
143
if (object == &smack_known_star)
security/smack/smack_access.c
149
if (subject->smk_known == object->smk_known)
security/smack/smack_access.c
157
if (object == &smack_known_floor)
security/smack/smack_access.c
170
may = smk_access_entry(subject->smk_known, object->smk_known,
security/smack/smack_access.c
192
if (object == smack_unconfined)
security/smack/smack_access.c
201
smack_log(subject->smk_known, object->smk_known,
security/smack/smack_access.c
329
audit_log_untrustedstring(ab, sad->object);
security/smack/smack_access.c
372
sad->object = object_label;
security/smack/smack_lsm.c
2725
struct smack_known *object,
security/smack/smack_lsm.c
2743
rc = smk_access(subject, object, MAY_WRITE, &ad);
security/smack/smack_lsm.c
2744
rc = smk_bu_note("IPv6 check", subject, object, MAY_WRITE, rc);
security/smack/smack_lsm.c
2852
struct smack_known *object;
security/smack/smack_lsm.c
2856
object = ssp->smk_in;
security/smack/smack_lsm.c
2859
object = smack_ipv6host_label(address);
security/smack/smack_lsm.c
2865
if (skp != NULL && object != NULL)
security/smack/smack_lsm.c
2866
return smk_ipv6_check(skp, object, address, act);
security/smack/smack_lsm.c
2869
if (object == NULL)
security/smack/smack_lsm.c
2870
object = smack_net_ambient;
security/smack/smack_lsm.c
2876
return smk_ipv6_check(skp, object, address, act);
security/smack/smack_lsm.c
2889
object = spp->smk_in;
security/smack/smack_lsm.c
2896
return smk_ipv6_check(skp, object, address, act);
security/smack/smack_lsm.c
972
const struct smack_known *object)
security/smack/smack_lsm.c
977
may = smk_access_entry(subject->smk_known, object->smk_known,
security/smack/smackfs.c
316
static int smk_fill_rule(const char *subject, const char *object,
security/smack/smackfs.c
329
rule->smk_object = smk_import_entry(object, len);
security/smack/smackfs.c
342
cp = smk_parse_smack(object, len);
sound/pci/asihpi/hpi6000.c
1708
switch (phm->object) {
sound/pci/asihpi/hpi6000.c
1745
hm.object = phm->u.d.u.stream.object_type;
sound/pci/asihpi/hpi6000.c
325
if (phm->object != HPI_OBJ_SUBSYSTEM) {
sound/pci/asihpi/hpi6000.c
328
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpi6000.c
337
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpi6000.c
346
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpi6000.c
351
switch (phm->object) {
sound/pci/asihpi/hpi6000.c
531
hm.object = HPI_OBJ_ADAPTER;
sound/pci/asihpi/hpi6205.c
1023
hpi_init_response(phr, phm->object, phm->function, 0);
sound/pci/asihpi/hpi6205.c
1191
hpi_init_response(phr, phm->object, phm->function, 0);
sound/pci/asihpi/hpi6205.c
1240
hpi_init_response(phr, phm->object, phm->function, 0);
sound/pci/asihpi/hpi6205.c
377
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpi6205.c
379
HPI_DEBUG_LOG(WARNING, " %d,%d dsp crashed.\n", phm->object,
sound/pci/asihpi/hpi6205.c
391
switch (phm->object) {
sound/pci/asihpi/hpi6205.c
428
if (phm->object != HPI_OBJ_SUBSYSTEM) {
sound/pci/asihpi/hpi6205.c
440
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpi6205.c
633
hm.object = HPI_OBJ_ADAPTER;
sound/pci/asihpi/hpi6205.c
737
hpi_init_response(phr, phm->object, phm->function, 0);
sound/pci/asihpi/hpi6205.c
906
hpi_init_response(phr, phm->object, phm->function, 0);
sound/pci/asihpi/hpi6205.c
978
hpi_init_response(phr, phm->object, phm->function, 0);
sound/pci/asihpi/hpi_internal.h
1002
u16 object; /* HPI_OBJ_* */
sound/pci/asihpi/hpi_internal.h
1053
u16 object; /* HPI_OBJ_* */
sound/pci/asihpi/hpi_internal.h
1064
u16 object; /* HPI_OBJ_* */
sound/pci/asihpi/hpi_internal.h
991
u16 object; /* HPI_OBJ_* */
sound/pci/asihpi/hpicmn.c
44
if (phr->object != phm->object) {
sound/pci/asihpi/hpicmn.c
46
phr->object);
sound/pci/asihpi/hpicmn.c
522
phr->object = phm->object;
sound/pci/asihpi/hpicmn.c
701
switch (phm->object) {
sound/pci/asihpi/hpimsginit.c
103
u16 object, u16 function)
sound/pci/asihpi/hpimsginit.c
105
(void)object;
sound/pci/asihpi/hpimsginit.c
115
struct hpi_response_header *phr, u16 res_size, u16 object,
sound/pci/asihpi/hpimsginit.c
118
hpi_init_messageV1(phm, msg_size, object, function);
sound/pci/asihpi/hpimsginit.c
119
hpi_init_responseV1(phr, res_size, object, function);
sound/pci/asihpi/hpimsginit.c
27
static void hpi_init_message(struct hpi_message *phm, u16 object,
sound/pci/asihpi/hpimsginit.c
32
if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) {
sound/pci/asihpi/hpimsginit.c
33
object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1);
sound/pci/asihpi/hpimsginit.c
34
size = msg_size[object];
sound/pci/asihpi/hpimsginit.c
46
phm->object = object;
sound/pci/asihpi/hpimsginit.c
56
void hpi_init_response(struct hpi_response *phr, u16 object, u16 function,
sound/pci/asihpi/hpimsginit.c
61
if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) {
sound/pci/asihpi/hpimsginit.c
62
object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1);
sound/pci/asihpi/hpimsginit.c
63
size = res_size[object];
sound/pci/asihpi/hpimsginit.c
71
phr->object = object;
sound/pci/asihpi/hpimsginit.c
79
struct hpi_response *phr, u16 object, u16 function)
sound/pci/asihpi/hpimsginit.c
81
hpi_init_message(phm, object, function);
sound/pci/asihpi/hpimsginit.c
84
hpi_init_response(phr, object, function,
sound/pci/asihpi/hpimsginit.c
89
u16 object, u16 function)
sound/pci/asihpi/hpimsginit.c
92
if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) {
sound/pci/asihpi/hpimsginit.c
95
phm->object = object;
sound/pci/asihpi/hpimsginit.h
22
void hpi_init_response(struct hpi_response *phr, u16 object, u16 function,
sound/pci/asihpi/hpimsginit.h
26
struct hpi_response *phr, u16 object, u16 function);
sound/pci/asihpi/hpimsginit.h
29
u16 object, u16 function);
sound/pci/asihpi/hpimsginit.h
32
struct hpi_response_header *phr, u16 res_size, u16 object,
sound/pci/asihpi/hpimsgx.c
305
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpimsgx.c
312
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpimsgx.c
317
switch (phm->object) {
sound/pci/asihpi/hpimsgx.c
68
hpi_init_response(phr, phm->object, phm->function,
sound/pci/asihpi/hpimsgx.c
700
hpi_init_response(&hr, phm->object, phm->function,
sound/pci/asihpi/hpioctl.c
173
if (hm->h.object == HPI_OBJ_SUBSYSTEM) {
sound/pci/asihpi/hpioctl.c
187
hpi_init_response(&hr->r0, hm->h.object,
sound/pci/asihpi/hpioctl.c
62
&& (phm->object != HPI_OBJ_SUBSYSTEM))
sound/soc/intel/avs/topology.c
127
int (*parse)(struct snd_soc_component *comp, void *elem, void *object, u32 offset);
sound/soc/intel/avs/topology.c
131
avs_parse_uuid_token(struct snd_soc_component *comp, void *elem, void *object, u32 offset)
sound/soc/intel/avs/topology.c
134
guid_t *val = (guid_t *)((u8 *)object + offset);
sound/soc/intel/avs/topology.c
142
avs_parse_bool_token(struct snd_soc_component *comp, void *elem, void *object, u32 offset)
sound/soc/intel/avs/topology.c
145
bool *val = (bool *)((u8 *)object + offset);
sound/soc/intel/avs/topology.c
153
avs_parse_byte_token(struct snd_soc_component *comp, void *elem, void *object, u32 offset)
sound/soc/intel/avs/topology.c
156
u8 *val = ((u8 *)object + offset);
sound/soc/intel/avs/topology.c
164
avs_parse_short_token(struct snd_soc_component *comp, void *elem, void *object, u32 offset)
sound/soc/intel/avs/topology.c
167
u16 *val = (u16 *)((u8 *)object + offset);
sound/soc/intel/avs/topology.c
175
avs_parse_word_token(struct snd_soc_component *comp, void *elem, void *object, u32 offset)
sound/soc/intel/avs/topology.c
178
u32 *val = (u32 *)((u8 *)object + offset);
sound/soc/intel/avs/topology.c
186
avs_parse_string_token(struct snd_soc_component *comp, void *elem, void *object, u32 offset)
sound/soc/intel/avs/topology.c
189
char *val = (char *)((u8 *)object + offset);
sound/soc/intel/avs/topology.c
196
static int avs_parse_uuid_tokens(struct snd_soc_component *comp, void *object,
sound/soc/intel/avs/topology.c
213
ret = parsers[j].parse(comp, tuple, object, parsers[j].offset);
sound/soc/intel/avs/topology.c
222
static int avs_parse_string_tokens(struct snd_soc_component *comp, void *object,
sound/soc/intel/avs/topology.c
239
ret = parsers[j].parse(comp, tuple, object, parsers[j].offset);
sound/soc/intel/avs/topology.c
248
static int avs_parse_word_tokens(struct snd_soc_component *comp, void *object,
sound/soc/intel/avs/topology.c
270
ret = parsers[j].parse(comp, tuple, object, parsers[j].offset);
sound/soc/intel/avs/topology.c
279
static int avs_parse_tokens(struct snd_soc_component *comp, void *object,
sound/soc/intel/avs/topology.c
302
ret = avs_parse_uuid_tokens(comp, object, parsers, count, tuples);
sound/soc/intel/avs/topology.c
305
ret = avs_parse_string_tokens(comp, object, parsers, count, tuples);
sound/soc/intel/avs/topology.c
311
ret = avs_parse_word_tokens(comp, object, parsers, count, tuples);
sound/soc/intel/avs/topology.c
332
avs_parse_##name##_ptr(struct snd_soc_component *comp, void *elem, void *object, u32 offset) \
sound/soc/intel/avs/topology.c
336
type **val = (type **)(object + offset); \
sound/soc/intel/avs/topology.c
356
parse_audio_format_bitfield(struct snd_soc_component *comp, void *elem, void *object, u32 offset)
sound/soc/intel/avs/topology.c
359
struct avs_audio_format *audio_format = object;
sound/soc/intel/avs/topology.c
399
void *object, u32 offset)
sound/soc/intel/avs/topology.c
403
char *val = (char *)((u8 *)object + offset);
sound/soc/intel/avs/topology.c
411
return avs_parse_string_token(comp, elem, object, offset);
sound/soc/intel/avs/topology.c
415
return avs_parse_string_token(comp, elem, object, offset);
sound/soc/intel/avs/topology.c
424
static int avs_parse_nhlt_config_size(struct snd_soc_component *comp, void *elem, void *object,
sound/soc/intel/avs/topology.c
428
struct acpi_nhlt_config **blob = (struct acpi_nhlt_config **)((u8 *)object + offset);
sound/soc/sof/ipc3-topology.c
64
static int get_token_process_type(void *elem, void *object, u32 offset)
sound/soc/sof/ipc3-topology.c
66
u32 *val = (u32 *)((u8 *)object + offset);
sound/soc/sof/ipc4-topology.c
62
static int get_token_comp_domain(void *elem, void *object, u32 offset)
sound/soc/sof/ipc4-topology.c
64
u32 *val = (u32 *)((u8 *)object + offset);
sound/soc/sof/sof-audio.h
308
int (*get_token)(void *elem, void *object, u32 offset);
sound/soc/sof/sof-audio.h
695
int get_token_u32(void *elem, void *object, u32 offset);
sound/soc/sof/sof-audio.h
696
int get_token_u16(void *elem, void *object, u32 offset);
sound/soc/sof/sof-audio.h
697
int get_token_comp_format(void *elem, void *object, u32 offset);
sound/soc/sof/sof-audio.h
698
int get_token_dai_type(void *elem, void *object, u32 offset);
sound/soc/sof/sof-audio.h
699
int get_token_uuid(void *elem, void *object, u32 offset);
sound/soc/sof/sof-audio.h
700
int get_token_string(void *elem, void *object, u32 offset);
sound/soc/sof/sof-audio.h
701
int sof_update_ipc_object(struct snd_soc_component *scomp, void *object, enum sof_tokens token_id,
sound/soc/sof/topology.c
104
u16 *val = (u16 *)((u8 *)object + tokens[i].offset +
sound/soc/sof/topology.c
119
tokens[i].get_token((void *)tuples[j].value.s, object,
sound/soc/sof/topology.c
1394
static int get_w_no_wname_in_long_name(void *elem, void *object, u32 offset)
sound/soc/sof/topology.c
1397
struct snd_soc_dapm_widget *w = object;
sound/soc/sof/topology.c
349
int get_token_u32(void *elem, void *object, u32 offset)
sound/soc/sof/topology.c
352
u32 *val = (u32 *)((u8 *)object + offset);
sound/soc/sof/topology.c
358
int get_token_u16(void *elem, void *object, u32 offset)
sound/soc/sof/topology.c
361
u16 *val = (u16 *)((u8 *)object + offset);
sound/soc/sof/topology.c
367
int get_token_uuid(void *elem, void *object, u32 offset)
sound/soc/sof/topology.c
370
u8 *dst = (u8 *)object + offset;
sound/soc/sof/topology.c
381
int get_token_string(void *elem, void *object, u32 offset)
sound/soc/sof/topology.c
384
char **dst = (char **)((u8 *)object + offset);
sound/soc/sof/topology.c
392
int get_token_comp_format(void *elem, void *object, u32 offset)
sound/soc/sof/topology.c
394
u32 *val = (u32 *)((u8 *)object + offset);
sound/soc/sof/topology.c
400
int get_token_dai_type(void *elem, void *object, u32 offset)
sound/soc/sof/topology.c
402
u32 *val = (u32 *)((u8 *)object + offset);
sound/soc/sof/topology.c
457
void *object, size_t offset,
sound/soc/sof/topology.c
480
tokens[j].get_token(elem, object,
sound/soc/sof/topology.c
56
int sof_update_ipc_object(struct snd_soc_component *scomp, void *object, enum sof_tokens token_id,
sound/soc/sof/topology.c
628
void *object, int offset,
sound/soc/sof/topology.c
651
ret = tokens[j].get_token(elem->string, object, offset + tokens[j].offset);
sound/soc/sof/topology.c
674
void *object, int offset,
sound/soc/sof/topology.c
700
tokens[j].get_token(elem, object, offset + tokens[j].offset);
sound/soc/sof/topology.c
725
void *object, const struct sof_topology_token *tokens,
sound/soc/sof/topology.c
756
found += sof_parse_uuid_tokens(scomp, object, offset, tokens, count,
sound/soc/sof/topology.c
761
ret = sof_parse_string_tokens(scomp, object, offset, tokens, count,
sound/soc/sof/topology.c
774
found += sof_parse_word_tokens(scomp, object, offset, tokens, count,
sound/soc/sof/topology.c
810
static int sof_parse_tokens(struct snd_soc_component *scomp, void *object,
sound/soc/sof/topology.c
822
return sof_parse_token_sets(scomp, object, tokens, num_tokens, array,
sound/soc/sof/topology.c
95
u32 *val = (u32 *)((u8 *)object + tokens[i].offset +
tools/bpf/resolve_btfids/main.c
1041
static int build_btf2btf_context(struct object *obj, struct btf2btf_context *ctx)
tools/bpf/resolve_btfids/main.c
1213
static int btf2btf(struct object *obj)
tools/bpf/resolve_btfids/main.c
1306
static int finalize_btf(struct object *obj)
tools/bpf/resolve_btfids/main.c
1471
struct object obj = {
tools/bpf/resolve_btfids/main.c
312
static struct btf_id *add_set(struct object *obj, char *name, enum btf_id_kind kind)
tools/bpf/resolve_btfids/main.c
383
static int elf_collect(struct object *obj)
tools/bpf/resolve_btfids/main.c
483
static int symbols_collect(struct object *obj)
tools/bpf/resolve_btfids/main.c
581
static int load_btf(struct object *obj)
tools/bpf/resolve_btfids/main.c
617
static int symbols_resolve(struct object *obj)
tools/bpf/resolve_btfids/main.c
687
static int id_patch(struct object *obj, struct btf_id *id)
tools/bpf/resolve_btfids/main.c
719
static int __symbols_patch(struct object *obj, struct rb_root *root)
tools/bpf/resolve_btfids/main.c
743
static int sets_patch(struct object *obj)
tools/bpf/resolve_btfids/main.c
795
static int symbols_patch(struct object *obj)
tools/bpf/resolve_btfids/main.c
835
static int dump_raw_btf_ids(struct object *obj, const char *out_path)
tools/bpf/resolve_btfids/main.c
954
static u32 find_kfunc_flags(struct object *obj, struct btf_id *kfunc_id)
tools/bpf/resolve_btfids/main.c
992
static int collect_kfuncs(struct object *obj, struct btf2btf_context *ctx)
tools/lib/bpf/libbpf.h
328
LIBBPF_API int bpf_object__pin(struct bpf_object *object, const char *path);
tools/lib/bpf/libbpf.h
329
LIBBPF_API int bpf_object__unpin(struct bpf_object *object, const char *path);
tools/lib/perf/include/internal/rc_check.h
101
if (object) { \
tools/lib/perf/include/internal/rc_check.h
102
object->orig = NULL; \
tools/lib/perf/include/internal/rc_check.h
103
free(object); \
tools/lib/perf/include/internal/rc_check.h
43
#define ADD_RC_CHK(result, object) (result = object, object)
tools/lib/perf/include/internal/rc_check.h
46
#define RC_CHK_ACCESS(object) object
tools/lib/perf/include/internal/rc_check.h
49
#define RC_CHK_FREE(object) free(object)
tools/lib/perf/include/internal/rc_check.h
52
#define RC_CHK_GET(result, object) ADD_RC_CHK(result, object)
tools/lib/perf/include/internal/rc_check.h
55
#define RC_CHK_PUT(object) {}
tools/lib/perf/include/internal/rc_check.h
77
#define ADD_RC_CHK(result, object) \
tools/lib/perf/include/internal/rc_check.h
79
object ? (result = malloc(sizeof(*result)), \
tools/lib/perf/include/internal/rc_check.h
80
result ? (result->orig = object, result) \
tools/lib/perf/include/internal/rc_check.h
86
#define RC_CHK_ACCESS(object) object->orig
tools/lib/perf/include/internal/rc_check.h
89
#define RC_CHK_FREE(object) \
tools/lib/perf/include/internal/rc_check.h
91
zfree(&object->orig); \
tools/lib/perf/include/internal/rc_check.h
92
free(object); \
tools/lib/perf/include/internal/rc_check.h
96
#define RC_CHK_GET(result, object) ADD_RC_CHK(result, (object ? object->orig : NULL))
tools/lib/perf/include/internal/rc_check.h
99
#define RC_CHK_PUT(object) \
tools/perf/util/llvm-c-helpers.cpp
166
object::SectionedAddress sectioned_addr = {
tools/perf/util/llvm-c-helpers.cpp
168
object::SectionedAddress::UndefSection
tools/perf/util/llvm-c-helpers.cpp
184
object::SectionedAddress sectioned_addr = {
tools/perf/util/llvm-c-helpers.cpp
186
object::SectionedAddress::UndefSection
tools/perf/util/llvm-c-helpers.cpp
80
object::SectionedAddress sectioned_addr = {
tools/perf/util/llvm-c-helpers.cpp
82
object::SectionedAddress::UndefSection
tools/power/x86/intel-speed-select/isst-config.c
1280
#define _get_tdp_level(desc, suffix, object, help, str0, str1) \
tools/power/x86/intel-speed-select/isst-config.c
1281
static void get_tdp_##object(int arg) \
tools/power/x86/intel-speed-select/isst-config.c
1297
&ctdp, desc, &ctdp.object); \
tools/power/x86/intel-speed-select/isst-config.c
1302
&ctdp.object); \
tools/testing/selftests/bpf/usdt.h
393
__usdt_asm2( .type sema, @object) \
tools/testing/selftests/kselftest_harness.h
474
struct __test_metadata *object = mmap(NULL, sizeof(*object), \
tools/testing/selftests/kselftest_harness.h
476
object->name = #test_name; \
tools/testing/selftests/kselftest_harness.h
477
object->fn = &wrapper_##fixture_name##_##test_name; \
tools/testing/selftests/kselftest_harness.h
478
object->fixture = &_##fixture_name##_fixture_object; \
tools/testing/selftests/kselftest_harness.h
479
object->teardown_fn = &wrapper_##fixture_name##_##test_name##_teardown; \
tools/testing/selftests/kselftest_harness.h
480
object->termsig = signal; \
tools/testing/selftests/kselftest_harness.h
481
object->timeout = tmout; \
tools/testing/selftests/kselftest_harness.h
482
_##fixture_name##_##test_name##_object = object; \
tools/testing/selftests/kselftest_harness.h
483
__register_test(object); \
tools/usb/usbip/libsrc/usbip_common.c
199
#define READ_ATTR(object, type, dev, name, format) \
tools/usb/usbip/libsrc/usbip_common.c
201
(object)->name = (type) read_attr_value(dev, to_string(name), \