Symbol: cli
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1043
struct nouveau_drm *drm = chan->cli->drm;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1099
struct nouveau_drm *drm = chan->cli->drm;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1148
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1158
cli = chan->cli;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1171
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1220
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1236
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1288
struct nouveau_cli *cli = &nouveau_drm(dev)->client;
drivers/gpu/drm/nouveau/dispnv04/crtc.c
1322
ret = nouveau_bo_new_map(cli, NOUVEAU_GEM_DOMAIN_VRAM, 64 * 64 * 4, &nv_crtc->cursor.nvbo);
drivers/gpu/drm/nouveau/nouveau_abi16.c
214
struct nouveau_cli *cli = abi16->cli;
drivers/gpu/drm/nouveau/nouveau_abi16.c
228
kfree(cli->abi16);
drivers/gpu/drm/nouveau/nouveau_abi16.c
229
cli->abi16 = NULL;
drivers/gpu/drm/nouveau/nouveau_abi16.c
249
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_abi16.c
329
NV_PRINTK(dbg, cli, "unknown parameter %lld\n", getparam->param);
drivers/gpu/drm/nouveau/nouveau_abi16.c
340
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_abi16.c
344
struct nvif_device *device = &cli->device;
drivers/gpu/drm/nouveau/nouveau_abi16.c
359
__nouveau_cli_disable_uvmm_noinit(cli);
drivers/gpu/drm/nouveau/nouveau_abi16.c
408
ret = nouveau_channel_new(cli, false, runm, init->fb_ctxdma_handle,
drivers/gpu/drm/nouveau/nouveau_abi16.c
417
if (nouveau_cli_uvmm(cli)) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
44
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_abi16.c
45
if (!cli->abi16) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
469
ret = nouveau_gem_new(cli, PAGE_SIZE, 0, NOUVEAU_GEM_DOMAIN_GART,
drivers/gpu/drm/nouveau/nouveau_abi16.c
47
cli->abi16 = abi16 = kzalloc_obj(*abi16);
drivers/gpu/drm/nouveau/nouveau_abi16.c
48
if (cli->abi16) {
drivers/gpu/drm/nouveau/nouveau_abi16.c
49
abi16->cli = cli;
drivers/gpu/drm/nouveau/nouveau_abi16.c
54
return cli->abi16;
drivers/gpu/drm/nouveau/nouveau_abi16.c
60
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_abi16.c
61
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_abi16.c
628
device = &abi16->cli->device;
drivers/gpu/drm/nouveau/nouveau_abi16.c
63
return cli->abi16;
drivers/gpu/drm/nouveau/nouveau_abi16.c
64
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_abi16.c
71
struct nouveau_cli *cli = abi16->cli;
drivers/gpu/drm/nouveau/nouveau_abi16.c
715
struct nouveau_cli *cli = abi16->cli;
drivers/gpu/drm/nouveau/nouveau_abi16.c
72
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_abi16.c
737
info = &cli->device.info;
drivers/gpu/drm/nouveau/nouveau_abi16.h
33
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/nouveau_bo.c
211
nouveau_bo_alloc(struct nouveau_cli *cli, u64 *size, int *align, u32 domain,
drivers/gpu/drm/nouveau/nouveau_bo.c
214
struct nouveau_drm *drm = cli->drm;
drivers/gpu/drm/nouveau/nouveau_bo.c
216
struct nvif_mmu *mmu = &cli->mmu;
drivers/gpu/drm/nouveau/nouveau_bo.c
217
struct nvif_vmm *vmm = &nouveau_cli_vmm(cli)->vmm;
drivers/gpu/drm/nouveau/nouveau_bo.c
248
if (cli->device.info.family >= NV_DEVICE_INFO_V0_FERMI) {
drivers/gpu/drm/nouveau/nouveau_bo.c
256
} else if (cli->device.info.family >= NV_DEVICE_INFO_V0_TESLA) {
drivers/gpu/drm/nouveau/nouveau_bo.c
268
if (!nouveau_cli_uvmm(cli) || internal) {
drivers/gpu/drm/nouveau/nouveau_bo.c
278
if (cli->device.info.family > NV_DEVICE_INFO_V0_CURIE &&
drivers/gpu/drm/nouveau/nouveau_bo.c
374
nouveau_bo_new(struct nouveau_cli *cli, u64 size, int align,
drivers/gpu/drm/nouveau/nouveau_bo.c
382
nvbo = nouveau_bo_alloc(cli, &size, &align, domain, tile_mode,
drivers/gpu/drm/nouveau/nouveau_bo.c
420
nouveau_bo_new_pin(struct nouveau_cli *cli, u32 domain, u32 size, struct nouveau_bo **pnvbo)
drivers/gpu/drm/nouveau/nouveau_bo.c
425
ret = nouveau_bo_new(cli, size, 0, domain, 0, 0, NULL, NULL, &nvbo);
drivers/gpu/drm/nouveau/nouveau_bo.c
440
nouveau_bo_new_map(struct nouveau_cli *cli, u32 domain, u32 size, struct nouveau_bo **pnvbo)
drivers/gpu/drm/nouveau/nouveau_bo.c
445
ret = nouveau_bo_new_pin(cli, domain, size, &nvbo);
drivers/gpu/drm/nouveau/nouveau_bo.c
460
nouveau_bo_new_map_gpu(struct nouveau_cli *cli, u32 domain, u32 size,
drivers/gpu/drm/nouveau/nouveau_bo.c
463
struct nouveau_vmm *vmm = nouveau_cli_vmm(cli);
drivers/gpu/drm/nouveau/nouveau_bo.c
467
ret = nouveau_bo_new_map(cli, domain, size, &nvbo);
drivers/gpu/drm/nouveau/nouveau_bo.c
942
struct nouveau_cli *cli = chan->cli;
drivers/gpu/drm/nouveau/nouveau_bo.c
957
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_bo.c
959
mutex_lock_nested(&cli->mutex, SINGLE_DEPTH_NESTING);
drivers/gpu/drm/nouveau/nouveau_bo.c
987
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_bo0039.c
107
PUSH_MTHD(push, NV039, SET_CONTEXT_DMA_NOTIFIES, chan->cli->drm->ntfy.handle);
drivers/gpu/drm/nouveau/nouveau_bo5039.c
147
PUSH_MTHD(push, NV5039, SET_CONTEXT_DMA_NOTIFY, chan->cli->drm->ntfy.handle,
drivers/gpu/drm/nouveau/nouveau_chan.c
142
nouveau_channel_prep(struct nouveau_cli *cli,
drivers/gpu/drm/nouveau/nouveau_chan.c
145
struct nouveau_drm *drm = cli->drm;
drivers/gpu/drm/nouveau/nouveau_chan.c
146
struct nvif_device *device = &cli->device;
drivers/gpu/drm/nouveau/nouveau_chan.c
156
chan->cli = cli;
drivers/gpu/drm/nouveau/nouveau_chan.c
157
chan->vmm = nouveau_cli_vmm(cli);
drivers/gpu/drm/nouveau/nouveau_chan.c
165
ret = nouveau_bo_new_map(cli, target, size, &chan->push.buffer);
drivers/gpu/drm/nouveau/nouveau_chan.c
171
chan->chan.push.mem.object.parent = cli->base.object.parent;
drivers/gpu/drm/nouveau/nouveau_chan.c
172
chan->chan.push.mem.object.client = &cli->base;
drivers/gpu/drm/nouveau/nouveau_chan.c
195
return nouveau_bo_new_map_gpu(cli, NOUVEAU_GEM_DOMAIN_GART, PAGE_SIZE,
drivers/gpu/drm/nouveau/nouveau_chan.c
248
nouveau_channel_ctor(struct nouveau_cli *cli, bool priv, u64 runm,
drivers/gpu/drm/nouveau/nouveau_chan.c
273
struct nvif_device *device = &cli->device;
drivers/gpu/drm/nouveau/nouveau_chan.c
291
ret = nouveau_channel_prep(cli, size, &chan);
drivers/gpu/drm/nouveau/nouveau_chan.c
322
ret = nvif_mem_ctor(&cli->mmu, "abi16ChanUSERD", NVIF_CLASS_MEM_GF100,
drivers/gpu/drm/nouveau/nouveau_chan.c
356
struct nouveau_cli *cli = chan->cli;
drivers/gpu/drm/nouveau/nouveau_chan.c
357
struct nouveau_drm *drm = cli->drm;
drivers/gpu/drm/nouveau/nouveau_chan.c
358
struct nvif_device *device = &cli->device;
drivers/gpu/drm/nouveau/nouveau_chan.c
494
nouveau_channel_new(struct nouveau_cli *cli,
drivers/gpu/drm/nouveau/nouveau_chan.c
499
ret = nouveau_channel_ctor(cli, priv, runm, pchan);
drivers/gpu/drm/nouveau/nouveau_chan.c
501
NV_PRINTK(dbg, cli, "channel create, %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_chan.c
507
NV_PRINTK(err, cli, "channel failed to initialise, %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_chan.c
55
struct nouveau_cli *cli = chan->cli;
drivers/gpu/drm/nouveau/nouveau_chan.c
57
NV_PRINTK(warn, cli, "channel %d killed!\n", chan->chid);
drivers/gpu/drm/nouveau/nouveau_chan.c
69
struct nouveau_cli *cli = chan->cli;
drivers/gpu/drm/nouveau/nouveau_chan.c
80
NV_PRINTK(err, cli, "failed to idle channel %d [%s]\n",
drivers/gpu/drm/nouveau/nouveau_chan.c
81
chan->chid, cli->name);
drivers/gpu/drm/nouveau/nouveau_chan.c
94
nouveau_fence(chan->cli->drm)->context_del(chan);
drivers/gpu/drm/nouveau/nouveau_chan.h
12
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/nouveau_debugfs.c
224
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/nouveau_debugfs.c
227
list_for_each_entry(cli, &drm->clients, head) {
drivers/gpu/drm/nouveau/nouveau_debugfs.c
228
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(cli);
drivers/gpu/drm/nouveau/nouveau_display.c
806
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_display.c
821
ret = nouveau_gem_new(cli, args->size, 0, domain, 0, 0, &bo);
drivers/gpu/drm/nouveau/nouveau_drm.c
1190
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/nouveau_drm.c
1206
if (!(cli = kzalloc_obj(*cli))) {
drivers/gpu/drm/nouveau/nouveau_drm.c
1211
ret = nouveau_cli_init(drm, name, cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
1215
fpriv->driver_priv = cli;
drivers/gpu/drm/nouveau/nouveau_drm.c
1218
list_add(&cli->head, &drm->clients);
drivers/gpu/drm/nouveau/nouveau_drm.c
1222
if (ret && cli) {
drivers/gpu/drm/nouveau/nouveau_drm.c
1223
nouveau_cli_fini(cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
1224
kfree(cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
1235
struct nouveau_cli *cli = nouveau_cli(fpriv);
drivers/gpu/drm/nouveau/nouveau_drm.c
1250
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drm.c
1251
if (cli->abi16)
drivers/gpu/drm/nouveau/nouveau_drm.c
1252
nouveau_abi16_fini(cli->abi16);
drivers/gpu/drm/nouveau/nouveau_drm.c
1253
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drm.c
1256
list_del(&cli->head);
drivers/gpu/drm/nouveau/nouveau_drm.c
1259
nouveau_cli_fini(cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
1260
kfree(cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
174
struct nouveau_cli *cli = container_of(w, typeof(*cli), work);
drivers/gpu/drm/nouveau/nouveau_drm.c
176
mutex_lock(&cli->lock);
drivers/gpu/drm/nouveau/nouveau_drm.c
177
list_for_each_entry_safe(work, wtmp, &cli->worker, head) {
drivers/gpu/drm/nouveau/nouveau_drm.c
183
mutex_unlock(&cli->lock);
drivers/gpu/drm/nouveau/nouveau_drm.c
190
schedule_work(&work->cli->work);
drivers/gpu/drm/nouveau/nouveau_drm.c
194
nouveau_cli_work_queue(struct nouveau_cli *cli, struct dma_fence *fence,
drivers/gpu/drm/nouveau/nouveau_drm.c
198
work->cli = cli;
drivers/gpu/drm/nouveau/nouveau_drm.c
199
mutex_lock(&cli->lock);
drivers/gpu/drm/nouveau/nouveau_drm.c
200
list_add_tail(&work->head, &cli->worker);
drivers/gpu/drm/nouveau/nouveau_drm.c
203
mutex_unlock(&cli->lock);
drivers/gpu/drm/nouveau/nouveau_drm.c
207
nouveau_cli_fini(struct nouveau_cli *cli)
drivers/gpu/drm/nouveau/nouveau_drm.c
209
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm_locked(cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
216
flush_work(&cli->work);
drivers/gpu/drm/nouveau/nouveau_drm.c
217
WARN_ON(!list_empty(&cli->worker));
drivers/gpu/drm/nouveau/nouveau_drm.c
219
if (cli->sched)
drivers/gpu/drm/nouveau/nouveau_drm.c
220
nouveau_sched_destroy(&cli->sched);
drivers/gpu/drm/nouveau/nouveau_drm.c
223
nouveau_vmm_fini(&cli->svm);
drivers/gpu/drm/nouveau/nouveau_drm.c
224
nouveau_vmm_fini(&cli->vmm);
drivers/gpu/drm/nouveau/nouveau_drm.c
225
nvif_mmu_dtor(&cli->mmu);
drivers/gpu/drm/nouveau/nouveau_drm.c
226
cli->device.object.map.ptr = NULL;
drivers/gpu/drm/nouveau/nouveau_drm.c
227
nvif_device_dtor(&cli->device);
drivers/gpu/drm/nouveau/nouveau_drm.c
228
mutex_lock(&cli->drm->client_mutex);
drivers/gpu/drm/nouveau/nouveau_drm.c
229
nvif_client_dtor(&cli->base);
drivers/gpu/drm/nouveau/nouveau_drm.c
230
mutex_unlock(&cli->drm->client_mutex);
drivers/gpu/drm/nouveau/nouveau_drm.c
235
struct nouveau_cli *cli)
drivers/gpu/drm/nouveau/nouveau_drm.c
255
snprintf(cli->name, sizeof(cli->name), "%s", sname);
drivers/gpu/drm/nouveau/nouveau_drm.c
256
cli->drm = drm;
drivers/gpu/drm/nouveau/nouveau_drm.c
257
mutex_init(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drm.c
259
INIT_WORK(&cli->work, nouveau_cli_work);
drivers/gpu/drm/nouveau/nouveau_drm.c
260
INIT_LIST_HEAD(&cli->worker);
drivers/gpu/drm/nouveau/nouveau_drm.c
261
mutex_init(&cli->lock);
drivers/gpu/drm/nouveau/nouveau_drm.c
264
ret = nvif_client_ctor(&drm->_client, cli->name, &cli->base);
drivers/gpu/drm/nouveau/nouveau_drm.c
267
NV_PRINTK(err, cli, "Client allocation failed: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_drm.c
271
ret = nvif_device_ctor(&cli->base, "drmDevice", &cli->device);
drivers/gpu/drm/nouveau/nouveau_drm.c
273
NV_PRINTK(err, cli, "Device allocation failed: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_drm.c
277
cli->device.object.map.ptr = drm->device.object.map.ptr;
drivers/gpu/drm/nouveau/nouveau_drm.c
279
ret = nvif_mmu_ctor(&cli->device.object, "drmMmu", drm->mmu.object.oclass,
drivers/gpu/drm/nouveau/nouveau_drm.c
280
&cli->mmu);
drivers/gpu/drm/nouveau/nouveau_drm.c
282
NV_PRINTK(err, cli, "MMU allocation failed: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_drm.c
286
ret = nvif_mclass(&cli->mmu.object, vmms);
drivers/gpu/drm/nouveau/nouveau_drm.c
288
NV_PRINTK(err, cli, "No supported VMM class\n");
drivers/gpu/drm/nouveau/nouveau_drm.c
292
ret = nouveau_vmm_init(cli, vmms[ret].oclass, &cli->vmm);
drivers/gpu/drm/nouveau/nouveau_drm.c
294
NV_PRINTK(err, cli, "VMM allocation failed: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_drm.c
298
ret = nvif_mclass(&cli->mmu.object, mems);
drivers/gpu/drm/nouveau/nouveau_drm.c
300
NV_PRINTK(err, cli, "No supported MEM class\n");
drivers/gpu/drm/nouveau/nouveau_drm.c
304
cli->mem = &mems[ret];
drivers/gpu/drm/nouveau/nouveau_drm.c
316
ret = nouveau_sched_create(&cli->sched, drm, NULL, 1);
drivers/gpu/drm/nouveau/nouveau_drm.c
323
nouveau_cli_fini(cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
582
struct nouveau_cli *cli, *temp_cli;
drivers/gpu/drm/nouveau/nouveau_drm.c
612
list_for_each_entry_safe(cli, temp_cli, &drm->clients, head) {
drivers/gpu/drm/nouveau/nouveau_drm.c
613
list_del(&cli->head);
drivers/gpu/drm/nouveau/nouveau_drm.c
614
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drm.c
615
if (cli->abi16)
drivers/gpu/drm/nouveau/nouveau_drm.c
616
nouveau_abi16_fini(cli->abi16);
drivers/gpu/drm/nouveau/nouveau_drm.c
617
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drm.c
618
nouveau_cli_fini(cli);
drivers/gpu/drm/nouveau/nouveau_drm.c
619
kfree(cli);
drivers/gpu/drm/nouveau/nouveau_drv.h
121
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/nouveau_drv.h
129
nouveau_cli_uvmm(struct nouveau_cli *cli)
drivers/gpu/drm/nouveau/nouveau_drv.h
131
return cli ? cli->uvmm.ptr : NULL;
drivers/gpu/drm/nouveau/nouveau_drv.h
135
nouveau_cli_uvmm_locked(struct nouveau_cli *cli)
drivers/gpu/drm/nouveau/nouveau_drv.h
139
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drv.h
140
uvmm = nouveau_cli_uvmm(cli);
drivers/gpu/drm/nouveau/nouveau_drv.h
141
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drv.h
147
nouveau_cli_vmm(struct nouveau_cli *cli)
drivers/gpu/drm/nouveau/nouveau_drv.h
151
uvmm = nouveau_cli_uvmm(cli);
drivers/gpu/drm/nouveau/nouveau_drv.h
155
if (cli->svm.cli)
drivers/gpu/drm/nouveau/nouveau_drv.h
156
return &cli->svm;
drivers/gpu/drm/nouveau/nouveau_drv.h
158
return &cli->vmm;
drivers/gpu/drm/nouveau/nouveau_drv.h
162
__nouveau_cli_disable_uvmm_noinit(struct nouveau_cli *cli)
drivers/gpu/drm/nouveau/nouveau_drv.h
164
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(cli);
drivers/gpu/drm/nouveau/nouveau_drv.h
167
cli->uvmm.disabled = true;
drivers/gpu/drm/nouveau/nouveau_drv.h
171
nouveau_cli_disable_uvmm_noinit(struct nouveau_cli *cli)
drivers/gpu/drm/nouveau/nouveau_drv.h
173
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_drv.h
174
__nouveau_cli_disable_uvmm_noinit(cli);
drivers/gpu/drm/nouveau/nouveau_drv.h
175
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_exec.c
138
NV_PRINTK(err, job->cli, "nv50cal_space: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_exec.c
154
NV_PRINTK(err, job->cli, "error fencing pushbuf: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_exec.c
189
NV_PRINTK(warn, job->cli, "job timeout, channel %d killed!\n",
drivers/gpu/drm/nouveau/nouveau_exec.c
355
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_exec.c
366
if (unlikely(!nouveau_cli_uvmm(cli)))
drivers/gpu/drm/nouveau/nouveau_exec.c
387
NV_PRINTK(err, cli, "pushbuf push count exceeds limit: %d max %d\n",
drivers/gpu/drm/nouveau/nouveau_exec.c
93
struct nouveau_cli *cli = job->cli;
drivers/gpu/drm/nouveau/nouveau_exec.c
94
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(cli);
drivers/gpu/drm/nouveau/nouveau_fence.c
170
struct nouveau_cli *cli = chan->cli;
drivers/gpu/drm/nouveau/nouveau_fence.c
171
struct nouveau_drm *drm = cli->drm;
drivers/gpu/drm/nouveau/nouveau_fence.c
190
strscpy(fctx->name, cli->name);
drivers/gpu/drm/nouveau/nouveau_fence.c
211
struct nouveau_fence_priv *priv = (void*)chan->cli->drm->fence;
drivers/gpu/drm/nouveau/nouveau_fence.c
368
f = nouveau_local_fence(fence, chan->cli->drm);
drivers/gpu/drm/nouveau/nouveau_fence.c
376
local = prev && prev->cli->drm == chan->cli->drm;
drivers/gpu/drm/nouveau/nouveau_gem.c
103
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(cli);
drivers/gpu/drm/nouveau/nouveau_gem.c
104
struct nouveau_vmm *vmm = nouveau_cli_vmm(cli);
drivers/gpu/drm/nouveau/nouveau_gem.c
126
if (!nouveau_cli_uvmm(cli))
drivers/gpu/drm/nouveau/nouveau_gem.c
179
nouveau_cli_work_queue(vma->vmm->cli, fence, &work->work);
drivers/gpu/drm/nouveau/nouveau_gem.c
185
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_gem.c
189
struct nouveau_vmm *vmm = nouveau_cli_vmm(cli);
drivers/gpu/drm/nouveau/nouveau_gem.c
196
if (nouveau_cli_uvmm(cli))
drivers/gpu/drm/nouveau/nouveau_gem.c
232
nouveau_gem_new(struct nouveau_cli *cli, u64 size, int align, uint32_t domain,
drivers/gpu/drm/nouveau/nouveau_gem.c
236
struct nouveau_drm *drm = cli->drm;
drivers/gpu/drm/nouveau/nouveau_gem.c
237
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(cli);
drivers/gpu/drm/nouveau/nouveau_gem.c
252
nvbo = nouveau_bo_alloc(cli, &size, &align, domain, tile_mode,
drivers/gpu/drm/nouveau/nouveau_gem.c
302
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_gem.c
304
struct nouveau_vmm *vmm = nouveau_cli_vmm(cli);
drivers/gpu/drm/nouveau/nouveau_gem.c
315
!nouveau_cli_uvmm(cli)) {
drivers/gpu/drm/nouveau/nouveau_gem.c
328
if (cli->device.info.family >= NV_DEVICE_INFO_V0_FERMI)
drivers/gpu/drm/nouveau/nouveau_gem.c
331
if (cli->device.info.family >= NV_DEVICE_INFO_V0_TESLA)
drivers/gpu/drm/nouveau/nouveau_gem.c
342
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_gem.c
350
nouveau_cli_disable_uvmm_noinit(cli);
drivers/gpu/drm/nouveau/nouveau_gem.c
352
ret = nouveau_gem_new(cli, req->info.size, req->align,
drivers/gpu/drm/nouveau/nouveau_gem.c
460
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_gem.c
471
NV_PRINTK(err, cli, "%s failed and gave up.\n", __func__);
drivers/gpu/drm/nouveau/nouveau_gem.c
482
NV_PRINTK(err, cli, "Unknown handle 0x%08x\n", b->handle);
drivers/gpu/drm/nouveau/nouveau_gem.c
494
NV_PRINTK(err, cli, "multiple instances of buffer %d on "
drivers/gpu/drm/nouveau/nouveau_gem.c
515
NV_PRINTK(err, cli, "fail reserve\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
524
NV_PRINTK(err, cli, "vma not found!\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
546
NV_PRINTK(err, cli, "invalid valid domains: 0x%08x\n",
drivers/gpu/drm/nouveau/nouveau_gem.c
570
struct nouveau_cli *cli = chan->cli;
drivers/gpu/drm/nouveau/nouveau_gem.c
571
struct nouveau_drm *drm = cli->drm;
drivers/gpu/drm/nouveau/nouveau_gem.c
582
NV_PRINTK(err, cli, "fail set_domain\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
589
NV_PRINTK(err, cli, "fail ttm_validate\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
596
NV_PRINTK(err, cli, "fail post-validate sync\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
628
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_gem.c
639
NV_PRINTK(err, cli, "validate_init\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
646
NV_PRINTK(err, cli, "validating bo list\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
657
nouveau_gem_pushbuf_reloc_apply(struct nouveau_cli *cli,
drivers/gpu/drm/nouveau/nouveau_gem.c
673
NV_PRINTK(err, cli, "reloc bo index invalid\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
683
NV_PRINTK(err, cli, "reloc container bo index invalid\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
691
NV_PRINTK(err, cli, "reloc outside of bo\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
700
NV_PRINTK(err, cli, "failed kmap for reloc\n");
drivers/gpu/drm/nouveau/nouveau_gem.c
732
NV_PRINTK(err, cli, "reloc wait_idle failed: %d\n",
drivers/gpu/drm/nouveau/nouveau_gem.c
748
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_gem.c
764
if (unlikely(nouveau_cli_uvmm(cli)))
drivers/gpu/drm/nouveau/nouveau_gem.c
787
NV_PRINTK(err, cli, "pushbuf push count exceeds limit: %d max %d\n",
drivers/gpu/drm/nouveau/nouveau_gem.c
793
NV_PRINTK(err, cli, "pushbuf bo count exceeds limit: %d max %d\n",
drivers/gpu/drm/nouveau/nouveau_gem.c
799
NV_PRINTK(err, cli, "pushbuf reloc count exceeds limit: %d max %d\n",
drivers/gpu/drm/nouveau/nouveau_gem.c
817
NV_PRINTK(err, cli, "push %d buffer not in list\n", i);
drivers/gpu/drm/nouveau/nouveau_gem.c
829
NV_PRINTK(err, cli, "validate: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_gem.c
846
ret = nouveau_gem_pushbuf_reloc_apply(cli, req, reloc, bo);
drivers/gpu/drm/nouveau/nouveau_gem.c
848
NV_PRINTK(err, cli, "reloc apply: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_gem.c
856
NV_PRINTK(err, cli, "nv50cal_space: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_gem.c
875
NV_PRINTK(err, cli, "cal_space: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_gem.c
889
NV_PRINTK(err, cli, "jmp_space: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_gem.c
925
NV_PRINTK(err, cli, "error fencing pushbuf: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_gem.c
99
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_sched.c
174
NV_PRINTK(warn, job->cli,
drivers/gpu/drm/nouveau/nouveau_sched.c
224
NV_PRINTK(warn, job->cli,
drivers/gpu/drm/nouveau/nouveau_sched.c
278
.vm = &nouveau_cli_uvmm(job->cli)->base,
drivers/gpu/drm/nouveau/nouveau_sched.c
35
job->cli = nouveau_cli(args->file_priv);
drivers/gpu/drm/nouveau/nouveau_sched.c
379
NV_PRINTK(warn, job->cli, "Generic job timeout.\n");
drivers/gpu/drm/nouveau/nouveau_sched.h
57
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/nouveau_svm.c
105
NV_DEBUG((s)->vmm->cli->drm, "svm-%p: "f"\n", (s), ##a)
drivers/gpu/drm/nouveau/nouveau_svm.c
107
NV_WARN((s)->vmm->cli->drm, "svm-%p: "f"\n", (s), ##a)
drivers/gpu/drm/nouveau/nouveau_svm.c
113
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_svm.c
169
if (!cli->svm.svmm) {
drivers/gpu/drm/nouveau/nouveau_svm.c
186
nouveau_dmem_migrate_vma(cli->drm, cli->svm.svmm, vma, addr,
drivers/gpu/drm/nouveau/nouveau_svm.c
210
mutex_lock(&svmm->vmm->cli->drm->svm->mutex);
drivers/gpu/drm/nouveau/nouveau_svm.c
211
ivmm = nouveau_ivmm_find(svmm->vmm->cli->drm->svm, inst);
drivers/gpu/drm/nouveau/nouveau_svm.c
216
mutex_unlock(&svmm->vmm->cli->drm->svm->mutex);
drivers/gpu/drm/nouveau/nouveau_svm.c
231
mutex_lock(&svmm->vmm->cli->drm->svm->mutex);
drivers/gpu/drm/nouveau/nouveau_svm.c
232
list_add(&ivmm->head, &svmm->vmm->cli->drm->svm->inst);
drivers/gpu/drm/nouveau/nouveau_svm.c
233
mutex_unlock(&svmm->vmm->cli->drm->svm->mutex);
drivers/gpu/drm/nouveau/nouveau_svm.c
274
update->owner == svmm->vmm->cli->drm->dev)
drivers/gpu/drm/nouveau/nouveau_svm.c
319
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_svm.c
325
if (!cli->drm->svm)
drivers/gpu/drm/nouveau/nouveau_svm.c
331
svmm->vmm = &cli->svm;
drivers/gpu/drm/nouveau/nouveau_svm.c
337
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_svm.c
338
if (cli->svm.cli) {
drivers/gpu/drm/nouveau/nouveau_svm.c
349
ret = nvif_vmm_ctor(&cli->mmu, "svmVmm",
drivers/gpu/drm/nouveau/nouveau_svm.c
350
cli->vmm.vmm.object.oclass, MANAGED,
drivers/gpu/drm/nouveau/nouveau_svm.c
354
}, sizeof(struct gp100_vmm_v0), &cli->svm.vmm);
drivers/gpu/drm/nouveau/nouveau_svm.c
365
cli->svm.svmm = svmm;
drivers/gpu/drm/nouveau/nouveau_svm.c
366
cli->svm.cli = cli;
drivers/gpu/drm/nouveau/nouveau_svm.c
368
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_svm.c
374
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_svm.c
514
range->owner == sn->svmm->vmm->cli->drm->dev)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1064
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(job->cli);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1116
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(job->cli);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1263
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(job->cli);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1515
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(job->cli);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1541
NV_PRINTK(err, job->cli, "bind job failed: %d\n", ret);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1549
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(job->cli);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1802
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1807
if (unlikely(!nouveau_cli_uvmm_locked(cli)))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1814
args.sched = cli->sched;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1892
struct nouveau_cli *cli = nouveau_cli(file_priv);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1893
struct drm_device *drm = cli->drm->dev;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1907
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1909
if (unlikely(cli->uvmm.disabled)) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1931
drm_gpuvm_init(&uvmm->base, cli->name, 0, drm, r_obj,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1940
ret = nvif_vmm_ctor(&cli->mmu, "uvmm",
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1941
cli->vmm.vmm.object.oclass, RAW,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1948
uvmm->vmm.cli = cli;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1949
cli->uvmm.ptr = uvmm;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1950
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1957
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1966
struct nouveau_cli *cli = uvmm->vmm.cli;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
2001
mutex_lock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
2004
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nouveau_vmm.c
128
vmm->cli = NULL;
drivers/gpu/drm/nouveau/nouveau_vmm.c
132
nouveau_vmm_init(struct nouveau_cli *cli, s32 oclass, struct nouveau_vmm *vmm)
drivers/gpu/drm/nouveau/nouveau_vmm.c
134
int ret = nvif_vmm_ctor(&cli->mmu, "drmVmm", oclass, UNMANAGED,
drivers/gpu/drm/nouveau/nouveau_vmm.c
139
vmm->cli = cli;
drivers/gpu/drm/nouveau/nouveau_vmm.h
26
struct nouveau_cli *cli;
drivers/gpu/drm/nouveau/nv17_fence.c
39
struct nouveau_cli *cli = prev->cli;
drivers/gpu/drm/nouveau/nv17_fence.c
40
struct nv10_fence_priv *priv = cli->drm->fence;
drivers/gpu/drm/nouveau/nv17_fence.c
47
if (!mutex_trylock(&cli->mutex))
drivers/gpu/drm/nouveau/nv17_fence.c
72
mutex_unlock(&cli->mutex);
drivers/gpu/drm/nouveau/nv17_fence.c
79
struct nv10_fence_priv *priv = chan->cli->drm->fence;
drivers/gpu/drm/nouveau/nv50_fence.c
38
struct nv10_fence_priv *priv = chan->cli->drm->fence;
drivers/gpu/drm/nouveau/nv84_fence.c
108
struct nv84_fence_priv *priv = chan->cli->drm->fence;
drivers/gpu/drm/nouveau/nv84_fence.c
115
struct nv84_fence_priv *priv = chan->cli->drm->fence;
drivers/gpu/drm/nouveau/nv84_fence.c
130
struct nv84_fence_priv *priv = chan->cli->drm->fence;
drivers/gpu/drm/nouveau/nv84_fence.c
82
return chan->cli->drm->runl[chan->runlist].chan_id_base + chan->chid;
drivers/i2c/busses/i2c-nomadik.c
205
struct i2c_nmk_client cli;
drivers/i2c/busses/i2c-nomadik.c
333
priv->cli.operation = I2C_NO_OPERATION;
drivers/i2c/busses/i2c-nomadik.c
355
mcr |= FIELD_PREP(I2C_MCR_A7, priv->cli.slave_adr);
drivers/i2c/busses/i2c-nomadik.c
367
priv->cli.slave_adr);
drivers/i2c/busses/i2c-nomadik.c
379
if (priv->cli.operation == I2C_WRITE)
drivers/i2c/busses/i2c-nomadik.c
390
mcr |= FIELD_PREP(I2C_MCR_LENGTH, priv->cli.count);
drivers/i2c/busses/i2c-nomadik.c
552
(priv->cli.count != 0);
drivers/i2c/busses/i2c-nomadik.c
555
nmk_i2c_writeb(priv, *priv->cli.buffer, I2C_TFR);
drivers/i2c/busses/i2c-nomadik.c
556
priv->cli.buffer++;
drivers/i2c/busses/i2c-nomadik.c
557
priv->cli.count--;
drivers/i2c/busses/i2c-nomadik.c
558
priv->cli.xfer_bytes++;
drivers/i2c/busses/i2c-nomadik.c
596
if (priv->cli.count != 0)
drivers/i2c/busses/i2c-nomadik.c
619
priv->cli.slave_adr);
drivers/i2c/busses/i2c-nomadik.c
637
priv->cli.operation = I2C_READ;
drivers/i2c/busses/i2c-nomadik.c
641
priv->cli.operation = I2C_WRITE;
drivers/i2c/busses/i2c-nomadik.c
729
priv->cli.slave_adr = msgs[i].addr;
drivers/i2c/busses/i2c-nomadik.c
730
priv->cli.buffer = msgs[i].buf;
drivers/i2c/busses/i2c-nomadik.c
731
priv->cli.count = msgs[i].len;
drivers/i2c/busses/i2c-nomadik.c
797
if (priv->cli.operation == I2C_READ) {
drivers/i2c/busses/i2c-nomadik.c
809
if (priv->cli.count == 0)
drivers/i2c/busses/i2c-nomadik.c
824
*priv->cli.buffer = nmk_i2c_readb(priv, I2C_RFR);
drivers/i2c/busses/i2c-nomadik.c
825
priv->cli.buffer++;
drivers/i2c/busses/i2c-nomadik.c
827
priv->cli.count -= rft;
drivers/i2c/busses/i2c-nomadik.c
828
priv->cli.xfer_bytes += rft;
drivers/i2c/busses/i2c-nomadik.c
834
*priv->cli.buffer = nmk_i2c_readb(priv, I2C_RFR);
drivers/i2c/busses/i2c-nomadik.c
835
priv->cli.buffer++;
drivers/i2c/busses/i2c-nomadik.c
837
priv->cli.count -= MAX_I2C_FIFO_THRESHOLD;
drivers/i2c/busses/i2c-nomadik.c
838
priv->cli.xfer_bytes += MAX_I2C_FIFO_THRESHOLD;
drivers/i2c/busses/i2c-nomadik.c
844
if (priv->cli.operation == I2C_READ) {
drivers/i2c/busses/i2c-nomadik.c
847
if (priv->cli.count == 0)
drivers/i2c/busses/i2c-nomadik.c
849
*priv->cli.buffer =
drivers/i2c/busses/i2c-nomadik.c
851
priv->cli.buffer++;
drivers/i2c/busses/i2c-nomadik.c
852
priv->cli.count--;
drivers/i2c/busses/i2c-nomadik.c
853
priv->cli.xfer_bytes++;
drivers/i2c/busses/i2c-nomadik.c
860
if (priv->cli.count) {
drivers/i2c/busses/i2c-nomadik.c
863
priv->cli.count);
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
104
int ms_sensors_convert_and_read(void *cli, u8 conv, u8 rd,
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
109
struct i2c_client *client = cli;
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
47
int ms_sensors_reset(void *cli, u8 cmd, unsigned int delay)
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
50
struct i2c_client *client = cli;
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
73
int ms_sensors_read_prom_word(void *cli, int cmd, u16 *word)
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
76
struct i2c_client *client = cli;
drivers/iio/common/ms_sensors/ms_sensors_i2c.h
54
int ms_sensors_reset(void *cli, u8 cmd, unsigned int delay);
drivers/iio/common/ms_sensors/ms_sensors_i2c.h
55
int ms_sensors_read_prom_word(void *cli, int cmd, u16 *word);
drivers/iio/common/ms_sensors/ms_sensors_i2c.h
56
int ms_sensors_convert_and_read(void *cli, u8 conv, u8 rd,
drivers/iio/magnetometer/hmc5843_i2c.c
55
static int hmc5843_i2c_probe(struct i2c_client *cli)
drivers/iio/magnetometer/hmc5843_i2c.c
57
const struct i2c_device_id *id = i2c_client_get_device_id(cli);
drivers/iio/magnetometer/hmc5843_i2c.c
58
struct regmap *regmap = devm_regmap_init_i2c(cli,
drivers/iio/magnetometer/hmc5843_i2c.c
63
return hmc5843_common_probe(&cli->dev,
drivers/iio/temperature/tsys01.c
34
int (*reset)(void *cli, u8 cmd, unsigned int delay);
drivers/iio/temperature/tsys01.c
35
int (*convert_and_read)(void *cli, u8 conv, u8 rd,
drivers/iio/temperature/tsys01.c
37
int (*read_prom_word)(void *cli, int cmd, u16 *word);
drivers/net/ethernet/broadcom/cnic.c
4927
u32 cli = cp->ethdev->iscsi_l2_client_id;
drivers/net/ethernet/broadcom/cnic.c
4973
if (cli < MAX_STAT_COUNTER_ID) {
drivers/net/ethernet/broadcom/cnic.c
4976
data->general.statistics_counter_id = cli;
drivers/net/ethernet/broadcom/cnic.c
4995
u32 cli = cp->ethdev->iscsi_l2_client_id;
drivers/net/ethernet/broadcom/cnic.c
4996
int cl_qzone_id = BNX2X_CL_QZONE_ID(bp, cli);
drivers/net/ethernet/broadcom/cnic.c
5001
data->general.client_id = cli;
drivers/net/ethernet/broadcom/cnic.c
5003
data->general.sp_client_id = cli;
drivers/net/ethernet/broadcom/cnic.c
5179
u32 cli = cp->ethdev->iscsi_l2_client_id;
drivers/net/ethernet/broadcom/cnic.c
5191
cl_qzone_id = BNX2X_CL_QZONE_ID(bp, cli);
drivers/net/ethernet/broadcom/cnic.c
5196
USTORM_RX_PRODS_E1X_OFFSET(BP_PORT(bp), cli));
drivers/net/ethernet/broadcom/cnic.c
5230
cnic_ring_ctl(dev, cid, cli, 1);
drivers/net/ethernet/broadcom/cnic.c
5249
u32 cli = cp->ethdev->iscsi_l2_client_id;
drivers/net/ethernet/broadcom/cnic.c
5254
cnic_ring_ctl(dev, cid, cli, 0);
drivers/net/ethernet/broadcom/cnic.c
5258
l5_data.phy_address.lo = cli;
drivers/net/ethernet/broadcom/cnic.h
411
#define BNX2X_CL_QZONE_ID(bp, cli) \
drivers/net/ethernet/broadcom/cnic.h
412
(BNX2X_CHIP_IS_E2_PLUS(bp) ? cli : \
drivers/net/ethernet/broadcom/cnic.h
413
cli + (BP_PORT(bp) * ETH_MAX_RX_CLIENTS_E1H))
drivers/net/ethernet/qlogic/qed/qed_cxt.c
45
#define ILT_CFG_REG(cli, reg) PSWRQ2_REG_ ## cli ## _ ## reg ## _RT_OFFSET
drivers/scsi/esas2r/atioctl.h
1162
struct atto_ioctl_vda_cli_cmd cli;
drivers/scsi/esas2r/atvda.h
1222
struct atto_vda_cli_req cli;
drivers/scsi/esas2r/esas2r_vda.c
126
rq->vrq->cli.cmd_rsp_len =
drivers/scsi/esas2r/esas2r_vda.c
127
cpu_to_le32(vi->cmd.cli.cmd_rsp_len);
drivers/scsi/esas2r/esas2r_vda.c
128
rq->vrq->cli.length = cpu_to_le32(datalen);
drivers/scsi/esas2r/esas2r_vda.c
130
firstsg = rq->vrq->cli.sge;
drivers/scsi/esas2r/esas2r_vda.c
335
vi->cmd.cli.cmd_rsp_len =
drivers/video/backlight/ili9320.h
69
struct ili9320_client *cli);
tools/testing/selftests/kvm/x86/hyperv_ipi.c
68
cli();
tools/testing/selftests/kvm/x86/xapic_ipi_test.c
110
cli();
tools/testing/selftests/kvm/x86/xapic_state_test.c
21
cli();
tools/testing/selftests/kvm/x86/xapic_state_test.c
41
cli();
tools/testing/selftests/kvm/x86/xapic_tpr_test.c
117
cli();