lib/libcurses/base/lib_color.c
849
rgb_bits_t *work = &(sp->_direct_color);
lib/libcurses/base/lib_color.c
851
#define max_direct_color(name) ((1 << work->bits.name) - 1)
lib/libcurses/base/lib_color.c
861
bitoff += work->bits.blue;
lib/libcurses/base/lib_color.c
864
bitoff += work->bits.green;
lib/libcurses/base/lib_mouse.c
325
MEVENT *work;
lib/libcurses/base/lib_mouse.c
337
work = &(sp->_sysmouse_fifo[sp->_sysmouse_tail]);
lib/libcurses/base/lib_mouse.c
338
memset(work, 0, sizeof(*work));
lib/libcurses/base/lib_mouse.c
339
work->id = NORMAL_EVENT; /* there's only one mouse... */
lib/libcurses/base/lib_mouse.c
346
work->bstate |= BUTTON1_PRESSED;
lib/libcurses/base/lib_mouse.c
348
work->bstate |= BUTTON2_PRESSED;
lib/libcurses/base/lib_mouse.c
350
work->bstate |= BUTTON3_PRESSED;
lib/libcurses/base/lib_mouse.c
353
work->bstate |= BUTTON1_RELEASED;
lib/libcurses/base/lib_mouse.c
355
work->bstate |= BUTTON2_RELEASED;
lib/libcurses/base/lib_mouse.c
357
work->bstate |= BUTTON3_RELEASED;
lib/libcurses/base/lib_mouse.c
373
work->x = the_mouse.u.data.x / sp->_sysmouse_char_width;
lib/libcurses/base/lib_mouse.c
374
work->y = the_mouse.u.data.y / sp->_sysmouse_char_height;
lib/libcurses/tinfo/lib_win32con.c
676
MEVENT work;
lib/libcurses/tinfo/lib_win32con.c
689
memset(&work, 0, sizeof(work));
lib/libcurses/tinfo/lib_win32con.c
692
work.bstate |= decode_mouse(sp, sp->_drv_mouse_new_buttons);
lib/libcurses/tinfo/lib_win32con.c
695
work.bstate |= (decode_mouse(sp, sp->_drv_mouse_old_buttons)
lib/libcurses/tinfo/lib_win32con.c
700
work.x = mer.dwMousePosition.X;
lib/libcurses/tinfo/lib_win32con.c
701
work.y = mer.dwMousePosition.Y - AdjustY();
lib/libcurses/tinfo/lib_win32con.c
703
sp->_drv_mouse_fifo[sp->_drv_mouse_tail] = work;
lib/libz/infback.c
338
&(state->lenbits), state->work);
lib/libz/infback.c
431
&(state->lenbits), state->work);
lib/libz/infback.c
444
&(state->next), &(state->distbits), state->work);
lib/libz/inflate.c
850
&(state->lenbits), state->work);
lib/libz/inflate.c
944
&(state->lenbits), state->work);
lib/libz/inflate.c
957
&(state->next), &(state->distbits), state->work);
lib/libz/inflate.h
121
unsigned short work[288]; /* work area for code table building */
lib/libz/inftrees.c
154
if (lens[sym] != 0) work[offs[lens[sym]]++] = (unsigned short)sym;
lib/libz/inftrees.c
222
if (work[sym] + 1U < match) {
lib/libz/inftrees.c
224
here.val = work[sym];
lib/libz/inftrees.c
226
else if (work[sym] >= match) {
lib/libz/inftrees.c
227
here.op = (unsigned char)(extra[work[sym] - match]);
lib/libz/inftrees.c
228
here.val = base[work[sym] - match];
lib/libz/inftrees.c
259
len = lens[work[sym]];
lib/libz/inftrees.c
328
unsigned short lens[288], work[288];
lib/libz/inftrees.c
339
inflate_table(LENS, lens, 288, &(next), &(bits), work);
lib/libz/inftrees.c
346
inflate_table(DISTS, lens, 32, &(next), &(bits), work);
lib/libz/inftrees.c
46
unsigned FAR *bits, unsigned short FAR *work) {
lib/libz/inftrees.h
62
unsigned FAR *bits, unsigned short FAR *work);
regress/lib/libc/sys/t_getrusage.c
219
work();
regress/lib/libc/sys/t_getrusage.c
223
work();
regress/lib/libc/sys/t_getrusage.c
259
work();
regress/lib/libc/sys/t_getrusage.c
50
static void work(void);
regress/lib/libz/infcover.c
621
unsigned short lens[16], work[16];
regress/lib/libz/infcover.c
631
ret = inflate_table(DISTS, lens, 16, &next, &bits, work);
regress/lib/libz/infcover.c
635
ret = inflate_table(DISTS, lens, 16, &next, &bits, work);
sbin/iked/smult_curve25519_ref.c
124
static void mainloop(unsigned int work[64],const unsigned char e[32])
sbin/iked/smult_curve25519_ref.c
145
for (j = 0;j < 32;++j) xzm1[j] = work[j];
sbin/iked/smult_curve25519_ref.c
173
mult(xzn1b + 32,r,work);
sbin/iked/smult_curve25519_ref.c
177
for (j = 0;j < 64;++j) work[j] = xzm[j];
sbin/iked/smult_curve25519_ref.c
251
unsigned int work[96];
sbin/iked/smult_curve25519_ref.c
258
for (i = 0;i < 32;++i) work[i] = p[i];
sbin/iked/smult_curve25519_ref.c
259
mainloop(work,e);
sbin/iked/smult_curve25519_ref.c
260
recip(work + 32,work + 32);
sbin/iked/smult_curve25519_ref.c
261
mult(work + 64,work,work + 32);
sbin/iked/smult_curve25519_ref.c
262
freeze(work + 64);
sbin/iked/smult_curve25519_ref.c
263
for (i = 0;i < 32;++i) q[i] = work[64 + i];
sys/arch/octeon/dev/if_cnmac.c
1120
cnmac_recv_mbuf(struct cnmac_softc *sc, uint64_t *work,
sys/arch/octeon/dev/if_cnmac.c
1125
uint64_t word1 = work[1];
sys/arch/octeon/dev/if_cnmac.c
1126
uint64_t word2 = work[2];
sys/arch/octeon/dev/if_cnmac.c
1127
uint64_t word3 = work[3];
sys/arch/octeon/dev/if_cnmac.c
1131
cn30xxfpa_buf_put_paddr(cnmac_fb_wqe, XKPHYS_TO_PHYS(work));
sys/arch/octeon/dev/if_cnmac.c
1211
cnmac_recv(struct cnmac_softc *sc, uint64_t *work, struct mbuf_list *ml)
sys/arch/octeon/dev/if_cnmac.c
1218
word2 = work[2];
sys/arch/octeon/dev/if_cnmac.c
1229
if (__predict_false(cnmac_recv_mbuf(sc, work, &m, &nmbuf) != 0)) {
sys/arch/octeon/dev/if_cnmac.c
1254
cnmac_buf_free_work(sc, work);
sys/arch/octeon/dev/if_cnmac.c
1264
uint64_t *work;
sys/arch/octeon/dev/if_cnmac.c
1277
work = (uint64_t *)cn30xxpow_work_response_async(
sys/arch/octeon/dev/if_cnmac.c
1279
if (work == NULL)
sys/arch/octeon/dev/if_cnmac.c
1286
port = (work[1] & PIP_WQE_WORD1_IPRT) >> 42;
sys/arch/octeon/dev/if_cnmac.c
1293
nmbuf += cnmac_recv(sc, work, &ml);
sys/arch/octeon/dev/if_cnmac.c
1307
printf("word0: 0x%016llx\n", work[0]);
sys/arch/octeon/dev/if_cnmac.c
1308
printf("word1: 0x%016llx\n", work[1]);
sys/arch/octeon/dev/if_cnmac.c
1309
printf("word2: 0x%016llx\n", work[2]);
sys/arch/octeon/dev/if_cnmac.c
1310
printf("word3: 0x%016llx\n", work[3]);
sys/arch/octeon/dev/if_cnmac.c
606
cnmac_buf_free_work(struct cnmac_softc *sc, uint64_t *work)
sys/arch/octeon/dev/if_cnmac.c
612
nbufs = (work[2] & PIP_WQE_WORD2_IP_BUFS) >>
sys/arch/octeon/dev/if_cnmac.c
614
word3 = work[3];
sys/arch/octeon/dev/if_cnmac.c
629
cn30xxfpa_buf_put_paddr(cnmac_fb_wqe, XKPHYS_TO_PHYS(work));
sys/arch/octeon/dev/if_ogx.c
1030
uint64_t *work;
sys/arch/octeon/dev/if_ogx.c
1045
work = ogx_get_work(sc->sc_node, PORT_GROUP_RX(sc));
sys/arch/octeon/dev/if_ogx.c
1046
if (work == NULL)
sys/arch/octeon/dev/if_ogx.c
1050
pkind = (work[0] & PKI_WORD0_PKIND_M) >> PKI_WORD0_PKIND_S;
sys/arch/octeon/dev/if_ogx.c
1058
nsegs = (work[0] & PKI_WORD0_BUFS_M) >> PKI_WORD0_BUFS_S;
sys/arch/octeon/dev/if_ogx.c
1059
word3 = work[3];
sys/arch/octeon/dev/if_ogx.c
1061
errlevel = (work[2] & PKI_WORD2_ERR_LEVEL_M) >>
sys/arch/octeon/dev/if_ogx.c
1063
errcode = (work[2] & PKI_WORD2_ERR_CODE_M) >>
sys/arch/octeon/dev/if_ogx.c
1073
pktlen = (work[1] & PKI_WORD1_LEN_M) >> PKI_WORD1_LEN_S;
sys/arch/octeon/dev/if_ogx.c
1136
printf("work0: %016llx\n", work[0]);
sys/arch/octeon/dev/if_ogx.c
1137
printf("work1: %016llx\n", work[1]);
sys/arch/octeon/dev/if_ogx.c
1138
printf("work2: %016llx\n", work[2]);
sys/arch/octeon/dev/if_ogx.c
1139
printf("work3: %016llx\n", work[3]);
sys/arch/octeon/dev/if_ogx.c
1140
printf("work4: %016llx\n", work[4]);
sys/arch/octeon/dev/if_ogx.c
1152
uint64_t *work;
sys/arch/octeon/dev/if_ogx.c
1159
work = ogx_get_work(node, PORT_GROUP_TX(sc));
sys/arch/octeon/dev/if_ogx.c
1160
if (work == NULL)
sys/arch/octeon/dev/if_ogx.c
1167
m = *(struct mbuf **)work;
sys/dev/pci/drm/amd/amdgpu/aldebaran.c
122
static void aldebaran_async_reset(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/aldebaran.c
126
container_of(work, struct amdgpu_reset_control, reset_work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd.c
130
static void amdgpu_amdkfd_reset_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd.c
132
struct amdgpu_device *adev = container_of(work, struct amdgpu_device,
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2781
static void amdgpu_amdkfd_restore_userptr_worker(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2783
struct delayed_work *dwork = to_delayed_work(work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
75
static void amdgpu_amdkfd_restore_userptr_worker(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_device.c
3792
static void amdgpu_device_delayed_init_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_device.c
3795
container_of(work, struct amdgpu_device, delayed_init_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_device.c
3803
static void amdgpu_device_delay_enable_gfx_off(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_device.c
3806
container_of(work, struct amdgpu_device, gfx.gfx_off_delay_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_device.c
4607
INIT_DELAYED_WORK(&adev->gfx.enforce_isolation[i].work,
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
101
static bool amdgpu_display_flip_handle_fence(struct amdgpu_flip_work *work,
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
111
if (!dma_fence_add_callback(fence, &work->cb,
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
122
container_of(__work, struct delayed_work, work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
123
struct amdgpu_flip_work *work =
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
125
struct amdgpu_device *adev = work->adev;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
126
struct amdgpu_crtc *amdgpu_crtc = adev->mode_info.crtcs[work->crtc_id];
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
133
for (i = 0; i < work->shared_count; ++i)
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
134
if (amdgpu_display_flip_handle_fence(work, &work->shared[i]))
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
141
(amdgpu_display_get_crtc_scanoutpos(adev_to_drm(adev), work->crtc_id, 0,
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
146
(int)(work->target_vblank -
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
148
schedule_delayed_work(&work->flip_work, usecs_to_jiffies(1000));
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
156
adev->mode_info.funcs->page_flip(adev, work->crtc_id, work->base, work->async);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
165
amdgpu_crtc->crtc_id, amdgpu_crtc, work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
174
struct amdgpu_flip_work *work =
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
179
r = amdgpu_bo_reserve(work->old_abo, true);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
181
amdgpu_bo_unpin(work->old_abo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
182
amdgpu_bo_unreserve(work->old_abo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
186
amdgpu_bo_unref(&work->old_abo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
187
kfree(work->shared);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
188
kfree(work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
201
struct amdgpu_flip_work *work;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
207
work = kzalloc(sizeof(*work), GFP_KERNEL);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
208
if (work == NULL)
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
211
INIT_DELAYED_WORK(&work->flip_work, amdgpu_display_flip_work_func);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
212
INIT_WORK(&work->unpin_work, amdgpu_display_unpin_work_func);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
214
work->event = event;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
215
work->adev = adev;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
216
work->crtc_id = amdgpu_crtc->crtc_id;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
217
work->async = (page_flip_flags & DRM_MODE_PAGE_FLIP_ASYNC) != 0;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
223
work->old_abo = gem_to_amdgpu_bo(obj);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
224
amdgpu_bo_ref(work->old_abo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
253
&work->shared_count,
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
254
&work->shared);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
264
work->base = amdgpu_bo_gpu_offset(new_abo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
265
work->target_vblank = target - (uint32_t)drm_crtc_vblank_count(crtc) +
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
278
amdgpu_crtc->pflip_works = work;
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
282
amdgpu_crtc->crtc_id, amdgpu_crtc, work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
286
amdgpu_display_flip_work_func(&work->flip_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
302
amdgpu_bo_unref(&work->old_abo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
303
for (i = 0; i < work->shared_count; ++i)
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
304
dma_fence_put(work->shared[i]);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
305
kfree(work->shared);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
306
kfree(work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
67
void amdgpu_display_hotplug_work_func(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
69
struct amdgpu_device *adev = container_of(work, struct amdgpu_device,
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
70
hotplug_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
94
struct amdgpu_flip_work *work =
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.c
98
schedule_work(&work->flip_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_display.h
40
void amdgpu_display_hotplug_work_func(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_eviction_fence.c
103
amdgpu_eviction_fence_suspend_worker(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_eviction_fence.c
105
struct amdgpu_eviction_fence_mgr *evf_mgr = work_to_evf_mgr(work, suspend_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_fence.c
1062
static void amdgpu_debugfs_reset_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_fence.c
1064
struct amdgpu_device *adev = container_of(work, struct amdgpu_device,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
2004
schedule_delayed_work(&adev->gfx.enforce_isolation[idx].work,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
2009
cancel_delayed_work_sync(&adev->gfx.enforce_isolation[idx].work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
2037
void amdgpu_gfx_enforce_isolation_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
2040
container_of(work, struct amdgpu_isolation_work, work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
2063
schedule_delayed_work(&adev->gfx.enforce_isolation[idx].work,
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
2211
void amdgpu_gfx_profile_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.c
2214
container_of(work, struct amdgpu_device, gfx.idle_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
362
struct work_struct work;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
403
struct delayed_work work;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
638
void amdgpu_gfx_enforce_isolation_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gfx.h
642
void amdgpu_gfx_profile_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_irq.c
190
static void amdgpu_irq_handle_ih1(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_irq.c
192
struct amdgpu_device *adev = container_of(work, struct amdgpu_device,
sys/dev/pci/drm/amd/amdgpu/amdgpu_irq.c
205
static void amdgpu_irq_handle_ih2(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_irq.c
207
struct amdgpu_device *adev = container_of(work, struct amdgpu_device,
sys/dev/pci/drm/amd/amdgpu/amdgpu_irq.c
220
static void amdgpu_irq_handle_ih_soft(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_irq.c
222
struct amdgpu_device *adev = container_of(work, struct amdgpu_device,
sys/dev/pci/drm/amd/amdgpu/amdgpu_jpeg.c
109
static void amdgpu_jpeg_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_jpeg.c
112
container_of(work, struct amdgpu_device, jpeg.idle_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_jpeg.c
35
static void amdgpu_jpeg_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2426
static void amdgpu_ras_interrupt_process_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2429
container_of(work, struct ras_ih_data, ih_work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2732
static void amdgpu_ras_do_recovery(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
2735
container_of(work, struct amdgpu_ras, recovery_work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
3414
static void amdgpu_ras_do_page_retirement(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
3416
struct amdgpu_ras *con = container_of(work, struct amdgpu_ras,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
3417
page_retirement_dwork.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4006
static void amdgpu_ras_counte_dw(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4008
struct amdgpu_ras *con = container_of(work, struct amdgpu_ras,
sys/dev/pci/drm/amd/amdgpu/amdgpu_ras.c
4009
ras_counte_delay_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_reset.h
135
struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_reset.h
137
return queue_work(domain->wq, work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_reset.h
86
void (*async_reset)(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
870
static void amdgpu_userq_restore_worker(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq.c
872
struct amdgpu_userq_mgr *uq_mgr = work_to_uq_mgr(work, resume_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
1262
static void amdgpu_uvd_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
1265
container_of(work, struct amdgpu_device, uvd.idle_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
137
static void amdgpu_uvd_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
332
static void amdgpu_vce_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
335
container_of(work, struct amdgpu_device, vce.idle_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
84
static void amdgpu_vce_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vcn.c
457
static void amdgpu_vcn_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vcn.c
460
container_of(work, struct amdgpu_vcn_inst, idle_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vcn.c
94
static void amdgpu_vcn_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_virt.c
631
static void amdgpu_virt_update_vf2pf_work_item(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_virt.c
633
struct amdgpu_device *adev = container_of(work, struct amdgpu_device, virt.vf2pf_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.h
623
void amdgpu_vm_pt_free_work(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_tlb_fence.c
100
INIT_WORK(&f->work, amdgpu_tlb_fence_work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_tlb_fence.c
108
schedule_work(&f->work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_tlb_fence.c
35
struct work_struct work;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_tlb_fence.c
51
static void amdgpu_tlb_fence_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm_tlb_fence.c
53
struct amdgpu_tlb_fence *f = container_of(work, typeof(*f), work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vpe.c
345
static void vpe_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vpe.c
348
container_of(work, struct amdgpu_device, vpe.idle_work.work);
sys/dev/pci/drm/amd/amdgpu/amdgpu_xgmi.c
1668
static void amdgpu_xgmi_reset_on_init_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/amdgpu_xgmi.c
1671
container_of(work, struct amdgpu_hive_info, reset_on_init_work);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
1884
static void gfx_v8_0_sq_irq_work_func(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
1944
INIT_WORK(&adev->gfx.sq_work.work, gfx_v8_0_sq_irq_work_func);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
6705
static void gfx_v8_0_sq_irq_work_func(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
6708
struct amdgpu_device *adev = container_of(work, struct amdgpu_device, gfx.sq_work.work);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
6709
struct sq_work *sq_work = container_of(work, struct sq_work, work);
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
6725
if (work_pending(&adev->gfx.sq_work.work)) {
sys/dev/pci/drm/amd/amdgpu/gfx_v8_0.c
6729
schedule_work(&adev->gfx.sq_work.work);
sys/dev/pci/drm/amd/amdgpu/mxgpu_ai.c
273
static void xgpu_ai_mailbox_flr_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/mxgpu_ai.c
275
struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work);
sys/dev/pci/drm/amd/amdgpu/mxgpu_ai.c
295
static void xgpu_ai_mailbox_req_bad_pages_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/mxgpu_ai.c
297
struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, req_bad_pages_work);
sys/dev/pci/drm/amd/amdgpu/mxgpu_ai.c
314
static void xgpu_ai_mailbox_handle_bad_pages_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/mxgpu_ai.c
316
struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, handle_bad_pages_work);
sys/dev/pci/drm/amd/amdgpu/mxgpu_nv.c
337
static void xgpu_nv_mailbox_flr_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/mxgpu_nv.c
339
struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work);
sys/dev/pci/drm/amd/amdgpu/mxgpu_nv.c
362
static void xgpu_nv_mailbox_req_bad_pages_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/mxgpu_nv.c
364
struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, req_bad_pages_work);
sys/dev/pci/drm/amd/amdgpu/mxgpu_nv.c
381
static void xgpu_nv_mailbox_handle_bad_pages_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/mxgpu_nv.c
383
struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, handle_bad_pages_work);
sys/dev/pci/drm/amd/amdgpu/mxgpu_vi.c
513
static void xgpu_vi_mailbox_flr_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/mxgpu_vi.c
515
struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work);
sys/dev/pci/drm/amd/amdgpu/sienna_cichlid.c
110
static void sienna_cichlid_async_reset(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/sienna_cichlid.c
114
container_of(work, struct amdgpu_reset_control, reset_work);
sys/dev/pci/drm/amd/amdgpu/smu_v13_0_10.c
109
static void smu_v13_0_10_async_reset(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/smu_v13_0_10.c
113
container_of(work, struct amdgpu_reset_control, reset_work);
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
155
adev->vcn.inst[0].idle_work.work.func = vcn_v1_0_idle_work_handler;
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
157
task_set(&adev->vcn.inst[0].idle_work.work.task,
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
159
&adev->vcn.inst[0].idle_work.work);
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
1874
static void vcn_v1_0_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
1877
container_of(work, struct amdgpu_vcn_inst, idle_work.work);
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
93
static void vcn_v1_0_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdgpu/vcn_v2_5.c
112
static void vcn_v2_5_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdgpu/vcn_v2_5.c
115
container_of(work, struct amdgpu_vcn_inst, idle_work.work);
sys/dev/pci/drm/amd/amdgpu/vcn_v2_5.c
310
adev->vcn.inst[j].idle_work.work.func = vcn_v2_5_idle_work_handler;
sys/dev/pci/drm/amd/amdgpu/vcn_v2_5.c
312
task_set(&adev->vcn.inst[j].idle_work.work.task,
sys/dev/pci/drm/amd/amdgpu/vcn_v2_5.c
314
&adev->vcn.inst[j].idle_work.work);
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
103
process = container_of(work,
sys/dev/pci/drm/amd/amdkfd/kfd_debug.c
96
void debug_event_write_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_debug.h
86
void debug_event_write_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3243
static void copy_context_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_device_queue_manager.c
3253
workarea = container_of(work,
sys/dev/pci/drm/amd/amdkfd/kfd_interrupt.c
136
static void interrupt_wq(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_interrupt.c
138
struct kfd_node *dev = container_of(work, struct kfd_node, interrupt_work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
100
static void kfd_sdma_activity_worker(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
113
workarea = container_of(work, struct kfd_sdma_activity_handler_workarea,
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
1159
static void kfd_process_wq_release(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
1161
struct kfd_process *p = container_of(work, struct kfd_process,
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2003
static void evict_process_worker(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2009
dwork = to_delayed_work(work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2056
static void restore_process_worker(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2062
dwork = to_delayed_work(work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2213
struct work_struct work;
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2219
static void send_exception_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2230
workarea = container_of(work,
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2232
work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2268
INIT_WORK_ONSTACK(&worker.work, send_exception_work_handler);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2274
schedule_work(&worker.work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2275
flush_work(&worker.work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
2276
destroy_work_on_stack(&worker.work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
73
static void evict_process_worker(struct work_struct *work);
sys/dev/pci/drm/amd/amdkfd/kfd_process.c
74
static void restore_process_worker(struct work_struct *work);
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
1846
static void svm_range_restore_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
1848
struct delayed_work *dwork = to_delayed_work(work);
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2365
static void svm_range_deferred_list_work(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2371
svms = container_of(work, struct svm_range_list, deferred_list_work);
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
3593
static void svm_range_evict_svm_bo_worker(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
3599
svm_bo = container_of(work, struct svm_range_bo, eviction_work);
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
438
static void svm_range_bo_wq_release(struct work_struct *work)
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
442
svm_bo = container_of(work, struct svm_range_bo, release_work);
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
72
static void svm_range_evict_svm_bo_worker(struct work_struct *work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
1562
static void dm_handle_hpd_rx_offload_work(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
1574
offload_work = container_of(work, struct hpd_rx_irq_offload_work, work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
2241
flush_work(&adev->dm.idle_workqueue->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
3913
static void hdmi_hpd_debounce_work(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
3916
container_of(to_delayed_work(work), struct amdgpu_dm_connector,
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4087
INIT_WORK(&offload_work->work, dm_handle_hpd_rx_offload_work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
4092
queue_work(offload_wq->wq, &offload_work->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
536
struct vupdate_offload_work *work = container_of(offload_work, struct vupdate_offload_work, work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
537
struct amdgpu_device *adev = work->adev;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
538
struct dc_stream_state *stream = work->stream;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
539
struct dc_crtc_timing_adjust *adjust = work->adjust;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
546
kfree(work->adjust);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
547
kfree(work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
570
INIT_WORK(&offload_work->work, dm_handle_vmin_vmax_update);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
575
queue_work(system_wq, &offload_work->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
941
static void dm_handle_hpd_work(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.c
945
dmub_hpd_wrk = container_of(work, struct dmub_hpd_work, handle_hpd_work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.h
139
struct work_struct work;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.h
154
struct work_struct work;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.h
168
struct work_struct work;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm.h
295
struct work_struct work;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crc.c
338
static void amdgpu_dm_crtc_notify_ta_to_read(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crc.c
354
crtc_ctx = container_of(work, struct secure_display_crtc_context, notify_ta_work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crc.c
421
amdgpu_dm_forward_crc_window(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crc.c
432
crtc_ctx = container_of(work, struct secure_display_crtc_context, forward_roi_work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
163
schedule_work(&dm->idle_workqueue->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
195
static void amdgpu_dm_idle_worker(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
199
idle_work = container_of(work, struct idle_workqueue, work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
241
INIT_WORK(&idle_work->work, amdgpu_dm_idle_worker);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
246
static void amdgpu_dm_crtc_vblank_control_worker(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
249
container_of(work, struct vblank_control_work, work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
296
struct vblank_control_work *work;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
395
work = kzalloc(sizeof(*work), GFP_ATOMIC);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
396
if (!work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
399
INIT_WORK(&work->work, amdgpu_dm_crtc_vblank_control_worker);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
400
work->dm = dm;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
401
work->acrtc = acrtc;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
402
work->enable = enable;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
406
work->stream = acrtc_state->stream;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_crtc.c
409
queue_work(dm->vblank_control_workqueue, &work->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
182
static void link_lock(struct hdcp_workqueue *work, bool lock)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
186
for (i = 0; i < work->max_link; i++) {
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
188
mutex_lock(&work[i].mutex);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
190
mutex_unlock(&work[i].mutex);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
309
static void event_callback(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
313
hdcp_work = container_of(to_delayed_work(work), struct hdcp_workqueue,
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
326
static void event_property_update(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
328
struct hdcp_workqueue *hdcp_work = container_of(work, struct hdcp_workqueue,
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
396
static void event_property_validate(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
399
container_of(to_delayed_work(work), struct hdcp_workqueue, property_validate_dwork);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
446
static void event_watchdog_timer(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
450
hdcp_work = container_of(to_delayed_work(work),
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
465
static void event_cpirq(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
469
hdcp_work = container_of(work, struct hdcp_workqueue, cpirq_work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
648
struct hdcp_workqueue *work;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
651
work = container_of(bin_attr, struct hdcp_workqueue, attr);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
652
link_lock(work, true);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
654
memcpy(work->srm_temp + pos, buffer, count);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
656
if (!psp_set_srm(work->hdcp.config.psp.handle, work->srm_temp, pos + count, &srm_version)) {
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
658
memcpy(work->srm, work->srm_temp, pos + count);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
659
work->srm_size = pos + count;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
660
work->srm_version = srm_version;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
663
link_lock(work, false);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
672
struct hdcp_workqueue *work;
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
678
work = container_of(bin_attr, struct hdcp_workqueue, attr);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
680
link_lock(work, true);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
682
srm = psp_get_srm(work->hdcp.config.psp.handle, &srm_version, &srm_size);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.c
701
link_lock(work, false);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.h
84
void hdcp_reset_display(struct hdcp_workqueue *work, unsigned int link_index);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.h
85
void hdcp_handle_cpirq(struct hdcp_workqueue *work, unsigned int link_index);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_hdcp.h
86
void hdcp_destroy(struct kobject *kobj, struct hdcp_workqueue *work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c
1361
schedule_work(&adev->dm.idle_workqueue->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
115
static void dm_irq_work_func(struct work_struct *work)
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
118
container_of(work, struct amdgpu_dm_irq_handler_data, work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
338
INIT_WORK(&handler_data->work, dm_irq_work_func);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
469
flush_work(&handler->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
508
flush_work(&handler->work);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
577
if (queue_work(system_highpri_wq, &handler_data->work)) {
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
603
INIT_WORK(&handler_data_add->work, dm_irq_work_func);
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
605
if (queue_work(system_highpri_wq, &handler_data_add->work))
sys/dev/pci/drm/amd/display/amdgpu_dm/amdgpu_dm_irq.c
87
struct work_struct work;
sys/dev/pci/drm/amd/pm/inc/amdgpu_dpm.h
77
struct work_struct work;
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
3011
INIT_WORK(&adev->pm.dpm.thermal.work, amdgpu_dpm_thermal_work_handler);
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
3032
flush_work(&adev->pm.dpm.thermal.work);
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
3074
cancel_work_sync(&adev->pm.dpm.thermal.work);
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
3190
schedule_work(&adev->pm.dpm.thermal.work);
sys/dev/pci/drm/amd/pm/legacy-dpm/legacy_dpm.c
975
void amdgpu_dpm_thermal_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/pm/legacy-dpm/legacy_dpm.c
978
container_of(work, struct amdgpu_device,
sys/dev/pci/drm/amd/pm/legacy-dpm/legacy_dpm.c
979
pm.dpm.thermal.work);
sys/dev/pci/drm/amd/pm/legacy-dpm/legacy_dpm.h
36
void amdgpu_dpm_thermal_work_handler(struct work_struct *work);
sys/dev/pci/drm/amd/pm/legacy-dpm/si_dpm.c
7671
schedule_work(&adev->pm.dpm.thermal.work);
sys/dev/pci/drm/amd/pm/legacy-dpm/si_dpm.c
7804
INIT_WORK(&adev->pm.dpm.thermal.work, amdgpu_dpm_thermal_work_handler);
sys/dev/pci/drm/amd/pm/legacy-dpm/si_dpm.c
7825
flush_work(&adev->pm.dpm.thermal.work);
sys/dev/pci/drm/amd/pm/legacy-dpm/si_dpm.c
7867
cancel_work_sync(&adev->pm.dpm.thermal.work);
sys/dev/pci/drm/amd/pm/powerplay/amd_powerplay.c
102
container_of(work, struct pp_hwmgr, swctf_delayed_work.work);
sys/dev/pci/drm/amd/pm/powerplay/amd_powerplay.c
99
static void pp_swctf_delayed_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1239
static void smu_throttling_logging_work_fn(struct work_struct *work)
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1241
struct smu_context *smu = container_of(work, struct smu_context,
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1247
static void smu_interrupt_work_fn(struct work_struct *work)
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1249
struct smu_context *smu = container_of(work, struct smu_context,
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1256
static void smu_swctf_delayed_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1259
container_of(work, struct smu_context, swctf_delayed_work.work);
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1573
static void smu_wbrf_delayed_work_handler(struct work_struct *work)
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
1575
struct smu_context *smu = container_of(work, struct smu_context, wbrf_delayed_work.work);
sys/dev/pci/drm/apple/afk.c
16
struct work_struct work;
sys/dev/pci/drm/apple/afk.c
634
struct afk_receive_message_work *work;
sys/dev/pci/drm/apple/afk.c
637
work = container_of(work_, struct afk_receive_message_work, work);
sys/dev/pci/drm/apple/afk.c
639
type = FIELD_GET(RBEP_TYPE, work->message);
sys/dev/pci/drm/apple/afk.c
645
complete_all(&work->ep->started);
sys/dev/pci/drm/apple/afk.c
649
complete_all(&work->ep->stopped);
sys/dev/pci/drm/apple/afk.c
653
afk_getbuf(work->ep, work->message);
sys/dev/pci/drm/apple/afk.c
657
afk_init_rxtx(work->ep, work->message, &work->ep->txbfr);
sys/dev/pci/drm/apple/afk.c
661
afk_init_rxtx(work->ep, work->message, &work->ep->rxbfr);
sys/dev/pci/drm/apple/afk.c
665
while (afk_recv(work->ep))
sys/dev/pci/drm/apple/afk.c
670
dev_err(work->ep->dcp->dev,
sys/dev/pci/drm/apple/afk.c
674
kfree(work);
sys/dev/pci/drm/apple/afk.c
679
struct afk_receive_message_work *work;
sys/dev/pci/drm/apple/afk.c
682
work = kzalloc(sizeof(*work), GFP_KERNEL);
sys/dev/pci/drm/apple/afk.c
683
if (!work)
sys/dev/pci/drm/apple/afk.c
686
work->ep = ep;
sys/dev/pci/drm/apple/afk.c
687
work->message = message;
sys/dev/pci/drm/apple/afk.c
688
INIT_WORK(&work->work, afk_receive_message_worker);
sys/dev/pci/drm/apple/afk.c
689
queue_work(ep->wq, &work->work);
sys/dev/pci/drm/apple/dcp.c
104
static void dcp_delayed_vblank(struct work_struct *work)
sys/dev/pci/drm/apple/dcp.c
108
dcp = container_of(work, struct apple_dcp, vblank_wq);
sys/dev/pci/drm/apple/dcp.c
497
static void dcp_work_register_backlight(struct work_struct *work)
sys/dev/pci/drm/apple/dcp.c
502
dcp = container_of(work, struct apple_dcp, bl_register_wq);
sys/dev/pci/drm/apple/dcp.c
519
static void dcp_work_update_backlight(struct work_struct *work)
sys/dev/pci/drm/apple/dcp.c
523
dcp = container_of(work, struct apple_dcp, bl_update_wq);
sys/dev/pci/drm/apple/dcp.h
25
void dcp_hotplug(struct work_struct *work);
sys/dev/pci/drm/apple/iomfb.c
233
void dcp_hotplug(struct work_struct *work)
sys/dev/pci/drm/apple/iomfb.c
238
connector = container_of(work, struct apple_connector, hotplug_wq);
sys/dev/pci/drm/apple/systemep.c
28
struct work_struct work;
sys/dev/pci/drm/apple/systemep.c
33
struct systemep_work *work =
sys/dev/pci/drm/apple/systemep.c
34
container_of(work_, struct systemep_work, work);
sys/dev/pci/drm/apple/systemep.c
36
afk_send_command(work->service, SYSTEM_SET_PROPERTY,
sys/dev/pci/drm/apple/systemep.c
40
complete(&work->service->ep->dcp->systemep_done);
sys/dev/pci/drm/apple/systemep.c
41
kfree(work);
sys/dev/pci/drm/apple/systemep.c
47
struct systemep_work *work;
sys/dev/pci/drm/apple/systemep.c
56
work = kzalloc(sizeof(*work), GFP_KERNEL);
sys/dev/pci/drm/apple/systemep.c
57
if (!work)
sys/dev/pci/drm/apple/systemep.c
60
work->service = service;
sys/dev/pci/drm/apple/systemep.c
61
INIT_WORK(&work->work, system_log_work);
sys/dev/pci/drm/apple/systemep.c
62
schedule_work(&work->work);
sys/dev/pci/drm/display/drm_dp_helper.c
2278
static void drm_dp_aux_crc_work(struct work_struct *work)
sys/dev/pci/drm/display/drm_dp_helper.c
2280
struct drm_dp_aux *aux = container_of(work, struct drm_dp_aux,
sys/dev/pci/drm/display/drm_dp_mst_topology.c
2664
static void drm_dp_mst_link_probe_work(struct work_struct *work)
sys/dev/pci/drm/display/drm_dp_mst_topology.c
2667
container_of(work, struct drm_dp_mst_topology_mgr, work);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
2714
queue_work(system_long_wq, &mgr->work);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
3776
flush_work(&mgr->work);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
3857
flush_work(&mgr->work);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
4102
queue_work(system_long_wq, &mgr->work);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
4106
static void drm_dp_mst_up_req_work(struct work_struct *work)
sys/dev/pci/drm/display/drm_dp_mst_topology.c
4109
container_of(work, struct drm_dp_mst_topology_mgr,
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5019
static void drm_dp_tx_work(struct work_struct *work)
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5021
struct drm_dp_mst_topology_mgr *mgr = container_of(work, struct drm_dp_mst_topology_mgr, tx_work);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5075
static void drm_dp_delayed_destroy_work(struct work_struct *work)
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5078
container_of(work, struct drm_dp_mst_topology_mgr,
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5745
INIT_WORK(&mgr->work, drm_dp_mst_link_probe_work);
sys/dev/pci/drm/display/drm_dp_mst_topology.c
5781
flush_work(&mgr->work);
sys/dev/pci/drm/drm_atomic_helper.c
2090
static void commit_work(struct work_struct *work)
sys/dev/pci/drm/drm_atomic_helper.c
2092
struct drm_atomic_state *state = container_of(work,
sys/dev/pci/drm/drm_connector.c
205
void drm_connector_free_work_fn(struct work_struct *work)
sys/dev/pci/drm/drm_connector.c
209
container_of(work, struct drm_device, mode_config.connector_free_work);
sys/dev/pci/drm/drm_crtc_internal.h
197
void drm_connector_free_work_fn(struct work_struct *work);
sys/dev/pci/drm/drm_fb_helper.c
357
static void drm_fb_helper_resume_worker(struct work_struct *work)
sys/dev/pci/drm/drm_fb_helper.c
360
struct drm_fb_helper *helper = container_of(work, struct drm_fb_helper,
sys/dev/pci/drm/drm_fb_helper.c
405
static void drm_fb_helper_damage_work(struct work_struct *work)
sys/dev/pci/drm/drm_fb_helper.c
407
struct drm_fb_helper *helper = container_of(work, struct drm_fb_helper, damage_work);
sys/dev/pci/drm/drm_flip_work.c
104
struct drm_flip_work *work = container_of(w, struct drm_flip_work, worker);
sys/dev/pci/drm/drm_flip_work.c
112
spin_lock_irqsave(&work->lock, flags);
sys/dev/pci/drm/drm_flip_work.c
113
list_splice_tail(&work->commited, &tasks);
sys/dev/pci/drm/drm_flip_work.c
114
INIT_LIST_HEAD(&work->commited);
sys/dev/pci/drm/drm_flip_work.c
115
spin_unlock_irqrestore(&work->lock, flags);
sys/dev/pci/drm/drm_flip_work.c
121
work->func(work, task->data);
sys/dev/pci/drm/drm_flip_work.c
135
void drm_flip_work_init(struct drm_flip_work *work,
sys/dev/pci/drm/drm_flip_work.c
138
work->name = name;
sys/dev/pci/drm/drm_flip_work.c
139
INIT_LIST_HEAD(&work->queued);
sys/dev/pci/drm/drm_flip_work.c
140
INIT_LIST_HEAD(&work->commited);
sys/dev/pci/drm/drm_flip_work.c
141
mtx_init(&work->lock, IPL_TTY);
sys/dev/pci/drm/drm_flip_work.c
142
work->func = func;
sys/dev/pci/drm/drm_flip_work.c
144
INIT_WORK(&work->worker, flip_worker);
sys/dev/pci/drm/drm_flip_work.c
154
void drm_flip_work_cleanup(struct drm_flip_work *work)
sys/dev/pci/drm/drm_flip_work.c
156
WARN_ON(!list_empty(&work->queued) || !list_empty(&work->commited));
sys/dev/pci/drm/drm_flip_work.c
47
static void drm_flip_work_queue_task(struct drm_flip_work *work, struct drm_flip_task *task)
sys/dev/pci/drm/drm_flip_work.c
51
spin_lock_irqsave(&work->lock, flags);
sys/dev/pci/drm/drm_flip_work.c
52
list_add_tail(&task->node, &work->queued);
sys/dev/pci/drm/drm_flip_work.c
53
spin_unlock_irqrestore(&work->lock, flags);
sys/dev/pci/drm/drm_flip_work.c
64
void drm_flip_work_queue(struct drm_flip_work *work, void *val)
sys/dev/pci/drm/drm_flip_work.c
71
drm_flip_work_queue_task(work, task);
sys/dev/pci/drm/drm_flip_work.c
73
DRM_ERROR("%s could not allocate task!\n", work->name);
sys/dev/pci/drm/drm_flip_work.c
74
work->func(work, val);
sys/dev/pci/drm/drm_flip_work.c
89
void drm_flip_work_commit(struct drm_flip_work *work,
sys/dev/pci/drm/drm_flip_work.c
94
spin_lock_irqsave(&work->lock, flags);
sys/dev/pci/drm/drm_flip_work.c
95
list_splice_tail(&work->queued, &work->commited);
sys/dev/pci/drm/drm_flip_work.c
96
INIT_LIST_HEAD(&work->queued);
sys/dev/pci/drm/drm_flip_work.c
97
spin_unlock_irqrestore(&work->lock, flags);
sys/dev/pci/drm/drm_flip_work.c
98
queue_work(wq, &work->worker);
sys/dev/pci/drm/drm_framebuffer.c
379
struct work_struct work;
sys/dev/pci/drm/drm_framebuffer.c
385
struct drm_mode_rmfb_work *arg = container_of(w, typeof(*arg), work);
sys/dev/pci/drm/drm_framebuffer.c
464
INIT_WORK_ONSTACK(&arg.work, drm_mode_rmfb_work_fn);
sys/dev/pci/drm/drm_framebuffer.c
469
schedule_work(&arg.work);
sys/dev/pci/drm/drm_framebuffer.c
470
flush_work(&arg.work);
sys/dev/pci/drm/drm_framebuffer.c
471
destroy_work_on_stack(&arg.work);
sys/dev/pci/drm/drm_framebuffer.c
819
INIT_WORK_ONSTACK(&arg.work, drm_mode_rmfb_work_fn);
sys/dev/pci/drm/drm_framebuffer.c
821
schedule_work(&arg.work);
sys/dev/pci/drm/drm_framebuffer.c
822
flush_work(&arg.work);
sys/dev/pci/drm/drm_framebuffer.c
823
destroy_work_on_stack(&arg.work);
sys/dev/pci/drm/drm_linux.c
229
flush_work(struct work_struct *work)
sys/dev/pci/drm/drm_linux.c
234
if (work->tq)
sys/dev/pci/drm/drm_linux.c
235
taskq_barrier(work->tq);
sys/dev/pci/drm/drm_linux.c
326
kthread_init_work(struct kthread_work *work, void (*func)(struct kthread_work *))
sys/dev/pci/drm/drm_linux.c
328
work->tq = NULL;
sys/dev/pci/drm/drm_linux.c
329
task_set(&work->task, (void (*)(void *))func, work);
sys/dev/pci/drm/drm_linux.c
333
kthread_queue_work(struct kthread_worker *worker, struct kthread_work *work)
sys/dev/pci/drm/drm_linux.c
335
work->tq = worker->tq;
sys/dev/pci/drm/drm_linux.c
336
return task_add(work->tq, &work->task);
sys/dev/pci/drm/drm_linux.c
340
kthread_cancel_work_sync(struct kthread_work *work)
sys/dev/pci/drm/drm_linux.c
342
return task_del(work->tq, &work->task);
sys/dev/pci/drm/drm_linux.c
346
kthread_flush_work(struct kthread_work *work)
sys/dev/pci/drm/drm_linux.c
351
if (work->tq)
sys/dev/pci/drm/drm_linux.c
352
taskq_barrier(work->tq);
sys/dev/pci/drm/drm_probe_helper.c
756
static void output_poll_execute(struct work_struct *work)
sys/dev/pci/drm/drm_probe_helper.c
758
struct delayed_work *delayed_work = to_delayed_work(work);
sys/dev/pci/drm/drm_probe_helper.c
870
struct work_struct *work = current_work();
sys/dev/pci/drm/drm_probe_helper.c
872
return work && work->func == output_poll_execute;
sys/dev/pci/drm/drm_self_refresh_helper.c
68
static void drm_self_refresh_helper_entry_work(struct work_struct *work)
sys/dev/pci/drm/drm_self_refresh_helper.c
71
to_delayed_work(work),
sys/dev/pci/drm/drm_vblank_work.c
113
int drm_vblank_work_schedule(struct drm_vblank_work *work,
sys/dev/pci/drm/drm_vblank_work.c
116
struct drm_vblank_crtc *vblank = work->vblank;
sys/dev/pci/drm/drm_vblank_work.c
124
if (work->cancelling)
sys/dev/pci/drm/drm_vblank_work.c
133
if (list_empty(&work->node)) {
sys/dev/pci/drm/drm_vblank_work.c
137
} else if (work->count == count) {
sys/dev/pci/drm/drm_vblank_work.c
144
work->count = count;
sys/dev/pci/drm/drm_vblank_work.c
154
ret = kthread_queue_work(vblank->worker, &work->base);
sys/dev/pci/drm/drm_vblank_work.c
157
list_del_init(&work->node);
sys/dev/pci/drm/drm_vblank_work.c
162
list_add_tail(&work->node, &vblank->pending_work);
sys/dev/pci/drm/drm_vblank_work.c
189
bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work)
sys/dev/pci/drm/drm_vblank_work.c
191
struct drm_vblank_crtc *vblank = work->vblank;
sys/dev/pci/drm/drm_vblank_work.c
196
if (!list_empty(&work->node)) {
sys/dev/pci/drm/drm_vblank_work.c
197
list_del_init(&work->node);
sys/dev/pci/drm/drm_vblank_work.c
202
work->cancelling++;
sys/dev/pci/drm/drm_vblank_work.c
207
if (kthread_cancel_work_sync(&work->base))
sys/dev/pci/drm/drm_vblank_work.c
211
work->cancelling--;
sys/dev/pci/drm/drm_vblank_work.c
225
void drm_vblank_work_flush(struct drm_vblank_work *work)
sys/dev/pci/drm/drm_vblank_work.c
227
struct drm_vblank_crtc *vblank = work->vblank;
sys/dev/pci/drm/drm_vblank_work.c
231
wait_event_lock_irq(vblank->work_wait_queue, list_empty(&work->node),
sys/dev/pci/drm/drm_vblank_work.c
235
kthread_flush_work(&work->base);
sys/dev/pci/drm/drm_vblank_work.c
269
void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
sys/dev/pci/drm/drm_vblank_work.c
270
void (*func)(struct kthread_work *work))
sys/dev/pci/drm/drm_vblank_work.c
272
kthread_init_work(&work->base, func);
sys/dev/pci/drm/drm_vblank_work.c
273
INIT_LIST_HEAD(&work->node);
sys/dev/pci/drm/drm_vblank_work.c
274
work->vblank = drm_crtc_vblank_crtc(crtc);
sys/dev/pci/drm/drm_vblank_work.c
52
struct drm_vblank_work *work, *next;
sys/dev/pci/drm/drm_vblank_work.c
58
list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
sys/dev/pci/drm/drm_vblank_work.c
59
if (!drm_vblank_passed(count, work->count))
sys/dev/pci/drm/drm_vblank_work.c
62
list_del_init(&work->node);
sys/dev/pci/drm/drm_vblank_work.c
64
kthread_queue_work(vblank->worker, &work->base);
sys/dev/pci/drm/drm_vblank_work.c
76
struct drm_vblank_work *work, *next;
sys/dev/pci/drm/drm_vblank_work.c
83
list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
sys/dev/pci/drm/drm_vblank_work.c
84
list_del_init(&work->node);
sys/dev/pci/drm/i915/display/intel_connector.c
42
static void intel_connector_modeset_retry_work_fn(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_connector.c
44
struct intel_connector *connector = container_of(work, typeof(*connector),
sys/dev/pci/drm/i915/display/intel_crtc.c
433
struct drm_vblank_work *work = to_drm_vblank_work(base);
sys/dev/pci/drm/i915/display/intel_crtc.c
435
container_of(work, typeof(*crtc_state), vblank_work);
sys/dev/pci/drm/i915/display/intel_cursor.c
790
struct drm_vblank_work *work = to_drm_vblank_work(base);
sys/dev/pci/drm/i915/display/intel_cursor.c
792
container_of(work, typeof(*plane_state), unpin_work);
sys/dev/pci/drm/i915/display/intel_display.c
7147
static void intel_atomic_cleanup_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_display.c
7150
container_of(work, struct intel_atomic_state, cleanup_work);
sys/dev/pci/drm/i915/display/intel_display.c
7566
static void intel_atomic_commit_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_display.c
7569
container_of(work, struct intel_atomic_state, base.commit_work);
sys/dev/pci/drm/i915/display/intel_display_irq.c
1705
static void intel_display_vblank_notify_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_display_irq.c
1708
container_of(work, typeof(*display), irq.vblank_notify_work);
sys/dev/pci/drm/i915/display/intel_display_power.c
675
intel_display_power_put_async_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_display_power.c
677
struct intel_display *display = container_of(work, struct intel_display,
sys/dev/pci/drm/i915/display/intel_display_power.c
678
power.domains.async_put_work.work);
sys/dev/pci/drm/i915/display/intel_display_types.h
1440
struct delayed_work work;
sys/dev/pci/drm/i915/display/intel_display_types.h
1664
struct work_struct work;
sys/dev/pci/drm/i915/display/intel_dmc.c
1317
static void dmc_load_work_fn(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_dmc.c
1319
struct intel_dmc *dmc = container_of(work, typeof(*dmc), work);
sys/dev/pci/drm/i915/display/intel_dmc.c
1398
INIT_WORK(&dmc->work, dmc_load_work_fn);
sys/dev/pci/drm/i915/display/intel_dmc.c
1419
queue_work(display->wq.unordered, &dmc->work);
sys/dev/pci/drm/i915/display/intel_dmc.c
1443
flush_work(&dmc->work);
sys/dev/pci/drm/i915/display/intel_dmc.c
1458
flush_work(&dmc->work);
sys/dev/pci/drm/i915/display/intel_dmc.c
66
struct work_struct work;
sys/dev/pci/drm/i915/display/intel_dmc_wl.c
161
queue_delayed_work(display->wq.unordered, &wl->work,
sys/dev/pci/drm/i915/display/intel_dmc_wl.c
165
static void intel_dmc_wl_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_dmc_wl.c
168
container_of(work, struct intel_dmc_wl, work.work);
sys/dev/pci/drm/i915/display/intel_dmc_wl.c
336
INIT_DELAYED_WORK(&wl->work, intel_dmc_wl_work);
sys/dev/pci/drm/i915/display/intel_dmc_wl.c
432
flush_delayed_work(&wl->work);
sys/dev/pci/drm/i915/display/intel_dmc_wl.c
455
cancel_delayed_work(&wl->work);
sys/dev/pci/drm/i915/display/intel_dmc_wl.h
29
struct delayed_work work;
sys/dev/pci/drm/i915/display/intel_drrs.c
129
mod_delayed_work(display->wq.unordered, &crtc->drrs.work, msecs_to_jiffies(1000));
sys/dev/pci/drm/i915/display/intel_drrs.c
209
cancel_delayed_work_sync(&crtc->drrs.work);
sys/dev/pci/drm/i915/display/intel_drrs.c
212
static void intel_drrs_downclock_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_drrs.c
214
struct intel_crtc *crtc = container_of(work, typeof(*crtc), drrs.work.work);
sys/dev/pci/drm/i915/display/intel_drrs.c
256
cancel_delayed_work(&crtc->drrs.work);
sys/dev/pci/drm/i915/display/intel_drrs.c
306
INIT_DELAYED_WORK(&crtc->drrs.work, intel_drrs_downclock_work);
sys/dev/pci/drm/i915/display/intel_encoder.c
13
static void intel_encoder_link_check_work_fn(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_encoder.c
16
container_of(work, typeof(*encoder), link_check_work.work);
sys/dev/pci/drm/i915/display/intel_fbc.c
1952
static void intel_fbc_underrun_work_fn(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_fbc.c
1954
struct intel_fbc *fbc = container_of(work, typeof(*fbc), underrun_work);
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
211
static void intel_frontbuffer_flush_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_frontbuffer.c
214
container_of(work, struct intel_frontbuffer, flush_work);
sys/dev/pci/drm/i915/display/intel_hdcp.c
1190
static void intel_hdcp_prop_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_hdcp.c
1192
struct intel_hdcp *hdcp = container_of(work, struct intel_hdcp,
sys/dev/pci/drm/i915/display/intel_hdcp.c
2234
static void intel_hdcp_check_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_hdcp.c
2236
struct intel_hdcp *hdcp = container_of(to_delayed_work(work),
sys/dev/pci/drm/i915/display/intel_hotplug.c
195
mod_delayed_detection_work(struct intel_display *display, struct delayed_work *work, int delay)
sys/dev/pci/drm/i915/display/intel_hotplug.c
202
return mod_delayed_work(display->wq.unordered, work, delay);
sys/dev/pci/drm/i915/display/intel_hotplug.c
206
queue_delayed_detection_work(struct intel_display *display, struct delayed_work *work, int delay)
sys/dev/pci/drm/i915/display/intel_hotplug.c
213
return queue_delayed_work(display->wq.unordered, work, delay);
sys/dev/pci/drm/i915/display/intel_hotplug.c
217
queue_detection_work(struct intel_display *display, struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_hotplug.c
224
return queue_work(display->wq.unordered, work);
sys/dev/pci/drm/i915/display/intel_hotplug.c
269
static void intel_hpd_irq_storm_reenable_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_hotplug.c
272
container_of(work, typeof(*display), hotplug.reenable_work.work);
sys/dev/pci/drm/i915/display/intel_hotplug.c
390
static void i915_digport_work_func(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_hotplug.c
393
container_of(work, struct intel_display, hotplug.dig_port_work);
sys/dev/pci/drm/i915/display/intel_hotplug.c
468
static void i915_hotplug_work_func(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_hotplug.c
471
container_of(work, struct intel_display, hotplug.hotplug_work.work);
sys/dev/pci/drm/i915/display/intel_hotplug.c
782
static void i915_hpd_poll_init_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_hotplug.c
785
container_of(work, typeof(*display), hotplug.poll_init_work);
sys/dev/pci/drm/i915/display/intel_opregion.c
615
static void asle_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_opregion.c
618
container_of(work, struct intel_opregion, asle_work);
sys/dev/pci/drm/i915/display/intel_psr.c
1167
static void tgl_dc3co_disable_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_psr.c
1170
container_of(work, typeof(*intel_dp), psr.dc3co_work.work);
sys/dev/pci/drm/i915/display/intel_psr.c
2241
cancel_work_sync(&intel_dp->psr.work);
sys/dev/pci/drm/i915/display/intel_psr.c
2272
cancel_work_sync(&psr->work);
sys/dev/pci/drm/i915/display/intel_psr.c
3296
static void intel_psr_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_psr.c
3299
container_of(work, typeof(*intel_dp), psr.work);
sys/dev/pci/drm/i915/display/intel_psr.c
3479
queue_work(display->wq.unordered, &intel_dp->psr.work);
sys/dev/pci/drm/i915/display/intel_psr.c
3583
INIT_WORK(&intel_dp->psr.work, intel_psr_work);
sys/dev/pci/drm/i915/display/intel_psr.c
3828
static void psr_dc5_dc6_wa_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_psr.c
3830
struct intel_display *display = container_of(work, typeof(*display),
sys/dev/pci/drm/i915/display/intel_psr.c
477
queue_work(display->wq.unordered, &intel_dp->psr.work);
sys/dev/pci/drm/i915/display/intel_tc.c
1815
static void intel_tc_port_link_reset_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_tc.c
1818
container_of(work, struct intel_tc_port, link_reset_work.work);
sys/dev/pci/drm/i915/display/intel_tc.c
1885
static void intel_tc_port_disconnect_phy_work(struct work_struct *work)
sys/dev/pci/drm/i915/display/intel_tc.c
1888
container_of(work, struct intel_tc_port, disconnect_phy_work.work);
sys/dev/pci/drm/i915/gem/i915_gem_clflush.c
45
.work = clflush_work,
sys/dev/pci/drm/i915/gem/i915_gem_context.c
1289
static void i915_gem_context_release_work(struct work_struct *work)
sys/dev/pci/drm/i915/gem/i915_gem_context.c
1291
struct i915_gem_context *ctx = container_of(work, typeof(*ctx),
sys/dev/pci/drm/i915/gem/i915_gem_object.c
449
static void __i915_gem_free_work(struct work_struct *work)
sys/dev/pci/drm/i915/gem/i915_gem_object.c
452
container_of(work, struct drm_i915_private, mm.free_work);
sys/dev/pci/drm/i915/gem/i915_gem_pm.h
16
void i915_gem_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
296
struct work_struct work;
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
351
static void __memcpy_work(struct work_struct *work)
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
354
container_of(work, typeof(*copy_work), work);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
413
INIT_WORK(©_work->work, __memcpy_work);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
414
queue_work(system_unbound_wq, ©_work->work);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
437
i915_ttm_memcpy_work_arm(struct i915_ttm_memcpy_work *work,
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
442
mtx_init(&work->lock, IPL_TTY);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
443
dma_fence_init(&work->fence, &dma_fence_memcpy_ops, &work->lock, 0, 0);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
444
dma_fence_get(&work->fence);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
445
ret = dma_fence_add_callback(dep, &work->cb, __memcpy_cb);
sys/dev/pci/drm/i915/gem/i915_gem_ttm_move.c
454
return &work->fence;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
185
struct kthread_work work;
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
190
static void __live_parallel_switch1(struct kthread_work *work)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
193
container_of(work, typeof(*arg), work);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
239
static void __live_parallel_switchN(struct kthread_work *work)
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
242
container_of(work, typeof(*arg), work);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
394
kthread_init_work(&data[n].work, *fn);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
395
kthread_queue_work(data[n].worker, &data[n].work);
sys/dev/pci/drm/i915/gem/selftests/i915_gem_context.c
400
kthread_flush_work(&data[n].work);
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
172
static void signal_irq_work(struct irq_work *work)
sys/dev/pci/drm/i915/gt/intel_breadcrumbs.c
174
struct intel_breadcrumbs *b = container_of(work, typeof(*b), irq_work);
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
140
container_of(wrk, typeof(*engine), heartbeat.work.work);
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
247
if (cancel_delayed_work(&engine->heartbeat.work))
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
272
INIT_DELAYED_WORK(&engine->heartbeat.work, heartbeat);
sys/dev/pci/drm/i915/gt/intel_engine_heartbeat.c
62
mod_delayed_work(system_highpri_wq, &engine->heartbeat.work, delay + 1);
sys/dev/pci/drm/i915/gt/intel_engine_types.h
446
struct delayed_work work;
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2215
struct work_struct work;
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2220
static void execlists_capture_work(struct work_struct *work)
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2222
struct execlists_capture *cap = container_of(work, typeof(*cap), work);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2376
INIT_WORK(&cap->work, execlists_capture_work);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
2377
queue_work(i915->unordered_wq, &cap->work);
sys/dev/pci/drm/i915/gt/intel_execlists_submission.c
3611
container_of(wrk, typeof(*ve), rcu.work);
sys/dev/pci/drm/i915/gt/intel_gt.c
48
INIT_WORK(>->watchdog.work, intel_gt_watchdog_work);
sys/dev/pci/drm/i915/gt/intel_gt.h
201
void intel_gt_watchdog_work(struct work_struct *work);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
122
queue_delayed_work(gt->i915->unordered_wq, &pool->work,
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
229
INIT_DELAYED_WORK(&pool->work, pool_free_work);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
239
} while (cancel_delayed_work_sync(&pool->work));
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
91
container_of(wrk, typeof(*pool), work.work);
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c
95
queue_delayed_work(gt->i915->unordered_wq, &pool->work,
sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool_types.h
19
struct delayed_work work;
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
205
static void retire_work_handler(struct work_struct *work)
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
208
container_of(work, typeof(*gt), requests.retire_work.work);
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
236
flush_work(>->watchdog.work);
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
239
void intel_gt_watchdog_work(struct work_struct *work)
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
242
container_of(work, typeof(*gt), watchdog.work);
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
51
flush_delayed_work(&engine->wakeref.work);
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
60
static void engine_retire(struct work_struct *work)
sys/dev/pci/drm/i915/gt/intel_gt_requests.c
63
container_of(work, typeof(*engine), retire_work);
sys/dev/pci/drm/i915/gt/intel_gt_types.h
146
struct work_struct work;
sys/dev/pci/drm/i915/gt/intel_gtt.c
234
static void __i915_vm_release(struct work_struct *work)
sys/dev/pci/drm/i915/gt/intel_gtt.c
237
container_of(work, struct i915_address_space, release_work);
sys/dev/pci/drm/i915/gt/intel_reset.c
1691
static void intel_wedge_me(struct work_struct *work)
sys/dev/pci/drm/i915/gt/intel_reset.c
1693
struct intel_wedge_me *w = container_of(work, typeof(*w), work.work);
sys/dev/pci/drm/i915/gt/intel_reset.c
1707
INIT_DELAYED_WORK_ONSTACK(&w->work, intel_wedge_me);
sys/dev/pci/drm/i915/gt/intel_reset.c
1708
queue_delayed_work(gt->i915->unordered_wq, &w->work, timeout);
sys/dev/pci/drm/i915/gt/intel_reset.c
1713
cancel_delayed_work_sync(&w->work);
sys/dev/pci/drm/i915/gt/intel_reset.c
1714
destroy_delayed_work_on_stack(&w->work);
sys/dev/pci/drm/i915/gt/intel_reset.h
65
struct delayed_work work;
sys/dev/pci/drm/i915/gt/intel_rps.c
1071
queue_work(rps_to_gt(rps)->i915->unordered_wq, &rps->work);
sys/dev/pci/drm/i915/gt/intel_rps.c
144
queue_work(gt->i915->unordered_wq, &rps->work);
sys/dev/pci/drm/i915/gt/intel_rps.c
149
queue_work(gt->i915->unordered_wq, &rps->work);
sys/dev/pci/drm/i915/gt/intel_rps.c
171
cancel_work_sync(&rps->work);
sys/dev/pci/drm/i915/gt/intel_rps.c
1823
static void rps_work(struct work_struct *work)
sys/dev/pci/drm/i915/gt/intel_rps.c
1825
struct intel_rps *rps = container_of(work, typeof(*rps), work);
sys/dev/pci/drm/i915/gt/intel_rps.c
1928
queue_work(gt->i915->unordered_wq, &rps->work);
sys/dev/pci/drm/i915/gt/intel_rps.c
1945
queue_work(gt->i915->unordered_wq, &rps->work);
sys/dev/pci/drm/i915/gt/intel_rps.c
1999
INIT_WORK(&rps->work, rps_work);
sys/dev/pci/drm/i915/gt/intel_rps.c
260
cancel_work_sync(&rps->work);
sys/dev/pci/drm/i915/gt/intel_rps.c
986
queue_work(rps_to_gt(rps)->i915->unordered_wq, &rps->work);
sys/dev/pci/drm/i915/gt/intel_rps_types.h
63
struct work_struct work;
sys/dev/pci/drm/i915/gt/selftest_engine_heartbeat.c
203
flush_delayed_work(&engine->heartbeat.work);
sys/dev/pci/drm/i915/gt/selftest_engine_heartbeat.c
204
if (!delayed_work_pending(&engine->heartbeat.work)) {
sys/dev/pci/drm/i915/gt/selftest_engine_heartbeat.c
216
flush_delayed_work(&engine->heartbeat.work);
sys/dev/pci/drm/i915/gt/selftest_engine_heartbeat.c
217
if (delayed_work_pending(&engine->heartbeat.work)) {
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3470
struct kthread_work work;
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3533
static void smoke_crescendo_work(struct kthread_work *work)
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3535
struct preempt_smoke *smoke = container_of(work, typeof(*smoke), work);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3583
kthread_init_work(&arg[id].work, smoke_crescendo_work);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3584
kthread_queue_work(worker[id], &arg[id].work);
sys/dev/pci/drm/i915/gt/selftest_execlists.c
3592
kthread_flush_work(&arg[id].work);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1039
kthread_init_work(&threads[tmp].work, active_engine);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1041
&threads[tmp].work);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
1195
kthread_flush_work(&threads[tmp].work);
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
855
struct kthread_work work;
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
891
static void active_engine(struct kthread_work *work)
sys/dev/pci/drm/i915/gt/selftest_hangcheck.c
894
struct active_engine *arg = container_of(work, typeof(*arg), work);
sys/dev/pci/drm/i915/gt/selftest_rps.c
1052
saved_work = rps->work.func;
sys/dev/pci/drm/i915/gt/selftest_rps.c
1053
rps->work.func = dummy_rps_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
1093
rps->work.func = saved_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
1159
saved_work = rps->work.func;
sys/dev/pci/drm/i915/gt/selftest_rps.c
1160
rps->work.func = dummy_rps_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
1232
rps->work.func = saved_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
237
saved_work = rps->work.func;
sys/dev/pci/drm/i915/gt/selftest_rps.c
238
rps->work.func = dummy_rps_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
364
rps->work.func = saved_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
400
saved_work = rps->work.func;
sys/dev/pci/drm/i915/gt/selftest_rps.c
401
rps->work.func = dummy_rps_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
499
rps->work.func = saved_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
630
saved_work = rps->work.func;
sys/dev/pci/drm/i915/gt/selftest_rps.c
631
rps->work.func = dummy_rps_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
735
rps->work.func = saved_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
769
saved_work = rps->work.func;
sys/dev/pci/drm/i915/gt/selftest_rps.c
770
rps->work.func = dummy_rps_work;
sys/dev/pci/drm/i915/gt/selftest_rps.c
873
rps->work.func = saved_work;
sys/dev/pci/drm/i915/gt/selftest_slpc.c
21
struct kthread_work work;
sys/dev/pci/drm/i915/gt/selftest_slpc.c
488
static void slpc_spinner_thread(struct kthread_work *work)
sys/dev/pci/drm/i915/gt/selftest_slpc.c
490
struct slpc_thread *thread = container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/gt/selftest_slpc.c
515
kthread_init_work(&threads[i].work, slpc_spinner_thread);
sys/dev/pci/drm/i915/gt/selftest_slpc.c
516
kthread_queue_work(threads[i].worker, &threads[i].work);
sys/dev/pci/drm/i915/gt/selftest_slpc.c
525
kthread_flush_work(&threads[i].work);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_proxy.c
307
queue_work(gsc->wq, &gsc->work);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
128
INIT_WORK(&gsc->work, gsc_work);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
16
static void gsc_work(struct work_struct *work)
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
18
struct intel_gsc_uc *gsc = container_of(work, typeof(*gsc), work);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
258
flush_work(&gsc->work);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
279
flush_work(&gsc->work);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.c
315
queue_work(gsc->wq, &gsc->work);
sys/dev/pci/drm/i915/gt/uc/intel_gsc_uc.h
45
struct work_struct work;
sys/dev/pci/drm/i915/gt/uc/intel_guc.h
273
struct delayed_work work;
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
492
static void copy_debug_logs_work(struct work_struct *work)
sys/dev/pci/drm/i915/gt/uc/intel_guc_log.c
495
container_of(work, struct intel_guc_log, relay.flush_work);
sys/dev/pci/drm/i915/gt/uc/intel_guc_slpc.c
255
static void slpc_boost_work(struct work_struct *work)
sys/dev/pci/drm/i915/gt/uc/intel_guc_slpc.c
257
struct intel_guc_slpc *slpc = container_of(work, typeof(*slpc), boost_work);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
1388
mod_delayed_work(system_highpri_wq, &guc->timestamp.work, guc->timestamp.ping_delay);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
1432
cancel_delayed_work(&guc->timestamp.work);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
1434
cancel_delayed_work_sync(&guc->timestamp.work);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
1523
timestamp.work.work);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
3344
container_of(wrk, typeof(*ce), guc_state.sched_disable_delay_work.work);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
3402
__delay_sched_disable(&ce->guc_state.sched_disable_delay_work.work);
sys/dev/pci/drm/i915/gt/uc/intel_guc_submission.c
4871
INIT_DELAYED_WORK(&guc->timestamp.work, guc_timestamp_ping);
sys/dev/pci/drm/i915/i915_active.c
183
struct i915_active *ref = container_of(wrk, typeof(*ref), work);
sys/dev/pci/drm/i915/i915_active.c
200
queue_work(system_unbound_wq, &ref->work);
sys/dev/pci/drm/i915/i915_active.c
370
INIT_WORK(&ref->work, active_work);
sys/dev/pci/drm/i915/i915_active.c
372
lockdep_init_map(&ref->work.lockdep_map, "i915_active.work", wkey, 0);
sys/dev/pci/drm/i915/i915_active.c
616
flush_work(&ref->work);
sys/dev/pci/drm/i915/i915_active.c
755
GEM_BUG_ON(work_pending(&ref->work));
sys/dev/pci/drm/i915/i915_active_types.h
42
struct work_struct work;
sys/dev/pci/drm/i915/i915_irq.c
156
static void ivb_parity_work(struct work_struct *work)
sys/dev/pci/drm/i915/i915_irq.c
159
container_of(work, typeof(*dev_priv), l3_parity.error_work);
sys/dev/pci/drm/i915/i915_request.c
195
struct execute_cb *cb = container_of(wrk, typeof(*cb), work);
sys/dev/pci/drm/i915/i915_request.c
215
work.node.llist)
sys/dev/pci/drm/i915/i915_request.c
216
fn(&cb->work);
sys/dev/pci/drm/i915/i915_request.c
305
queue_work(gt->i915->unordered_wq, >->watchdog.work);
sys/dev/pci/drm/i915/i915_request.c
323
queue_work(gt->i915->unordered_wq, >->watchdog.work);
sys/dev/pci/drm/i915/i915_request.c
53
struct irq_work work;
sys/dev/pci/drm/i915/i915_request.c
568
init_irq_work(&cb->work, irq_execute_cb);
sys/dev/pci/drm/i915/i915_request.c
583
if (llist_add(&cb->work.node.llist, &signal->execute_cb)) {
sys/dev/pci/drm/i915/i915_sw_fence.c
414
struct irq_work work;
sys/dev/pci/drm/i915/i915_sw_fence.c
466
irq_work_queue(&cb->work);
sys/dev/pci/drm/i915/i915_sw_fence.c
472
container_of(wrk, typeof(*cb), work);
sys/dev/pci/drm/i915/i915_sw_fence.c
522
init_irq_work(&timer->work, irq_i915_sw_fence_work);
sys/dev/pci/drm/i915/i915_sw_fence_work.c
16
static void fence_work(struct work_struct *work)
sys/dev/pci/drm/i915/i915_sw_fence_work.c
18
struct dma_fence_work *f = container_of(work, typeof(*f), work);
sys/dev/pci/drm/i915/i915_sw_fence_work.c
20
f->ops->work(f);
sys/dev/pci/drm/i915/i915_sw_fence_work.c
39
fence_work(&f->work);
sys/dev/pci/drm/i915/i915_sw_fence_work.c
41
queue_work(system_unbound_wq, &f->work);
sys/dev/pci/drm/i915/i915_sw_fence_work.c
90
INIT_WORK(&f->work, fence_work);
sys/dev/pci/drm/i915/i915_sw_fence_work.h
20
void (*work)(struct dma_fence_work *f);
sys/dev/pci/drm/i915/i915_sw_fence_work.h
31
struct work_struct work;
sys/dev/pci/drm/i915/i915_vma.c
1445
struct i915_vma_work *work = NULL;
sys/dev/pci/drm/i915/i915_vma.c
1483
work = i915_vma_work();
sys/dev/pci/drm/i915/i915_vma.c
1484
if (!work) {
sys/dev/pci/drm/i915/i915_vma.c
1489
work->vm = vma->vm;
sys/dev/pci/drm/i915/i915_vma.c
1495
dma_fence_work_chain(&work->base, moving);
sys/dev/pci/drm/i915/i915_vma.c
1500
&work->stash,
sys/dev/pci/drm/i915/i915_vma.c
1505
err = i915_vm_map_pt_stash(vma->vm, &work->stash);
sys/dev/pci/drm/i915/i915_vma.c
1579
flags, work, vma_res);
sys/dev/pci/drm/i915/i915_vma.c
1608
if (work) {
sys/dev/pci/drm/i915/i915_vma.c
1618
dma_fence_work_commit(&work->base);
sys/dev/pci/drm/i915/i915_vma.c
1620
dma_fence_work_commit_imm(&work->base);
sys/dev/pci/drm/i915/i915_vma.c
365
static void __vma_bind(struct dma_fence_work *work)
sys/dev/pci/drm/i915/i915_vma.c
367
struct i915_vma_work *vw = container_of(work, typeof(*vw), base);
sys/dev/pci/drm/i915/i915_vma.c
383
static void __vma_release(struct dma_fence_work *work)
sys/dev/pci/drm/i915/i915_vma.c
385
struct i915_vma_work *vw = container_of(work, typeof(*vw), base);
sys/dev/pci/drm/i915/i915_vma.c
397
.work = __vma_bind,
sys/dev/pci/drm/i915/i915_vma.c
485
struct i915_vma_work *work,
sys/dev/pci/drm/i915/i915_vma.c
523
if (work && bind_flags & vma->vm->bind_async_flags)
sys/dev/pci/drm/i915/i915_vma.c
525
&work->base.chain,
sys/dev/pci/drm/i915/i915_vma.c
549
if (work && bind_flags & vma->vm->bind_async_flags) {
sys/dev/pci/drm/i915/i915_vma.c
552
work->vma_res = i915_vma_resource_get(vma->resource);
sys/dev/pci/drm/i915/i915_vma.c
553
work->pat_index = pat_index;
sys/dev/pci/drm/i915/i915_vma.c
554
work->flags = bind_flags;
sys/dev/pci/drm/i915/i915_vma.c
565
prev = i915_active_set_exclusive(&vma->active, &work->base.dma);
sys/dev/pci/drm/i915/i915_vma.c
567
__i915_sw_fence_await_dma_fence(&work->base.chain,
sys/dev/pci/drm/i915/i915_vma.c
569
&work->cb);
sys/dev/pci/drm/i915/i915_vma.c
573
work->base.dma.error = 0; /* enable the queue_work() */
sys/dev/pci/drm/i915/i915_vma.c
574
work->obj = i915_gem_object_get(vma->obj);
sys/dev/pci/drm/i915/i915_vma.h
259
struct i915_vma_work *work,
sys/dev/pci/drm/i915/i915_vma_resource.c
243
static void i915_vma_resource_unbind_work(struct work_struct *work)
sys/dev/pci/drm/i915/i915_vma_resource.c
246
container_of(work, typeof(*vma_res), work);
sys/dev/pci/drm/i915/i915_vma_resource.c
272
i915_vma_resource_unbind_work(&vma_res->work);
sys/dev/pci/drm/i915/i915_vma_resource.c
274
INIT_WORK(&vma_res->work, i915_vma_resource_unbind_work);
sys/dev/pci/drm/i915/i915_vma_resource.c
275
queue_work(system_unbound_wq, &vma_res->work);
sys/dev/pci/drm/i915/i915_vma_resource.h
105
struct work_struct work;
sys/dev/pci/drm/i915/intel_wakeref.c
116
INIT_DELAYED_WORK(&wf->work, __intel_wakeref_put_work);
sys/dev/pci/drm/i915/intel_wakeref.c
117
lockdep_init_map(&wf->work.work.lockdep_map,
sys/dev/pci/drm/i915/intel_wakeref.c
118
"wakeref.work", &key->work, 0);
sys/dev/pci/drm/i915/intel_wakeref.c
76
INTEL_WAKEREF_BUG_ON(delayed_work_pending(&wf->work));
sys/dev/pci/drm/i915/intel_wakeref.c
80
mod_delayed_work(wf->i915->unordered_wq, &wf->work,
sys/dev/pci/drm/i915/intel_wakeref.c
90
struct intel_wakeref *wf = container_of(wrk, typeof(*wf), work.work);
sys/dev/pci/drm/i915/intel_wakeref.h
229
flush_delayed_work(&wf->work);
sys/dev/pci/drm/i915/intel_wakeref.h
49
struct delayed_work work;
sys/dev/pci/drm/i915/intel_wakeref.h
58
struct lock_class_key work;
sys/dev/pci/drm/i915/pxp/intel_pxp_session.c
148
static void pxp_session_work(struct work_struct *work)
sys/dev/pci/drm/i915/pxp/intel_pxp_session.c
150
struct intel_pxp *pxp = container_of(work, typeof(*pxp), session_work);
sys/dev/pci/drm/i915/selftests/i915_active.c
351
flush_work(&ref->work);
sys/dev/pci/drm/i915/selftests/i915_request.c
1458
struct kthread_work work;
sys/dev/pci/drm/i915/selftests/i915_request.c
1463
static void __live_parallel_engine1(struct kthread_work *work)
sys/dev/pci/drm/i915/selftests/i915_request.c
1466
container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/selftests/i915_request.c
1501
static void __live_parallel_engineN(struct kthread_work *work)
sys/dev/pci/drm/i915/selftests/i915_request.c
1504
container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/selftests/i915_request.c
1553
static void __live_parallel_spin(struct kthread_work *work)
sys/dev/pci/drm/i915/selftests/i915_request.c
1556
container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/selftests/i915_request.c
1658
kthread_init_work(&threads[idx].work, *fn);
sys/dev/pci/drm/i915/selftests/i915_request.c
1659
kthread_queue_work(worker, &threads[idx].work);
sys/dev/pci/drm/i915/selftests/i915_request.c
1670
kthread_flush_work(&threads[idx].work);
sys/dev/pci/drm/i915/selftests/i915_request.c
1817
kthread_init_work(&threads[i].work,
sys/dev/pci/drm/i915/selftests/i915_request.c
1819
kthread_queue_work(worker, &threads[i].work);
sys/dev/pci/drm/i915/selftests/i915_request.c
1840
kthread_flush_work(&threads[i].work);
sys/dev/pci/drm/i915/selftests/i915_request.c
2958
struct kthread_work work;
sys/dev/pci/drm/i915/selftests/i915_request.c
2963
static void p_sync0(struct kthread_work *work)
sys/dev/pci/drm/i915/selftests/i915_request.c
2965
struct p_thread *thread = container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/selftests/i915_request.c
303
struct kthread_work work;
sys/dev/pci/drm/i915/selftests/i915_request.c
3037
static void p_sync1(struct kthread_work *work)
sys/dev/pci/drm/i915/selftests/i915_request.c
3039
struct p_thread *thread = container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/selftests/i915_request.c
309
static void __igt_breadcrumbs_smoketest(struct kthread_work *work)
sys/dev/pci/drm/i915/selftests/i915_request.c
311
struct smoke_thread *thread = container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/selftests/i915_request.c
3114
static void p_many(struct kthread_work *work)
sys/dev/pci/drm/i915/selftests/i915_request.c
3116
struct p_thread *thread = container_of(work, typeof(*thread), work);
sys/dev/pci/drm/i915/selftests/i915_request.c
3233
kthread_init_work(&engines[idx].work, *fn);
sys/dev/pci/drm/i915/selftests/i915_request.c
3234
kthread_queue_work(worker, &engines[idx].work);
sys/dev/pci/drm/i915/selftests/i915_request.c
3245
kthread_flush_work(&engines[idx].work);
sys/dev/pci/drm/i915/selftests/i915_request.c
506
kthread_init_work(&threads[n].work,
sys/dev/pci/drm/i915/selftests/i915_request.c
508
kthread_queue_work(worker, &threads[n].work);
sys/dev/pci/drm/i915/selftests/i915_request.c
517
kthread_flush_work(&threads[n].work);
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
506
struct work_struct work;
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
512
static void task_ipc(struct work_struct *work)
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
514
struct task_ipc *ipc = container_of(work, typeof(*ipc), work);
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
549
INIT_WORK_ONSTACK(&ipc.work, task_ipc);
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
550
queue_work(wq, &ipc.work);
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
568
flush_work(&ipc.work);
sys/dev/pci/drm/i915/selftests/i915_sw_fence.c
569
destroy_work_on_stack(&ipc.work);
sys/dev/pci/drm/include/drm/display/drm_dp_mst_helper.h
766
struct work_struct work;
sys/dev/pci/drm/include/drm/drm_flip_work.h
51
typedef void (*drm_flip_func_t)(struct drm_flip_work *work, void *val);
sys/dev/pci/drm/include/drm/drm_flip_work.h
71
void drm_flip_work_queue(struct drm_flip_work *work, void *val);
sys/dev/pci/drm/include/drm/drm_flip_work.h
72
void drm_flip_work_commit(struct drm_flip_work *work,
sys/dev/pci/drm/include/drm/drm_flip_work.h
74
void drm_flip_work_init(struct drm_flip_work *work,
sys/dev/pci/drm/include/drm/drm_flip_work.h
76
void drm_flip_work_cleanup(struct drm_flip_work *work);
sys/dev/pci/drm/include/drm/drm_vblank_work.h
65
int drm_vblank_work_schedule(struct drm_vblank_work *work,
sys/dev/pci/drm/include/drm/drm_vblank_work.h
67
void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
sys/dev/pci/drm/include/drm/drm_vblank_work.h
68
void (*func)(struct kthread_work *work));
sys/dev/pci/drm/include/drm/drm_vblank_work.h
69
bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work);
sys/dev/pci/drm/include/drm/drm_vblank_work.h
70
void drm_vblank_work_flush(struct drm_vblank_work *work);
sys/dev/pci/drm/include/drm/gpu_scheduler.h
379
struct work_struct work;
sys/dev/pci/drm/include/linux/irq_work.h
44
init_irq_work(struct irq_work *work, irq_work_func_t func)
sys/dev/pci/drm/include/linux/irq_work.h
46
work->tq = (struct taskq *)system_wq;
sys/dev/pci/drm/include/linux/irq_work.h
47
task_set(&work->task, (void (*)(void *))func, work);
sys/dev/pci/drm/include/linux/irq_work.h
51
irq_work_queue(struct irq_work *work)
sys/dev/pci/drm/include/linux/irq_work.h
53
return task_add(work->tq, &work->task);
sys/dev/pci/drm/include/linux/irq_work.h
57
irq_work_sync(struct irq_work *work)
sys/dev/pci/drm/include/linux/irq_work.h
59
taskq_barrier(work->tq);
sys/dev/pci/drm/include/linux/workqueue.h
102
cancel_work(struct work_struct *work)
sys/dev/pci/drm/include/linux/workqueue.h
104
if (work->tq != NULL)
sys/dev/pci/drm/include/linux/workqueue.h
105
task_del(work->tq, &work->task);
sys/dev/pci/drm/include/linux/workqueue.h
109
cancel_work_sync(struct work_struct *work)
sys/dev/pci/drm/include/linux/workqueue.h
111
if (work->tq != NULL)
sys/dev/pci/drm/include/linux/workqueue.h
112
return task_del(work->tq, &work->task);
sys/dev/pci/drm/include/linux/workqueue.h
116
#define work_pending(work) task_pending(&(work)->task)
sys/dev/pci/drm/include/linux/workqueue.h
119
struct work_struct work;
sys/dev/pci/drm/include/linux/workqueue.h
127
to_delayed_work(struct work_struct *work)
sys/dev/pci/drm/include/linux/workqueue.h
129
return container_of(work, struct delayed_work, work);
sys/dev/pci/drm/include/linux/workqueue.h
137
task_add(dwork->tq, &dwork->work.task);
sys/dev/pci/drm/include/linux/workqueue.h
143
INIT_WORK(&dwork->work, func);
sys/dev/pci/drm/include/linux/workqueue.h
144
timeout_set(&dwork->to, __delayed_work_tick, &dwork->work);
sys/dev/pci/drm/include/linux/workqueue.h
150
INIT_WORK(&dwork->work, func);
sys/dev/pci/drm/include/linux/workqueue.h
151
timeout_set(&dwork->to, __delayed_work_tick, &dwork->work);
sys/dev/pci/drm/include/linux/workqueue.h
157
.work.tq = NULL, \
sys/dev/pci/drm/include/linux/workqueue.h
158
.work.task = TASK_INITIALIZER((void (*)(void *))(fn), &(dw).work) \
sys/dev/pci/drm/include/linux/workqueue.h
162
schedule_work(struct work_struct *work)
sys/dev/pci/drm/include/linux/workqueue.h
164
work->tq = (struct taskq *)system_wq;
sys/dev/pci/drm/include/linux/workqueue.h
165
return task_add(work->tq, &work->task);
sys/dev/pci/drm/include/linux/workqueue.h
198
return task_del(dwork->tq, &dwork->work.task);
sys/dev/pci/drm/include/linux/workqueue.h
208
return task_del(dwork->tq, &dwork->work.task);
sys/dev/pci/drm/include/linux/workqueue.h
216
return task_pending(&dwork->work.task);
sys/dev/pci/drm/include/linux/workqueue.h
236
destroy_work_on_stack(struct work_struct *work)
sys/dev/pci/drm/include/linux/workqueue.h
238
if (work->tq)
sys/dev/pci/drm/include/linux/workqueue.h
239
task_del(work->tq, &work->task);
sys/dev/pci/drm/include/linux/workqueue.h
248
struct work_struct work;
sys/dev/pci/drm/include/linux/workqueue.h
253
INIT_RCU_WORK(struct rcu_work *work, work_func_t func)
sys/dev/pci/drm/include/linux/workqueue.h
255
INIT_WORK(&work->work, func);
sys/dev/pci/drm/include/linux/workqueue.h
259
queue_rcu_work(struct workqueue_struct *wq, struct rcu_work *work)
sys/dev/pci/drm/include/linux/workqueue.h
261
return queue_work(wq, &work->work);
sys/dev/pci/drm/include/linux/workqueue.h
80
INIT_WORK(struct work_struct *work, work_func_t func)
sys/dev/pci/drm/include/linux/workqueue.h
82
work->tq = NULL;
sys/dev/pci/drm/include/linux/workqueue.h
83
task_set(&work->task, (void (*)(void *))func, work);
sys/dev/pci/drm/include/linux/workqueue.h
89
queue_work(struct workqueue_struct *wq, struct work_struct *work)
sys/dev/pci/drm/include/linux/workqueue.h
91
work->tq = (struct taskq *)wq;
sys/dev/pci/drm/include/linux/workqueue.h
92
return task_add(work->tq, &work->task);
sys/dev/pci/drm/include/linux/workqueue.h
96
queue_work_node(int node, struct workqueue_struct *wq, struct work_struct *work)
sys/dev/pci/drm/include/linux/workqueue.h
98
return queue_work(wq, work);
sys/dev/pci/drm/radeon/cik.c
8096
schedule_work(&rdev->pm.dpm.thermal.work);
sys/dev/pci/drm/radeon/evergreen.c
4922
schedule_work(&rdev->pm.dpm.thermal.work);
sys/dev/pci/drm/radeon/r600.c
4331
schedule_work(&rdev->pm.dpm.thermal.work);
sys/dev/pci/drm/radeon/r600_hdmi.c
115
void r600_audio_update_hdmi(struct work_struct *work)
sys/dev/pci/drm/radeon/r600_hdmi.c
117
struct radeon_device *rdev = container_of(work, struct radeon_device,
sys/dev/pci/drm/radeon/radeon.h
1361
struct work_struct work;
sys/dev/pci/drm/radeon/radeon.h
2917
void r600_audio_update_hdmi(struct work_struct *work);
sys/dev/pci/drm/radeon/radeon_display.c
266
struct radeon_flip_work *work =
sys/dev/pci/drm/radeon/radeon_display.c
271
r = radeon_bo_reserve(work->old_rbo, false);
sys/dev/pci/drm/radeon/radeon_display.c
273
radeon_bo_unpin(work->old_rbo);
sys/dev/pci/drm/radeon/radeon_display.c
274
radeon_bo_unreserve(work->old_rbo);
sys/dev/pci/drm/radeon/radeon_display.c
278
drm_gem_object_put(&work->old_rbo->tbo.base);
sys/dev/pci/drm/radeon/radeon_display.c
279
kfree(work);
sys/dev/pci/drm/radeon/radeon_display.c
366
struct radeon_flip_work *work;
sys/dev/pci/drm/radeon/radeon_display.c
374
work = radeon_crtc->flip_work;
sys/dev/pci/drm/radeon/radeon_display.c
389
if (work->event)
sys/dev/pci/drm/radeon/radeon_display.c
390
drm_crtc_send_vblank_event(&radeon_crtc->base, work->event);
sys/dev/pci/drm/radeon/radeon_display.c
395
radeon_irq_kms_pflip_irq_put(rdev, work->crtc_id);
sys/dev/pci/drm/radeon/radeon_display.c
396
queue_work(radeon_crtc->flip_queue, &work->unpin_work);
sys/dev/pci/drm/radeon/radeon_display.c
408
struct radeon_flip_work *work =
sys/dev/pci/drm/radeon/radeon_display.c
410
struct radeon_device *rdev = work->rdev;
sys/dev/pci/drm/radeon/radeon_display.c
412
struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[work->crtc_id];
sys/dev/pci/drm/radeon/radeon_display.c
420
if (work->fence) {
sys/dev/pci/drm/radeon/radeon_display.c
423
fence = to_radeon_fence(work->fence);
sys/dev/pci/drm/radeon/radeon_display.c
434
r = dma_fence_wait(work->fence, false);
sys/dev/pci/drm/radeon/radeon_display.c
444
dma_fence_put(work->fence);
sys/dev/pci/drm/radeon/radeon_display.c
445
work->fence = NULL;
sys/dev/pci/drm/radeon/radeon_display.c
454
(radeon_get_crtc_scanoutpos(dev, work->crtc_id, 0,
sys/dev/pci/drm/radeon/radeon_display.c
460
((int) (work->target_vblank -
sys/dev/pci/drm/radeon/radeon_display.c
471
radeon_page_flip(rdev, radeon_crtc->crtc_id, work->base, work->async);
sys/dev/pci/drm/radeon/radeon_display.c
489
struct radeon_flip_work *work;
sys/dev/pci/drm/radeon/radeon_display.c
496
work = kzalloc(sizeof *work, GFP_KERNEL);
sys/dev/pci/drm/radeon/radeon_display.c
497
if (work == NULL)
sys/dev/pci/drm/radeon/radeon_display.c
500
INIT_WORK(&work->flip_work, radeon_flip_work_func);
sys/dev/pci/drm/radeon/radeon_display.c
501
INIT_WORK(&work->unpin_work, radeon_unpin_work_func);
sys/dev/pci/drm/radeon/radeon_display.c
503
work->rdev = rdev;
sys/dev/pci/drm/radeon/radeon_display.c
504
work->crtc_id = radeon_crtc->crtc_id;
sys/dev/pci/drm/radeon/radeon_display.c
505
work->event = event;
sys/dev/pci/drm/radeon/radeon_display.c
506
work->async = (page_flip_flags & DRM_MODE_PAGE_FLIP_ASYNC) != 0;
sys/dev/pci/drm/radeon/radeon_display.c
513
work->old_rbo = gem_to_radeon_bo(obj);
sys/dev/pci/drm/radeon/radeon_display.c
520
work->old_rbo, new_rbo);
sys/dev/pci/drm/radeon/radeon_display.c
537
&work->fence);
sys/dev/pci/drm/radeon/radeon_display.c
581
work->base = base;
sys/dev/pci/drm/radeon/radeon_display.c
582
work->target_vblank = target - (uint32_t)drm_crtc_vblank_count(crtc) +
sys/dev/pci/drm/radeon/radeon_display.c
595
radeon_crtc->flip_work = work;
sys/dev/pci/drm/radeon/radeon_display.c
602
queue_work(radeon_crtc->flip_queue, &work->flip_work);
sys/dev/pci/drm/radeon/radeon_display.c
614
drm_gem_object_put(&work->old_rbo->tbo.base);
sys/dev/pci/drm/radeon/radeon_display.c
615
dma_fence_put(work->fence);
sys/dev/pci/drm/radeon/radeon_display.c
616
kfree(work);
sys/dev/pci/drm/radeon/radeon_fence.c
267
static void radeon_fence_check_lockup(struct work_struct *work)
sys/dev/pci/drm/radeon/radeon_fence.c
273
fence_drv = container_of(work, struct radeon_fence_driver,
sys/dev/pci/drm/radeon/radeon_fence.c
274
lockup_work.work);
sys/dev/pci/drm/radeon/radeon_irq_kms.c
103
static void radeon_dp_work_func(struct work_struct *work)
sys/dev/pci/drm/radeon/radeon_irq_kms.c
105
struct radeon_device *rdev = container_of(work, struct radeon_device,
sys/dev/pci/drm/radeon/radeon_irq_kms.c
82
static void radeon_hotplug_work_func(struct work_struct *work)
sys/dev/pci/drm/radeon/radeon_irq_kms.c
84
struct radeon_device *rdev = container_of(work, struct radeon_device,
sys/dev/pci/drm/radeon/radeon_irq_kms.c
85
hotplug_work.work);
sys/dev/pci/drm/radeon/radeon_pm.c
1458
INIT_WORK(&rdev->pm.dpm.thermal.work, radeon_dpm_thermal_work_handler);
sys/dev/pci/drm/radeon/radeon_pm.c
1875
static void radeon_dynpm_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/radeon/radeon_pm.c
1879
rdev = container_of(work, struct radeon_device,
sys/dev/pci/drm/radeon/radeon_pm.c
1880
pm.dynpm_idle_work.work);
sys/dev/pci/drm/radeon/radeon_pm.c
50
static void radeon_dynpm_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/radeon/radeon_pm.c
893
static void radeon_dpm_thermal_work_handler(struct work_struct *work)
sys/dev/pci/drm/radeon/radeon_pm.c
896
container_of(work, struct radeon_device,
sys/dev/pci/drm/radeon/radeon_pm.c
897
pm.dpm.thermal.work);
sys/dev/pci/drm/radeon/radeon_uvd.c
64
static void radeon_uvd_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/radeon/radeon_uvd.c
859
static void radeon_uvd_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/radeon/radeon_uvd.c
862
container_of(work, struct radeon_device, uvd.idle_work.work);
sys/dev/pci/drm/radeon/radeon_vce.c
291
static void radeon_vce_idle_work_handler(struct work_struct *work)
sys/dev/pci/drm/radeon/radeon_vce.c
294
container_of(work, struct radeon_device, vce.idle_work.work);
sys/dev/pci/drm/radeon/radeon_vce.c
47
static void radeon_vce_idle_work_handler(struct work_struct *work);
sys/dev/pci/drm/radeon/si.c
6426
schedule_work(&rdev->pm.dpm.thermal.work);
sys/dev/pci/drm/scheduler/sched_entity.c
185
struct drm_sched_job *job = container_of(wrk, typeof(*job), work);
sys/dev/pci/drm/scheduler/sched_entity.c
231
INIT_WORK(&job->work, drm_sched_entity_kill_jobs_work);
sys/dev/pci/drm/scheduler/sched_entity.c
232
schedule_work(&job->work);
sys/dev/pci/drm/scheduler/sched_main.c
543
static void drm_sched_job_timedout(struct work_struct *work)
sys/dev/pci/drm/scheduler/sched_main.c
549
sched = container_of(work, struct drm_gpu_scheduler, work_tdr.work);
sys/dev/pci/drm/ttm/ttm_bo.c
234
static void ttm_bo_delayed_delete(struct work_struct *work)
sys/dev/pci/drm/ttm/ttm_bo.c
238
bo = container_of(work, typeof(*bo), delayed_delete);
sys/dev/usb/dwc2/dwc2.c
1415
task_set(&dw->work, dw->dw_fn, dw->dw_arg);
sys/dev/usb/dwc2/dwc2.c
1416
task_add(dw->dw_wq, &dw->work);
sys/dev/usb/dwc2/dwc2.h
262
struct task work;
sys/dev/usb/dwc2/dwc2_hcd.c
4363
static void dwc2_hcd_phy_reset_func(struct work_struct *work)
sys/dev/usb/dwc2/dwc2_hcd.c
4365
struct dwc2_hsotg *hsotg = container_of(work, struct dwc2_hsotg,
sys/kern/kern_task.c
394
taskq_next_work(struct taskq *tq, struct task *work)
sys/kern/kern_task.c
411
*work = *next; /* copy to caller to avoid races */
sys/kern/kern_task.c
427
struct task work;
sys/kern/kern_task.c
439
while (taskq_next_work(tq, &work)) {
sys/kern/kern_task.c
442
kcov_remote_enter(KCOV_REMOTE_COMMON, work.t_process);
sys/kern/kern_task.c
444
(*work.t_func)(work.t_arg);
sys/kern/kern_task.c
446
kcov_remote_leave(KCOV_REMOTE_COMMON, work.t_process);
sys/lib/libz/infback.c
338
&(state->lenbits), state->work);
sys/lib/libz/infback.c
431
&(state->lenbits), state->work);
sys/lib/libz/infback.c
444
&(state->next), &(state->distbits), state->work);
sys/lib/libz/inflate.c
850
&(state->lenbits), state->work);
sys/lib/libz/inflate.c
944
&(state->lenbits), state->work);
sys/lib/libz/inflate.c
957
&(state->next), &(state->distbits), state->work);
sys/lib/libz/inflate.h
121
unsigned short work[288]; /* work area for code table building */
sys/lib/libz/inftrees.c
154
if (lens[sym] != 0) work[offs[lens[sym]]++] = (unsigned short)sym;
sys/lib/libz/inftrees.c
222
if (work[sym] + 1U < match) {
sys/lib/libz/inftrees.c
224
here.val = work[sym];
sys/lib/libz/inftrees.c
226
else if (work[sym] >= match) {
sys/lib/libz/inftrees.c
227
here.op = (unsigned char)(extra[work[sym] - match]);
sys/lib/libz/inftrees.c
228
here.val = base[work[sym] - match];
sys/lib/libz/inftrees.c
259
len = lens[work[sym]];
sys/lib/libz/inftrees.c
328
unsigned short lens[288], work[288];
sys/lib/libz/inftrees.c
339
inflate_table(LENS, lens, 288, &(next), &(bits), work);
sys/lib/libz/inftrees.c
346
inflate_table(DISTS, lens, 32, &(next), &(bits), work);
sys/lib/libz/inftrees.c
46
unsigned FAR *bits, unsigned short FAR *work) {
sys/lib/libz/inftrees.h
62
unsigned FAR *bits, unsigned short FAR *work);
sys/net/ifq.c
105
work = *t; /* copy to caller to avoid races */
sys/net/ifq.c
109
(*work.t_func)(work.t_arg);
sys/net/ifq.c
88
struct task work;
sys/uvm/uvm_pmemrange.c
636
struct vm_page **work, int is_desperate)
sys/uvm/uvm_pmemrange.c
716
if (work != NULL && *work != NULL &&
sys/uvm/uvm_pmemrange.c
717
atop(VM_PAGE_TO_PHYS(inserted)) <= atop(VM_PAGE_TO_PHYS(*work)) &&
sys/uvm/uvm_pmemrange.c
719
atop(VM_PAGE_TO_PHYS(*work)))
sys/uvm/uvm_pmemrange.c
720
*work = inserted;
usr.bin/deroff/deroff.c
235
void work(void);
usr.bin/deroff/deroff.c
363
work();
usr.bin/ssh/smult_curve25519_ref.c
124
static void mainloop(unsigned int work[64],const unsigned char e[32])
usr.bin/ssh/smult_curve25519_ref.c
145
for (j = 0;j < 32;++j) xzm1[j] = work[j];
usr.bin/ssh/smult_curve25519_ref.c
173
mult(xzn1b + 32,r,work);
usr.bin/ssh/smult_curve25519_ref.c
177
for (j = 0;j < 64;++j) work[j] = xzm[j];
usr.bin/ssh/smult_curve25519_ref.c
251
unsigned int work[96];
usr.bin/ssh/smult_curve25519_ref.c
258
for (i = 0;i < 32;++i) work[i] = p[i];
usr.bin/ssh/smult_curve25519_ref.c
259
mainloop(work,e);
usr.bin/ssh/smult_curve25519_ref.c
260
recip(work + 32,work + 32);
usr.bin/ssh/smult_curve25519_ref.c
261
mult(work + 64,work,work + 32);
usr.bin/ssh/smult_curve25519_ref.c
262
freeze(work + 64);
usr.bin/ssh/smult_curve25519_ref.c
263
for (i = 0;i < 32;++i) q[i] = work[64 + i];