worker
bn_mul_mont_f worker = funcs[num / 16 - 1];
if ((*worker)(rp, ap, bp, np, n0))
if ((*worker)(rp, ap, bp, np, n0))
static int test_multi_shared_pkey_common(void (*worker)(void))
|| !start_threads(1, worker))
static struct worker *wp;
struct worker *twp;
static void worker(int, int);
worker(cmd[0], i);
kthread_destroy_worker(struct kthread_worker *worker)
taskqueue_drain_all(worker->tq);
taskqueue_free(worker->tq);
free(worker, M_KMALLOC);
kthread_queue_work(struct kthread_worker *worker, struct kthread_work *work)
error = taskqueue_enqueue_flags(worker->tq, &work->task,
work->tq = worker->tq;
kthread_flush_worker(struct kthread_worker *worker)
taskqueue_drain_all(worker->tq);
#define kthread_init_worker(worker) \
struct kthread_worker *worker = context;
worker->task = current;
struct delayed_work worker;
worker.work);
schedule_delayed_work(&rdev->worker, msecs_to_jiffies(1000));
cancel_delayed_work_sync(&rdev->worker);
INIT_DELAYED_WORK(&rdev->worker, bnxt_re_worker);
schedule_delayed_work(&rdev->worker, msecs_to_jiffies(1000));
if (vbus_ext->worker.ta_context)
TASK_ENQUEUE(&vbus_ext->worker);
TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext);
TASK_ENQUEUE(&vbus_ext->worker);
vbus_ext->worker.ta_context = 0;
struct task worker;
if (vbus_ext->worker.ta_context)
TASK_ENQUEUE(&vbus_ext->worker);
TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext);
TASK_ENQUEUE(&vbus_ext->worker);
vbus_ext->worker.ta_context = 0;
struct task worker;
if (vbus_ext->worker.ta_context)
TASK_ENQUEUE(&vbus_ext->worker);
TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext);
TASK_ENQUEUE(&vbus_ext->worker);
vbus_ext->worker.ta_context = 0;
struct task worker;
struct mlx5e_rl_worker *worker;
return (mlx5e_rl_modify(channel->worker, channel, params->rate_limit.max_rate));
return (mlx5e_rl_query(channel->worker, channel, params));
mlx5e_rl_free(channel->worker, channel);
channel->worker = rlw;
struct thread *worker;
if (nmk->worker)
&nmk->worker, RFNOWAIT /* to be checked */, 0, "nm-kthread-%ld",
nm_prinf("nm_kthread started td %p", nmk->worker);
nmk->worker = NULL;
if (!nmk->worker)
kthread_resume(nmk->worker);
nmk->worker = NULL;
if (nmk->worker)
g_raid_destroy_node(struct g_raid_softc *sc, int worker)
if (worker) {
int g_raid_destroy_node(struct g_raid_softc *sc, int worker);
struct kthread_worker *worker; /* per device cq worker */
if (pthread_create(&threads[i], NULL, worker, (void*)&thread_data[i])){
struct worker *w = arg;
struct worker *workers;
struct pci_vtscsi_worker *worker = (struct pci_vtscsi_worker *)arg;
struct pci_vtscsi_queue *q = worker->vsw_queue;
while (STAILQ_EMPTY(&q->vsq_requests) && !worker->vsw_exiting)
if (worker->vsw_exiting) {