kick
goto kick;
kick:
goto kick;
.kick = div4_kick,
static void virtio_pcidev_irq_vq_addbuf(struct virtqueue *vq, void *buf, bool kick)
else if (kick)
pv_ops_lock.kick = hv_qlock_kick;
void (*kick)(int cpu);
PVOP_VCALL1(pv_ops_lock, kick, cpu);
pv_ops_lock.kick = kvm_kick_cpu;
.kick = paravirt_nop,
pv_ops_lock.kick = xen_qlock_kick;
bool kick;
kick = virtqueue_kick_prepare(vq->vq);
if (kick)
bool kick;
kick = virtqueue_kick_prepare(vq->vq);
if (kick)
kick = readl(zclk->kick_reg);
kick |= CPG_FRQCRB_KICK;
writel(kick, zclk->kick_reg);
u32 val, kick;
bool kick = true;
kick = false;
if (kick)
dmac->push.kick = nv50_dmac_kick;
void (*kick)(struct nvif_chan *);
void (*kick)(struct nvif_push *push);
push->kick(push);
chan->chan.push.kick = nouveau_channel_kick;
chan->push.kick = nvif_chan_gpfifo_push_kick;
chan->func->gpfifo.kick(chan);
.gpfifo.kick = nvif_chan506f_gpfifo_kick,
.gpfifo.kick = nvif_chan506f_gpfifo_kick,
.gpfifo.kick = nvif_chanc36f_gpfifo_kick,
int timeout, bitout, sched, i, kick, start, strobe;
kick = id ? gameport_time(gameport, SW_KICK) : 0; /* Set up kick timeout for ID packet */
if (kick && (~v & u & 0x01)) { /* Falling edge on axis 0 */
sched = kick; /* Schedule second trigger */
kick = 0; /* Don't schedule next time on falling edge */
bool kick;
kick = __netdev_tx_sent_queue(txring_txq(tx_ring), first->bytecount,
if (!kick)
bool kick;
kick = list_empty(&tasklet_ctx->list);
if (kick)
int cluster_start, curr, count, kick;
kick = -1;
kick = curr;
if (kick >= 0) {
writel(kick, gp->regs + RXDMA_KICK);
bool kick = false;
kick = true;
if (kick && virtqueue_kick_prepare(sq->vq) && virtqueue_notify(sq->vq))
goto kick;
kick:
bool kick;
kick = use_napi ? __netdev_tx_sent_queue(txq, skb->len, xmit_more) :
if (kick) {
if (use_napi && kick && unlikely(!virtqueue_enable_cb_delayed(sq->vq)))
.kick = mt76_dma_kick_queue,
int timeout, int kick);
#define mt76_queue_kick(dev, ...) (dev)->mt76.queue_ops->kick(&((dev)->mt76), __VA_ARGS__)
void (*kick)(struct mt76_dev *dev, struct mt76_queue *q);
.kick = mt76s_tx_kick,
dev->queue_ops->kick(dev, q);
dev->queue_ops->kick(dev, hwq);
dev->queue_ops->kick(dev, q);
dev->queue_ops->kick(dev, q);
.kick = mt76u_tx_kick,
.kick = da8xx_rproc_kick,
.kick = imx_dsp_rproc_kick,
rproc->ops->kick(rproc, 0);
.kick = imx_rproc_kick,
.kick = ingenic_rproc_kick,
.kick = keystone_rproc_kick,
.kick = omap_rproc_kick,
if (rproc->ops->kick == NULL) {
rproc->ops->kick(rproc, notifyid);
.kick = st_rproc_kick,
.kick = stm32_rproc_kick,
.kick = k3_rproc_kick,
.kick = k3_rproc_kick,
.kick = k3_rproc_kick,
.kick = zynqmp_r5_rproc_kick,
glink->tx_pipe->kick(glink->tx_pipe);
void (*kick)(struct qcom_glink_pipe *glink_pipe);
rpm->tx_pipe.native.kick = glink_rpm_tx_kick;
tx_pipe->native.kick = glink_smem_tx_kick;
bool kick)
if (!err && kick)
bool kick;
kick = (sc->flags & SCMD_LAST) != 0;
ret = virtscsi_add_cmd(req_vq, cmd, req_size, sizeof(cmd->resp.cmd), kick);
int kick;
kick = _scsifront_put_rqid(info, id);
if (kick)
int kick;
kick = _scsifront_put_rqid(info, id);
if (kick)
if (dt->kick)
dt->kick(clk);
INIT_WORK(&dev->vqs[i]->kick, vduse_vq_kick_work);
flush_work(&vq->kick);
struct vduse_virtqueue, kick);
schedule_work(&vq->kick);
struct work_struct kick;
r = vhost_poll_start(&vq->poll, vq->kick);
if (dev->vqs[i]->kick && dev->vqs[i]->handle_kick)
if (dev->vqs[i]->kick)
fput(dev->vqs[i]->kick);
if (eventfp != vq->kick) {
pollstop = (filep = vq->kick) != NULL;
pollstart = (vq->kick = eventfp) != NULL;
r = vhost_poll_start(&vq->poll, vq->kick);
vq->kick = NULL;
if (!vhost_vq_get_backend(vq) && !vq->kick) {
struct file *kick;
int kick = 0;
kick = 1;
kick = 1;
if (kick)
void (*kick)(struct rproc *rproc, int vqid);
void (*kick)(struct clk *clk);
bool kick;
kick = (defer_count - 1) == (defer_max >> 1);
if (unlikely(kick))
void kick(void);
kick();
kick();
r = write(info->kick, &v, sizeof(v));
file.fd = info->kick;
info->kick = eventfd(0, EFD_NONBLOCK);
int kick;
info->kick = eventfd(0, EFD_NONBLOCK);
int kick;
r = write(info->kick, &v, sizeof v);
file.fd = info->kick;