deliver
int deliver = 1;
deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit;
deliver = !kvmhv_is_nestedv2() && (kvmppc_get_msr(vcpu) & MSR_EE) && !crit;
deliver = 0;
if (deliver)
return deliver;
bool deliver;
deliver = false;
deliver = true;
return deliver;
int deliver = 0;
deliver = 1;
deliver = 1;
if (deliver) {
} deliver;
} deliver;
if (!evtchnfd->deliver.port.port ||
evtchnfd->deliver.port.port != data->u.evtchn.deliver.port.port)
if (data->u.evtchn.deliver.port.priority != KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL)
evtchnfd->deliver.port.priority = data->u.evtchn.deliver.port.priority;
if (evtchnfd->deliver.port.vcpu_id != data->u.evtchn.deliver.port.vcpu) {
evtchnfd->deliver.port.vcpu_id = data->u.evtchn.deliver.port.vcpu;
evtchnfd->deliver.port.vcpu_idx = -1;
if (data->u.evtchn.deliver.port.port != data->u.evtchn.send_port)
if (data->u.evtchn.deliver.port.port) {
if (data->u.evtchn.deliver.port.port >= max_evtchn_port(kvm))
eventfd = eventfd_ctx_fdget(data->u.evtchn.deliver.eventfd.fd);
evtchnfd->deliver.eventfd.ctx = eventfd;
if (data->u.evtchn.deliver.port.priority != KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL)
evtchnfd->deliver.port.port = data->u.evtchn.deliver.port.port;
evtchnfd->deliver.port.vcpu_id = data->u.evtchn.deliver.port.vcpu;
evtchnfd->deliver.port.vcpu_idx = -1;
evtchnfd->deliver.port.priority = data->u.evtchn.deliver.port.priority;
if (!evtchnfd->deliver.port.port)
eventfd_ctx_put(evtchnfd->deliver.eventfd.ctx);
if (!evtchnfd->deliver.port.port)
eventfd_ctx_put(evtchnfd->deliver.eventfd.ctx);
if (evtchnfd->deliver.port.port) {
int ret = kvm_xen_set_evtchn(&evtchnfd->deliver.port, vcpu->kvm);
eventfd_signal(evtchnfd->deliver.eventfd.ctx);
if (!evtchnfd->deliver.port.port)
eventfd_ctx_put(evtchnfd->deliver.eventfd.ctx);
struct nct6694_i2c_deliver deliver;
struct nct6694_i2c_deliver *deliver = &data->deliver;
.len = cpu_to_le16(sizeof(*deliver))
memset(deliver, 0, sizeof(*deliver));
deliver->port = data->port;
deliver->br = data->br;
deliver->addr = i2c_8bit_addr_from_msg(msg_temp);
deliver->r_cnt = msg_temp->len;
ret = nct6694_write_msg(data->nct6694, &cmd_hd, deliver);
memcpy(msg_temp->buf, deliver->read_data, msg_temp->len);
deliver->w_cnt = msg_temp->len;
memcpy(deliver->write_data, msg_temp->buf, msg_temp->len);
ret = nct6694_write_msg(data->nct6694, &cmd_hd, deliver);
struct list_head *list = &fault->common.deliver;
list_add(&group->node, &fault->common.deliver);
list_for_each_entry_safe(cur, next, &eventq->deliver, node) {
struct list_head *list = &eventq->deliver;
struct list_head *list = &eventq->deliver;
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
if (!list_empty(&eventq->deliver))
INIT_LIST_HEAD(&eventq->deliver);
list_add_tail(&group->node, &fault->common.deliver);
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
struct list_head deliver;
if (list_is_last(&veventq->lost_events_header.node, &eventq->deliver))
list_add_tail(&vevent->node, &eventq->deliver);
if (act->deliver)
act->deliver = 1;
act->deliver = 1;
act->deliver = 1;
act->deliver = 1;
act->deliver = 1;
act->deliver = 1;
act->deliver = 1;
return !act->deliver;
u16 deliver:1;
goto deliver;
deliver:
u32 dtag, deliver = 0;
deliver = 1;
deliver = 1;
deliver = 1;
deliver = lpfc_els_rcv_fpin_cgn(phba, tlv);
if (deliver)
.deliver = afs_deliver_cb_callback,
.deliver = afs_deliver_cb_init_call_back_state,
.deliver = afs_deliver_cb_init_call_back_state3,
.deliver = afs_deliver_cb_probe,
.deliver = afs_deliver_cb_probe_uuid,
.deliver = afs_deliver_cb_tell_me_about_yourself,
.deliver = afs_deliver_yfs_cb_callback,
.deliver = afs_deliver_fs_store_data,
.deliver = afs_deliver_fs_store_data,
.deliver = afs_deliver_fs_store_data,
.deliver = afs_deliver_fs_store_data,
.deliver = afs_deliver_fs_store_data,
.deliver = afs_deliver_fs_get_volume_status,
.deliver = afs_deliver_fs_xxxx_lock,
.deliver = afs_deliver_fs_xxxx_lock,
.deliver = afs_deliver_fs_xxxx_lock,
.deliver = afs_deliver_fs_give_up_all_callbacks,
.deliver = afs_deliver_fs_get_capabilities,
.deliver = afs_deliver_fs_inline_bulk_status,
.deliver = afs_deliver_fs_fetch_acl,
.deliver = afs_deliver_fs_file_status_and_vol,
.deliver = afs_deliver_fs_fetch_status,
.deliver = afs_deliver_fs_fetch_data,
.deliver = afs_deliver_fs_fetch_data,
.deliver = afs_deliver_fs_create_vnode,
.deliver = afs_deliver_fs_create_vnode,
.deliver = afs_deliver_fs_file_status_and_vol,
.deliver = afs_deliver_fs_file_status_and_vol,
.deliver = afs_deliver_fs_link,
.deliver = afs_deliver_fs_symlink,
.deliver = afs_deliver_fs_rename,
int (*deliver)(struct afs_call *call);
.deliver = afs_deliver_cm_op_id,
ret = call->type->deliver(call);
return call->type->deliver(call);
.deliver = afs_deliver_vl_get_entry_by_name_u,
.deliver = afs_deliver_vl_get_addrs_u,
.deliver = afs_deliver_vl_get_capabilities,
.deliver = afs_deliver_yfsvl_get_endpoints,
.deliver = afs_deliver_yfsvl_get_cell_name,
.deliver = yfs_deliver_fs_rename,
.deliver = yfs_deliver_fs_rename_2,
.deliver = yfs_deliver_fs_rename_1,
.deliver = yfs_deliver_fs_rename_2,
.deliver = yfs_deliver_status_and_volsync,
.deliver = yfs_deliver_status_and_volsync,
.deliver = yfs_deliver_status_and_volsync,
.deliver = yfs_deliver_fs_get_volume_status,
.deliver = yfs_deliver_status_and_volsync,
.deliver = yfs_deliver_status_and_volsync,
.deliver = yfs_deliver_status_and_volsync,
.deliver = yfs_deliver_fs_fetch_status,
.deliver = yfs_deliver_fs_inline_bulk_status,
.deliver = yfs_deliver_fs_fetch_opaque_acl,
.deliver = yfs_deliver_status_and_volsync,
.deliver = yfs_deliver_fs_fetch_data64,
.deliver = yfs_deliver_fs_create_vnode,
.deliver = yfs_deliver_fs_create_vnode,
.deliver = yfs_deliver_fs_remove_file2,
.deliver = yfs_deliver_fs_remove,
.deliver = yfs_deliver_fs_remove,
.deliver = yfs_deliver_fs_link,
.deliver = yfs_deliver_fs_symlink,
deliver(skb, rcv);
deliver(skb, rcv);
deliver(skb, rcv);
deliver(skb, rcv);
deliver(skb, rcv);
deliver(skb, rcv);
bool deliver;
deliver = ipv6_chk_mcast_addr(dev, &hdr->daddr, NULL);
deliver = false;
deliver = true;
if (deliver) {
if (likely(deliver)) {
bool deliver;
if (params->deliver) {
{ .hdr = RX_HDR(1, 10, 8, FL_S | FL_E | FL_TO), .type = 0, .deliver = true },
{ .hdr = RX_HDR(1, 10, 8, FL_S | FL_E | FL_TO), .type = 1, .deliver = false },
{ .hdr = RX_HDR(1, 10, 8, FL_S | FL_E), .type = 0, .deliver = false },
{ .hdr = RX_HDR(1, 10, 8, FL_E | FL_TO), .type = 0, .deliver = false },
{ .hdr = RX_HDR(1, 10, 8, FL_TO), .type = 0, .deliver = false },
{ .hdr = RX_HDR(1, 10, 8, 0), .type = 0, .deliver = false },
bool deliver;
if (params->deliver)
.deliver = true,
.deliver = false,
.deliver = false,
.deliver = false,
.deliver = false,
.deliver = true,
.deliver = true,
call->deliver = rxperf_deliver_param_block;
ret = call->deliver(call);
call->deliver = rxperf_deliver_request;
return call->deliver(call);
int (*deliver)(struct rxperf_call *call);
deliver:
goto deliver;
goto deliver;
goto deliver;
enum sctp_verb deliver;
deliver = SCTP_CMD_CHUNK_ULP;
deliver = SCTP_CMD_RENEGE;
deliver = SCTP_CMD_RENEGE;
sctp_add_cmd_sf(commands, deliver, SCTP_CHUNK(chunk));
bool ack, deliver, update, leave = false;
deliver = true;
deliver = false;
deliver = false;
if (deliver)
} deliver;
ret = oe->deliver(oe, iter);
void ordered_events__init(struct ordered_events *oe, ordered_events__deliver_t deliver,
oe->deliver = deliver;
ordered_events__deliver_t old_deliver = oe->deliver;
ordered_events__deliver_t deliver;
void ordered_events__init(struct ordered_events *oe, ordered_events__deliver_t deliver,
.u.evtchn.deliver.port.port = EVTCHN_TEST1,
.u.evtchn.deliver.port.vcpu = vcpu->id + 1,
.u.evtchn.deliver.port.priority = KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL,
inj.u.evtchn.deliver.port.vcpu = vcpu->id;
inj.u.evtchn.deliver.eventfd.port = 0;
inj.u.evtchn.deliver.eventfd.fd = irq_fd[1];