event_dev
static struct clock_event_device event_dev = {
DRIVER_NAME, &event_dev);
event_dev.cpumask = cpumask_of(0);
clockevents_config_and_register(&event_dev, rate, 0x2c00, 0xfffffffe);
struct clock_event_device *event_dev = &timer.event_dev;
TIMER_NAME, event_dev);
event_dev->features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT;
event_dev->set_next_event = keystone_set_next_event;
event_dev->set_state_shutdown = keystone_shutdown;
event_dev->set_state_periodic = keystone_set_periodic;
event_dev->set_state_oneshot = keystone_shutdown;
event_dev->cpumask = cpu_possible_mask;
event_dev->owner = THIS_MODULE;
event_dev->name = TIMER_NAME;
event_dev->irq = irq;
clockevents_config_and_register(event_dev, rate, 1, ULONG_MAX);
struct clock_event_device event_dev;
struct net_device *event_dev = netdev_notifier_info_to_dev(ptr);
if (event_dev == mc->ports[i]) {
struct event_dev *edev;
sizeof(struct event_dev) + keymapnamelen + 1,
struct event_dev *edev = dev_id;
static void events_import_bits(struct event_dev *edev,
static void events_import_abs_params(struct event_dev *edev)
struct net_device *event_dev = netdev_notifier_info_to_dev(ptr);
netdev_dbg(event_dev, "%s received %s\n",
if (!(event_dev->priv_flags & IFF_BONDING))
if (event_dev->flags & IFF_MASTER) {
ret = bond_master_netdev_event(event, event_dev);
if (event_dev->flags & IFF_SLAVE)
return bond_slave_netdev_event(event, event_dev);
struct net_device *event_dev = ifa->idev->dev;
if (is_vlan_dev(event_dev))
event_dev = vlan_dev_real_dev(event_dev);
if (event_dev->flags & IFF_MASTER) {
if (event_dev)
parent = event_dev->dev.parent;
cxgb4_clip_get(event_dev, (const u32 *)ifa, 1);
cxgb4_clip_release(event_dev, (const u32 *)ifa, 1);
struct net_device *event_dev = netdev_notifier_info_to_dev(ptr);
if (event_dev->netdev_ops == &device_ops && event == NETDEV_REGISTER) {
ndev_ctx = netdev_priv(event_dev);
ret = check_dev_is_matching_vf(event_dev);
return netvsc_prepare_bonding(event_dev);
return netvsc_register_vf(event_dev, VF_REG_IN_NOTIFIER);
return netvsc_unregister_vf(event_dev);
return netvsc_vf_changed(event_dev, event);
struct net_device *event_dev = netdev_notifier_info_to_dev(ptr);
if (netif_is_failover(event_dev))
return failover_slave_register(event_dev);
return failover_slave_unregister(event_dev);
return failover_slave_link_change(event_dev);
return failover_slave_name_change(event_dev);
struct net_device *event_dev = netdev_notifier_info_to_dev(ptr);
if ((iucv->hs_dev == event_dev) &&
struct net_device *event_dev = netdev_notifier_info_to_dev(ptr);
struct net *net = dev_net(event_dev);
smc_pnet_remove_by_ndev(event_dev);
smc_ib_ndev_change(event_dev, event);
smc_pnet_add_by_ndev(event_dev);
smc_ib_ndev_change(event_dev, event);
smc_pnet_add_base_pnetid(net, event_dev, ndev_pnetid);
event_dev = __pnet_find_base_ndev(event_dev);
if (!smc_pnetid_by_dev_port(event_dev->dev.parent,
event_dev->dev_port, ndev_pnetid)) {