add_list
static bool hclge_sync_from_add_list(struct list_head *add_list,
list_for_each_entry_safe(mac_node, tmp, add_list, node) {
static void hclgevf_sync_from_add_list(struct list_head *add_list,
list_for_each_entry_safe(mac_node, tmp, add_list, node) {
struct list_head *add_list,
if (!list_empty(add_list)) {
list_for_each_entry_safe(f, ftmp, add_list, list) {
struct i40e_aqc_add_macvlan_element_data *add_list,
if (add_list[i].match_method == I40E_AQC_MM_ERR_NO_RES) {
struct i40e_aqc_add_macvlan_element_data *add_list;
add_list = kzalloc(list_size, GFP_ATOMIC);
if (!add_list)
ether_addr_copy(add_list[num_add].mac_addr,
add_list[num_add].vlan_tag = 0;
add_list[num_add].vlan_tag =
add_list[num_add].queue_number = 0;
add_list[num_add].match_method = I40E_AQC_MM_ERR_NO_RES;
add_list[num_add].flags = cpu_to_le16(cmd_flags);
i40e_aqc_add_filters(vsi, vsi_name, add_list,
memset(add_list, 0, list_size);
i40e_aqc_add_filters(vsi, vsi_name, add_list, add_head,
kfree(add_list);
add_list = NULL;
struct hlist_head *add_list,
hlist_add_head(&ep->list_entry, add_list);
LIST_HEAD(add_list);
&add_list);
__pci_bus_assign_resources(bus, &add_list, NULL);
struct list_head *add_list,
if (add_list)
pci_dev_res_remove_from_list(add_list, res);
struct list_head *add_list,
adjust_bridge_window(bridge, res, add_list,
pci_bus_distribute_available_resources(b, add_list, available);
struct list_head *add_list)
add_list, available);
struct list_head *add_list)
pci_bridge_distribute_available_resources(dev, add_list);
pci_root_bus_distribute_available_resources(b, add_list);
struct list_head *add_list = NULL;
add_list = &realloc_head;
__pci_bus_size_bridges(bus, add_list);
pci_root_bus_distribute_available_resources(bus, add_list);
__pci_bus_assign_resources(bus, add_list, &fail_head);
if (WARN_ON_ONCE(add_list && !list_empty(add_list)))
pci_dev_res_free_list(add_list);
LIST_HEAD(add_list);
__pci_bus_size_bridges(parent, &add_list);
pci_bridge_distribute_available_resources(bridge, &add_list);
__pci_bridge_assign_resources(bridge, &add_list, &fail_head);
if (WARN_ON_ONCE(!list_empty(&add_list)))
pci_dev_res_free_list(&add_list);
LIST_HEAD(add_list);
__pci_bus_size_bridges(dev->subordinate, &add_list);
__pci_bus_assign_resources(bus, &add_list, NULL);
if (WARN_ON_ONCE(!list_empty(&add_list)))
pci_dev_res_free_list(&add_list);
LIST_HEAD(add_list);
list_add_tail(&device->add_list_entry, &add_list);
list_for_each_entry_safe(device, next, &add_list, add_list_entry) {
if (!list_empty(&ref->add_list))
list_del(&ref->add_list);
struct btrfs_delayed_ref_node, add_list);
ASSERT(list_empty(&ref->add_list));
list_add_tail(&ref->add_list, &href->ref_add_list);
list_add_tail(&exist->add_list,
ASSERT(!list_empty(&exist->add_list));
list_del_init(&exist->add_list);
INIT_LIST_HEAD(&ref->add_list);
struct list_head add_list;
if (!list_empty(&ref->add_list))
list_del(&ref->add_list);
if (!list_empty(&node->add_list))
list_del_init(&node->add_list);
if (!add_list(buf, buf_len, ext_buf))