group_type
MLX5_SET(create_flow_group_in, in, group_type,
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
va_hdr->group_type = cmd->group_type;
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SELF);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SELF);
cmd.group_type = cpu_to_le16(VIRTIO_ADMIN_GROUP_TYPE_SRIOV);
pr_err("\tgroup_type %d (%s)\n", ch->group_type,
dbg_gtype(ch->group_type));
ch->group_type = UBIFS_NO_NODE_GROUP;
ch->group_type = UBIFS_NO_NODE_GROUP;
ch->group_type = UBIFS_LAST_OF_NODE_GROUP;
ch->group_type = UBIFS_IN_NODE_GROUP;
if (ch->group_type != UBIFS_IN_NODE_GROUP)
__u8 group_type;
u8 group_type[0x4];
__le16 group_type;
__le16 group_type;
enum group_type group_type;
.group_type = group_has_spare,
group_type group_classify(unsigned int imbalance_pct,
sgs->group_type = group_classify(env->sd->imbalance_pct, group, sgs);
if (sgs->group_type == group_overloaded)
(sgs->group_type == group_misfit_task) &&
sds->local_stat.group_type != group_has_spare))
if (sgs->group_type > busiest->group_type)
if (sgs->group_type < busiest->group_type)
switch (sgs->group_type) {
(sgs->group_type <= group_fully_busy) &&
sgs->group_type = group_classify(sd->imbalance_pct, group, sgs);
if (sgs->group_type == group_fully_busy ||
sgs->group_type == group_overloaded)
if (sgs->group_type < idlest_sgs->group_type)
if (sgs->group_type > idlest_sgs->group_type)
switch (sgs->group_type) {
.group_type = group_overloaded,
if (local_sgs.group_type < idlest_sgs.group_type)
if (local_sgs.group_type > idlest_sgs.group_type)
switch (local_sgs.group_type) {
if (busiest->group_type == group_misfit_task) {
if (busiest->group_type == group_asym_packing) {
if (busiest->group_type == group_smt_balance) {
if (busiest->group_type == group_imbalanced) {
if (local->group_type == group_has_spare) {
if ((busiest->group_type > group_fully_busy) &&
if (local->group_type < group_overloaded) {
if (busiest->group_type == group_misfit_task)
if (busiest->group_type == group_asym_packing)
if (busiest->group_type == group_imbalanced)
if (local->group_type > busiest->group_type)
if (local->group_type == group_overloaded) {
if (sds.prefer_sibling && local->group_type == group_has_spare &&
if (busiest->group_type != group_overloaded) {
if (busiest->group_type == group_smt_balance &&
u16 group_type = 0;
group_type = NEXTHOP_GRP_TYPE_MPATH;
group_type = NEXTHOP_GRP_TYPE_RES;
if (nla_put_u16(skb, NHA_GROUP_TYPE, group_type))
int group_type;
group_type = ipv6_addr_type(group);
if (group_type != IPV6_ADDR_ANY &&
!(group_type&IPV6_ADDR_MULTICAST))
if (group_type == IPV6_ADDR_ANY) { /* general query */
if (group_type == IPV6_ADDR_ANY) {