less
size_t less = outlen - usedpages;
if (used < less) {
used -= less;
outlen -= less;
unsigned int less;
less = min(reqlen, sg_dma_len(sg) - skip_len);
nents += DIV_ROUND_UP(less, entlen);
reqlen -= less;
int offset = 0, soffset = 0, less, sless = 0;
less = min_t(unsigned int, sg_dma_len(dst) - offset -
dstlen += less;
offset += less;
.less = mapping_is_less_than,
.less = slab_status_is_less_than,
b = func->less(data + c, data + d, args) ? c : d;
while (b != a && func->less(data + a, data + b, args))
if (func->less(data + b, data + a, args))
bool (*less)(const void *lhs, const void *rhs, void *args);
bool (*less)(struct rb_node *, const struct rb_node *))
if (less(node, parent)) {
bool (*less)(struct rb_node *, const struct rb_node *))
if (less(node, parent))
bool (*less)(struct rb_node *, const struct rb_node *),
if (less(node, parent)) {
__lt_insert(node, root, 0, ops->less);
__lt_insert(node, root, 1, ops->less);
bool (*less)(struct latch_tree_node *a, struct latch_tree_node *b);
bool (*less)(struct latch_tree_node *a, struct latch_tree_node *b))
if (less(ltn, ltp))
.less = bpf_tree_less,
void *less, struct btf_record *rec, u64 off)
bpf_callback_t cb = (bpf_callback_t)less;
bool (less)(struct bpf_rb_node *a, const struct bpf_rb_node *b),
return __bpf_rbtree_add(root, n, (void *)less, meta ? meta->record : NULL, off);
.less = perf_less_group_idx,
.less = mod_tree_less,
.less = params->min_heap ? less_than : greater_than,
.less = params->min_heap ? less_than : greater_than,
.less = params->min_heap ? less_than : greater_than,
.less = params->min_heap ? less_than : greater_than,
return !less(val, min) && !more(val, max);
if (grp->bc_ackers && less(m->bc_acked, grp->bc_snd_nxt - 1))
if (!less(bc_seqno, msg_grp_bc_seqno(_hdr)))
if (less(msg_grp_bc_seqno(hdr), m->bc_rcv_nxt))
if (link_is_bc_sndlink(l) && !less(r->acked, msg_seqno(hdr)))
if (unlikely(less(seqno, l->drop_point))) {
} else if (unlikely(less(seqno, expect))) {
if (less(acked, start)) {
} else if (less(acked, end)) {
if (unlikely(less(seqno, rcv_nxt) || more(seqno, win_lim))) {
return !less(session, curr_session);
return !less(msg_seqno(hdr), l->rcv_nxt_state);
if (less(l->session, msg_dest_session(hdr)))
if (less(acked, r->acked) || (acked == r->acked && !gap && !ga))
if (more(peers_snd_nxt, l->rcv_nxt) && !less(l->rcv_nxt, from))
if (skb_queue_empty(list) || less(seqno, buf_seqno(skb_peek(list)))) {
if (less(seqno, *rcv_nxt)) {
if ((usr != LINK_PROTOCOL) && less(oseqno, rcv_nxt))
if (less(syncpt, n->sync_point))
bool (*less)(struct rb_node *, const struct rb_node *))
if (less(node, parent)) {
bool (*less)(struct rb_node *, const struct rb_node *))
if (less(node, parent))
bool (less)(struct bpf_rb_node *a, const struct bpf_rb_node *b),
#define bpf_rbtree_add(head, node, less) bpf_rbtree_add_impl(head, node, less, NULL, 0)
bool (less)(struct bpf_rb_node *a, const struct bpf_rb_node *b),
#define bpf_rbtree_add(head, node, less) bpf_rbtree_add_impl(head, node, less, NULL, 0)
bpf_rbtree_add(&res->r, &n->node, less);
bpf_rbtree_add(&groot, &n->node, less);
bpf_rbtree_add(&groot, &m->node, less);
bpf_rbtree_add(&groot_array[i], &nodes[i][j]->node, less);
bpf_rbtree_add(&groot_array_one[0], &nodes[2][j]->node, less);
bpf_rbtree_add(&groot, &n->node, less);
bpf_rbtree_add(&groot, &m->node, less);
bpf_rbtree_add(&groot, &o->node, less);
bpf_rbtree_add(&groot, &n->node, less);
static bool less(struct bpf_rb_node *a, const struct bpf_rb_node *b)
bpf_rbtree_add(root, &n->node, less);
bpf_rbtree_add(root, &m->node, less);
bpf_rbtree_add(root, &n->node, less);
bpf_rbtree_add(&groot, &n->node, less);
bpf_rbtree_add(&groot2, &n->node, less);
bpf_rbtree_add(&groot, res, less);
bpf_rbtree_add(&groot, &n->node, less);
static bool less(struct bpf_rb_node *a, const struct bpf_rb_node *b)
bpf_rbtree_add(&groot, &node_a->node, less);
bpf_rbtree_add(&groot, &n->node, less);
bpf_rbtree_add(&groot, &n->node, less);
bpf_rbtree_add(&groot, &n->node, less);
if (bpf_rbtree_add(root, &m->r, less)) {
bpf_rbtree_add(&broot, &n->r, less);
static bool less(struct bpf_rb_node *node_a, const struct bpf_rb_node *node_b)
bpf_rbtree_add(&root, &n->r, less);
bpf_rbtree_add(&root, &n->r, less);
if (bpf_rbtree_add(root, &n->r, less)) {
bpf_rbtree_add(&groot, &n->node, less);
static bool less(struct bpf_rb_node *a, const struct bpf_rb_node *b)
bpf_rbtree_add(&groot, &n->node, less);
bpf_rbtree_add(&groot, &n->node, less);