arch/m68k/sun3/sun3dvma.c
102
hole = list_entry(cur, struct hole, list);
arch/m68k/sun3/sun3dvma.c
105
prev = hole;
arch/m68k/sun3/sun3dvma.c
109
if(hole->end == prev->start) {
arch/m68k/sun3/sun3dvma.c
110
hole->size += prev->size;
arch/m68k/sun3/sun3dvma.c
111
hole->end = prev->end;
arch/m68k/sun3/sun3dvma.c
121
static inline struct hole *rmcache(void)
arch/m68k/sun3/sun3dvma.c
123
struct hole *ret;
arch/m68k/sun3/sun3dvma.c
132
ret = list_entry(hole_cache.next, struct hole, list);
arch/m68k/sun3/sun3dvma.c
143
struct hole *hole;
arch/m68k/sun3/sun3dvma.c
157
hole = list_entry(cur, struct hole, list);
arch/m68k/sun3/sun3dvma.c
160
newlen = len + ((hole->end - len) & (align-1));
arch/m68k/sun3/sun3dvma.c
164
if(hole->size > newlen) {
arch/m68k/sun3/sun3dvma.c
165
hole->end -= newlen;
arch/m68k/sun3/sun3dvma.c
166
hole->size -= newlen;
arch/m68k/sun3/sun3dvma.c
167
dvma_entry_use(hole->end) = newlen;
arch/m68k/sun3/sun3dvma.c
172
return hole->end;
arch/m68k/sun3/sun3dvma.c
173
} else if(hole->size == newlen) {
arch/m68k/sun3/sun3dvma.c
174
list_move(&(hole->list), &hole_cache);
arch/m68k/sun3/sun3dvma.c
175
dvma_entry_use(hole->start) = newlen;
arch/m68k/sun3/sun3dvma.c
180
return hole->start;
arch/m68k/sun3/sun3dvma.c
194
struct hole *hole;
arch/m68k/sun3/sun3dvma.c
208
hole = list_entry(cur, struct hole, list);
arch/m68k/sun3/sun3dvma.c
210
if(hole->end == baddr) {
arch/m68k/sun3/sun3dvma.c
211
hole->end += len;
arch/m68k/sun3/sun3dvma.c
212
hole->size += len;
arch/m68k/sun3/sun3dvma.c
214
} else if(hole->start == (baddr + len)) {
arch/m68k/sun3/sun3dvma.c
215
hole->start = baddr;
arch/m68k/sun3/sun3dvma.c
216
hole->size += len;
arch/m68k/sun3/sun3dvma.c
222
hole = rmcache();
arch/m68k/sun3/sun3dvma.c
224
hole->start = baddr;
arch/m68k/sun3/sun3dvma.c
225
hole->end = baddr + len;
arch/m68k/sun3/sun3dvma.c
226
hole->size = len;
arch/m68k/sun3/sun3dvma.c
229
list_add(&(hole->list), cur);
arch/m68k/sun3/sun3dvma.c
238
struct hole *hole;
arch/m68k/sun3/sun3dvma.c
248
hole = rmcache();
arch/m68k/sun3/sun3dvma.c
249
hole->start = DVMA_START;
arch/m68k/sun3/sun3dvma.c
250
hole->end = DVMA_END;
arch/m68k/sun3/sun3dvma.c
251
hole->size = DVMA_SIZE;
arch/m68k/sun3/sun3dvma.c
253
list_add(&(hole->list), &hole_list);
arch/m68k/sun3/sun3dvma.c
38
static struct hole initholes[64];
arch/m68k/sun3/sun3dvma.c
76
struct hole *hole;
arch/m68k/sun3/sun3dvma.c
80
hole = list_entry(cur, struct hole, list);
arch/m68k/sun3/sun3dvma.c
82
if((hole->start == 0) && (hole->end == 0) && (hole->size == 0))
arch/m68k/sun3/sun3dvma.c
86
hole->start, hole->end, hole->size);
arch/m68k/sun3/sun3dvma.c
96
struct hole *hole;
arch/m68k/sun3/sun3dvma.c
97
struct hole *prev = NULL;
drivers/gpu/drm/drm_mm.c
453
struct drm_mm_node *hole;
drivers/gpu/drm/drm_mm.c
463
hole = find_hole_addr(mm, node->start, 0);
drivers/gpu/drm/drm_mm.c
464
if (!hole)
drivers/gpu/drm/drm_mm.c
467
adj_start = hole_start = __drm_mm_hole_node_start(hole);
drivers/gpu/drm/drm_mm.c
468
adj_end = hole_end = hole_start + hole->hole_size;
drivers/gpu/drm/drm_mm.c
471
mm->color_adjust(hole, node->color, &adj_start, &adj_end);
drivers/gpu/drm/drm_mm.c
479
list_add(&node->node_list, &hole->node_list);
drivers/gpu/drm/drm_mm.c
480
drm_mm_interval_tree_add_node(hole, node);
drivers/gpu/drm/drm_mm.c
483
rm_hole(hole);
drivers/gpu/drm/drm_mm.c
485
add_hole(hole);
drivers/gpu/drm/drm_mm.c
522
struct drm_mm_node *hole;
drivers/gpu/drm/drm_mm.c
541
for (hole = first_hole(mm, range_start, range_end, size, mode);
drivers/gpu/drm/drm_mm.c
542
hole;
drivers/gpu/drm/drm_mm.c
543
hole = once ? NULL : next_hole(mm, hole, size, mode)) {
drivers/gpu/drm/drm_mm.c
544
u64 hole_start = __drm_mm_hole_node_start(hole);
drivers/gpu/drm/drm_mm.c
545
u64 hole_end = hole_start + hole->hole_size;
drivers/gpu/drm/drm_mm.c
558
mm->color_adjust(hole, color, &col_start, &col_end);
drivers/gpu/drm/drm_mm.c
598
list_add(&node->node_list, &hole->node_list);
drivers/gpu/drm/drm_mm.c
599
drm_mm_interval_tree_add_node(hole, node);
drivers/gpu/drm/drm_mm.c
601
rm_hole(hole);
drivers/gpu/drm/drm_mm.c
603
add_hole(hole);
drivers/gpu/drm/drm_mm.c
750
struct drm_mm_node *hole;
drivers/gpu/drm/drm_mm.c
766
hole = list_prev_entry(node, node_list);
drivers/gpu/drm/drm_mm.c
767
DRM_MM_BUG_ON(list_next_entry(hole, node_list) != node);
drivers/gpu/drm/drm_mm.c
770
hole_start = __drm_mm_hole_node_start(hole);
drivers/gpu/drm/drm_mm.c
771
hole_end = __drm_mm_hole_node_end(hole);
drivers/gpu/drm/drm_mm.c
776
mm->color_adjust(hole, scan->color, &col_start, &col_end);
drivers/gpu/drm/drm_mm.c
881
struct drm_mm_node *hole;
drivers/gpu/drm/drm_mm.c
894
list_for_each_entry(hole, &mm->hole_stack, hole_stack) {
drivers/gpu/drm/drm_mm.c
895
hole_start = __drm_mm_hole_node_start(hole);
drivers/gpu/drm/drm_mm.c
896
hole_end = hole_start + hole->hole_size;
drivers/gpu/drm/drm_mm.c
904
DRM_MM_BUG_ON(&hole->hole_stack == &mm->hole_stack);
drivers/gpu/drm/drm_mm.c
905
if (unlikely(&hole->hole_stack == &mm->hole_stack))
drivers/gpu/drm/drm_mm.c
911
mm->color_adjust(hole, scan->color, &hole_start, &hole_end);
drivers/gpu/drm/drm_mm.c
913
return hole;
drivers/gpu/drm/drm_mm.c
915
return list_next_entry(hole, node_list);
drivers/gpu/drm/drm_suballoc.c
103
sa_manager->hole = &sa_manager->olist;
drivers/gpu/drm/drm_suballoc.c
120
if (sa_manager->hole == &sa->olist)
drivers/gpu/drm/drm_suballoc.c
121
sa_manager->hole = sa->olist.prev;
drivers/gpu/drm/drm_suballoc.c
133
if (sa_manager->hole->next == &sa_manager->olist)
drivers/gpu/drm/drm_suballoc.c
136
sa = list_entry(sa_manager->hole->next, struct drm_suballoc, olist);
drivers/gpu/drm/drm_suballoc.c
147
struct list_head *hole = sa_manager->hole;
drivers/gpu/drm/drm_suballoc.c
149
if (hole != &sa_manager->olist)
drivers/gpu/drm/drm_suballoc.c
150
return list_entry(hole, struct drm_suballoc, olist)->eoffset;
drivers/gpu/drm/drm_suballoc.c
157
struct list_head *hole = sa_manager->hole;
drivers/gpu/drm/drm_suballoc.c
159
if (hole->next != &sa_manager->olist)
drivers/gpu/drm/drm_suballoc.c
160
return list_entry(hole->next, struct drm_suballoc, olist)->soffset;
drivers/gpu/drm/drm_suballoc.c
180
list_add(&sa->olist, sa_manager->hole);
drivers/gpu/drm/drm_suballoc.c
182
sa_manager->hole = &sa->olist;
drivers/gpu/drm/drm_suballoc.c
235
if (sa_manager->hole->next == &sa_manager->olist) {
drivers/gpu/drm/drm_suballoc.c
237
sa_manager->hole = &sa_manager->olist;
drivers/gpu/drm/drm_suballoc.c
283
sa_manager->hole = best_bo->olist.prev;
drivers/gpu/drm/drm_suballoc.c
437
if (&i->olist == sa_manager->hole)
drivers/gpu/drm/drm_suballoc.c
80
sa_manager->hole = &sa_manager->olist;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
107
u64 hole;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
538
struct drm_mm_node hole;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
562
memset(&hole, 0, sizeof(hole));
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
563
err = drm_mm_insert_node_in_range(&t->ce->vm->mm, &hole,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
569
drm_mm_remove_node(&hole);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
576
t->hole = hole.start + t->align;
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
577
pr_info("Using hole at %llx\n", t->hole);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
618
&t->buffers[i], t->hole + offset,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
619
&t->scratch, t->hole);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
640
&t->buffers[1], t->hole + offset / 2,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
641
&t->buffers[0], t->hole + 2 * offset);
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
650
&t->buffers[2], t->hole + t->align,
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c
651
&t->buffers[1], t->hole + 3 * offset / 2);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
663
struct drm_mm_node *hole, *next;
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
677
list_for_each_entry_safe(hole, next, &mm->hole_stack, hole_stack) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
686
resv->start = drm_mm_hole_node_start(hole) + loop;
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
687
resv->size = hole->hole_size - loop;
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
766
drm_mm_for_each_node_safe(hole, next, mm) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
767
if (hole->color != -1ul)
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
770
drm_mm_remove_node(hole);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
771
kfree(hole);
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
388
struct drm_mm_node hole;
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
409
memset(&hole, 0, sizeof(hole));
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
411
err = i915_gem_gtt_insert(&ggtt->vm, NULL, &hole,
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
444
drm_mm_remove_node(&hole);
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
529
if (drm_mm_node_allocated(&hole))
drivers/gpu/drm/i915/selftests/i915_gem_evict.c
530
drm_mm_remove_node(&hole);
drivers/gpu/drm/tests/drm_mm_test.c
41
struct drm_mm_node *hole;
drivers/gpu/drm/tests/drm_mm_test.c
46
drm_mm_for_each_hole(hole, mm, hole_start, hole_end)
drivers/gpu/drm/tests/drm_mm_test.c
54
drm_mm_for_each_node(hole, mm) {
drivers/gpu/drm/tests/drm_mm_test.c
55
if (drm_mm_hole_follows(hole)) {
drivers/gpu/drm/tests/drm_mm_test.c
66
struct drm_mm_node *hole;
drivers/gpu/drm/tests/drm_mm_test.c
75
drm_mm_for_each_hole(hole, mm, hole_start, hole_end) {
drivers/gpu/drm/xe/xe_ggtt.c
401
struct drm_mm_node *hole;
drivers/gpu/drm/xe/xe_ggtt.c
406
drm_mm_for_each_hole(hole, &ggtt->mm, start, end)
drivers/md/dm-vdo/int-map.c
453
static struct bucket *move_empty_bucket(struct bucket *hole)
drivers/md/dm-vdo/int-map.c
463
for (bucket = &hole[1 - NEIGHBORHOOD]; bucket < hole; bucket++) {
drivers/md/dm-vdo/int-map.c
482
if (hole < new_hole)
drivers/md/dm-vdo/int-map.c
498
hole->key = new_hole->key;
drivers/md/dm-vdo/int-map.c
499
hole->value = new_hole->value;
drivers/md/dm-vdo/int-map.c
503
insert_in_hop_list(bucket, hole);
drivers/md/dm-vdo/int-map.c
560
struct bucket *hole = find_empty_bucket(map, neighborhood, MAX_PROBES);
drivers/md/dm-vdo/int-map.c
566
while (hole != NULL) {
drivers/md/dm-vdo/int-map.c
567
int distance = hole - neighborhood;
drivers/md/dm-vdo/int-map.c
574
return hole;
drivers/md/dm-vdo/int-map.c
581
hole = move_empty_bucket(hole);
drivers/net/virtio_net.c
2769
unsigned int len, hole;
drivers/net/virtio_net.c
2791
hole = alloc_frag->size - alloc_frag->offset;
drivers/net/virtio_net.c
2792
if (hole < len + room) {
drivers/net/virtio_net.c
2800
len += hole;
drivers/net/virtio_net.c
2801
alloc_frag->offset += hole;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
824
u8 hole[AMPDU_MAX_MPDU];
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
825
memset(hole, 0, sizeof(hole));
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3631
int hole = api_ver >= 3 ? 0 : 2;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3638
rx_pn[2 + hole] = seq.tkip.iv32;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3639
rx_pn[3 + hole] = seq.tkip.iv32 >> 8;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3640
rx_pn[4 + hole] = seq.tkip.iv32 >> 16;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3641
rx_pn[5 + hole] = seq.tkip.iv32 >> 24;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3648
rx_pn[2 + hole] = seq.ccmp.pn[2];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3649
rx_pn[3 + hole] = seq.ccmp.pn[3];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3650
rx_pn[4 + hole] = seq.ccmp.pn[4];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3651
rx_pn[5 + hole] = seq.ccmp.pn[5];
drivers/pcmcia/rsrc_nonstatic.c
188
u_char *b, hole, most;
drivers/pcmcia/rsrc_nonstatic.c
203
hole = inb(i);
drivers/pcmcia/rsrc_nonstatic.c
205
if (inb(i+j) != hole)
drivers/pcmcia/rsrc_nonstatic.c
208
if ((j == 8) && (++b[hole] > b[most]))
drivers/pcmcia/rsrc_nonstatic.c
209
most = hole;
fs/nfs/nfs42xdr.c
1064
} hole;
fs/nfs/nfs42xdr.c
1075
return seg->type == NFS4_CONTENT_DATA ? seg->data.length : seg->hole.length;
fs/nfs/nfs42xdr.c
1103
xdr_decode_hyper(p, &seg->hole.length);
fs/ntfs3/attrib.c
2266
CLST svcn, evcn1, vcn, len, end, alen, hole, next_svcn;
fs/ntfs3/attrib.c
2332
hole = 0;
fs/ntfs3/attrib.c
2359
CLST vcn1, zero, hole2 = hole;
fs/ntfs3/attrib.c
2378
if (hole2 == hole)
fs/ntfs3/attrib.c
2411
run_deallocate_ex(sbi, &run2, vcn1, zero, &hole, true,
fs/ntfs3/attrib.c
2433
if (!hole)
fs/ntfs3/attrib.c
2445
total_size -= (u64)hole << sbi->cluster_bits;
fs/ocfs2/dir.c
1220
unsigned int hole;
fs/ocfs2/dir.c
1223
hole = le16_to_cpu(de->rec_len);
fs/ocfs2/dir.c
1225
hole = le16_to_cpu(de->rec_len) -
fs/ocfs2/dir.c
1228
return hole;
fs/ubifs/file.c
595
int i = 0, nn = *n, offs = bu->zbranch[0].offs, hole = 0, read = 0;
fs/ubifs/file.c
607
hole = 1;
fs/ubifs/file.c
617
hole = 1;
fs/ubifs/file.c
656
hole = 1;
fs/ubifs/file.c
673
if (hole) {
fs/xfs/scrub/ialloc.c
102
if (fhole != hole)
fs/xfs/scrub/ialloc.c
116
if (hole)
fs/xfs/scrub/ialloc.c
150
bool free, hole;
fs/xfs/scrub/ialloc.c
155
hole = irec->ir_holemask & (1U << hole_idx);
fs/xfs/scrub/ialloc.c
157
error = xchk_inobt_xref_finobt(sc, irec, i, free, hole);
fs/xfs/scrub/ialloc.c
177
bool free, hole;
fs/xfs/scrub/ialloc.c
201
hole = irec.ir_holemask & (1U << hole_idx);
fs/xfs/scrub/ialloc.c
205
if (fhole != hole)
fs/xfs/scrub/ialloc.c
70
bool hole)
fs/xfs/xfs_mount.c
69
int hole, i;
fs/xfs/xfs_mount.c
83
for (i = 0, hole = -1; i < xfs_uuid_table_size; i++) {
fs/xfs/xfs_mount.c
85
hole = i;
fs/xfs/xfs_mount.c
92
if (hole < 0) {
fs/xfs/xfs_mount.c
96
hole = xfs_uuid_table_size++;
fs/xfs/xfs_mount.c
98
xfs_uuid_table[hole] = *uuid;
include/drm/drm_suballoc.h
26
struct list_head *hole;
kernel/bpf/core.c
1080
u32 size, hole, start;
kernel/bpf/core.c
1103
hole = min_t(unsigned int, size - (proglen + sizeof(*hdr)),
kernel/bpf/core.c
1105
start = get_random_u32_below(hole) & ~(alignment - 1);
kernel/bpf/core.c
1137
u32 size, hole, start;
kernel/bpf/core.c
1164
hole = min_t(unsigned int, size - (proglen + sizeof(*ro_header)),
kernel/bpf/core.c
1166
start = get_random_u32_below(hole) & ~(alignment - 1);
mm/numa_emulation.c
204
static u64 uniform_size(u64 max_addr, u64 base, u64 hole, int nr_nodes)
mm/numa_emulation.c
208
unsigned long hole_pfns = PHYS_PFN(hole);
net/ipv4/tcp_output.c
3730
struct sk_buff *skb, *rtx_head, *hole = NULL;
net/ipv4/tcp_output.c
3750
if (!hole)
net/ipv4/tcp_output.c
3765
if (!hole && !(sacked & (TCPCB_SACKED_RETRANS|TCPCB_SACKED_ACKED)))
net/ipv4/tcp_output.c
3766
hole = skb;
tools/objtool/check.c
2617
insn->hole = 1;
tools/objtool/check.c
4344
insn->hole || (func && func->ignore))
tools/objtool/elf.c
229
struct symbol_hole hole = {
tools/objtool/elf.c
239
n = rb_find(&hole, &sec->symbol_tree.rb_root, symbol_hole_by_offset);
tools/objtool/elf.c
251
if (hole.sym)
tools/objtool/elf.c
252
n = rb_next(&hole.sym->node);
tools/objtool/include/objtool/check.h
81
hole : 1,
tools/perf/util/block-range.c
271
struct block_range *hole = malloc(sizeof(struct block_range));
tools/perf/util/block-range.c
272
if (!hole)
tools/perf/util/block-range.c
275
*hole = (struct block_range){
tools/perf/util/block-range.c
282
rb_link_left_of_node(&hole->node, &next->node);
tools/perf/util/block-range.c
283
rb_insert_color(&hole->node, &block_ranges.root);