local_read
long c = local_read(l);
long c = local_read(l);
long c = local_read(l);
sum += local_read(pv);
index = local_read(&bb->head);
old_head = local_read(&bb->head);
if (old_head == local_read(&bb->head))
to = local_read(&buf->data_size);
long c = local_read(l);
long c = local_read(l);
u32 this_met = local_read(&stat->missed[rw].nr_met);
u32 this_missed = local_read(&stat->missed[rw].nr_missed);
mr->hw_mr.local_read = 1;
mr->hw_mr.local_read = 1;
mr->hw_mr.local_read = 1;
params->local_read);
#define __local_inc(l) local_set((l), local_read(l) + 1)
#define __local_dec(l) local_set((l), local_read(l) - 1)
#define __local_add(i,l) local_set((l), local_read(l) + (i))
#define __local_sub(i,l) local_set((l), local_read(l) - (i))
#define local64_read(l) local_read(&(l)->a)
local_read(&(part_stat_get(part, field)))
local_read(&(part_stat_get_cpu(part, field, cpu)))
bool local_read;
if (local_read(&ctx->nr_no_switch_fast) ||
local_read(&next_ctx->nr_no_switch_fast)) {
if (unlikely(head != local_read(&rb->head))) {
if (handle->wakeup != local_read(&rb->wakeup))
have_lost = local_read(&rb->lost);
offset = local_read(&rb->head);
if (unlikely(head - local_read(&rb->wakeup) > rb->watermark))
handle->wakeup = local_read(&rb->wakeup);
head = local_read(&rb->head);
if (local_read(&rq->scx.reenq_local_deferred)) {
if ((unsigned)local_read(&subbuf->commit) > subbuf_size) {
tail = local_read(&dpage->commit);
entry_bytes += local_read(&cpu_buffer->reader_page->page->commit);
entry_bytes += local_read(&head_page->page->commit);
return local_read(&bpage->entries) & RB_WRITE_MASK;
return local_read(&bpage->write) & RB_WRITE_MASK;
unsigned int nest = local_read(&cpu_buffer->committing) - 1;
local_read(&bpage->write) & ~RB_WRITE_MASK;
return local_read(&bpage->page->commit);
local_read(&cpu_buffer->commit_page->page->commit) &
!local_read(&cpu_buffer->committing)))
commits = local_read(&cpu_buffer->commits);
if (local_read(&cpu_buffer->committing) == 1)
if (unlikely(local_read(&cpu_buffer->commits) != commits) &&
!local_read(&cpu_buffer->committing)) {
if (cpu_buffer->last_pages_touch == local_read(&cpu_buffer->pages_touched))
cpu_buffer->last_pages_touch = local_read(&cpu_buffer->pages_touched);
tail = local_read(&bpage->commit);
if (tail <= 8 || tail > local_read(&bpage->commit))
/*A*/ w = local_read(&tail_page->write) & RB_WRITE_MASK;
/*F*/ if (write == (local_read(&tail_page->write) & RB_WRITE_MASK) &&
RB_WARN_ON(buffer, !local_read(&cpu_buffer->committing));
return local_read(&cpu_buffer->entries) -
(local_read(&cpu_buffer->overrun) + cpu_buffer->read);
ret = local_read(&cpu_buffer->entries_bytes) - cpu_buffer->read_bytes;
ret = local_read(&cpu_buffer->overrun);
ret = local_read(&cpu_buffer->commit_overrun);
ret = local_read(&cpu_buffer->dropped_events);
overruns += local_read(&cpu_buffer->overrun);
overwrite = local_read(&(cpu_buffer->overrun));
meta->entries = local_read(&cpu_buffer->entries);
meta->overrun = local_read(&cpu_buffer->overrun);
if (RB_WARN_ON(cpu_buffer, local_read(&cpu_buffer->committing)))
if (local_read(&cpu_buffer_a->committing))
if (local_read(&cpu_buffer_b->committing))
commit = local_read(&bpage->commit);
commit = local_read(&page->page->commit);
write = local_read(&page->write);
nest = local_read(&cpu_buffer->committing);
read = local_read(&buffer->buffers[cpu]->pages_read);
lost = local_read(&buffer->buffers[cpu]->pages_lost);
cnt = local_read(&buffer->buffers[cpu]->pages_touched);
commit = local_read(&rpage->commit) & 0xfffff;
if (unlikely(!data) || local_read(&data->disabled))
!data->critical_start || local_read(&data->disabled))
int_counter = local_read(&osn_var->int_counter);
} while (int_counter != local_read(&osn_var->int_counter));
int_counter = local_read(&osn_var->int_counter);
} while (int_counter != local_read(&osn_var->int_counter));
int_counter = local_read(&osn_var->int_counter);
} while (int_counter != local_read(&osn_var->int_counter));