readable
return trans->readable;
bool readable;
((!write && !out.readable) || (write && !out.writable)))
out->readable = desc & KVM_PTE_LEAF_ATTR_LO_S2_S2AP_R;
(!write_fault && !trans->readable));
if (((filp->f_mode & FMODE_READ) && !chan->readable) ||
unsigned readable:1;
unsigned int log2_element_size = chan->readable ?
chan->readable = 1;
.readable = (_r), \
unsigned int readable : 1,
if (output_parms[j].readable) {
if (!aat2870->reg_cache[addr].readable) {
u16 readable; /* Mask of readable bits */
return wm8350_reg_io_map[reg].readable;
return do_mem_probe(s, m->base, m->num, readable, checksum);
ok += do_mem_probe(s, mm.base, mm.num, readable,
readable, checksum);
ok += do_mem_probe(s, mm.base, mm.num, readable, checksum);
ceph_decode_8_safe(&p, end, cap_auths[i].readable, bad);
cap_auths[i].readable, cap_auths[i].writeable);
((mask & MAY_READ) && !s->readable))
rw_perms_s->readable, rw_perms_s->writeable,
bool readable;
bool readable;
int readable, chunk;
readable = atomic_read(&conn->bytes_to_rcv);
if (readable >= conn->rmb_desc->len)
if (len < readable)
readable = smc_rx_data_available(conn, peeked_bytes);
if (!readable || (msg && splbytes)) {
readable > 1)
readable--; /* always stop at urgent Byte */
copylen = min_t(size_t, read_remaining, readable);
u32 readable = hv_get_bytes_to_read(&chan->inbound);
return readable >= HVS_PKT_LEN(0);
u32 readable = hv_get_bytes_to_read(&chan->inbound);
if (readable > HVS_PKT_LEN(0)) {
if (readable == HVS_PKT_LEN(0)) {
int hvs_notify_poll_in(struct vsock_sock *vsk, size_t target, bool *readable)
*readable = hvs_channel_readable(hvs->chan);
bool readable, writeable;
readable = uac_v2v3_control_is_readable(bmControls,
if (!readable)
uint64_t readable;
#define PTE_READABLE_MASK(mmu) ((mmu)->arch.pte_masks.readable)
.readable = BIT_ULL(0),