write_bytes
.write_bytes = tpm_tcg_write_bytes,
int (*write_bytes)(struct tpm_tis_data *data, u32 addr, u16 len,
return data->phy_ops->write_bytes(data, addr, len, value,
return data->phy_ops->write_bytes(data, addr, 1, &value,
rc = data->phy_ops->write_bytes(data, addr, sizeof(u32),
.write_bytes = tpm_tis_i2c_write_bytes,
.write_bytes = tpm_tis_spi_cr50_write_bytes,
.write_bytes = tpm_tis_spi_write_bytes,
.write_bytes = tpm_tis_synquacer_write_bytes,
u32 write_bytes;
static bool stm_addr_unaligned(const void *addr, u8 write_bytes)
return ((unsigned long)addr & (write_bytes - 1));
u32 size, u8 write_bytes)
if (stm_addr_unaligned(data, write_bytes)) {
if (size > drvdata->write_bytes)
size = drvdata->write_bytes;
stm_send(ch_addr, payload, 1, drvdata->write_bytes);
drvdata->write_bytes);
drvdata->write_bytes = stm_fundamental_data_size(drvdata);
u64 write_bytes;
result->rdma_write_stats.write_bytes = rws->write_bytes;
u64 write_bytes;
stats->value[EFA_RDMA_WRITE_BYTES] = rws->write_bytes;
unsigned int write_bytes;
write_bytes = remaining;
write_words = write_bytes / 4;
mod_bytes = write_bytes % 4;
remaining -= write_bytes;
unsigned int write_bytes;
write_bytes = nbytes;
tspi->cur_tx_pos += write_bytes;
unsigned int write_bytes;
write_bytes = consume;
tspi->cur_tx_pos += write_bytes;
unsigned int write_bytes;
write_bytes = len;
tqspi->cur_tx_pos += write_bytes;
unsigned int i, count, consume, write_bytes;
write_bytes = consume;
tqspi->cur_tx_pos += write_bytes;
this_cpu_add(se_cmd->se_dev->stats->write_bytes,
this_cpu_add(deve->stats->write_bytes,
auth_show_per_cpu_stat(auth_write_mbytes, write_bytes, 20);
lu_show_per_cpu_stat(lu_write_mbytes, write_bytes, 20);
unsigned long write_bytes;
timing_stats.write_bytes += length + 2;
timing_stats.write_bytes,
timing_stats.write_bytes = 0;
*write_bytes = min_t(size_t, *write_bytes, cur_offset - pos);
size_t write_bytes = calc_write_bytes(inode, iter, start);
if (unlikely(fault_in_iov_iter_readable(iter, write_bytes)))
ret = reserve_space(inode, data_reserved, start, &write_bytes, nowait,
ASSERT(start + write_bytes <= reserved_start + reserved_len);
ret = prepare_one_folio(&inode->vfs_inode, &folio, start, write_bytes, false);
write_bytes = last_block - start;
write_bytes, &lockstart,
write_bytes, iter);
if (unlikely(copied < write_bytes)) {
size_t write_bytes = reserved_space;
if (btrfs_check_nocow_lock(inode, page_start, &write_bytes, false) <= 0)
if (write_bytes < reserved_space)
size_t write_bytes, struct extent_state **cached, bool noreserve)
const u64 end_pos = pos + write_bytes;
if (write_bytes == 0)
loff_t pos, size_t write_bytes,
fgf_set_order(write_bytes);
ret = prepare_uptodate_folio(inode, folio, pos, write_bytes);
loff_t pos, size_t write_bytes,
last_pos = round_up(pos + write_bytes, fs_info->sectorsize) - 1;
size_t *write_bytes, bool nowait)
lockend = round_up(pos + *write_bytes,
size_t write_bytes, struct extent_state **cached, bool noreserve);
size_t *write_bytes, bool nowait);
size_t write_bytes = blocksize;
if (btrfs_check_nocow_lock(inode, block_start, &write_bytes, false) > 0) {
ASSERT(write_bytes == blocksize, "write_bytes=%zu blocksize=%u",
write_bytes, blocksize);
p = xdr_encode_hyper(p, devinfo->write_bytes);
devinfo->write_bytes =
p = xdr_encode_hyper(p, devinfo->write_bytes);
(unsigned long long)acct.write_bytes,
__u64 write_bytes;
u64 write_bytes;
current->ioac.write_bytes += bytes;
return p->ioac.write_bytes >> 9;
dst->write_bytes += src->write_bytes;
u64 write_bytes;
u64 write_bytes;
__u64 write_bytes; /* bytes of write I/O */
stats->write_bytes = p->ioac.write_bytes & KB_MASK;
stats->write_bytes = 0;
(unsigned long long)t->write_bytes,
write_bytes(addr);
write_bytes(addr, length);
write_bytes(addr, maplength);
write_bytes(addr, length);