i_data
truncate_inode_pages_final(&inode->i_data);
inode->i_data.a_ops = &def_blk_aops;
mapping_set_gfp_mask(&inode->i_data, GFP_USER);
bdev->bd_mapping = &inode->i_data;
mapping_set_gfp_mask(&inode->i_data, GFP_USER);
void *k_data, *i_data;
i_data = (void *)data;
memcpy(k_data, i_data, size);
i_data += size;
i_data = (void *)data + i * stride;
memcpy(k_data, i_data, size);
i_data += size;
u8 buf[5], i_data[4], q_data[4];
i_data[i] = (((buf[2] >> 0) & 0x3) << 6) | (buf[0] & 0x3f);
swap(i_data[2], i_data[3]);
ret = regmap_bulk_write(dev->regmap, 0x60, i_data, 4);
data = (struct ec_params_get_cmd_versions *)mock->i_data;
data = (u8 *)mock->i_data;
data = (struct ec_params_motion_sense *)mock->i_data;
data = (struct ec_params_motion_sense *)mock->i_data;
data = (struct ec_params_motion_sense *)mock->i_data;
data = (u8 *)mock->i_data;
data = (struct ec_params_get_cmd_versions *)mock->i_data;
data = (struct ec_params_get_cmd_versions *)mock->i_data;
data = (struct ec_params_hello *)mock->i_data;
data = (struct ec_params_hello *)mock->i_data;
data = (struct ec_params_get_cmd_versions *)mock->i_data;
data = (struct ec_params_get_cmd_versions *)mock->i_data;
mock->i_data = kunit_kzalloc(mock->test, msg->outsize, GFP_KERNEL);
if (mock->i_data)
memcpy(mock->i_data, msg->data, msg->outsize);
void *i_data;
invalidate_mapping_pages(&inode->i_data, 0, -1);
invalidate_mapping_pages(&inode->i_data, 0, -1);
invalidate_mapping_pages(&inode->i_data, 0, -1);
truncate_inode_pages_final(&inode->i_data);
filemap_fdatawrite(&inode->i_data);
inode->i_data.a_ops = &affs_symlink_aops;
truncate_inode_pages_final(&inode->i_data);
inode->i_data.a_ops = &affs_symlink_aops;
truncate_inode_pages_final(&inode->i_data);
if (!mapping_tagged(&vnode->netfs.inode.i_data, PAGECACHE_TAG_WRITEBACK) &&
!mapping_tagged(&vnode->netfs.inode.i_data, PAGECACHE_TAG_DIRTY)) {
} i_data;
befs_data_stream *ds = &BEFS_I(inode)->i_data.ds;
const befs_data_stream *ds = &BEFS_I(dir)->i_data.ds;
const befs_data_stream *ds = &BEFS_I(inode)->i_data.ds;
strscpy(befs_ino->i_data.symlink, raw_inode->data.symlink,
befs_ino->i_data.ds =
num_blks = befs_count_blocks(sb, &befs_ino->i_data.ds);
inode->i_size = befs_ino->i_data.ds.size;
inode->i_link = befs_ino->i_data.symlink;
i_data.symlink),
i_data.symlink),
befs_data_stream *data = &befs_ino->i_data.ds;
truncate_inode_pages_final(&inode->i_data);
lockdep_set_class(&(&inode->i_data)->invalidate_lock,
truncate_inode_pages_final(&inode->i_data);
WARN_ON(vfs_inode->i_data.nrpages);
truncate_inode_pages_range(&sctx->cur_inode->i_data,
truncate_inode_pages_range(&sctx->cur_inode->i_data,
return !list_empty(&inode->i_data.i_private_list);
struct address_space *mapping = &inode->i_data;
struct address_space *mapping = &inode->i_data;
mapping_set_error(&inode->i_data, err);
invalidate_mapping_pages(&inode->i_data, 0, -1);
if (inode->i_data.nrpages == 0 &&
inode->i_data.nrpages && /* have cached pages */
if (inode->i_data.nrpages > 0)
mapping_set_error(&inode->i_data, -EIO);
ci->netfs.inode.i_data.nrpages))
cache_ctl->folio = filemap_lock_folio(&dir->i_data, ptr_pgoff);
ctl->folio = __filemap_get_folio(&dir->i_data, pgoff,
fgf, mapping_gfp_mask(&dir->i_data));
filemap_write_and_wait_range(&inode->i_data, 0,
filemap_fdatawrite(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
inode->i_mapping = &inode->i_data;
inode->i_data.a_ops = &coda_symlink_aops;
inode->i_mapping = &inode->i_data;
if (coda_inode->i_mapping == &coda_inode->i_data)
if (coda_inode->i_mapping == &host_inode->i_data) {
coda_inode->i_mapping = &coda_inode->i_data;
truncate_inode_pages_final(&inode->i_data);
inode->i_data.a_ops = &cramfs_aops;
inode->i_data.a_ops = &cramfs_aops;
truncate_inode_pages_final(&inode->i_data);
inode->i_data.a_ops = &efs_aops;
inode->i_data.a_ops = &efs_symlink_aops;
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages(&inode->i_data, 0);
__le32 i_data[15];
memset(ei->i_data, 0, sizeof(ei->i_data));
__le32 *i_data = EXT2_I(inode)->i_data;
ext2_free_data(inode, i_data+offsets[0],
i_data + EXT2_NDIR_BLOCKS);
nr = i_data[EXT2_IND_BLOCK];
i_data[EXT2_IND_BLOCK] = 0;
nr = i_data[EXT2_DIND_BLOCK];
i_data[EXT2_DIND_BLOCK] = 0;
nr = i_data[EXT2_TIND_BLOCK];
i_data[EXT2_TIND_BLOCK] = 0;
ei->i_data[n] = raw_inode->i_block[n];
inode->i_link = (char *)ei->i_data;
nd_terminate_link(ei->i_data, inode->i_size,
sizeof(ei->i_data) - 1);
raw_inode->i_block[n] = ei->i_data[n];
add_chain (chain, NULL, EXT2_I(inode)->i_data + *offsets);
__le32 *start = ind->bh ? (__le32 *) ind->bh->b_data : ei->i_data;
truncate_inode_pages_final(&inode->i_data);
if (l > sizeof (EXT2_I(inode)->i_data)) {
inode->i_link = (char*)EXT2_I(inode)->i_data;
offsetof(struct ext2_inode_info, i_data),
sizeof_field(struct ext2_inode_info, i_data),
__le32 i_data[15]; /* unconverted */
EXT4_I(inode)->i_data, \
return (struct ext4_extent_header *) EXT4_I(inode)->i_data;
eh = (struct ext4_extent_header *)k_ctx.k_ei->i_data;
ext_size = sizeof(EXT4_I(inode)->i_data);
memmove(bh->b_data, EXT4_I(inode)->i_data, ext_size);
size = sizeof(EXT4_I(inode)->i_data);
size = sizeof(EXT4_I(inode)->i_data);
memset(ei->i_data, 0, sizeof(ei->i_data));
__le32 *i_data = ei->i_data;
ext4_free_data(handle, inode, NULL, i_data+offsets[0],
i_data + EXT4_NDIR_BLOCKS);
nr = i_data[EXT4_IND_BLOCK];
i_data[EXT4_IND_BLOCK] = 0;
nr = i_data[EXT4_DIND_BLOCK];
i_data[EXT4_DIND_BLOCK] = 0;
nr = i_data[EXT4_TIND_BLOCK];
i_data[EXT4_TIND_BLOCK] = 0;
__le32 *i_data = ei->i_data;
ext4_free_data(handle, inode, NULL, i_data + offsets[0],
i_data + offsets2[0]);
ext4_free_data(handle, inode, NULL, i_data + offsets[0],
i_data + EXT4_NDIR_BLOCKS);
nr = i_data[EXT4_IND_BLOCK];
i_data[EXT4_IND_BLOCK] = 0;
nr = i_data[EXT4_DIND_BLOCK];
i_data[EXT4_DIND_BLOCK] = 0;
nr = i_data[EXT4_TIND_BLOCK];
i_data[EXT4_TIND_BLOCK] = 0;
add_chain(chain, NULL, EXT4_I(inode)->i_data + *offsets);
__le32 *start = ind->bh ? (__le32 *) ind->bh->b_data : ei->i_data;
memset(ei->i_data, 0, EXT4_MIN_INLINE_DATA_SIZE);
mapping_tagged(&inode->i_data, PAGECACHE_TAG_DIRTY)))
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
memset(EXT4_I(inode)->i_data, 0, sizeof(EXT4_I(inode)->i_data));
raw_inode->i_block[block] = ei->i_data[block];
ei->i_data[block] = raw_inode->i_block[block];
inode->i_size >= sizeof(ei->i_data) ||
strnlen((char *)ei->i_data, inode->i_size + 1) !=
inode_set_cached_link(inode, (char *)ei->i_data,
memswap(ei1->i_data, ei2->i_data, sizeof(ei1->i_data));
truncate_inode_pages(&inode->i_data, 0);
truncate_inode_pages(&inode_bl->i_data, 0);
memset(ei_bl->i_data, 0, sizeof(ei_bl->i_data));
i_data = (__le32 *)bh->b_data;
if (i_data[i]) {
le32_to_cpu(i_data[i]), lb);
__le32 *i_data;
i_data = (__le32 *)bh->b_data;
if (i_data[i]) {
le32_to_cpu(i_data[i]), lb);
__le32 *i_data;
i_data = (__le32 *)bh->b_data;
if (i_data[i]) {
le32_to_cpu(i_data[i]), lb);
struct inode *inode, __le32 i_data)
bh = ext4_sb_bread(sb, le32_to_cpu(i_data), 0);
ext4_free_blocks(handle, inode, NULL, le32_to_cpu(i_data), 1,
struct inode *inode, __le32 i_data)
bh = ext4_sb_bread(inode->i_sb, le32_to_cpu(i_data), 0);
ext4_free_blocks(handle, inode, NULL, le32_to_cpu(i_data), 1,
static int free_ind_block(handle_t *handle, struct inode *inode, __le32 *i_data)
if (i_data[0]) {
le32_to_cpu(i_data[0]), 1,
if (i_data[1]) {
retval = free_dind_blocks(handle, inode, i_data[1]);
if (i_data[2]) {
retval = free_tind_blocks(handle, inode, i_data[2]);
__le32 i_data[3];
i_data[0] = ei->i_data[EXT4_IND_BLOCK];
i_data[1] = ei->i_data[EXT4_DIND_BLOCK];
i_data[2] = ei->i_data[EXT4_TIND_BLOCK];
memcpy(ei->i_data, tmp_ei->i_data, sizeof(ei->i_data));
retval = free_ind_block(handle, inode, i_data);
struct ext4_extent_header *eh = (struct ext4_extent_header *)ei->i_data;
__le32 *i_data;
i_data = ei->i_data;
if (i_data[i]) {
le32_to_cpu(i_data[i]), &lb);
if (i_data[EXT4_IND_BLOCK]) {
le32_to_cpu(i_data[EXT4_IND_BLOCK]), &lb);
if (i_data[EXT4_DIND_BLOCK]) {
le32_to_cpu(i_data[EXT4_DIND_BLOCK]), &lb);
if (i_data[EXT4_TIND_BLOCK]) {
le32_to_cpu(i_data[EXT4_TIND_BLOCK]), &lb);
memset(ei->i_data, 0, sizeof(ei->i_data));
ei->i_data[i] = cpu_to_le32(blk++);
__le32 *i_data;
memcpy((char *)&EXT4_I(inode)->i_data, disk_link.name,
inode_set_cached_link(inode, (char *)&EXT4_I(inode)->i_data,
data = EXT4_I(inode)->i_data + EXT4_DIND_BLOCK;
if (ei->i_data[i])
if (ei->i_data[i])
nr = le32_to_cpu(ei->i_data[EXT4_DIND_BLOCK]);
ei->i_data[EXT4_DIND_BLOCK] = 0;
data = EXT4_I(inode)->i_data + EXT4_DIND_BLOCK;
.useroffset = offsetof(struct ext4_inode_info, i_data),
.usersize = sizeof_field(struct ext4_inode_info, i_data),
caddr = EXT4_I(inode)->i_data;
max_size = sizeof(EXT4_I(inode)->i_data);
truncate_inode_pages(&dir->i_data, 0);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages(&dqopt->files[type]->i_data, 0);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&ip->i_data);
inode->i_data.a_ops = &fuse_dax_file_aops;
inode->i_data.a_ops = &fuse_symlink_aops;
inode->i_data.a_ops = &fuse_file_aops;
mapping_set_writeback_may_deadlock_on_reclaim(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
inode->i_data.nrpages);
truncate_inode_pages(&inode->i_data, 0);
truncate_inode_pages(&inode->i_data, 0);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
result->i_data.a_ops = &hpfs_aops;
i->i_data.a_ops = &hpfs_aops;
truncate_inode_pages_final(&inode->i_data);
i->i_data.a_ops = &hpfs_symlink_aops;
result->i_data.a_ops = &hpfs_aops;
result->i_data.a_ops = &hpfs_symlink_aops;
struct address_space *mapping = &inode->i_data;
resv_map = (struct resv_map *)(&inode->i_data)->i_private_data;
reap = invalidate_mapping_pages(&inode->i_data, 0, -1);
struct address_space *const mapping = &inode->i_data;
__address_space_init_once(&inode->i_data);
if (!mapping_shrinkable(&inode->i_data))
xa_lock_irq(&inode->i_data.i_pages);
BUG_ON(inode->i_data.nrpages);
xa_unlock_irq(&inode->i_data.i_pages);
BUG_ON(!list_empty(&inode->i_data.i_private_list));
truncate_inode_pages_final(&inode->i_data);
!mapping_shrinkable(&inode->i_data)) {
if (inode_has_buffers(inode) || !mapping_empty(&inode->i_data)) {
inode->i_data.a_ops = &zisofs_aops;
inode->i_data.a_ops = &isofs_aops;
inode->i_data.a_ops = &isofs_symlink_aops;
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
block_t *idata = i_data(inode);
add_chain (chain, NULL, i_data(inode) + *offsets);
truncate_inode_pages_final(&inode->i_data);
inode->i_data.a_ops = &nfs_file_aops;
inode->i_data.a_ops = &nfs_dir_aops;
truncate_inode_pages_range(&dst_inode->i_data, dst_off, dst_off + count - 1);
truncate_inode_pages_final(&inode->i_data);
folio = read_cache_folio(&inode->i_data, 0, nfs_symlink_filler,
truncate_inode_pages(&ii->vfs_inode.i_data, 0);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages(&ii->vfs_inode.i_data, 0);
truncate_inode_pages_final(&inode->i_data);
struct ocfs2_inline_data *idata = &di->id2.i_data;
struct ocfs2_inline_data *idata = &di->id2.i_data;
if (new_size <= le16_to_cpu(di->id2.i_data.id_count))
memcpy_from_folio(di->id2.i_data.id_data + pos, wc->w_target_folio,
le16_to_cpu(di->id2.i_data.id_count),
folio_fill_tail(folio, 0, di->id2.i_data.id_data, size);
data = &di->id2.i_data;
data_start = di->id2.i_data.id_data;
data = &di->id2.i_data;
struct ocfs2_inline_data *data = &di->id2.i_data;
de_buf = di->id2.i_data.id_data;
memcpy(dirdata_bh->b_data, di->id2.i_data.id_data, i_size_read(dir));
first_de = di->id2.i_data.id_data;
data = &di->id2.i_data;
id_count = le16_to_cpu(di->id2.i_data.id_count);
id2.i_data.id_data);
truncate_inode_pages_range(&inode_out->i_data,
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
struct ocfs2_inline_data *data = &di->id2.i_data;
fe->id2.i_data.id_count = cpu_to_le16(
offsetof(struct ocfs2_dinode, id2.i_data.id_data) -
offsetof(struct ocfs2_dinode, id2.i_data.id_data);
offsetof(struct ocfs2_dinode, id2.i_data.id_data) -
offsetof(struct ocfs2_dinode, id2.i_data.id_data);
struct ocfs2_inline_data i_data;
t_di->id2.i_data.id_count = s_di->id2.i_data.id_count;
memcpy(t_di->id2.i_data.id_data, s_di->id2.i_data.id_data,
le16_to_cpu(s_di->id2.i_data.id_count));
struct ocfs2_inline_data *idata = &di->id2.i_data;
struct ocfs2_inline_data *idata = &di->id2.i_data;
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages_final(&inode->i_data);
truncate_inode_pages(&dqopt->files[cnt]->i_data, 0);
truncate_inode_pages(&dqopt->files[cnt]->i_data, 0);
i->i_data.a_ops = &romfs_aops;
i->i_data.a_ops = &romfs_aops;
truncate_inode_pages_range(&target_inode->i_data, fstart, fend);
truncate_inode_pages_final(&inode->i_data);
inode->i_data.a_ops = &cifs_addr_ops_smallbuf;
inode->i_data.a_ops = &cifs_addr_ops;
inode->i_data.a_ops = &squashfs_aops;
inode->i_data.a_ops = &squashfs_aops;
inode->i_data.a_ops = &squashfs_symlink_aops;
truncate_inode_pages_final(&inode->i_data);
buf1 = iinfo->i_data + iinfo->i_lenEAttr;
memcpy(&iter->fi, iinfo->i_data + iinfo->i_lenEAttr + iter->pos,
iter->name = iinfo->i_data + iinfo->i_lenEAttr + iter->pos +
UDF_I(inode)->i_data,
iinfo->i_data = kzalloc(inode->i_sb->s_blocksize -
iinfo->i_data = kzalloc(inode->i_sb->s_blocksize -
if (!iinfo->i_data) {
memset(iinfo->i_data + iinfo->i_lenEAttr + newsize,
memcpy(iinfo->i_data,
memcpy(iinfo->i_data,
memcpy(iinfo->i_data,
truncate_inode_pages_final(&inode->i_data);
kfree(iinfo->i_data);
iinfo->i_data = NULL;
inode->i_data.a_ops = &udf_aops;
inode->i_data.a_ops = &udf_symlink_aops;
iinfo->i_data = kmalloc(size, GFP_KERNEL);
if (!iinfo->i_data)
iinfo->i_data, inode->i_sb->s_blocksize -
iinfo->i_data,
iinfo->i_data,
memcpy_from_file_folio(iinfo->i_data + iinfo->i_lenEAttr, folio,
ptr = iinfo->i_data + epos->offset -
folio_fill_tail(folio, 0, iinfo->i_data + iinfo->i_lenEAttr, isize);
ptr = iinfo->i_data + epos->offset -
memset(iinfo->i_data + iinfo->i_lenEAttr, 0x00,
memcpy_from_folio(iinfo->i_data + iinfo->i_lenEAttr,
ea = iinfo->i_data;
ea = iinfo->i_data;
ad = iinfo->i_data + iinfo->i_lenEAttr;
memcpy(dbh->b_data, iinfo->i_data, inode->i_size);
memset(iinfo->i_data + iinfo->i_lenEAttr, 0, iinfo->i_lenAlloc);
inode->i_data.a_ops = &udf_aops;
inode->i_data.a_ops = &udf_aops;
inode->i_data.a_ops = &udf_symlink_aops;
ea = iinfo->i_data + iinfo->i_lenEAttr;
loc = le32_to_cpu(((__le32 *)(iinfo->i_data +
vati->i_data;
ei->i_data = NULL;
symlink = iinfo->i_data + iinfo->i_lenEAttr;
__u8 *i_data;
if (!grow_chain32(ufsi, NULL, &ufsi->i_u1.i_data[*p++], chain, q))
memcpy(ufsi->i_u1.i_data, &ufs_inode->ui_u2.ui_addr,
ufs_inode->ui_u2.ui_addr.ui_db[0] = ufsi->i_u1.i_data[0];
memcpy(&ufs_inode->ui_u2.ui_addr, ufsi->i_u1.i_data,
truncate_inode_pages_final(&inode->i_data);
__fs32 i_data[15];
fs32 = fs32_to_cpu(sb, ufsi->i_u1.i_data[1]);
fs32 = fs32_to_cpu(sb, ufsi->i_u1.i_data[0]);
ufsi->i_u1.i_data[1] = cpu_to_fs32(sb, fs32);
ufsi->i_u1.i_data[0] = cpu_to_fs32(sb, fs32);
(void *)&ufsi->i_u1.i_data[blk];
truncate_inode_pages_final(&inode->i_data);
struct address_space i_data;
return (struct resv_map *)(&inode->i_data)->i_private_data;