group_data
if (!memcmp(tg->group_data, topology, 8)) {
memcpy(tg->group_data, topology, 8);
struct ext4_new_group_data *group_data)
bh = ext4_get_bitmap(sb, group_data->inode_bitmap);
bh = ext4_get_bitmap(sb, group_data->block_bitmap);
struct ext4_new_group_data *group_data = flex_gd->groups;
for (i = 0; i < flex_gd->count; i++, group_data++, bg_flags++) {
group = group_data->group;
ext4_block_bitmap_set(sb, gdp, group_data->block_bitmap);
ext4_inode_bitmap_set(sb, gdp, group_data->inode_bitmap);
err = ext4_set_bitmap_checksums(sb, gdp, group_data);
ext4_inode_table_set(sb, gdp, group_data->inode_table);
group_data->free_clusters_count);
struct ext4_new_group_data *group_data = flex_gd->groups;
BUG_ON(flex_gd->count == 0 || group_data == NULL);
blocks_count += group_data[i].blocks_count;
free_blocks += EXT4_C2B(sbi, group_data[i].free_clusters_count);
flex_group = ext4_flex_group(sbi, group_data[0].group);
struct ext4_new_group_data *group_data = flex_gd->groups;
group_data[i].group = group + i;
group_data[i].blocks_count = EXT4_BLOCKS_PER_GROUP(sb);
group_data[i].mdata_blocks = overhead;
group_data[i].free_clusters_count = EXT4_CLUSTERS_PER_GROUP(sb);
group_data[i - 1].blocks_count = EXT4_C2B(sbi, last + 1);
group_data[i - 1].free_clusters_count -= clusters_per_group -
struct ext4_new_group_data *group_data = flex_gd->groups;
BUG_ON(flex_gd->count == 0 || group_data == NULL);
src_group = group_data[0].group;
group = group_data[0].group;
if (src_group >= group_data[0].group + flex_gd->count)
last_blk = start_blk + group_data[src_group - group].blocks_count;
last_blk += group_data[src_group - group].blocks_count;
group_data[bb_index].block_bitmap = start_blk++;
group -= group_data[0].group;
group_data[group].mdata_blocks++;
group_data[ib_index].inode_bitmap = start_blk++;
group -= group_data[0].group;
group_data[group].mdata_blocks++;
group_data[it_index].inode_table = start_blk;
group -= group_data[0].group;
group_data[group + 1].mdata_blocks += overhead;
group_data[group].mdata_blocks += itb;
group_data[i].free_clusters_count -=
group_data[i].mdata_blocks);
group = group_data[0].group;
group_data[i].blocks_count,
group_data[i].free_clusters_count,
group_data[i].mdata_blocks);
struct ext4_new_group_data *group_data = flex_gd->groups;
BUG_ON(!flex_gd->count || !group_data ||
group_data[0].group != sbi->s_groups_count);
group = group_data[0].group;
block = group_data[i].inode_table;
block = group_data[i].block_bitmap;
ext4_mark_bitmap_end(EXT4_B2C(sbi, group_data[i].blocks_count),
block = group_data[i].inode_bitmap;
start = (&group_data[0].block_bitmap)[j];
if (block == (&group_data[i].block_bitmap)[j]) {
start = (&group_data[i].block_bitmap)[j];
u8 group_data[8];
struct group_data groups[MAX_GROUPS]; /* Huffman coding tables */
struct group_data *hufGroup = NULL;
u64 *data = ps->group_data;
ps->group_data = data;
zfree(&ps->group_data);
u64 *group_data;