block_info
struct hl_special_block_info *block_info,
u32 fw_block_base_address = block_info->base_addr +
major * block_info->major_offset +
minor * block_info->minor_offset +
sub_minor * block_info->sub_minor_offset;
struct hl_special_block_info *block_info,
block_base_addr = hl_automated_get_block_base_addr(hdev, block_info,
struct ta_ras_trigger_error_input block_info = {
block_info.address =
block_info.address);
ret = block_obj->hw_ops->ras_error_inject(adev, &block_info,
ret = psp_ras_trigger_error(&adev->psp, &block_info, info->instance_mask);
struct ta_ras_trigger_error_input *block_info =
ret2 = psp_ras_trigger_error(&adev->psp, block_info, instance_mask);
struct ta_ras_trigger_error_input block_info = { 0 };
block_info.block_id = amdgpu_ras_block_to_ta(info->head.block);
block_info.sub_block_index =
block_info.inject_error_type = amdgpu_ras_error_to_ta(info->head.type);
block_info.address = info->address;
block_info.value = info->value;
ret = psp_ras_trigger_error(&adev->psp, &block_info, instance_mask);
struct ras_cmd_inject_error_req *block_info)
if (block_info->block_id == TA_RAS_BLOCK__XGMI_WAFL) {
struct ras_cmd_inject_error_req *block_info)
if (block_info->block_id == TA_RAS_BLOCK__XGMI_WAFL) {
aca_blk = &ras_aca->aca_blk[ip_func->block_info[i]->ras_block_id];
aca_blk->blk_info = ip_func->block_info[i];
const struct aca_block_info **block_info;
.block_info = aca_block_info_v1_0,
struct ras_ta_trigger_error_input block_info = {
ret = ras_psp_trigger_error(ras_core, &block_info, req->instance_mask);
output_data->address = block_info.address;
get_resources_id(blk->block_info, &pipe_id, &comp_id);
BLOCK_INFO_INPUT_ID(blk->block_info),
get_resources_id(blk->block_info, &pipe_id, &comp_id);
BLOCK_INFO_INPUT_ID(blk->block_info),
get_resources_id(blk->block_info, &pipe_id, &comp_id);
BLOCK_INFO_INPUT_ID(blk->block_info),
int blk_id = BLOCK_INFO_BLK_ID(blk->block_info);
switch (BLOCK_INFO_BLK_TYPE(blk->block_info)) {
blk->block_info);
get_resources_id(blk->block_info, &pipe_id, &layer_id);
BLOCK_INFO_INPUT_ID(blk->block_info),
get_resources_id(blk->block_info, &pipe_id, &layer_id);
layer_id, BLOCK_INFO_INPUT_ID(blk->block_info),
get_resources_id(blk->block_info, &pipe_id, &comp_id);
BLOCK_INFO_INPUT_ID(blk->block_info),
get_resources_id(blk->block_info, &pipe_id, &comp_id);
comp_id, BLOCK_INFO_INPUT_ID(blk->block_info),
pipe_id, BLOCK_INFO_BLK_ID(blk->block_info));
get_resources_id(blk->block_info, &pipe_id, &comp_id);
BLOCK_INFO_INPUT_ID(blk->block_info),
seq_printf(sf, "BLOCK_INFO:\t\t0x%X\n", hdr.block_info);
blk->block_info = malidp_read32(reg, BLK_BLOCK_INFO);
if (BLOCK_INFO_BLK_TYPE(blk->block_info) == D71_BLK_TYPE_RESERVED)
if (BLOCK_INFO_BLK_TYPE(blk.block_info) != D71_BLK_TYPE_RESERVED) {
u32 block_info;
return BLOCK_INFO_BLK_TYPE(blk->block_info);
uint32_t block_info;
uint32_t block_info;
struct btrfs_tree_block_info *block_info;
size += sizeof(*block_info);
block_info = (struct btrfs_tree_block_info *)(extent_item + 1);
btrfs_set_tree_block_key(leaf, block_info, &extent_op->key);
btrfs_set_tree_block_level(leaf, block_info, level);
iref = (struct btrfs_extent_inline_ref *)(block_info + 1);
struct btrfs_tree_block_info *block_info;
u32 size = sizeof(*item) + sizeof(*iref) + sizeof(*block_info);
block_info = (struct btrfs_tree_block_info *)(item + 1);
btrfs_set_tree_block_level(leaf, block_info, 0);
iref = (struct btrfs_extent_inline_ref *)(block_info + 1);
q->block_info.chain_head_change = clsact_chain_head_change;
q->block_info.chain_head_change_priv = &q->miniqp;
err = tcf_block_get_ext(&q->block, sch, &q->block_info, extack);
tcf_block_put_ext(q->block, sch, &q->block_info);
struct tcf_block_ext_info block_info;
q->block_info.block_index = block_index;
return q->block_info.block_index;
q->block_info.binder_type = FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS;
struct block_info *bi;
bi = block_he->block_info;
if (pair->block_info->num && he->block_info->num) {
pair->block_info->cycles_aggr / pair->block_info->num_aggr -
he->block_info->cycles_aggr / he->block_info->num_aggr;
for (int i = 0; i < pair->block_info->num; i++) {
if (i >= he->block_info->num || i >= NUM_SPARKS)
val = llabs(pair->block_info->cycles_spark[i] -
he->block_info->cycles_spark[i]);
static void init_block_info(struct block_info *bi, struct symbol *sym,
struct block_info *bi;
struct block_info *bi = he->block_info;
struct block_info *bi_l = left->block_info;
struct block_info *bi_r = right->block_info;
struct block_info *bi = he->block_info;
struct block_info *bi = he->block_info;
struct block_info *bi = he->block_info;
struct block_info *bi = he->block_info;
struct block_info *bi = he->block_info;
static struct block_info *block_info__new(unsigned int br_cntr_nr)
struct block_info *bi = zalloc(sizeof(struct block_info));
struct block_info *bi = he->block_info;
void block_info__delete(struct block_info *bi)
struct block_info *bi_l = left->block_info;
struct block_info *bi_r = right->block_info;
void block_info__delete(struct block_info *bi);
if (he->block_info)
block_info__delete(he->block_info);
block_info__delete(entry->block_info);
struct block_info *block_info,
.block_info = block_info,
struct block_info *block_info)
.block_info = block_info,
if (h->block_info)
struct block_info;
struct block_info *block_info;
struct block_info *bi);