block_id
static void ice_fill_tbl(struct ice_hw *hw, enum ice_block block_id, u32 sid)
sizeof(*hw->blk[block_id].xlt1.t);
dst = hw->blk[block_id].xlt1.t;
dst_len = hw->blk[block_id].xlt1.count *
sizeof(*hw->blk[block_id].xlt1.t);
sizeof(*hw->blk[block_id].xlt2.t);
dst = (u8 *)hw->blk[block_id].xlt2.t;
dst_len = hw->blk[block_id].xlt2.count *
sizeof(*hw->blk[block_id].xlt2.t);
sizeof(*hw->blk[block_id].prof.t);
dst = (u8 *)hw->blk[block_id].prof.t;
dst_len = hw->blk[block_id].prof.count *
sizeof(*hw->blk[block_id].prof.t);
sizeof(*hw->blk[block_id].prof_redir.t);
dst = hw->blk[block_id].prof_redir.t;
dst_len = hw->blk[block_id].prof_redir.count *
sizeof(*hw->blk[block_id].prof_redir.t);
hw->blk[block_id].es.fvw) *
sizeof(*hw->blk[block_id].es.t);
dst = (u8 *)hw->blk[block_id].es.t;
dst_len = (u32)(hw->blk[block_id].es.count *
hw->blk[block_id].es.fvw) *
sizeof(*hw->blk[block_id].es.t);
enum block_id block_id)
return (const struct dbg_bus_block *)&dbg_bus_blocks[block_id * MAX_CHIP_IDS + dev_data->chip_id];
enum block_id block_id)
block_bus = &dev_data->bus.blocks[block_id];
block_desc = get_dbg_bus_block_desc(p_hwfn, block_id);
if (dev_data->block_in_reset[storm->block_id])
enum block_id block_id;
u32 block_id;
if (dev_data->block_in_reset[storm->block_id])
if (dev_data->block_in_reset[storm->block_id] || ecore_rd(p_hwfn, p_ptt, storm->sem_sync_dbg_empty_addr)) {
if (dev_data->block_in_reset[storm->block_id])
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct block_defs *block = s_block_defs[block_id];
if (block->dbg_client_id[dev_data->chip_id] != MAX_DBG_BUS_CLIENTS && !dev_data->block_in_reset[block_id])
u32 block_id, offset = 0;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct dbg_bus_block_data *block_bus = &dev_data->bus.blocks[block_id];
struct block_defs *block = s_block_defs[block_id];
enum block_id block_id,
struct block_defs *block = s_block_defs[block_id];
u32 block_id, i;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct block_defs *block = s_block_defs[block_id];
static const struct dbg_attn_block_type_data* ecore_get_block_attn_data(enum block_id block_id,
return &base_attn_block_arr[block_id].per_type_data[attn_type];
static const struct dbg_attn_reg* ecore_get_block_attn_regs(enum block_id block_id,
const struct dbg_attn_block_type_data *block_type_data = ecore_get_block_attn_data(block_id, attn_type);
u32 block_id;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
if (dev_data->block_in_reset[block_id])
attn_reg_arr = ecore_get_block_attn_regs((enum block_id)block_id, ATTN_TYPE_PARITY, &num_attn_regs);
if (!mode_match || !block_enable[cond_hdr->block_id]) {
u32 block_id, offset = 0, num_reg_entries = 0;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
if (dev_data->block_in_reset[block_id] && dump)
attn_reg_arr = ecore_get_block_attn_regs((enum block_id)block_id, ATTN_TYPE_PARITY, &num_attn_regs);
if (dev_data->block_in_reset[storm->block_id] && dump)
if (!ecore_grc_is_mem_included(p_hwfn, (enum block_id)cond_hdr->block_id, mem_group_id))
if (s_block_defs[cond_hdr->block_id]->associated_to_storm) {
storm_letter = s_storm_defs[s_block_defs[cond_hdr->block_id]->storm_id].letter;
enum block_id block_id,
struct block_defs *block = s_block_defs[block_id];
u32 block_id, line_id, offset = 0;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct block_defs *block = s_block_defs[block_id];
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct block_defs *block = s_block_defs[block_id];
block_desc = get_dbg_bus_block_desc(p_hwfn, (enum block_id)block_id);
if (dev_data->block_in_reset[block_id]) {
ecore_config_dbg_line(p_hwfn, p_ptt, (enum block_id)block_id, (u8)line_id, 0xf, 0, 0, 0);
u32 block_id;
block_id = GET_FIELD(reg->data, DBG_IDLE_CHK_INFO_REG_BLOCK_ID);
if (block_id >= MAX_BLOCK_ID) {
if (!dev_data->block_in_reset[block_id]) {
u32 block_id = GET_FIELD(cond_regs[reg_id].data, DBG_IDLE_CHK_COND_REG_BLOCK_ID);
if (block_id >= MAX_BLOCK_ID) {
check_rule = !dev_data->block_in_reset[block_id];
if (dev_data->block_in_reset[storm->block_id])
u32 block_id;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct dbg_bus_block_data *block_bus = &dev_data->bus.blocks[block_id];
enum block_id block_id,
struct block_defs *block = s_block_defs[block_id];
block_bus = &dev_data->bus.blocks[block_id];
block_desc = get_dbg_bus_block_desc(p_hwfn, block_id);
DP_VERBOSE(p_hwfn, ECORE_MSG_DEBUG, "dbg_bus_enable_block: block = %d, line_num = %d, enable_mask = 0x%x, right_shift = %d, force_valid_mask = 0x%x, force_frame_mask = 0x%x\n", block_id, line_num, enable_mask, right_shift, force_valid_mask, force_frame_mask);
if (block_id >= MAX_BLOCK_ID)
if (dev_data->block_in_reset[block_id])
dev_data->bus.blocks[block_id].line_num = line_num;
if (dev_data->block_in_reset[storm->block_id])
enum block_id block_id,
DP_VERBOSE(p_hwfn, ECORE_MSG_DEBUG, "dbg_bus_enable_filter: block = %d, const_msg_len = %d\n", block_id, const_msg_len);
if (block_id >= MAX_BLOCK_ID)
if (!GET_FIELD(dev_data->bus.blocks[block_id].data, DBG_BUS_BLOCK_DATA_ENABLE_MASK))
enum block_id block_id,
DP_VERBOSE(p_hwfn, ECORE_MSG_DEBUG, "dbg_bus_add_trigger_state: block = %d, const_msg_len = %d, count_to_next = %d\n", block_id, const_msg_len, count_to_next);
block_bus = &bus->blocks[block_id];
if (block_id >= MAX_BLOCK_ID)
u32 block_id, client_mask = 0;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct dbg_bus_block_data *block_bus = &bus->blocks[block_id];
struct block_defs *block = s_block_defs[block_id];
if (GET_FIELD(block_bus->data, DBG_BUS_BLOCK_DATA_ENABLE_MASK) && block_id != BLOCK_DBG)
u32 block_id;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct dbg_bus_block_data *block_bus = &bus->blocks[block_id];
line_desc = get_dbg_bus_line_desc(p_hwfn, (enum block_id)block_id);
u32 block_id;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
struct dbg_bus_block_data *block_bus = &bus->blocks[block_id];
u32 block_id;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++) {
if (!GET_FIELD(dev_data->bus.blocks[block_id].data, DBG_BUS_BLOCK_DATA_ENABLE_MASK) || block_id == BLOCK_DBG)
ecore_config_dbg_line(p_hwfn, p_ptt, (enum block_id)block_id,
dev_data->bus.blocks[block_id].line_num,
GET_FIELD(dev_data->bus.blocks[block_id].data, DBG_BUS_BLOCK_DATA_ENABLE_MASK),
GET_FIELD(dev_data->bus.blocks[block_id].data, DBG_BUS_BLOCK_DATA_RIGHT_SHIFT),
GET_FIELD(dev_data->bus.blocks[block_id].data, DBG_BUS_BLOCK_DATA_FORCE_VALID_MASK),
GET_FIELD(dev_data->bus.blocks[block_id].data, DBG_BUS_BLOCK_DATA_FORCE_FRAME_MASK));
u32 min_buf_size_in_dwords, block_id, offset = 0;
for (block_id = 0; block_id < MAX_BLOCK_ID; block_id++)
enum block_id block_id,
attn_reg_arr = ecore_get_block_attn_regs(block_id, attn_type, &num_attn_regs);
results->block_id = (u8)block_id;
results->names_offset = ecore_get_block_attn_data(block_id, attn_type)->names_offset;
enum block_id block_id)
struct block_defs *block = s_block_defs[block_id];
enum block_id block,
enum block_id block,
enum block_id block,
enum block_id block,
enum block_id block);
u8 block_id /* Registers block ID */;
u8 block_id /* block ID */;
u16 block_id /* Blocks ID */;
enum block_id block_index;
enum block_id block,
enum block_id id, enum dbg_attn_type type,
u32 block_id = p_aeu->block_index, mask, val;
if (block_id != MAX_BLOCK_ID) {
ecore_int_attn_print(p_hwfn, block_id, ATTN_TYPE_PARITY, false);
if (block_id == BLOCK_BTB) {