root/arch/x86/events/intel/uncore_snbep.c
// SPDX-License-Identifier: GPL-2.0
/* SandyBridge-EP/IvyTown uncore support */
#include <asm/cpu_device_id.h>
#include <asm/msr.h>
#include "uncore.h"
#include "uncore_discovery.h"

/* SNB-EP pci bus to socket mapping */
#define SNBEP_CPUNODEID                 0x40
#define SNBEP_GIDNIDMAP                 0x54

/* SNB-EP Box level control */
#define SNBEP_PMON_BOX_CTL_RST_CTRL     (1 << 0)
#define SNBEP_PMON_BOX_CTL_RST_CTRS     (1 << 1)
#define SNBEP_PMON_BOX_CTL_FRZ          (1 << 8)
#define SNBEP_PMON_BOX_CTL_FRZ_EN       (1 << 16)
#define SNBEP_PMON_BOX_CTL_INT          (SNBEP_PMON_BOX_CTL_RST_CTRL | \
                                         SNBEP_PMON_BOX_CTL_RST_CTRS | \
                                         SNBEP_PMON_BOX_CTL_FRZ_EN)
/* SNB-EP event control */
#define SNBEP_PMON_CTL_EV_SEL_MASK      0x000000ff
#define SNBEP_PMON_CTL_UMASK_MASK       0x0000ff00
#define SNBEP_PMON_CTL_RST              (1 << 17)
#define SNBEP_PMON_CTL_EDGE_DET         (1 << 18)
#define SNBEP_PMON_CTL_EV_SEL_EXT       (1 << 21)
#define SNBEP_PMON_CTL_EN               (1 << 22)
#define SNBEP_PMON_CTL_INVERT           (1 << 23)
#define SNBEP_PMON_CTL_TRESH_MASK       0xff000000
#define SNBEP_PMON_RAW_EVENT_MASK       (SNBEP_PMON_CTL_EV_SEL_MASK | \
                                         SNBEP_PMON_CTL_UMASK_MASK | \
                                         SNBEP_PMON_CTL_EDGE_DET | \
                                         SNBEP_PMON_CTL_INVERT | \
                                         SNBEP_PMON_CTL_TRESH_MASK)

/* SNB-EP Ubox event control */
#define SNBEP_U_MSR_PMON_CTL_TRESH_MASK         0x1f000000
#define SNBEP_U_MSR_PMON_RAW_EVENT_MASK         \
                                (SNBEP_PMON_CTL_EV_SEL_MASK | \
                                 SNBEP_PMON_CTL_UMASK_MASK | \
                                 SNBEP_PMON_CTL_EDGE_DET | \
                                 SNBEP_PMON_CTL_INVERT | \
                                 SNBEP_U_MSR_PMON_CTL_TRESH_MASK)

#define SNBEP_CBO_PMON_CTL_TID_EN               (1 << 19)
#define SNBEP_CBO_MSR_PMON_RAW_EVENT_MASK       (SNBEP_PMON_RAW_EVENT_MASK | \
                                                 SNBEP_CBO_PMON_CTL_TID_EN)

/* SNB-EP PCU event control */
#define SNBEP_PCU_MSR_PMON_CTL_OCC_SEL_MASK     0x0000c000
#define SNBEP_PCU_MSR_PMON_CTL_TRESH_MASK       0x1f000000
#define SNBEP_PCU_MSR_PMON_CTL_OCC_INVERT       (1 << 30)
#define SNBEP_PCU_MSR_PMON_CTL_OCC_EDGE_DET     (1 << 31)
#define SNBEP_PCU_MSR_PMON_RAW_EVENT_MASK       \
                                (SNBEP_PMON_CTL_EV_SEL_MASK | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_SEL_MASK | \
                                 SNBEP_PMON_CTL_EDGE_DET | \
                                 SNBEP_PMON_CTL_INVERT | \
                                 SNBEP_PCU_MSR_PMON_CTL_TRESH_MASK | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_INVERT | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_EDGE_DET)

#define SNBEP_QPI_PCI_PMON_RAW_EVENT_MASK       \
                                (SNBEP_PMON_RAW_EVENT_MASK | \
                                 SNBEP_PMON_CTL_EV_SEL_EXT)

/* SNB-EP pci control register */
#define SNBEP_PCI_PMON_BOX_CTL                  0xf4
#define SNBEP_PCI_PMON_CTL0                     0xd8
/* SNB-EP pci counter register */
#define SNBEP_PCI_PMON_CTR0                     0xa0

/* SNB-EP home agent register */
#define SNBEP_HA_PCI_PMON_BOX_ADDRMATCH0        0x40
#define SNBEP_HA_PCI_PMON_BOX_ADDRMATCH1        0x44
#define SNBEP_HA_PCI_PMON_BOX_OPCODEMATCH       0x48
/* SNB-EP memory controller register */
#define SNBEP_MC_CHy_PCI_PMON_FIXED_CTL         0xf0
#define SNBEP_MC_CHy_PCI_PMON_FIXED_CTR         0xd0
/* SNB-EP QPI register */
#define SNBEP_Q_Py_PCI_PMON_PKT_MATCH0          0x228
#define SNBEP_Q_Py_PCI_PMON_PKT_MATCH1          0x22c
#define SNBEP_Q_Py_PCI_PMON_PKT_MASK0           0x238
#define SNBEP_Q_Py_PCI_PMON_PKT_MASK1           0x23c

/* SNB-EP Ubox register */
#define SNBEP_U_MSR_PMON_CTR0                   0xc16
#define SNBEP_U_MSR_PMON_CTL0                   0xc10

#define SNBEP_U_MSR_PMON_UCLK_FIXED_CTL         0xc08
#define SNBEP_U_MSR_PMON_UCLK_FIXED_CTR         0xc09

/* SNB-EP Cbo register */
#define SNBEP_C0_MSR_PMON_CTR0                  0xd16
#define SNBEP_C0_MSR_PMON_CTL0                  0xd10
#define SNBEP_C0_MSR_PMON_BOX_CTL               0xd04
#define SNBEP_C0_MSR_PMON_BOX_FILTER            0xd14
#define SNBEP_CBO_MSR_OFFSET                    0x20

#define SNBEP_CB0_MSR_PMON_BOX_FILTER_TID       0x1f
#define SNBEP_CB0_MSR_PMON_BOX_FILTER_NID       0x3fc00
#define SNBEP_CB0_MSR_PMON_BOX_FILTER_STATE     0x7c0000
#define SNBEP_CB0_MSR_PMON_BOX_FILTER_OPC       0xff800000

#define SNBEP_CBO_EVENT_EXTRA_REG(e, m, i) {    \
        .event = (e),                           \
        .msr = SNBEP_C0_MSR_PMON_BOX_FILTER,    \
        .config_mask = (m),                     \
        .idx = (i)                              \
}

/* SNB-EP PCU register */
#define SNBEP_PCU_MSR_PMON_CTR0                 0xc36
#define SNBEP_PCU_MSR_PMON_CTL0                 0xc30
#define SNBEP_PCU_MSR_PMON_BOX_CTL              0xc24
#define SNBEP_PCU_MSR_PMON_BOX_FILTER           0xc34
#define SNBEP_PCU_MSR_PMON_BOX_FILTER_MASK      0xffffffff
#define SNBEP_PCU_MSR_CORE_C3_CTR               0x3fc
#define SNBEP_PCU_MSR_CORE_C6_CTR               0x3fd

/* IVBEP event control */
#define IVBEP_PMON_BOX_CTL_INT          (SNBEP_PMON_BOX_CTL_RST_CTRL | \
                                         SNBEP_PMON_BOX_CTL_RST_CTRS)
#define IVBEP_PMON_RAW_EVENT_MASK               (SNBEP_PMON_CTL_EV_SEL_MASK | \
                                         SNBEP_PMON_CTL_UMASK_MASK | \
                                         SNBEP_PMON_CTL_EDGE_DET | \
                                         SNBEP_PMON_CTL_TRESH_MASK)
/* IVBEP Ubox */
#define IVBEP_U_MSR_PMON_GLOBAL_CTL             0xc00
#define IVBEP_U_PMON_GLOBAL_FRZ_ALL             (1 << 31)
#define IVBEP_U_PMON_GLOBAL_UNFRZ_ALL           (1 << 29)

#define IVBEP_U_MSR_PMON_RAW_EVENT_MASK \
                                (SNBEP_PMON_CTL_EV_SEL_MASK | \
                                 SNBEP_PMON_CTL_UMASK_MASK | \
                                 SNBEP_PMON_CTL_EDGE_DET | \
                                 SNBEP_U_MSR_PMON_CTL_TRESH_MASK)
/* IVBEP Cbo */
#define IVBEP_CBO_MSR_PMON_RAW_EVENT_MASK               (IVBEP_PMON_RAW_EVENT_MASK | \
                                                 SNBEP_CBO_PMON_CTL_TID_EN)

#define IVBEP_CB0_MSR_PMON_BOX_FILTER_TID               (0x1fULL << 0)
#define IVBEP_CB0_MSR_PMON_BOX_FILTER_LINK      (0xfULL << 5)
#define IVBEP_CB0_MSR_PMON_BOX_FILTER_STATE     (0x3fULL << 17)
#define IVBEP_CB0_MSR_PMON_BOX_FILTER_NID               (0xffffULL << 32)
#define IVBEP_CB0_MSR_PMON_BOX_FILTER_OPC               (0x1ffULL << 52)
#define IVBEP_CB0_MSR_PMON_BOX_FILTER_C6                (0x1ULL << 61)
#define IVBEP_CB0_MSR_PMON_BOX_FILTER_NC                (0x1ULL << 62)
#define IVBEP_CB0_MSR_PMON_BOX_FILTER_ISOC      (0x1ULL << 63)

/* IVBEP home agent */
#define IVBEP_HA_PCI_PMON_CTL_Q_OCC_RST         (1 << 16)
#define IVBEP_HA_PCI_PMON_RAW_EVENT_MASK                \
                                (IVBEP_PMON_RAW_EVENT_MASK | \
                                 IVBEP_HA_PCI_PMON_CTL_Q_OCC_RST)
/* IVBEP PCU */
#define IVBEP_PCU_MSR_PMON_RAW_EVENT_MASK       \
                                (SNBEP_PMON_CTL_EV_SEL_MASK | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_SEL_MASK | \
                                 SNBEP_PMON_CTL_EDGE_DET | \
                                 SNBEP_PCU_MSR_PMON_CTL_TRESH_MASK | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_INVERT | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_EDGE_DET)
/* IVBEP QPI */
#define IVBEP_QPI_PCI_PMON_RAW_EVENT_MASK       \
                                (IVBEP_PMON_RAW_EVENT_MASK | \
                                 SNBEP_PMON_CTL_EV_SEL_EXT)

#define __BITS_VALUE(x, i, n)  ((typeof(x))(((x) >> ((i) * (n))) & \
                                ((1ULL << (n)) - 1)))

/* Haswell-EP Ubox */
#define HSWEP_U_MSR_PMON_CTR0                   0x709
#define HSWEP_U_MSR_PMON_CTL0                   0x705
#define HSWEP_U_MSR_PMON_FILTER                 0x707

#define HSWEP_U_MSR_PMON_UCLK_FIXED_CTL         0x703
#define HSWEP_U_MSR_PMON_UCLK_FIXED_CTR         0x704

#define HSWEP_U_MSR_PMON_BOX_FILTER_TID         (0x1 << 0)
#define HSWEP_U_MSR_PMON_BOX_FILTER_CID         (0x1fULL << 1)
#define HSWEP_U_MSR_PMON_BOX_FILTER_MASK \
                                        (HSWEP_U_MSR_PMON_BOX_FILTER_TID | \
                                         HSWEP_U_MSR_PMON_BOX_FILTER_CID)

/* Haswell-EP CBo */
#define HSWEP_C0_MSR_PMON_CTR0                  0xe08
#define HSWEP_C0_MSR_PMON_CTL0                  0xe01
#define HSWEP_C0_MSR_PMON_BOX_CTL                       0xe00
#define HSWEP_C0_MSR_PMON_BOX_FILTER0           0xe05
#define HSWEP_CBO_MSR_OFFSET                    0x10


#define HSWEP_CB0_MSR_PMON_BOX_FILTER_TID               (0x3fULL << 0)
#define HSWEP_CB0_MSR_PMON_BOX_FILTER_LINK      (0xfULL << 6)
#define HSWEP_CB0_MSR_PMON_BOX_FILTER_STATE     (0x7fULL << 17)
#define HSWEP_CB0_MSR_PMON_BOX_FILTER_NID               (0xffffULL << 32)
#define HSWEP_CB0_MSR_PMON_BOX_FILTER_OPC               (0x1ffULL << 52)
#define HSWEP_CB0_MSR_PMON_BOX_FILTER_C6                (0x1ULL << 61)
#define HSWEP_CB0_MSR_PMON_BOX_FILTER_NC                (0x1ULL << 62)
#define HSWEP_CB0_MSR_PMON_BOX_FILTER_ISOC      (0x1ULL << 63)


/* Haswell-EP Sbox */
#define HSWEP_S0_MSR_PMON_CTR0                  0x726
#define HSWEP_S0_MSR_PMON_CTL0                  0x721
#define HSWEP_S0_MSR_PMON_BOX_CTL                       0x720
#define HSWEP_SBOX_MSR_OFFSET                   0xa
#define HSWEP_S_MSR_PMON_RAW_EVENT_MASK         (SNBEP_PMON_RAW_EVENT_MASK | \
                                                 SNBEP_CBO_PMON_CTL_TID_EN)

/* Haswell-EP PCU */
#define HSWEP_PCU_MSR_PMON_CTR0                 0x717
#define HSWEP_PCU_MSR_PMON_CTL0                 0x711
#define HSWEP_PCU_MSR_PMON_BOX_CTL              0x710
#define HSWEP_PCU_MSR_PMON_BOX_FILTER           0x715

/* KNL Ubox */
#define KNL_U_MSR_PMON_RAW_EVENT_MASK \
                                        (SNBEP_U_MSR_PMON_RAW_EVENT_MASK | \
                                                SNBEP_CBO_PMON_CTL_TID_EN)
/* KNL CHA */
#define KNL_CHA_MSR_OFFSET                      0xc
#define KNL_CHA_MSR_PMON_CTL_QOR                (1 << 16)
#define KNL_CHA_MSR_PMON_RAW_EVENT_MASK \
                                        (SNBEP_CBO_MSR_PMON_RAW_EVENT_MASK | \
                                         KNL_CHA_MSR_PMON_CTL_QOR)
#define KNL_CHA_MSR_PMON_BOX_FILTER_TID         0x1ff
#define KNL_CHA_MSR_PMON_BOX_FILTER_STATE       (7 << 18)
#define KNL_CHA_MSR_PMON_BOX_FILTER_OP          (0xfffffe2aULL << 32)
#define KNL_CHA_MSR_PMON_BOX_FILTER_REMOTE_NODE (0x1ULL << 32)
#define KNL_CHA_MSR_PMON_BOX_FILTER_LOCAL_NODE  (0x1ULL << 33)
#define KNL_CHA_MSR_PMON_BOX_FILTER_NNC         (0x1ULL << 37)

/* KNL EDC/MC UCLK */
#define KNL_UCLK_MSR_PMON_CTR0_LOW              0x400
#define KNL_UCLK_MSR_PMON_CTL0                  0x420
#define KNL_UCLK_MSR_PMON_BOX_CTL               0x430
#define KNL_UCLK_MSR_PMON_UCLK_FIXED_LOW        0x44c
#define KNL_UCLK_MSR_PMON_UCLK_FIXED_CTL        0x454
#define KNL_PMON_FIXED_CTL_EN                   0x1

/* KNL EDC */
#define KNL_EDC0_ECLK_MSR_PMON_CTR0_LOW         0xa00
#define KNL_EDC0_ECLK_MSR_PMON_CTL0             0xa20
#define KNL_EDC0_ECLK_MSR_PMON_BOX_CTL          0xa30
#define KNL_EDC0_ECLK_MSR_PMON_ECLK_FIXED_LOW   0xa3c
#define KNL_EDC0_ECLK_MSR_PMON_ECLK_FIXED_CTL   0xa44

/* KNL MC */
#define KNL_MC0_CH0_MSR_PMON_CTR0_LOW           0xb00
#define KNL_MC0_CH0_MSR_PMON_CTL0               0xb20
#define KNL_MC0_CH0_MSR_PMON_BOX_CTL            0xb30
#define KNL_MC0_CH0_MSR_PMON_FIXED_LOW          0xb3c
#define KNL_MC0_CH0_MSR_PMON_FIXED_CTL          0xb44

/* KNL IRP */
#define KNL_IRP_PCI_PMON_BOX_CTL                0xf0
#define KNL_IRP_PCI_PMON_RAW_EVENT_MASK         (SNBEP_PMON_RAW_EVENT_MASK | \
                                                 KNL_CHA_MSR_PMON_CTL_QOR)
/* KNL PCU */
#define KNL_PCU_PMON_CTL_EV_SEL_MASK            0x0000007f
#define KNL_PCU_PMON_CTL_USE_OCC_CTR            (1 << 7)
#define KNL_PCU_MSR_PMON_CTL_TRESH_MASK         0x3f000000
#define KNL_PCU_MSR_PMON_RAW_EVENT_MASK \
                                (KNL_PCU_PMON_CTL_EV_SEL_MASK | \
                                 KNL_PCU_PMON_CTL_USE_OCC_CTR | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_SEL_MASK | \
                                 SNBEP_PMON_CTL_EDGE_DET | \
                                 SNBEP_CBO_PMON_CTL_TID_EN | \
                                 SNBEP_PMON_CTL_INVERT | \
                                 KNL_PCU_MSR_PMON_CTL_TRESH_MASK | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_INVERT | \
                                 SNBEP_PCU_MSR_PMON_CTL_OCC_EDGE_DET)

/* SKX pci bus to socket mapping */
#define SKX_CPUNODEID                   0xc0
#define SKX_GIDNIDMAP                   0xd4

/*
 * The CPU_BUS_NUMBER MSR returns the values of the respective CPUBUSNO CSR
 * that BIOS programmed. MSR has package scope.
 * |  Bit  |  Default  |  Description
 * | [63]  |    00h    | VALID - When set, indicates the CPU bus
 *                       numbers have been initialized. (RO)
 * |[62:48]|    ---    | Reserved
 * |[47:40]|    00h    | BUS_NUM_5 - Return the bus number BIOS assigned
 *                       CPUBUSNO(5). (RO)
 * |[39:32]|    00h    | BUS_NUM_4 - Return the bus number BIOS assigned
 *                       CPUBUSNO(4). (RO)
 * |[31:24]|    00h    | BUS_NUM_3 - Return the bus number BIOS assigned
 *                       CPUBUSNO(3). (RO)
 * |[23:16]|    00h    | BUS_NUM_2 - Return the bus number BIOS assigned
 *                       CPUBUSNO(2). (RO)
 * |[15:8] |    00h    | BUS_NUM_1 - Return the bus number BIOS assigned
 *                       CPUBUSNO(1). (RO)
 * | [7:0] |    00h    | BUS_NUM_0 - Return the bus number BIOS assigned
 *                       CPUBUSNO(0). (RO)
 */
#define SKX_MSR_CPU_BUS_NUMBER          0x300
#define SKX_MSR_CPU_BUS_VALID_BIT       (1ULL << 63)
#define BUS_NUM_STRIDE                  8

/* SKX CHA */
#define SKX_CHA_MSR_PMON_BOX_FILTER_TID         (0x1ffULL << 0)
#define SKX_CHA_MSR_PMON_BOX_FILTER_LINK        (0xfULL << 9)
#define SKX_CHA_MSR_PMON_BOX_FILTER_STATE       (0x3ffULL << 17)
#define SKX_CHA_MSR_PMON_BOX_FILTER_REM         (0x1ULL << 32)
#define SKX_CHA_MSR_PMON_BOX_FILTER_LOC         (0x1ULL << 33)
#define SKX_CHA_MSR_PMON_BOX_FILTER_ALL_OPC     (0x1ULL << 35)
#define SKX_CHA_MSR_PMON_BOX_FILTER_NM          (0x1ULL << 36)
#define SKX_CHA_MSR_PMON_BOX_FILTER_NOT_NM      (0x1ULL << 37)
#define SKX_CHA_MSR_PMON_BOX_FILTER_OPC0        (0x3ffULL << 41)
#define SKX_CHA_MSR_PMON_BOX_FILTER_OPC1        (0x3ffULL << 51)
#define SKX_CHA_MSR_PMON_BOX_FILTER_C6          (0x1ULL << 61)
#define SKX_CHA_MSR_PMON_BOX_FILTER_NC          (0x1ULL << 62)
#define SKX_CHA_MSR_PMON_BOX_FILTER_ISOC        (0x1ULL << 63)

/* SKX IIO */
#define SKX_IIO0_MSR_PMON_CTL0          0xa48
#define SKX_IIO0_MSR_PMON_CTR0          0xa41
#define SKX_IIO0_MSR_PMON_BOX_CTL       0xa40
#define SKX_IIO_MSR_OFFSET              0x20

#define SKX_PMON_CTL_TRESH_MASK         (0xff << 24)
#define SKX_PMON_CTL_TRESH_MASK_EXT     (0xf)
#define SKX_PMON_CTL_CH_MASK            (0xff << 4)
#define SKX_PMON_CTL_FC_MASK            (0x7 << 12)
#define SKX_IIO_PMON_RAW_EVENT_MASK     (SNBEP_PMON_CTL_EV_SEL_MASK | \
                                         SNBEP_PMON_CTL_UMASK_MASK | \
                                         SNBEP_PMON_CTL_EDGE_DET | \
                                         SNBEP_PMON_CTL_INVERT | \
                                         SKX_PMON_CTL_TRESH_MASK)
#define SKX_IIO_PMON_RAW_EVENT_MASK_EXT (SKX_PMON_CTL_TRESH_MASK_EXT | \
                                         SKX_PMON_CTL_CH_MASK | \
                                         SKX_PMON_CTL_FC_MASK)

/* SKX IRP */
#define SKX_IRP0_MSR_PMON_CTL0          0xa5b
#define SKX_IRP0_MSR_PMON_CTR0          0xa59
#define SKX_IRP0_MSR_PMON_BOX_CTL       0xa58
#define SKX_IRP_MSR_OFFSET              0x20

/* SKX UPI */
#define SKX_UPI_PCI_PMON_CTL0           0x350
#define SKX_UPI_PCI_PMON_CTR0           0x318
#define SKX_UPI_PCI_PMON_BOX_CTL        0x378
#define SKX_UPI_CTL_UMASK_EXT           0xffefff

/* SKX M2M */
#define SKX_M2M_PCI_PMON_CTL0           0x228
#define SKX_M2M_PCI_PMON_CTR0           0x200
#define SKX_M2M_PCI_PMON_BOX_CTL        0x258

/* Memory Map registers device ID */
#define SNR_ICX_MESH2IIO_MMAP_DID               0x9a2
#define SNR_ICX_SAD_CONTROL_CFG         0x3f4

/* Getting I/O stack id in SAD_COTROL_CFG notation */
#define SAD_CONTROL_STACK_ID(data)              (((data) >> 4) & 0x7)

/* SNR Ubox */
#define SNR_U_MSR_PMON_CTR0                     0x1f98
#define SNR_U_MSR_PMON_CTL0                     0x1f91
#define SNR_U_MSR_PMON_UCLK_FIXED_CTL           0x1f93
#define SNR_U_MSR_PMON_UCLK_FIXED_CTR           0x1f94

/* SNR CHA */
#define SNR_CHA_RAW_EVENT_MASK_EXT              0x3ffffff
#define SNR_CHA_MSR_PMON_CTL0                   0x1c01
#define SNR_CHA_MSR_PMON_CTR0                   0x1c08
#define SNR_CHA_MSR_PMON_BOX_CTL                0x1c00
#define SNR_C0_MSR_PMON_BOX_FILTER0             0x1c05


/* SNR IIO */
#define SNR_IIO_MSR_PMON_CTL0                   0x1e08
#define SNR_IIO_MSR_PMON_CTR0                   0x1e01
#define SNR_IIO_MSR_PMON_BOX_CTL                0x1e00
#define SNR_IIO_MSR_OFFSET                      0x10
#define SNR_IIO_PMON_RAW_EVENT_MASK_EXT         0x7ffff

/* SNR IRP */
#define SNR_IRP0_MSR_PMON_CTL0                  0x1ea8
#define SNR_IRP0_MSR_PMON_CTR0                  0x1ea1
#define SNR_IRP0_MSR_PMON_BOX_CTL               0x1ea0
#define SNR_IRP_MSR_OFFSET                      0x10

/* SNR M2PCIE */
#define SNR_M2PCIE_MSR_PMON_CTL0                0x1e58
#define SNR_M2PCIE_MSR_PMON_CTR0                0x1e51
#define SNR_M2PCIE_MSR_PMON_BOX_CTL             0x1e50
#define SNR_M2PCIE_MSR_OFFSET                   0x10

/* SNR PCU */
#define SNR_PCU_MSR_PMON_CTL0                   0x1ef1
#define SNR_PCU_MSR_PMON_CTR0                   0x1ef8
#define SNR_PCU_MSR_PMON_BOX_CTL                0x1ef0
#define SNR_PCU_MSR_PMON_BOX_FILTER             0x1efc

/* SNR M2M */
#define SNR_M2M_PCI_PMON_CTL0                   0x468
#define SNR_M2M_PCI_PMON_CTR0                   0x440
#define SNR_M2M_PCI_PMON_BOX_CTL                0x438
#define SNR_M2M_PCI_PMON_UMASK_EXT              0xff

/* SNR PCIE3 */
#define SNR_PCIE3_PCI_PMON_CTL0                 0x508
#define SNR_PCIE3_PCI_PMON_CTR0                 0x4e8
#define SNR_PCIE3_PCI_PMON_BOX_CTL              0x4e0

/* SNR IMC */
#define SNR_IMC_MMIO_PMON_FIXED_CTL             0x54
#define SNR_IMC_MMIO_PMON_FIXED_CTR             0x38
#define SNR_IMC_MMIO_PMON_CTL0                  0x40
#define SNR_IMC_MMIO_PMON_CTR0                  0x8
#define SNR_IMC_MMIO_PMON_BOX_CTL               0x22800
#define SNR_IMC_MMIO_OFFSET                     0x4000
#define SNR_IMC_MMIO_SIZE                       0x4000
#define SNR_IMC_MMIO_BASE_OFFSET                0xd0
#define SNR_IMC_MMIO_BASE_MASK                  0x1FFFFFFF
#define SNR_IMC_MMIO_MEM0_OFFSET                0xd8
#define SNR_IMC_MMIO_MEM0_MASK                  0x7FF

/* ICX CHA */
#define ICX_C34_MSR_PMON_CTR0                   0xb68
#define ICX_C34_MSR_PMON_CTL0                   0xb61
#define ICX_C34_MSR_PMON_BOX_CTL                0xb60
#define ICX_C34_MSR_PMON_BOX_FILTER0            0xb65

/* ICX IIO */
#define ICX_IIO_MSR_PMON_CTL0                   0xa58
#define ICX_IIO_MSR_PMON_CTR0                   0xa51
#define ICX_IIO_MSR_PMON_BOX_CTL                0xa50

/* ICX IRP */
#define ICX_IRP0_MSR_PMON_CTL0                  0xa4d
#define ICX_IRP0_MSR_PMON_CTR0                  0xa4b
#define ICX_IRP0_MSR_PMON_BOX_CTL               0xa4a

/* ICX M2PCIE */
#define ICX_M2PCIE_MSR_PMON_CTL0                0xa46
#define ICX_M2PCIE_MSR_PMON_CTR0                0xa41
#define ICX_M2PCIE_MSR_PMON_BOX_CTL             0xa40

/* ICX UPI */
#define ICX_UPI_PCI_PMON_CTL0                   0x350
#define ICX_UPI_PCI_PMON_CTR0                   0x320
#define ICX_UPI_PCI_PMON_BOX_CTL                0x318
#define ICX_UPI_CTL_UMASK_EXT                   0xffffff
#define ICX_UBOX_DID                            0x3450

/* ICX M3UPI*/
#define ICX_M3UPI_PCI_PMON_CTL0                 0xd8
#define ICX_M3UPI_PCI_PMON_CTR0                 0xa8
#define ICX_M3UPI_PCI_PMON_BOX_CTL              0xa0

/* ICX IMC */
#define ICX_NUMBER_IMC_CHN                      3
#define ICX_IMC_MEM_STRIDE                      0x4

/* SPR */
#define SPR_RAW_EVENT_MASK_EXT                  0xffffff
#define SPR_UBOX_DID                            0x3250

/* SPR CHA */
#define SPR_CHA_EVENT_MASK_EXT                  0xffffffff
#define SPR_CHA_PMON_CTL_TID_EN                 (1 << 16)
#define SPR_CHA_PMON_EVENT_MASK                 (SNBEP_PMON_RAW_EVENT_MASK | \
                                                 SPR_CHA_PMON_CTL_TID_EN)
#define SPR_CHA_PMON_BOX_FILTER_TID             0x3ff

#define SPR_C0_MSR_PMON_BOX_FILTER0             0x200e

/* DMR */
#define DMR_IMH1_HIOP_MMIO_BASE                 0x1ffff6ae7000
#define DMR_HIOP_MMIO_SIZE                      0x8000
#define DMR_CXLCM_EVENT_MASK_EXT                0xf
#define DMR_HAMVF_EVENT_MASK_EXT                0xffffffff
#define DMR_PCIE4_EVENT_MASK_EXT                0xffffff

#define UNCORE_DMR_ITC                          0x30

#define DMR_IMC_PMON_FIXED_CTR                  0x18
#define DMR_IMC_PMON_FIXED_CTL                  0x10

DEFINE_UNCORE_FORMAT_ATTR(event, event, "config:0-7");
DEFINE_UNCORE_FORMAT_ATTR(event2, event, "config:0-6");
DEFINE_UNCORE_FORMAT_ATTR(event_ext, event, "config:0-7,21");
DEFINE_UNCORE_FORMAT_ATTR(use_occ_ctr, use_occ_ctr, "config:7");
DEFINE_UNCORE_FORMAT_ATTR(umask, umask, "config:8-15");
DEFINE_UNCORE_FORMAT_ATTR(umask_ext, umask, "config:8-15,32-43,45-55");
DEFINE_UNCORE_FORMAT_ATTR(umask_ext2, umask, "config:8-15,32-57");
DEFINE_UNCORE_FORMAT_ATTR(umask_ext3, umask, "config:8-15,32-39");
DEFINE_UNCORE_FORMAT_ATTR(umask_ext4, umask, "config:8-15,32-55");
DEFINE_UNCORE_FORMAT_ATTR(umask_ext5, umask, "config:8-15,32-63");
DEFINE_UNCORE_FORMAT_ATTR(qor, qor, "config:16");
DEFINE_UNCORE_FORMAT_ATTR(edge, edge, "config:18");
DEFINE_UNCORE_FORMAT_ATTR(tid_en, tid_en, "config:19");
DEFINE_UNCORE_FORMAT_ATTR(tid_en2, tid_en, "config:16");
DEFINE_UNCORE_FORMAT_ATTR(inv, inv, "config:23");
DEFINE_UNCORE_FORMAT_ATTR(inv2, inv, "config:21");
DEFINE_UNCORE_FORMAT_ATTR(thresh_ext, thresh_ext, "config:32-35");
DEFINE_UNCORE_FORMAT_ATTR(thresh10, thresh, "config:23-32");
DEFINE_UNCORE_FORMAT_ATTR(thresh9_2, thresh, "config:23-31");
DEFINE_UNCORE_FORMAT_ATTR(thresh9, thresh, "config:24-35");
DEFINE_UNCORE_FORMAT_ATTR(thresh8, thresh, "config:24-31");
DEFINE_UNCORE_FORMAT_ATTR(thresh6, thresh, "config:24-29");
DEFINE_UNCORE_FORMAT_ATTR(thresh5, thresh, "config:24-28");
DEFINE_UNCORE_FORMAT_ATTR(occ_sel, occ_sel, "config:14-15");
DEFINE_UNCORE_FORMAT_ATTR(occ_invert, occ_invert, "config:30");
DEFINE_UNCORE_FORMAT_ATTR(occ_edge, occ_edge, "config:14-51");
DEFINE_UNCORE_FORMAT_ATTR(occ_edge_det, occ_edge_det, "config:31");
DEFINE_UNCORE_FORMAT_ATTR(port_en, port_en, "config:32-35");
DEFINE_UNCORE_FORMAT_ATTR(rs3_sel, rs3_sel, "config:36");
DEFINE_UNCORE_FORMAT_ATTR(rx_sel, rx_sel, "config:37");
DEFINE_UNCORE_FORMAT_ATTR(tx_sel, tx_sel, "config:38");
DEFINE_UNCORE_FORMAT_ATTR(iep_sel, iep_sel, "config:39");
DEFINE_UNCORE_FORMAT_ATTR(vc_sel, vc_sel, "config:40-47");
DEFINE_UNCORE_FORMAT_ATTR(port_sel, port_sel, "config:48-55");
DEFINE_UNCORE_FORMAT_ATTR(ch_mask, ch_mask, "config:36-43");
DEFINE_UNCORE_FORMAT_ATTR(ch_mask2, ch_mask, "config:36-47");
DEFINE_UNCORE_FORMAT_ATTR(fc_mask, fc_mask, "config:44-46");
DEFINE_UNCORE_FORMAT_ATTR(fc_mask2, fc_mask, "config:48-50");
DEFINE_UNCORE_FORMAT_ATTR(filter_tid, filter_tid, "config1:0-4");
DEFINE_UNCORE_FORMAT_ATTR(filter_tid2, filter_tid, "config1:0");
DEFINE_UNCORE_FORMAT_ATTR(filter_tid3, filter_tid, "config1:0-5");
DEFINE_UNCORE_FORMAT_ATTR(filter_tid4, filter_tid, "config1:0-8");
DEFINE_UNCORE_FORMAT_ATTR(filter_tid5, filter_tid, "config1:0-9");
DEFINE_UNCORE_FORMAT_ATTR(filter_cid, filter_cid, "config1:5");
DEFINE_UNCORE_FORMAT_ATTR(filter_link, filter_link, "config1:5-8");
DEFINE_UNCORE_FORMAT_ATTR(filter_link2, filter_link, "config1:6-8");
DEFINE_UNCORE_FORMAT_ATTR(filter_link3, filter_link, "config1:12");
DEFINE_UNCORE_FORMAT_ATTR(filter_nid, filter_nid, "config1:10-17");
DEFINE_UNCORE_FORMAT_ATTR(filter_nid2, filter_nid, "config1:32-47");
DEFINE_UNCORE_FORMAT_ATTR(filter_state, filter_state, "config1:18-22");
DEFINE_UNCORE_FORMAT_ATTR(filter_state2, filter_state, "config1:17-22");
DEFINE_UNCORE_FORMAT_ATTR(filter_state3, filter_state, "config1:17-23");
DEFINE_UNCORE_FORMAT_ATTR(filter_state4, filter_state, "config1:18-20");
DEFINE_UNCORE_FORMAT_ATTR(filter_state5, filter_state, "config1:17-26");
DEFINE_UNCORE_FORMAT_ATTR(filter_rem, filter_rem, "config1:32");
DEFINE_UNCORE_FORMAT_ATTR(filter_loc, filter_loc, "config1:33");
DEFINE_UNCORE_FORMAT_ATTR(filter_nm, filter_nm, "config1:36");
DEFINE_UNCORE_FORMAT_ATTR(filter_not_nm, filter_not_nm, "config1:37");
DEFINE_UNCORE_FORMAT_ATTR(filter_local, filter_local, "config1:33");
DEFINE_UNCORE_FORMAT_ATTR(filter_all_op, filter_all_op, "config1:35");
DEFINE_UNCORE_FORMAT_ATTR(filter_nnm, filter_nnm, "config1:37");
DEFINE_UNCORE_FORMAT_ATTR(filter_opc, filter_opc, "config1:23-31");
DEFINE_UNCORE_FORMAT_ATTR(filter_opc2, filter_opc, "config1:52-60");
DEFINE_UNCORE_FORMAT_ATTR(filter_opc3, filter_opc, "config1:41-60");
DEFINE_UNCORE_FORMAT_ATTR(filter_opc_0, filter_opc0, "config1:41-50");
DEFINE_UNCORE_FORMAT_ATTR(filter_opc_1, filter_opc1, "config1:51-60");
DEFINE_UNCORE_FORMAT_ATTR(filter_nc, filter_nc, "config1:62");
DEFINE_UNCORE_FORMAT_ATTR(filter_c6, filter_c6, "config1:61");
DEFINE_UNCORE_FORMAT_ATTR(filter_isoc, filter_isoc, "config1:63");
DEFINE_UNCORE_FORMAT_ATTR(filter_band0, filter_band0, "config1:0-7");
DEFINE_UNCORE_FORMAT_ATTR(filter_band1, filter_band1, "config1:8-15");
DEFINE_UNCORE_FORMAT_ATTR(filter_band2, filter_band2, "config1:16-23");
DEFINE_UNCORE_FORMAT_ATTR(filter_band3, filter_band3, "config1:24-31");
DEFINE_UNCORE_FORMAT_ATTR(match_rds, match_rds, "config1:48-51");
DEFINE_UNCORE_FORMAT_ATTR(match_rnid30, match_rnid30, "config1:32-35");
DEFINE_UNCORE_FORMAT_ATTR(match_rnid4, match_rnid4, "config1:31");
DEFINE_UNCORE_FORMAT_ATTR(match_dnid, match_dnid, "config1:13-17");
DEFINE_UNCORE_FORMAT_ATTR(match_mc, match_mc, "config1:9-12");
DEFINE_UNCORE_FORMAT_ATTR(match_opc, match_opc, "config1:5-8");
DEFINE_UNCORE_FORMAT_ATTR(match_vnw, match_vnw, "config1:3-4");
DEFINE_UNCORE_FORMAT_ATTR(match0, match0, "config1:0-31");
DEFINE_UNCORE_FORMAT_ATTR(match1, match1, "config1:32-63");
DEFINE_UNCORE_FORMAT_ATTR(mask_rds, mask_rds, "config2:48-51");
DEFINE_UNCORE_FORMAT_ATTR(mask_rnid30, mask_rnid30, "config2:32-35");
DEFINE_UNCORE_FORMAT_ATTR(mask_rnid4, mask_rnid4, "config2:31");
DEFINE_UNCORE_FORMAT_ATTR(mask_dnid, mask_dnid, "config2:13-17");
DEFINE_UNCORE_FORMAT_ATTR(mask_mc, mask_mc, "config2:9-12");
DEFINE_UNCORE_FORMAT_ATTR(mask_opc, mask_opc, "config2:5-8");
DEFINE_UNCORE_FORMAT_ATTR(mask_vnw, mask_vnw, "config2:3-4");
DEFINE_UNCORE_FORMAT_ATTR(mask0, mask0, "config2:0-31");
DEFINE_UNCORE_FORMAT_ATTR(mask1, mask1, "config2:32-63");

static void snbep_uncore_pci_disable_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;
        int box_ctl = uncore_pci_box_ctl(box);
        u32 config = 0;

        if (!pci_read_config_dword(pdev, box_ctl, &config)) {
                config |= SNBEP_PMON_BOX_CTL_FRZ;
                pci_write_config_dword(pdev, box_ctl, config);
        }
}

static void snbep_uncore_pci_enable_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;
        int box_ctl = uncore_pci_box_ctl(box);
        u32 config = 0;

        if (!pci_read_config_dword(pdev, box_ctl, &config)) {
                config &= ~SNBEP_PMON_BOX_CTL_FRZ;
                pci_write_config_dword(pdev, box_ctl, config);
        }
}

static void snbep_uncore_pci_enable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;

        pci_write_config_dword(pdev, hwc->config_base, hwc->config | SNBEP_PMON_CTL_EN);
}

static void snbep_uncore_pci_disable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;

        pci_write_config_dword(pdev, hwc->config_base, hwc->config);
}

static u64 snbep_uncore_pci_read_counter(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;
        u64 count = 0;

        pci_read_config_dword(pdev, hwc->event_base, (u32 *)&count);
        pci_read_config_dword(pdev, hwc->event_base + 4, (u32 *)&count + 1);

        return count;
}

static void snbep_uncore_pci_init_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;
        int box_ctl = uncore_pci_box_ctl(box);

        pci_write_config_dword(pdev, box_ctl, SNBEP_PMON_BOX_CTL_INT);
}

static void snbep_uncore_msr_disable_box(struct intel_uncore_box *box)
{
        u64 config;
        unsigned msr;

        msr = uncore_msr_box_ctl(box);
        if (msr) {
                rdmsrq(msr, config);
                config |= SNBEP_PMON_BOX_CTL_FRZ;
                wrmsrq(msr, config);
        }
}

static void snbep_uncore_msr_enable_box(struct intel_uncore_box *box)
{
        u64 config;
        unsigned msr;

        msr = uncore_msr_box_ctl(box);
        if (msr) {
                rdmsrq(msr, config);
                config &= ~SNBEP_PMON_BOX_CTL_FRZ;
                wrmsrq(msr, config);
        }
}

static void snbep_uncore_msr_enable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;

        if (reg1->idx != EXTRA_REG_NONE)
                wrmsrq(reg1->reg, uncore_shared_reg_config(box, 0));

        wrmsrq(hwc->config_base, hwc->config | SNBEP_PMON_CTL_EN);
}

static void snbep_uncore_msr_disable_event(struct intel_uncore_box *box,
                                        struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;

        wrmsrq(hwc->config_base, hwc->config);
}

static void snbep_uncore_msr_init_box(struct intel_uncore_box *box)
{
        unsigned msr = uncore_msr_box_ctl(box);

        if (msr)
                wrmsrq(msr, SNBEP_PMON_BOX_CTL_INT);
}

static struct attribute *snbep_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static struct attribute *snbep_uncore_ubox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh5.attr,
        NULL,
};

static struct attribute *snbep_uncore_cbox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_filter_tid.attr,
        &format_attr_filter_nid.attr,
        &format_attr_filter_state.attr,
        &format_attr_filter_opc.attr,
        NULL,
};

static struct attribute *snbep_uncore_pcu_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_occ_sel.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh5.attr,
        &format_attr_occ_invert.attr,
        &format_attr_occ_edge.attr,
        &format_attr_filter_band0.attr,
        &format_attr_filter_band1.attr,
        &format_attr_filter_band2.attr,
        &format_attr_filter_band3.attr,
        NULL,
};

static struct attribute *snbep_uncore_qpi_formats_attr[] = {
        &format_attr_event_ext.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_match_rds.attr,
        &format_attr_match_rnid30.attr,
        &format_attr_match_rnid4.attr,
        &format_attr_match_dnid.attr,
        &format_attr_match_mc.attr,
        &format_attr_match_opc.attr,
        &format_attr_match_vnw.attr,
        &format_attr_match0.attr,
        &format_attr_match1.attr,
        &format_attr_mask_rds.attr,
        &format_attr_mask_rnid30.attr,
        &format_attr_mask_rnid4.attr,
        &format_attr_mask_dnid.attr,
        &format_attr_mask_mc.attr,
        &format_attr_mask_opc.attr,
        &format_attr_mask_vnw.attr,
        &format_attr_mask0.attr,
        &format_attr_mask1.attr,
        NULL,
};

static struct uncore_event_desc snbep_uncore_imc_events[] = {
        INTEL_UNCORE_EVENT_DESC(clockticks,      "event=0xff,umask=0x00"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read,  "event=0x04,umask=0x03"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.unit, "MiB"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write, "event=0x04,umask=0x0c"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.unit, "MiB"),
        { /* end: all zeroes */ },
};

static struct uncore_event_desc snbep_uncore_qpi_events[] = {
        INTEL_UNCORE_EVENT_DESC(clockticks,       "event=0x14"),
        INTEL_UNCORE_EVENT_DESC(txl_flits_active, "event=0x00,umask=0x06"),
        INTEL_UNCORE_EVENT_DESC(drs_data,         "event=0x102,umask=0x08"),
        INTEL_UNCORE_EVENT_DESC(ncb_data,         "event=0x103,umask=0x04"),
        { /* end: all zeroes */ },
};

static const struct attribute_group snbep_uncore_format_group = {
        .name = "format",
        .attrs = snbep_uncore_formats_attr,
};

static const struct attribute_group snbep_uncore_ubox_format_group = {
        .name = "format",
        .attrs = snbep_uncore_ubox_formats_attr,
};

static const struct attribute_group snbep_uncore_cbox_format_group = {
        .name = "format",
        .attrs = snbep_uncore_cbox_formats_attr,
};

static const struct attribute_group snbep_uncore_pcu_format_group = {
        .name = "format",
        .attrs = snbep_uncore_pcu_formats_attr,
};

static const struct attribute_group snbep_uncore_qpi_format_group = {
        .name = "format",
        .attrs = snbep_uncore_qpi_formats_attr,
};

#define __SNBEP_UNCORE_MSR_OPS_COMMON_INIT()                    \
        .disable_box    = snbep_uncore_msr_disable_box,         \
        .enable_box     = snbep_uncore_msr_enable_box,          \
        .disable_event  = snbep_uncore_msr_disable_event,       \
        .enable_event   = snbep_uncore_msr_enable_event,        \
        .read_counter   = uncore_msr_read_counter

#define SNBEP_UNCORE_MSR_OPS_COMMON_INIT()                      \
        __SNBEP_UNCORE_MSR_OPS_COMMON_INIT(),                   \
        .init_box       = snbep_uncore_msr_init_box             \

static struct intel_uncore_ops snbep_uncore_msr_ops = {
        SNBEP_UNCORE_MSR_OPS_COMMON_INIT(),
};

#define SNBEP_UNCORE_PCI_OPS_COMMON_INIT()                      \
        .init_box       = snbep_uncore_pci_init_box,            \
        .disable_box    = snbep_uncore_pci_disable_box,         \
        .enable_box     = snbep_uncore_pci_enable_box,          \
        .disable_event  = snbep_uncore_pci_disable_event,       \
        .read_counter   = snbep_uncore_pci_read_counter

static struct intel_uncore_ops snbep_uncore_pci_ops = {
        SNBEP_UNCORE_PCI_OPS_COMMON_INIT(),
        .enable_event   = snbep_uncore_pci_enable_event,        \
};

static struct event_constraint snbep_uncore_cbox_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x01, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x02, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x04, 0x5, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x07, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x09, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x11, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x12, 0x13, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x1b, 0x1e, 0xc),
        UNCORE_EVENT_CONSTRAINT(0x1f, 0xe),
        UNCORE_EVENT_CONSTRAINT(0x21, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x23, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x31, 0x35, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x36, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x37, 0x39, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x3b, 0x1),
        EVENT_CONSTRAINT_END
};

static struct event_constraint snbep_uncore_r2pcie_constraints[] = {
        UNCORE_EVENT_CONSTRAINT_RANGE(0x10, 0x11, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x12, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x23, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x24, 0x26, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x32, 0x34, 0x3),
        EVENT_CONSTRAINT_END
};

static struct event_constraint snbep_uncore_r3qpi_constraints[] = {
        UNCORE_EVENT_CONSTRAINT_RANGE(0x10, 0x12, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x13, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x20, 0x26, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x28, 0x34, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x36, 0x39, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type snbep_uncore_ubox = {
        .name           = "ubox",
        .num_counters   = 2,
        .num_boxes      = 1,
        .perf_ctr_bits  = 44,
        .fixed_ctr_bits = 48,
        .perf_ctr       = SNBEP_U_MSR_PMON_CTR0,
        .event_ctl      = SNBEP_U_MSR_PMON_CTL0,
        .event_mask     = SNBEP_U_MSR_PMON_RAW_EVENT_MASK,
        .fixed_ctr      = SNBEP_U_MSR_PMON_UCLK_FIXED_CTR,
        .fixed_ctl      = SNBEP_U_MSR_PMON_UCLK_FIXED_CTL,
        .ops            = &snbep_uncore_msr_ops,
        .format_group   = &snbep_uncore_ubox_format_group,
};

static struct extra_reg snbep_uncore_cbox_extra_regs[] = {
        SNBEP_CBO_EVENT_EXTRA_REG(SNBEP_CBO_PMON_CTL_TID_EN,
                                  SNBEP_CBO_PMON_CTL_TID_EN, 0x1),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0334, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4334, 0xffff, 0x6),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0534, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4534, 0xffff, 0x6),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0934, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4934, 0xffff, 0x6),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4134, 0xffff, 0x6),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0135, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0335, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4135, 0xffff, 0xa),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4335, 0xffff, 0xa),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4435, 0xffff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4835, 0xffff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4a35, 0xffff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x5035, 0xffff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0136, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0336, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4136, 0xffff, 0xa),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4336, 0xffff, 0xa),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4436, 0xffff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4836, 0xffff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4a36, 0xffff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4037, 0x40ff, 0x2),
        EVENT_EXTRA_END
};

static void snbep_cbox_put_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct intel_uncore_extra_reg *er = &box->shared_regs[0];
        int i;

        if (uncore_box_is_fake(box))
                return;

        for (i = 0; i < 5; i++) {
                if (reg1->alloc & (0x1 << i))
                        atomic_sub(1 << (i * 6), &er->ref);
        }
        reg1->alloc = 0;
}

static struct event_constraint *
__snbep_cbox_get_constraint(struct intel_uncore_box *box, struct perf_event *event,
                            u64 (*cbox_filter_mask)(int fields))
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct intel_uncore_extra_reg *er = &box->shared_regs[0];
        int i, alloc = 0;
        unsigned long flags;
        u64 mask;

        if (reg1->idx == EXTRA_REG_NONE)
                return NULL;

        raw_spin_lock_irqsave(&er->lock, flags);
        for (i = 0; i < 5; i++) {
                if (!(reg1->idx & (0x1 << i)))
                        continue;
                if (!uncore_box_is_fake(box) && (reg1->alloc & (0x1 << i)))
                        continue;

                mask = cbox_filter_mask(0x1 << i);
                if (!__BITS_VALUE(atomic_read(&er->ref), i, 6) ||
                    !((reg1->config ^ er->config) & mask)) {
                        atomic_add(1 << (i * 6), &er->ref);
                        er->config &= ~mask;
                        er->config |= reg1->config & mask;
                        alloc |= (0x1 << i);
                } else {
                        break;
                }
        }
        raw_spin_unlock_irqrestore(&er->lock, flags);
        if (i < 5)
                goto fail;

        if (!uncore_box_is_fake(box))
                reg1->alloc |= alloc;

        return NULL;
fail:
        for (; i >= 0; i--) {
                if (alloc & (0x1 << i))
                        atomic_sub(1 << (i * 6), &er->ref);
        }
        return &uncore_constraint_empty;
}

static u64 snbep_cbox_filter_mask(int fields)
{
        u64 mask = 0;

        if (fields & 0x1)
                mask |= SNBEP_CB0_MSR_PMON_BOX_FILTER_TID;
        if (fields & 0x2)
                mask |= SNBEP_CB0_MSR_PMON_BOX_FILTER_NID;
        if (fields & 0x4)
                mask |= SNBEP_CB0_MSR_PMON_BOX_FILTER_STATE;
        if (fields & 0x8)
                mask |= SNBEP_CB0_MSR_PMON_BOX_FILTER_OPC;

        return mask;
}

static struct event_constraint *
snbep_cbox_get_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        return __snbep_cbox_get_constraint(box, event, snbep_cbox_filter_mask);
}

static int snbep_cbox_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct extra_reg *er;
        int idx = 0;

        for (er = snbep_uncore_cbox_extra_regs; er->msr; er++) {
                if (er->event != (event->hw.config & er->config_mask))
                        continue;
                idx |= er->idx;
        }

        if (idx) {
                reg1->reg = SNBEP_C0_MSR_PMON_BOX_FILTER +
                        SNBEP_CBO_MSR_OFFSET * box->pmu->pmu_idx;
                reg1->config = event->attr.config1 & snbep_cbox_filter_mask(idx);
                reg1->idx = idx;
        }
        return 0;
}

static struct intel_uncore_ops snbep_uncore_cbox_ops = {
        SNBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .hw_config              = snbep_cbox_hw_config,
        .get_constraint         = snbep_cbox_get_constraint,
        .put_constraint         = snbep_cbox_put_constraint,
};

static struct intel_uncore_type snbep_uncore_cbox = {
        .name                   = "cbox",
        .num_counters           = 4,
        .num_boxes              = 8,
        .perf_ctr_bits          = 44,
        .event_ctl              = SNBEP_C0_MSR_PMON_CTL0,
        .perf_ctr               = SNBEP_C0_MSR_PMON_CTR0,
        .event_mask             = SNBEP_CBO_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_C0_MSR_PMON_BOX_CTL,
        .msr_offset             = SNBEP_CBO_MSR_OFFSET,
        .num_shared_regs        = 1,
        .constraints            = snbep_uncore_cbox_constraints,
        .ops                    = &snbep_uncore_cbox_ops,
        .format_group           = &snbep_uncore_cbox_format_group,
};

static u64 snbep_pcu_alter_er(struct perf_event *event, int new_idx, bool modify)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;
        u64 config = reg1->config;

        if (new_idx > reg1->idx)
                config <<= 8 * (new_idx - reg1->idx);
        else
                config >>= 8 * (reg1->idx - new_idx);

        if (modify) {
                hwc->config += new_idx - reg1->idx;
                reg1->config = config;
                reg1->idx = new_idx;
        }
        return config;
}

static struct event_constraint *
snbep_pcu_get_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct intel_uncore_extra_reg *er = &box->shared_regs[0];
        unsigned long flags;
        int idx = reg1->idx;
        u64 mask, config1 = reg1->config;
        bool ok = false;

        if (reg1->idx == EXTRA_REG_NONE ||
            (!uncore_box_is_fake(box) && reg1->alloc))
                return NULL;
again:
        mask = 0xffULL << (idx * 8);
        raw_spin_lock_irqsave(&er->lock, flags);
        if (!__BITS_VALUE(atomic_read(&er->ref), idx, 8) ||
            !((config1 ^ er->config) & mask)) {
                atomic_add(1 << (idx * 8), &er->ref);
                er->config &= ~mask;
                er->config |= config1 & mask;
                ok = true;
        }
        raw_spin_unlock_irqrestore(&er->lock, flags);

        if (!ok) {
                idx = (idx + 1) % 4;
                if (idx != reg1->idx) {
                        config1 = snbep_pcu_alter_er(event, idx, false);
                        goto again;
                }
                return &uncore_constraint_empty;
        }

        if (!uncore_box_is_fake(box)) {
                if (idx != reg1->idx)
                        snbep_pcu_alter_er(event, idx, true);
                reg1->alloc = 1;
        }
        return NULL;
}

static void snbep_pcu_put_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct intel_uncore_extra_reg *er = &box->shared_regs[0];

        if (uncore_box_is_fake(box) || !reg1->alloc)
                return;

        atomic_sub(1 << (reg1->idx * 8), &er->ref);
        reg1->alloc = 0;
}

static int snbep_pcu_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;
        int ev_sel = hwc->config & SNBEP_PMON_CTL_EV_SEL_MASK;

        if (ev_sel >= 0xb && ev_sel <= 0xe) {
                reg1->reg = SNBEP_PCU_MSR_PMON_BOX_FILTER;
                reg1->idx = ev_sel - 0xb;
                reg1->config = event->attr.config1 & (0xff << (reg1->idx * 8));
        }
        return 0;
}

static struct intel_uncore_ops snbep_uncore_pcu_ops = {
        SNBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .hw_config              = snbep_pcu_hw_config,
        .get_constraint         = snbep_pcu_get_constraint,
        .put_constraint         = snbep_pcu_put_constraint,
};

static struct intel_uncore_type snbep_uncore_pcu = {
        .name                   = "pcu",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNBEP_PCU_MSR_PMON_CTR0,
        .event_ctl              = SNBEP_PCU_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PCU_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCU_MSR_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &snbep_uncore_pcu_ops,
        .format_group           = &snbep_uncore_pcu_format_group,
};

static struct intel_uncore_type *snbep_msr_uncores[] = {
        &snbep_uncore_ubox,
        &snbep_uncore_cbox,
        &snbep_uncore_pcu,
        NULL,
};

void snbep_uncore_cpu_init(void)
{
        if (snbep_uncore_cbox.num_boxes > topology_num_cores_per_package())
                snbep_uncore_cbox.num_boxes = topology_num_cores_per_package();
        uncore_msr_uncores = snbep_msr_uncores;
}

enum {
        SNBEP_PCI_QPI_PORT0_FILTER,
        SNBEP_PCI_QPI_PORT1_FILTER,
        BDX_PCI_QPI_PORT2_FILTER,
};

static int snbep_qpi_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;
        struct hw_perf_event_extra *reg2 = &hwc->branch_reg;

        if ((hwc->config & SNBEP_PMON_CTL_EV_SEL_MASK) == 0x38) {
                reg1->idx = 0;
                reg1->reg = SNBEP_Q_Py_PCI_PMON_PKT_MATCH0;
                reg1->config = event->attr.config1;
                reg2->reg = SNBEP_Q_Py_PCI_PMON_PKT_MASK0;
                reg2->config = event->attr.config2;
        }
        return 0;
}

static void snbep_qpi_enable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;
        struct hw_perf_event_extra *reg2 = &hwc->branch_reg;

        if (reg1->idx != EXTRA_REG_NONE) {
                int idx = box->pmu->pmu_idx + SNBEP_PCI_QPI_PORT0_FILTER;
                int die = box->dieid;
                struct pci_dev *filter_pdev = uncore_extra_pci_dev[die].dev[idx];

                if (filter_pdev) {
                        pci_write_config_dword(filter_pdev, reg1->reg,
                                                (u32)reg1->config);
                        pci_write_config_dword(filter_pdev, reg1->reg + 4,
                                                (u32)(reg1->config >> 32));
                        pci_write_config_dword(filter_pdev, reg2->reg,
                                                (u32)reg2->config);
                        pci_write_config_dword(filter_pdev, reg2->reg + 4,
                                                (u32)(reg2->config >> 32));
                }
        }

        pci_write_config_dword(pdev, hwc->config_base, hwc->config | SNBEP_PMON_CTL_EN);
}

static struct intel_uncore_ops snbep_uncore_qpi_ops = {
        SNBEP_UNCORE_PCI_OPS_COMMON_INIT(),
        .enable_event           = snbep_qpi_enable_event,
        .hw_config              = snbep_qpi_hw_config,
        .get_constraint         = uncore_get_constraint,
        .put_constraint         = uncore_put_constraint,
};

#define SNBEP_UNCORE_PCI_COMMON_INIT()                          \
        .perf_ctr       = SNBEP_PCI_PMON_CTR0,                  \
        .event_ctl      = SNBEP_PCI_PMON_CTL0,                  \
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,            \
        .box_ctl        = SNBEP_PCI_PMON_BOX_CTL,               \
        .ops            = &snbep_uncore_pci_ops,                \
        .format_group   = &snbep_uncore_format_group

static struct intel_uncore_type snbep_uncore_ha = {
        .name           = "ha",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 48,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct intel_uncore_type snbep_uncore_imc = {
        .name           = "imc",
        .num_counters   = 4,
        .num_boxes      = 4,
        .perf_ctr_bits  = 48,
        .fixed_ctr_bits = 48,
        .fixed_ctr      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTR,
        .fixed_ctl      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTL,
        .event_descs    = snbep_uncore_imc_events,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct intel_uncore_type snbep_uncore_qpi = {
        .name                   = "qpi",
        .num_counters           = 4,
        .num_boxes              = 2,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNBEP_PCI_PMON_CTR0,
        .event_ctl              = SNBEP_PCI_PMON_CTL0,
        .event_mask             = SNBEP_QPI_PCI_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCI_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &snbep_uncore_qpi_ops,
        .event_descs            = snbep_uncore_qpi_events,
        .format_group           = &snbep_uncore_qpi_format_group,
};


static struct intel_uncore_type snbep_uncore_r2pcie = {
        .name           = "r2pcie",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 44,
        .constraints    = snbep_uncore_r2pcie_constraints,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct intel_uncore_type snbep_uncore_r3qpi = {
        .name           = "r3qpi",
        .num_counters   = 3,
        .num_boxes      = 2,
        .perf_ctr_bits  = 44,
        .constraints    = snbep_uncore_r3qpi_constraints,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

enum {
        SNBEP_PCI_UNCORE_HA,
        SNBEP_PCI_UNCORE_IMC,
        SNBEP_PCI_UNCORE_QPI,
        SNBEP_PCI_UNCORE_R2PCIE,
        SNBEP_PCI_UNCORE_R3QPI,
};

static struct intel_uncore_type *snbep_pci_uncores[] = {
        [SNBEP_PCI_UNCORE_HA]           = &snbep_uncore_ha,
        [SNBEP_PCI_UNCORE_IMC]          = &snbep_uncore_imc,
        [SNBEP_PCI_UNCORE_QPI]          = &snbep_uncore_qpi,
        [SNBEP_PCI_UNCORE_R2PCIE]       = &snbep_uncore_r2pcie,
        [SNBEP_PCI_UNCORE_R3QPI]        = &snbep_uncore_r3qpi,
        NULL,
};

static const struct pci_device_id snbep_uncore_pci_ids[] = {
        { /* Home Agent */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_HA),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_HA, 0),
        },
        { /* MC Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_IMC0),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_IMC, 0),
        },
        { /* MC Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_IMC1),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_IMC, 1),
        },
        { /* MC Channel 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_IMC2),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_IMC, 2),
        },
        { /* MC Channel 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_IMC3),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_IMC, 3),
        },
        { /* QPI Port 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_QPI0),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_QPI, 0),
        },
        { /* QPI Port 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_QPI1),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_QPI, 1),
        },
        { /* R2PCIe */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_R2PCIE),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_R2PCIE, 0),
        },
        { /* R3QPI Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_R3QPI0),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_R3QPI, 0),
        },
        { /* R3QPI Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_UNC_R3QPI1),
                .driver_data = UNCORE_PCI_DEV_DATA(SNBEP_PCI_UNCORE_R3QPI, 1),
        },
        { /* QPI Port 0 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3c86),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT0_FILTER),
        },
        { /* QPI Port 0 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3c96),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT1_FILTER),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver snbep_uncore_pci_driver = {
        .name           = "snbep_uncore",
        .id_table       = snbep_uncore_pci_ids,
};

#define NODE_ID_MASK    0x7

/* Each three bits from 0 to 23 of GIDNIDMAP register correspond Node ID. */
#define GIDNIDMAP(config, id)   (((config) >> (3 * (id))) & 0x7)

static int upi_nodeid_groupid(struct pci_dev *ubox_dev, int nodeid_loc, int idmap_loc,
                              int *nodeid, int *groupid)
{
        int ret;

        /* get the Node ID of the local register */
        ret = pci_read_config_dword(ubox_dev, nodeid_loc, nodeid);
        if (ret)
                goto err;

        *nodeid = *nodeid & NODE_ID_MASK;
        /* get the Node ID mapping */
        ret = pci_read_config_dword(ubox_dev, idmap_loc, groupid);
        if (ret)
                goto err;
err:
        return ret;
}

static int topology_gidnid_map(int nodeid, u32 gidnid)
{
        int i, die_id = -1;

        /*
         * every three bits in the Node ID mapping register maps
         * to a particular node.
         */
        for (i = 0; i < 8; i++) {
                if (nodeid == GIDNIDMAP(gidnid, i)) {
                        if (topology_max_dies_per_package() > 1)
                                die_id = i;
                        else
                                die_id = topology_phys_to_logical_pkg(i);
                        if (die_id < 0)
                                die_id = -ENODEV;
                        break;
                }
        }

        return die_id;
}

/*
 * build pci bus to socket mapping
 */
static int snbep_pci2phy_map_init(int devid, int nodeid_loc, int idmap_loc, bool reverse)
{
        struct pci_dev *ubox_dev = NULL;
        int i, bus, nodeid, segment, die_id;
        struct pci2phy_map *map;
        int err = 0;
        u32 config = 0;

        while (1) {
                /* find the UBOX device */
                ubox_dev = pci_get_device(PCI_VENDOR_ID_INTEL, devid, ubox_dev);
                if (!ubox_dev)
                        break;
                bus = ubox_dev->bus->number;
                /*
                 * The nodeid and idmap registers only contain enough
                 * information to handle 8 nodes.  On systems with more
                 * than 8 nodes, we need to rely on NUMA information,
                 * filled in from BIOS supplied information, to determine
                 * the topology.
                 */
                if (nr_node_ids <= 8) {
                        err = upi_nodeid_groupid(ubox_dev, nodeid_loc, idmap_loc,
                                                 &nodeid, &config);
                        if (err)
                                break;

                        segment = pci_domain_nr(ubox_dev->bus);
                        raw_spin_lock(&pci2phy_map_lock);
                        map = __find_pci2phy_map(segment);
                        if (!map) {
                                raw_spin_unlock(&pci2phy_map_lock);
                                err = -ENOMEM;
                                break;
                        }

                        map->pbus_to_dieid[bus] = topology_gidnid_map(nodeid, config);
                        raw_spin_unlock(&pci2phy_map_lock);
                } else {
                        segment = pci_domain_nr(ubox_dev->bus);
                        raw_spin_lock(&pci2phy_map_lock);
                        map = __find_pci2phy_map(segment);
                        if (!map) {
                                raw_spin_unlock(&pci2phy_map_lock);
                                err = -ENOMEM;
                                break;
                        }

                        map->pbus_to_dieid[bus] = die_id = uncore_device_to_die(ubox_dev);

                        raw_spin_unlock(&pci2phy_map_lock);

                        if (WARN_ON_ONCE(die_id == -1)) {
                                err = -EINVAL;
                                break;
                        }
                }
        }

        if (!err) {
                /*
                 * For PCI bus with no UBOX device, find the next bus
                 * that has UBOX device and use its mapping.
                 */
                raw_spin_lock(&pci2phy_map_lock);
                list_for_each_entry(map, &pci2phy_map_head, list) {
                        i = -1;
                        if (reverse) {
                                for (bus = 255; bus >= 0; bus--) {
                                        if (map->pbus_to_dieid[bus] != -1)
                                                i = map->pbus_to_dieid[bus];
                                        else
                                                map->pbus_to_dieid[bus] = i;
                                }
                        } else {
                                for (bus = 0; bus <= 255; bus++) {
                                        if (map->pbus_to_dieid[bus] != -1)
                                                i = map->pbus_to_dieid[bus];
                                        else
                                                map->pbus_to_dieid[bus] = i;
                                }
                        }
                }
                raw_spin_unlock(&pci2phy_map_lock);
        }

        pci_dev_put(ubox_dev);

        return pcibios_err_to_errno(err);
}

int snbep_uncore_pci_init(void)
{
        int ret = snbep_pci2phy_map_init(0x3ce0, SNBEP_CPUNODEID, SNBEP_GIDNIDMAP, true);
        if (ret)
                return ret;
        uncore_pci_uncores = snbep_pci_uncores;
        uncore_pci_driver = &snbep_uncore_pci_driver;
        return 0;
}
/* end of Sandy Bridge-EP uncore support */

/* IvyTown uncore support */
static void ivbep_uncore_msr_init_box(struct intel_uncore_box *box)
{
        unsigned msr = uncore_msr_box_ctl(box);
        if (msr)
                wrmsrq(msr, IVBEP_PMON_BOX_CTL_INT);
}

static void ivbep_uncore_pci_init_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;

        pci_write_config_dword(pdev, SNBEP_PCI_PMON_BOX_CTL, IVBEP_PMON_BOX_CTL_INT);
}

#define IVBEP_UNCORE_MSR_OPS_COMMON_INIT()                      \
        .init_box       = ivbep_uncore_msr_init_box,            \
        .disable_box    = snbep_uncore_msr_disable_box,         \
        .enable_box     = snbep_uncore_msr_enable_box,          \
        .disable_event  = snbep_uncore_msr_disable_event,       \
        .enable_event   = snbep_uncore_msr_enable_event,        \
        .read_counter   = uncore_msr_read_counter

static struct intel_uncore_ops ivbep_uncore_msr_ops = {
        IVBEP_UNCORE_MSR_OPS_COMMON_INIT(),
};

static struct intel_uncore_ops ivbep_uncore_pci_ops = {
        .init_box       = ivbep_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = snbep_uncore_pci_disable_event,
        .enable_event   = snbep_uncore_pci_enable_event,
        .read_counter   = snbep_uncore_pci_read_counter,
};

#define IVBEP_UNCORE_PCI_COMMON_INIT()                          \
        .perf_ctr       = SNBEP_PCI_PMON_CTR0,                  \
        .event_ctl      = SNBEP_PCI_PMON_CTL0,                  \
        .event_mask     = IVBEP_PMON_RAW_EVENT_MASK,            \
        .box_ctl        = SNBEP_PCI_PMON_BOX_CTL,               \
        .ops            = &ivbep_uncore_pci_ops,                        \
        .format_group   = &ivbep_uncore_format_group

static struct attribute *ivbep_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static struct attribute *ivbep_uncore_ubox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh5.attr,
        NULL,
};

static struct attribute *ivbep_uncore_cbox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_thresh8.attr,
        &format_attr_filter_tid.attr,
        &format_attr_filter_link.attr,
        &format_attr_filter_state2.attr,
        &format_attr_filter_nid2.attr,
        &format_attr_filter_opc2.attr,
        &format_attr_filter_nc.attr,
        &format_attr_filter_c6.attr,
        &format_attr_filter_isoc.attr,
        NULL,
};

static struct attribute *ivbep_uncore_pcu_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_occ_sel.attr,
        &format_attr_edge.attr,
        &format_attr_thresh5.attr,
        &format_attr_occ_invert.attr,
        &format_attr_occ_edge.attr,
        &format_attr_filter_band0.attr,
        &format_attr_filter_band1.attr,
        &format_attr_filter_band2.attr,
        &format_attr_filter_band3.attr,
        NULL,
};

static struct attribute *ivbep_uncore_qpi_formats_attr[] = {
        &format_attr_event_ext.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_thresh8.attr,
        &format_attr_match_rds.attr,
        &format_attr_match_rnid30.attr,
        &format_attr_match_rnid4.attr,
        &format_attr_match_dnid.attr,
        &format_attr_match_mc.attr,
        &format_attr_match_opc.attr,
        &format_attr_match_vnw.attr,
        &format_attr_match0.attr,
        &format_attr_match1.attr,
        &format_attr_mask_rds.attr,
        &format_attr_mask_rnid30.attr,
        &format_attr_mask_rnid4.attr,
        &format_attr_mask_dnid.attr,
        &format_attr_mask_mc.attr,
        &format_attr_mask_opc.attr,
        &format_attr_mask_vnw.attr,
        &format_attr_mask0.attr,
        &format_attr_mask1.attr,
        NULL,
};

static const struct attribute_group ivbep_uncore_format_group = {
        .name = "format",
        .attrs = ivbep_uncore_formats_attr,
};

static const struct attribute_group ivbep_uncore_ubox_format_group = {
        .name = "format",
        .attrs = ivbep_uncore_ubox_formats_attr,
};

static const struct attribute_group ivbep_uncore_cbox_format_group = {
        .name = "format",
        .attrs = ivbep_uncore_cbox_formats_attr,
};

static const struct attribute_group ivbep_uncore_pcu_format_group = {
        .name = "format",
        .attrs = ivbep_uncore_pcu_formats_attr,
};

static const struct attribute_group ivbep_uncore_qpi_format_group = {
        .name = "format",
        .attrs = ivbep_uncore_qpi_formats_attr,
};

static struct intel_uncore_type ivbep_uncore_ubox = {
        .name           = "ubox",
        .num_counters   = 2,
        .num_boxes      = 1,
        .perf_ctr_bits  = 44,
        .fixed_ctr_bits = 48,
        .perf_ctr       = SNBEP_U_MSR_PMON_CTR0,
        .event_ctl      = SNBEP_U_MSR_PMON_CTL0,
        .event_mask     = IVBEP_U_MSR_PMON_RAW_EVENT_MASK,
        .fixed_ctr      = SNBEP_U_MSR_PMON_UCLK_FIXED_CTR,
        .fixed_ctl      = SNBEP_U_MSR_PMON_UCLK_FIXED_CTL,
        .ops            = &ivbep_uncore_msr_ops,
        .format_group   = &ivbep_uncore_ubox_format_group,
};

static struct extra_reg ivbep_uncore_cbox_extra_regs[] = {
        SNBEP_CBO_EVENT_EXTRA_REG(SNBEP_CBO_PMON_CTL_TID_EN,
                                  SNBEP_CBO_PMON_CTL_TID_EN, 0x1),
        SNBEP_CBO_EVENT_EXTRA_REG(0x1031, 0x10ff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x1134, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4134, 0xffff, 0xc),
        SNBEP_CBO_EVENT_EXTRA_REG(0x5134, 0xffff, 0xc),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0334, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4334, 0xffff, 0xc),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0534, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4534, 0xffff, 0xc),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0934, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4934, 0xffff, 0xc),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0135, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0335, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2135, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2335, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4135, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4335, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4435, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4835, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4a35, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x5035, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8135, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8335, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0136, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0336, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2136, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2336, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4136, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4336, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4436, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4836, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4a36, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x5036, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8136, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8336, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4037, 0x40ff, 0x8),
        EVENT_EXTRA_END
};

static u64 ivbep_cbox_filter_mask(int fields)
{
        u64 mask = 0;

        if (fields & 0x1)
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_TID;
        if (fields & 0x2)
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_LINK;
        if (fields & 0x4)
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_STATE;
        if (fields & 0x8)
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_NID;
        if (fields & 0x10) {
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_OPC;
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_NC;
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_C6;
                mask |= IVBEP_CB0_MSR_PMON_BOX_FILTER_ISOC;
        }

        return mask;
}

static struct event_constraint *
ivbep_cbox_get_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        return __snbep_cbox_get_constraint(box, event, ivbep_cbox_filter_mask);
}

static int ivbep_cbox_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct extra_reg *er;
        int idx = 0;

        for (er = ivbep_uncore_cbox_extra_regs; er->msr; er++) {
                if (er->event != (event->hw.config & er->config_mask))
                        continue;
                idx |= er->idx;
        }

        if (idx) {
                reg1->reg = SNBEP_C0_MSR_PMON_BOX_FILTER +
                        SNBEP_CBO_MSR_OFFSET * box->pmu->pmu_idx;
                reg1->config = event->attr.config1 & ivbep_cbox_filter_mask(idx);
                reg1->idx = idx;
        }
        return 0;
}

static void ivbep_cbox_enable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;

        if (reg1->idx != EXTRA_REG_NONE) {
                u64 filter = uncore_shared_reg_config(box, 0);
                wrmsrq(reg1->reg, filter & 0xffffffff);
                wrmsrq(reg1->reg + 6, filter >> 32);
        }

        wrmsrq(hwc->config_base, hwc->config | SNBEP_PMON_CTL_EN);
}

static struct intel_uncore_ops ivbep_uncore_cbox_ops = {
        .init_box               = ivbep_uncore_msr_init_box,
        .disable_box            = snbep_uncore_msr_disable_box,
        .enable_box             = snbep_uncore_msr_enable_box,
        .disable_event          = snbep_uncore_msr_disable_event,
        .enable_event           = ivbep_cbox_enable_event,
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = ivbep_cbox_hw_config,
        .get_constraint         = ivbep_cbox_get_constraint,
        .put_constraint         = snbep_cbox_put_constraint,
};

static struct intel_uncore_type ivbep_uncore_cbox = {
        .name                   = "cbox",
        .num_counters           = 4,
        .num_boxes              = 15,
        .perf_ctr_bits          = 44,
        .event_ctl              = SNBEP_C0_MSR_PMON_CTL0,
        .perf_ctr               = SNBEP_C0_MSR_PMON_CTR0,
        .event_mask             = IVBEP_CBO_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_C0_MSR_PMON_BOX_CTL,
        .msr_offset             = SNBEP_CBO_MSR_OFFSET,
        .num_shared_regs        = 1,
        .constraints            = snbep_uncore_cbox_constraints,
        .ops                    = &ivbep_uncore_cbox_ops,
        .format_group           = &ivbep_uncore_cbox_format_group,
};

static struct intel_uncore_ops ivbep_uncore_pcu_ops = {
        IVBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .hw_config              = snbep_pcu_hw_config,
        .get_constraint         = snbep_pcu_get_constraint,
        .put_constraint         = snbep_pcu_put_constraint,
};

static struct intel_uncore_type ivbep_uncore_pcu = {
        .name                   = "pcu",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNBEP_PCU_MSR_PMON_CTR0,
        .event_ctl              = SNBEP_PCU_MSR_PMON_CTL0,
        .event_mask             = IVBEP_PCU_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCU_MSR_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &ivbep_uncore_pcu_ops,
        .format_group           = &ivbep_uncore_pcu_format_group,
};

static struct intel_uncore_type *ivbep_msr_uncores[] = {
        &ivbep_uncore_ubox,
        &ivbep_uncore_cbox,
        &ivbep_uncore_pcu,
        NULL,
};

void ivbep_uncore_cpu_init(void)
{
        if (ivbep_uncore_cbox.num_boxes > topology_num_cores_per_package())
                ivbep_uncore_cbox.num_boxes = topology_num_cores_per_package();
        uncore_msr_uncores = ivbep_msr_uncores;
}

static struct intel_uncore_type ivbep_uncore_ha = {
        .name           = "ha",
        .num_counters   = 4,
        .num_boxes      = 2,
        .perf_ctr_bits  = 48,
        IVBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct intel_uncore_type ivbep_uncore_imc = {
        .name           = "imc",
        .num_counters   = 4,
        .num_boxes      = 8,
        .perf_ctr_bits  = 48,
        .fixed_ctr_bits = 48,
        .fixed_ctr      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTR,
        .fixed_ctl      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTL,
        .event_descs    = snbep_uncore_imc_events,
        IVBEP_UNCORE_PCI_COMMON_INIT(),
};

/* registers in IRP boxes are not properly aligned */
static unsigned ivbep_uncore_irp_ctls[] = {0xd8, 0xdc, 0xe0, 0xe4};
static unsigned ivbep_uncore_irp_ctrs[] = {0xa0, 0xb0, 0xb8, 0xc0};

static void ivbep_uncore_irp_enable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;

        pci_write_config_dword(pdev, ivbep_uncore_irp_ctls[hwc->idx],
                               hwc->config | SNBEP_PMON_CTL_EN);
}

static void ivbep_uncore_irp_disable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;

        pci_write_config_dword(pdev, ivbep_uncore_irp_ctls[hwc->idx], hwc->config);
}

static u64 ivbep_uncore_irp_read_counter(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;
        u64 count = 0;

        pci_read_config_dword(pdev, ivbep_uncore_irp_ctrs[hwc->idx], (u32 *)&count);
        pci_read_config_dword(pdev, ivbep_uncore_irp_ctrs[hwc->idx] + 4, (u32 *)&count + 1);

        return count;
}

static struct intel_uncore_ops ivbep_uncore_irp_ops = {
        .init_box       = ivbep_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = ivbep_uncore_irp_disable_event,
        .enable_event   = ivbep_uncore_irp_enable_event,
        .read_counter   = ivbep_uncore_irp_read_counter,
};

static struct intel_uncore_type ivbep_uncore_irp = {
        .name                   = "irp",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .event_mask             = IVBEP_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCI_PMON_BOX_CTL,
        .ops                    = &ivbep_uncore_irp_ops,
        .format_group           = &ivbep_uncore_format_group,
};

static struct intel_uncore_ops ivbep_uncore_qpi_ops = {
        .init_box       = ivbep_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = snbep_uncore_pci_disable_event,
        .enable_event   = snbep_qpi_enable_event,
        .read_counter   = snbep_uncore_pci_read_counter,
        .hw_config      = snbep_qpi_hw_config,
        .get_constraint = uncore_get_constraint,
        .put_constraint = uncore_put_constraint,
};

static struct intel_uncore_type ivbep_uncore_qpi = {
        .name                   = "qpi",
        .num_counters           = 4,
        .num_boxes              = 3,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNBEP_PCI_PMON_CTR0,
        .event_ctl              = SNBEP_PCI_PMON_CTL0,
        .event_mask             = IVBEP_QPI_PCI_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCI_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &ivbep_uncore_qpi_ops,
        .format_group           = &ivbep_uncore_qpi_format_group,
};

static struct intel_uncore_type ivbep_uncore_r2pcie = {
        .name           = "r2pcie",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 44,
        .constraints    = snbep_uncore_r2pcie_constraints,
        IVBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct intel_uncore_type ivbep_uncore_r3qpi = {
        .name           = "r3qpi",
        .num_counters   = 3,
        .num_boxes      = 2,
        .perf_ctr_bits  = 44,
        .constraints    = snbep_uncore_r3qpi_constraints,
        IVBEP_UNCORE_PCI_COMMON_INIT(),
};

enum {
        IVBEP_PCI_UNCORE_HA,
        IVBEP_PCI_UNCORE_IMC,
        IVBEP_PCI_UNCORE_IRP,
        IVBEP_PCI_UNCORE_QPI,
        IVBEP_PCI_UNCORE_R2PCIE,
        IVBEP_PCI_UNCORE_R3QPI,
};

static struct intel_uncore_type *ivbep_pci_uncores[] = {
        [IVBEP_PCI_UNCORE_HA]   = &ivbep_uncore_ha,
        [IVBEP_PCI_UNCORE_IMC]  = &ivbep_uncore_imc,
        [IVBEP_PCI_UNCORE_IRP]  = &ivbep_uncore_irp,
        [IVBEP_PCI_UNCORE_QPI]  = &ivbep_uncore_qpi,
        [IVBEP_PCI_UNCORE_R2PCIE]       = &ivbep_uncore_r2pcie,
        [IVBEP_PCI_UNCORE_R3QPI]        = &ivbep_uncore_r3qpi,
        NULL,
};

static const struct pci_device_id ivbep_uncore_pci_ids[] = {
        { /* Home Agent 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe30),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_HA, 0),
        },
        { /* Home Agent 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe38),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_HA, 1),
        },
        { /* MC0 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xeb4),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 0),
        },
        { /* MC0 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xeb5),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 1),
        },
        { /* MC0 Channel 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xeb0),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 2),
        },
        { /* MC0 Channel 4 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xeb1),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 3),
        },
        { /* MC1 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xef4),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 4),
        },
        { /* MC1 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xef5),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 5),
        },
        { /* MC1 Channel 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xef0),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 6),
        },
        { /* MC1 Channel 4 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xef1),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IMC, 7),
        },
        { /* IRP */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe39),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_IRP, 0),
        },
        { /* QPI0 Port 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe32),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_QPI, 0),
        },
        { /* QPI0 Port 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe33),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_QPI, 1),
        },
        { /* QPI1 Port 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe3a),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_QPI, 2),
        },
        { /* R2PCIe */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe34),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_R2PCIE, 0),
        },
        { /* R3QPI0 Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe36),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_R3QPI, 0),
        },
        { /* R3QPI0 Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe37),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_R3QPI, 1),
        },
        { /* R3QPI1 Link 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe3e),
                .driver_data = UNCORE_PCI_DEV_DATA(IVBEP_PCI_UNCORE_R3QPI, 2),
        },
        { /* QPI Port 0 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe86),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT0_FILTER),
        },
        { /* QPI Port 0 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0xe96),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT1_FILTER),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver ivbep_uncore_pci_driver = {
        .name           = "ivbep_uncore",
        .id_table       = ivbep_uncore_pci_ids,
};

int ivbep_uncore_pci_init(void)
{
        int ret = snbep_pci2phy_map_init(0x0e1e, SNBEP_CPUNODEID, SNBEP_GIDNIDMAP, true);
        if (ret)
                return ret;
        uncore_pci_uncores = ivbep_pci_uncores;
        uncore_pci_driver = &ivbep_uncore_pci_driver;
        return 0;
}
/* end of IvyTown uncore support */

/* KNL uncore support */
static struct attribute *knl_uncore_ubox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_inv.attr,
        &format_attr_thresh5.attr,
        NULL,
};

static const struct attribute_group knl_uncore_ubox_format_group = {
        .name = "format",
        .attrs = knl_uncore_ubox_formats_attr,
};

static struct intel_uncore_type knl_uncore_ubox = {
        .name                   = "ubox",
        .num_counters           = 2,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = HSWEP_U_MSR_PMON_CTR0,
        .event_ctl              = HSWEP_U_MSR_PMON_CTL0,
        .event_mask             = KNL_U_MSR_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTR,
        .fixed_ctl              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTL,
        .ops                    = &snbep_uncore_msr_ops,
        .format_group           = &knl_uncore_ubox_format_group,
};

static struct attribute *knl_uncore_cha_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_qor.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_filter_tid4.attr,
        &format_attr_filter_link3.attr,
        &format_attr_filter_state4.attr,
        &format_attr_filter_local.attr,
        &format_attr_filter_all_op.attr,
        &format_attr_filter_nnm.attr,
        &format_attr_filter_opc3.attr,
        &format_attr_filter_nc.attr,
        &format_attr_filter_isoc.attr,
        NULL,
};

static const struct attribute_group knl_uncore_cha_format_group = {
        .name = "format",
        .attrs = knl_uncore_cha_formats_attr,
};

static struct event_constraint knl_uncore_cha_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x11, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x1f, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x36, 0x1),
        EVENT_CONSTRAINT_END
};

static struct extra_reg knl_uncore_cha_extra_regs[] = {
        SNBEP_CBO_EVENT_EXTRA_REG(SNBEP_CBO_PMON_CTL_TID_EN,
                                  SNBEP_CBO_PMON_CTL_TID_EN, 0x1),
        SNBEP_CBO_EVENT_EXTRA_REG(0x3d, 0xff, 0x2),
        SNBEP_CBO_EVENT_EXTRA_REG(0x35, 0xff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x36, 0xff, 0x4),
        EVENT_EXTRA_END
};

static u64 knl_cha_filter_mask(int fields)
{
        u64 mask = 0;

        if (fields & 0x1)
                mask |= KNL_CHA_MSR_PMON_BOX_FILTER_TID;
        if (fields & 0x2)
                mask |= KNL_CHA_MSR_PMON_BOX_FILTER_STATE;
        if (fields & 0x4)
                mask |= KNL_CHA_MSR_PMON_BOX_FILTER_OP;
        return mask;
}

static struct event_constraint *
knl_cha_get_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        return __snbep_cbox_get_constraint(box, event, knl_cha_filter_mask);
}

static int knl_cha_hw_config(struct intel_uncore_box *box,
                             struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct extra_reg *er;
        int idx = 0;

        for (er = knl_uncore_cha_extra_regs; er->msr; er++) {
                if (er->event != (event->hw.config & er->config_mask))
                        continue;
                idx |= er->idx;
        }

        if (idx) {
                reg1->reg = HSWEP_C0_MSR_PMON_BOX_FILTER0 +
                            KNL_CHA_MSR_OFFSET * box->pmu->pmu_idx;
                reg1->config = event->attr.config1 & knl_cha_filter_mask(idx);

                reg1->config |= KNL_CHA_MSR_PMON_BOX_FILTER_REMOTE_NODE;
                reg1->config |= KNL_CHA_MSR_PMON_BOX_FILTER_LOCAL_NODE;
                reg1->config |= KNL_CHA_MSR_PMON_BOX_FILTER_NNC;
                reg1->idx = idx;
        }
        return 0;
}

static void hswep_cbox_enable_event(struct intel_uncore_box *box,
                                    struct perf_event *event);

static struct intel_uncore_ops knl_uncore_cha_ops = {
        .init_box               = snbep_uncore_msr_init_box,
        .disable_box            = snbep_uncore_msr_disable_box,
        .enable_box             = snbep_uncore_msr_enable_box,
        .disable_event          = snbep_uncore_msr_disable_event,
        .enable_event           = hswep_cbox_enable_event,
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = knl_cha_hw_config,
        .get_constraint         = knl_cha_get_constraint,
        .put_constraint         = snbep_cbox_put_constraint,
};

static struct intel_uncore_type knl_uncore_cha = {
        .name                   = "cha",
        .num_counters           = 4,
        .num_boxes              = 38,
        .perf_ctr_bits          = 48,
        .event_ctl              = HSWEP_C0_MSR_PMON_CTL0,
        .perf_ctr               = HSWEP_C0_MSR_PMON_CTR0,
        .event_mask             = KNL_CHA_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_C0_MSR_PMON_BOX_CTL,
        .msr_offset             = KNL_CHA_MSR_OFFSET,
        .num_shared_regs        = 1,
        .constraints            = knl_uncore_cha_constraints,
        .ops                    = &knl_uncore_cha_ops,
        .format_group           = &knl_uncore_cha_format_group,
};

static struct attribute *knl_uncore_pcu_formats_attr[] = {
        &format_attr_event2.attr,
        &format_attr_use_occ_ctr.attr,
        &format_attr_occ_sel.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_inv.attr,
        &format_attr_thresh6.attr,
        &format_attr_occ_invert.attr,
        &format_attr_occ_edge_det.attr,
        NULL,
};

static const struct attribute_group knl_uncore_pcu_format_group = {
        .name = "format",
        .attrs = knl_uncore_pcu_formats_attr,
};

static struct intel_uncore_type knl_uncore_pcu = {
        .name                   = "pcu",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .perf_ctr               = HSWEP_PCU_MSR_PMON_CTR0,
        .event_ctl              = HSWEP_PCU_MSR_PMON_CTL0,
        .event_mask             = KNL_PCU_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_PCU_MSR_PMON_BOX_CTL,
        .ops                    = &snbep_uncore_msr_ops,
        .format_group           = &knl_uncore_pcu_format_group,
};

static struct intel_uncore_type *knl_msr_uncores[] = {
        &knl_uncore_ubox,
        &knl_uncore_cha,
        &knl_uncore_pcu,
        NULL,
};

void knl_uncore_cpu_init(void)
{
        uncore_msr_uncores = knl_msr_uncores;
}

static void knl_uncore_imc_enable_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;
        int box_ctl = uncore_pci_box_ctl(box);

        pci_write_config_dword(pdev, box_ctl, 0);
}

static void knl_uncore_imc_enable_event(struct intel_uncore_box *box,
                                        struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;

        if ((event->attr.config & SNBEP_PMON_CTL_EV_SEL_MASK)
                                                        == UNCORE_FIXED_EVENT)
                pci_write_config_dword(pdev, hwc->config_base,
                                       hwc->config | KNL_PMON_FIXED_CTL_EN);
        else
                pci_write_config_dword(pdev, hwc->config_base,
                                       hwc->config | SNBEP_PMON_CTL_EN);
}

static struct intel_uncore_ops knl_uncore_imc_ops = {
        .init_box       = snbep_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = knl_uncore_imc_enable_box,
        .read_counter   = snbep_uncore_pci_read_counter,
        .enable_event   = knl_uncore_imc_enable_event,
        .disable_event  = snbep_uncore_pci_disable_event,
};

static struct intel_uncore_type knl_uncore_imc_uclk = {
        .name                   = "imc_uclk",
        .num_counters           = 4,
        .num_boxes              = 2,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = KNL_UCLK_MSR_PMON_CTR0_LOW,
        .event_ctl              = KNL_UCLK_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = KNL_UCLK_MSR_PMON_UCLK_FIXED_LOW,
        .fixed_ctl              = KNL_UCLK_MSR_PMON_UCLK_FIXED_CTL,
        .box_ctl                = KNL_UCLK_MSR_PMON_BOX_CTL,
        .ops                    = &knl_uncore_imc_ops,
        .format_group           = &snbep_uncore_format_group,
};

static struct intel_uncore_type knl_uncore_imc_dclk = {
        .name                   = "imc",
        .num_counters           = 4,
        .num_boxes              = 6,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = KNL_MC0_CH0_MSR_PMON_CTR0_LOW,
        .event_ctl              = KNL_MC0_CH0_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = KNL_MC0_CH0_MSR_PMON_FIXED_LOW,
        .fixed_ctl              = KNL_MC0_CH0_MSR_PMON_FIXED_CTL,
        .box_ctl                = KNL_MC0_CH0_MSR_PMON_BOX_CTL,
        .ops                    = &knl_uncore_imc_ops,
        .format_group           = &snbep_uncore_format_group,
};

static struct intel_uncore_type knl_uncore_edc_uclk = {
        .name                   = "edc_uclk",
        .num_counters           = 4,
        .num_boxes              = 8,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = KNL_UCLK_MSR_PMON_CTR0_LOW,
        .event_ctl              = KNL_UCLK_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = KNL_UCLK_MSR_PMON_UCLK_FIXED_LOW,
        .fixed_ctl              = KNL_UCLK_MSR_PMON_UCLK_FIXED_CTL,
        .box_ctl                = KNL_UCLK_MSR_PMON_BOX_CTL,
        .ops                    = &knl_uncore_imc_ops,
        .format_group           = &snbep_uncore_format_group,
};

static struct intel_uncore_type knl_uncore_edc_eclk = {
        .name                   = "edc_eclk",
        .num_counters           = 4,
        .num_boxes              = 8,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = KNL_EDC0_ECLK_MSR_PMON_CTR0_LOW,
        .event_ctl              = KNL_EDC0_ECLK_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = KNL_EDC0_ECLK_MSR_PMON_ECLK_FIXED_LOW,
        .fixed_ctl              = KNL_EDC0_ECLK_MSR_PMON_ECLK_FIXED_CTL,
        .box_ctl                = KNL_EDC0_ECLK_MSR_PMON_BOX_CTL,
        .ops                    = &knl_uncore_imc_ops,
        .format_group           = &snbep_uncore_format_group,
};

static struct event_constraint knl_uncore_m2pcie_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x23, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type knl_uncore_m2pcie = {
        .name           = "m2pcie",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 48,
        .constraints    = knl_uncore_m2pcie_constraints,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct attribute *knl_uncore_irp_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_qor.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group knl_uncore_irp_format_group = {
        .name = "format",
        .attrs = knl_uncore_irp_formats_attr,
};

static struct intel_uncore_type knl_uncore_irp = {
        .name                   = "irp",
        .num_counters           = 2,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNBEP_PCI_PMON_CTR0,
        .event_ctl              = SNBEP_PCI_PMON_CTL0,
        .event_mask             = KNL_IRP_PCI_PMON_RAW_EVENT_MASK,
        .box_ctl                = KNL_IRP_PCI_PMON_BOX_CTL,
        .ops                    = &snbep_uncore_pci_ops,
        .format_group           = &knl_uncore_irp_format_group,
};

enum {
        KNL_PCI_UNCORE_MC_UCLK,
        KNL_PCI_UNCORE_MC_DCLK,
        KNL_PCI_UNCORE_EDC_UCLK,
        KNL_PCI_UNCORE_EDC_ECLK,
        KNL_PCI_UNCORE_M2PCIE,
        KNL_PCI_UNCORE_IRP,
};

static struct intel_uncore_type *knl_pci_uncores[] = {
        [KNL_PCI_UNCORE_MC_UCLK]        = &knl_uncore_imc_uclk,
        [KNL_PCI_UNCORE_MC_DCLK]        = &knl_uncore_imc_dclk,
        [KNL_PCI_UNCORE_EDC_UCLK]       = &knl_uncore_edc_uclk,
        [KNL_PCI_UNCORE_EDC_ECLK]       = &knl_uncore_edc_eclk,
        [KNL_PCI_UNCORE_M2PCIE]         = &knl_uncore_m2pcie,
        [KNL_PCI_UNCORE_IRP]            = &knl_uncore_irp,
        NULL,
};

/*
 * KNL uses a common PCI device ID for multiple instances of an Uncore PMU
 * device type. prior to KNL, each instance of a PMU device type had a unique
 * device ID.
 *
 *      PCI Device ID   Uncore PMU Devices
 *      ----------------------------------
 *      0x7841          MC0 UClk, MC1 UClk
 *      0x7843          MC0 DClk CH 0, MC0 DClk CH 1, MC0 DClk CH 2,
 *                      MC1 DClk CH 0, MC1 DClk CH 1, MC1 DClk CH 2
 *      0x7833          EDC0 UClk, EDC1 UClk, EDC2 UClk, EDC3 UClk,
 *                      EDC4 UClk, EDC5 UClk, EDC6 UClk, EDC7 UClk
 *      0x7835          EDC0 EClk, EDC1 EClk, EDC2 EClk, EDC3 EClk,
 *                      EDC4 EClk, EDC5 EClk, EDC6 EClk, EDC7 EClk
 *      0x7817          M2PCIe
 *      0x7814          IRP
*/

static const struct pci_device_id knl_uncore_pci_ids[] = {
        { /* MC0 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7841),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(10, 0, KNL_PCI_UNCORE_MC_UCLK, 0),
        },
        { /* MC1 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7841),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(11, 0, KNL_PCI_UNCORE_MC_UCLK, 1),
        },
        { /* MC0 DClk CH 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7843),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(8, 2, KNL_PCI_UNCORE_MC_DCLK, 0),
        },
        { /* MC0 DClk CH 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7843),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(8, 3, KNL_PCI_UNCORE_MC_DCLK, 1),
        },
        { /* MC0 DClk CH 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7843),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(8, 4, KNL_PCI_UNCORE_MC_DCLK, 2),
        },
        { /* MC1 DClk CH 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7843),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(9, 2, KNL_PCI_UNCORE_MC_DCLK, 3),
        },
        { /* MC1 DClk CH 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7843),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(9, 3, KNL_PCI_UNCORE_MC_DCLK, 4),
        },
        { /* MC1 DClk CH 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7843),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(9, 4, KNL_PCI_UNCORE_MC_DCLK, 5),
        },
        { /* EDC0 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(15, 0, KNL_PCI_UNCORE_EDC_UCLK, 0),
        },
        { /* EDC1 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(16, 0, KNL_PCI_UNCORE_EDC_UCLK, 1),
        },
        { /* EDC2 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(17, 0, KNL_PCI_UNCORE_EDC_UCLK, 2),
        },
        { /* EDC3 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(18, 0, KNL_PCI_UNCORE_EDC_UCLK, 3),
        },
        { /* EDC4 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(19, 0, KNL_PCI_UNCORE_EDC_UCLK, 4),
        },
        { /* EDC5 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(20, 0, KNL_PCI_UNCORE_EDC_UCLK, 5),
        },
        { /* EDC6 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(21, 0, KNL_PCI_UNCORE_EDC_UCLK, 6),
        },
        { /* EDC7 UClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7833),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(22, 0, KNL_PCI_UNCORE_EDC_UCLK, 7),
        },
        { /* EDC0 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(24, 2, KNL_PCI_UNCORE_EDC_ECLK, 0),
        },
        { /* EDC1 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(25, 2, KNL_PCI_UNCORE_EDC_ECLK, 1),
        },
        { /* EDC2 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(26, 2, KNL_PCI_UNCORE_EDC_ECLK, 2),
        },
        { /* EDC3 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(27, 2, KNL_PCI_UNCORE_EDC_ECLK, 3),
        },
        { /* EDC4 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(28, 2, KNL_PCI_UNCORE_EDC_ECLK, 4),
        },
        { /* EDC5 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(29, 2, KNL_PCI_UNCORE_EDC_ECLK, 5),
        },
        { /* EDC6 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(30, 2, KNL_PCI_UNCORE_EDC_ECLK, 6),
        },
        { /* EDC7 EClk */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7835),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(31, 2, KNL_PCI_UNCORE_EDC_ECLK, 7),
        },
        { /* M2PCIe */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7817),
                .driver_data = UNCORE_PCI_DEV_DATA(KNL_PCI_UNCORE_M2PCIE, 0),
        },
        { /* IRP */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x7814),
                .driver_data = UNCORE_PCI_DEV_DATA(KNL_PCI_UNCORE_IRP, 0),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver knl_uncore_pci_driver = {
        .name           = "knl_uncore",
        .id_table       = knl_uncore_pci_ids,
};

int knl_uncore_pci_init(void)
{
        int ret;

        /* All KNL PCI based PMON units are on the same PCI bus except IRP */
        ret = snb_pci2phy_map_init(0x7814); /* IRP */
        if (ret)
                return ret;
        ret = snb_pci2phy_map_init(0x7817); /* M2PCIe */
        if (ret)
                return ret;
        uncore_pci_uncores = knl_pci_uncores;
        uncore_pci_driver = &knl_uncore_pci_driver;
        return 0;
}

/* end of KNL uncore support */

/* Haswell-EP uncore support */
static struct attribute *hswep_uncore_ubox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh5.attr,
        &format_attr_filter_tid2.attr,
        &format_attr_filter_cid.attr,
        NULL,
};

static const struct attribute_group hswep_uncore_ubox_format_group = {
        .name = "format",
        .attrs = hswep_uncore_ubox_formats_attr,
};

static int hswep_ubox_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        reg1->reg = HSWEP_U_MSR_PMON_FILTER;
        reg1->config = event->attr.config1 & HSWEP_U_MSR_PMON_BOX_FILTER_MASK;
        reg1->idx = 0;
        return 0;
}

static struct intel_uncore_ops hswep_uncore_ubox_ops = {
        SNBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .hw_config              = hswep_ubox_hw_config,
        .get_constraint         = uncore_get_constraint,
        .put_constraint         = uncore_put_constraint,
};

static struct intel_uncore_type hswep_uncore_ubox = {
        .name                   = "ubox",
        .num_counters           = 2,
        .num_boxes              = 1,
        .perf_ctr_bits          = 44,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = HSWEP_U_MSR_PMON_CTR0,
        .event_ctl              = HSWEP_U_MSR_PMON_CTL0,
        .event_mask             = SNBEP_U_MSR_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTR,
        .fixed_ctl              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTL,
        .num_shared_regs        = 1,
        .ops                    = &hswep_uncore_ubox_ops,
        .format_group           = &hswep_uncore_ubox_format_group,
};

static struct attribute *hswep_uncore_cbox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_thresh8.attr,
        &format_attr_filter_tid3.attr,
        &format_attr_filter_link2.attr,
        &format_attr_filter_state3.attr,
        &format_attr_filter_nid2.attr,
        &format_attr_filter_opc2.attr,
        &format_attr_filter_nc.attr,
        &format_attr_filter_c6.attr,
        &format_attr_filter_isoc.attr,
        NULL,
};

static const struct attribute_group hswep_uncore_cbox_format_group = {
        .name = "format",
        .attrs = hswep_uncore_cbox_formats_attr,
};

static struct event_constraint hswep_uncore_cbox_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x01, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x09, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x11, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x36, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x38, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x3b, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x3e, 0x1),
        EVENT_CONSTRAINT_END
};

static struct extra_reg hswep_uncore_cbox_extra_regs[] = {
        SNBEP_CBO_EVENT_EXTRA_REG(SNBEP_CBO_PMON_CTL_TID_EN,
                                  SNBEP_CBO_PMON_CTL_TID_EN, 0x1),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0334, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0534, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0934, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x1134, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2134, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4134, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4037, 0x40ff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4028, 0x40ff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4032, 0x40ff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4029, 0x40ff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4033, 0x40ff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x402A, 0x40ff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0135, 0xffff, 0x12),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0335, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4135, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4435, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4835, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x5035, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4335, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4a35, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2335, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8335, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2135, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8135, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0136, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0336, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4136, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4436, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4836, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4336, 0xffff, 0x18),
        SNBEP_CBO_EVENT_EXTRA_REG(0x4a36, 0xffff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2336, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8336, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x2136, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x8136, 0xffff, 0x10),
        SNBEP_CBO_EVENT_EXTRA_REG(0x5036, 0xffff, 0x8),
        EVENT_EXTRA_END
};

static u64 hswep_cbox_filter_mask(int fields)
{
        u64 mask = 0;
        if (fields & 0x1)
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_TID;
        if (fields & 0x2)
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_LINK;
        if (fields & 0x4)
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_STATE;
        if (fields & 0x8)
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_NID;
        if (fields & 0x10) {
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_OPC;
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_NC;
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_C6;
                mask |= HSWEP_CB0_MSR_PMON_BOX_FILTER_ISOC;
        }
        return mask;
}

static struct event_constraint *
hswep_cbox_get_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        return __snbep_cbox_get_constraint(box, event, hswep_cbox_filter_mask);
}

static int hswep_cbox_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct extra_reg *er;
        int idx = 0;

        for (er = hswep_uncore_cbox_extra_regs; er->msr; er++) {
                if (er->event != (event->hw.config & er->config_mask))
                        continue;
                idx |= er->idx;
        }

        if (idx) {
                reg1->reg = HSWEP_C0_MSR_PMON_BOX_FILTER0 +
                            HSWEP_CBO_MSR_OFFSET * box->pmu->pmu_idx;
                reg1->config = event->attr.config1 & hswep_cbox_filter_mask(idx);
                reg1->idx = idx;
        }
        return 0;
}

static void hswep_cbox_enable_event(struct intel_uncore_box *box,
                                  struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;

        if (reg1->idx != EXTRA_REG_NONE) {
                u64 filter = uncore_shared_reg_config(box, 0);
                wrmsrq(reg1->reg, filter & 0xffffffff);
                wrmsrq(reg1->reg + 1, filter >> 32);
        }

        wrmsrq(hwc->config_base, hwc->config | SNBEP_PMON_CTL_EN);
}

static struct intel_uncore_ops hswep_uncore_cbox_ops = {
        .init_box               = snbep_uncore_msr_init_box,
        .disable_box            = snbep_uncore_msr_disable_box,
        .enable_box             = snbep_uncore_msr_enable_box,
        .disable_event          = snbep_uncore_msr_disable_event,
        .enable_event           = hswep_cbox_enable_event,
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = hswep_cbox_hw_config,
        .get_constraint         = hswep_cbox_get_constraint,
        .put_constraint         = snbep_cbox_put_constraint,
};

static struct intel_uncore_type hswep_uncore_cbox = {
        .name                   = "cbox",
        .num_counters           = 4,
        .num_boxes              = 18,
        .perf_ctr_bits          = 48,
        .event_ctl              = HSWEP_C0_MSR_PMON_CTL0,
        .perf_ctr               = HSWEP_C0_MSR_PMON_CTR0,
        .event_mask             = SNBEP_CBO_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_C0_MSR_PMON_BOX_CTL,
        .msr_offset             = HSWEP_CBO_MSR_OFFSET,
        .num_shared_regs        = 1,
        .constraints            = hswep_uncore_cbox_constraints,
        .ops                    = &hswep_uncore_cbox_ops,
        .format_group           = &hswep_uncore_cbox_format_group,
};

/*
 * Write SBOX Initialization register bit by bit to avoid spurious #GPs
 */
static void hswep_uncore_sbox_msr_init_box(struct intel_uncore_box *box)
{
        unsigned msr = uncore_msr_box_ctl(box);

        if (msr) {
                u64 init = SNBEP_PMON_BOX_CTL_INT;
                u64 flags = 0;
                int i;

                for_each_set_bit(i, (unsigned long *)&init, 64) {
                        flags |= (1ULL << i);
                        wrmsrq(msr, flags);
                }
        }
}

static struct intel_uncore_ops hswep_uncore_sbox_msr_ops = {
        __SNBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .init_box               = hswep_uncore_sbox_msr_init_box
};

static struct attribute *hswep_uncore_sbox_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group hswep_uncore_sbox_format_group = {
        .name = "format",
        .attrs = hswep_uncore_sbox_formats_attr,
};

static struct intel_uncore_type hswep_uncore_sbox = {
        .name                   = "sbox",
        .num_counters           = 4,
        .num_boxes              = 4,
        .perf_ctr_bits          = 44,
        .event_ctl              = HSWEP_S0_MSR_PMON_CTL0,
        .perf_ctr               = HSWEP_S0_MSR_PMON_CTR0,
        .event_mask             = HSWEP_S_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_S0_MSR_PMON_BOX_CTL,
        .msr_offset             = HSWEP_SBOX_MSR_OFFSET,
        .ops                    = &hswep_uncore_sbox_msr_ops,
        .format_group           = &hswep_uncore_sbox_format_group,
};

static int hswep_pcu_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;
        int ev_sel = hwc->config & SNBEP_PMON_CTL_EV_SEL_MASK;

        if (ev_sel >= 0xb && ev_sel <= 0xe) {
                reg1->reg = HSWEP_PCU_MSR_PMON_BOX_FILTER;
                reg1->idx = ev_sel - 0xb;
                reg1->config = event->attr.config1 & (0xff << reg1->idx);
        }
        return 0;
}

static struct intel_uncore_ops hswep_uncore_pcu_ops = {
        SNBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .hw_config              = hswep_pcu_hw_config,
        .get_constraint         = snbep_pcu_get_constraint,
        .put_constraint         = snbep_pcu_put_constraint,
};

static struct intel_uncore_type hswep_uncore_pcu = {
        .name                   = "pcu",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .perf_ctr               = HSWEP_PCU_MSR_PMON_CTR0,
        .event_ctl              = HSWEP_PCU_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PCU_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_PCU_MSR_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &hswep_uncore_pcu_ops,
        .format_group           = &snbep_uncore_pcu_format_group,
};

static struct intel_uncore_type *hswep_msr_uncores[] = {
        &hswep_uncore_ubox,
        &hswep_uncore_cbox,
        &hswep_uncore_sbox,
        &hswep_uncore_pcu,
        NULL,
};

#define HSWEP_PCU_DID                   0x2fc0
#define HSWEP_PCU_CAPID4_OFFET          0x94
#define hswep_get_chop(_cap)            (((_cap) >> 6) & 0x3)

static bool hswep_has_limit_sbox(unsigned int device)
{
        struct pci_dev *dev = pci_get_device(PCI_VENDOR_ID_INTEL, device, NULL);
        u32 capid4;

        if (!dev)
                return false;

        pci_read_config_dword(dev, HSWEP_PCU_CAPID4_OFFET, &capid4);
        pci_dev_put(dev);
        if (!hswep_get_chop(capid4))
                return true;

        return false;
}

void hswep_uncore_cpu_init(void)
{
        if (hswep_uncore_cbox.num_boxes > topology_num_cores_per_package())
                hswep_uncore_cbox.num_boxes = topology_num_cores_per_package();

        /* Detect 6-8 core systems with only two SBOXes */
        if (hswep_has_limit_sbox(HSWEP_PCU_DID))
                hswep_uncore_sbox.num_boxes = 2;

        uncore_msr_uncores = hswep_msr_uncores;
}

static struct intel_uncore_type hswep_uncore_ha = {
        .name           = "ha",
        .num_counters   = 4,
        .num_boxes      = 2,
        .perf_ctr_bits  = 48,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct uncore_event_desc hswep_uncore_imc_events[] = {
        INTEL_UNCORE_EVENT_DESC(clockticks,      "event=0x00,umask=0x00"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read,  "event=0x04,umask=0x03"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.unit, "MiB"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write, "event=0x04,umask=0x0c"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.unit, "MiB"),
        { /* end: all zeroes */ },
};

static struct intel_uncore_type hswep_uncore_imc = {
        .name           = "imc",
        .num_counters   = 4,
        .num_boxes      = 8,
        .perf_ctr_bits  = 48,
        .fixed_ctr_bits = 48,
        .fixed_ctr      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTR,
        .fixed_ctl      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTL,
        .event_descs    = hswep_uncore_imc_events,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static unsigned hswep_uncore_irp_ctrs[] = {0xa0, 0xa8, 0xb0, 0xb8};

static u64 hswep_uncore_irp_read_counter(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;
        u64 count = 0;

        pci_read_config_dword(pdev, hswep_uncore_irp_ctrs[hwc->idx], (u32 *)&count);
        pci_read_config_dword(pdev, hswep_uncore_irp_ctrs[hwc->idx] + 4, (u32 *)&count + 1);

        return count;
}

static struct intel_uncore_ops hswep_uncore_irp_ops = {
        .init_box       = snbep_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = ivbep_uncore_irp_disable_event,
        .enable_event   = ivbep_uncore_irp_enable_event,
        .read_counter   = hswep_uncore_irp_read_counter,
};

static struct intel_uncore_type hswep_uncore_irp = {
        .name                   = "irp",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCI_PMON_BOX_CTL,
        .ops                    = &hswep_uncore_irp_ops,
        .format_group           = &snbep_uncore_format_group,
};

static struct intel_uncore_type hswep_uncore_qpi = {
        .name                   = "qpi",
        .num_counters           = 4,
        .num_boxes              = 3,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNBEP_PCI_PMON_CTR0,
        .event_ctl              = SNBEP_PCI_PMON_CTL0,
        .event_mask             = SNBEP_QPI_PCI_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCI_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &snbep_uncore_qpi_ops,
        .format_group           = &snbep_uncore_qpi_format_group,
};

static struct event_constraint hswep_uncore_r2pcie_constraints[] = {
        UNCORE_EVENT_CONSTRAINT_RANGE(0x10, 0x11, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x13, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x23, 0x25, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x26, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x27, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x28, 0x29, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x2a, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x2b, 0x2d, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x32, 0x35, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type hswep_uncore_r2pcie = {
        .name           = "r2pcie",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 48,
        .constraints    = hswep_uncore_r2pcie_constraints,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct event_constraint hswep_uncore_r3qpi_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x01, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x7, 0x0a, 0x7),
        UNCORE_EVENT_CONSTRAINT(0x0e, 0x7),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x10, 0x12, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x13, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x14, 0x15, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x1f, 0x23, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x25, 0x26, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x28, 0x29, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x2c, 0x2f, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x31, 0x34, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x36, 0x39, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type hswep_uncore_r3qpi = {
        .name           = "r3qpi",
        .num_counters   = 3,
        .num_boxes      = 3,
        .perf_ctr_bits  = 44,
        .constraints    = hswep_uncore_r3qpi_constraints,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

enum {
        HSWEP_PCI_UNCORE_HA,
        HSWEP_PCI_UNCORE_IMC,
        HSWEP_PCI_UNCORE_IRP,
        HSWEP_PCI_UNCORE_QPI,
        HSWEP_PCI_UNCORE_R2PCIE,
        HSWEP_PCI_UNCORE_R3QPI,
};

static struct intel_uncore_type *hswep_pci_uncores[] = {
        [HSWEP_PCI_UNCORE_HA]   = &hswep_uncore_ha,
        [HSWEP_PCI_UNCORE_IMC]  = &hswep_uncore_imc,
        [HSWEP_PCI_UNCORE_IRP]  = &hswep_uncore_irp,
        [HSWEP_PCI_UNCORE_QPI]  = &hswep_uncore_qpi,
        [HSWEP_PCI_UNCORE_R2PCIE]       = &hswep_uncore_r2pcie,
        [HSWEP_PCI_UNCORE_R3QPI]        = &hswep_uncore_r3qpi,
        NULL,
};

static const struct pci_device_id hswep_uncore_pci_ids[] = {
        { /* Home Agent 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f30),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_HA, 0),
        },
        { /* Home Agent 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f38),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_HA, 1),
        },
        { /* MC0 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fb0),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 0),
        },
        { /* MC0 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fb1),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 1),
        },
        { /* MC0 Channel 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fb4),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 2),
        },
        { /* MC0 Channel 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fb5),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 3),
        },
        { /* MC1 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fd0),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 4),
        },
        { /* MC1 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fd1),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 5),
        },
        { /* MC1 Channel 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fd4),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 6),
        },
        { /* MC1 Channel 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2fd5),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IMC, 7),
        },
        { /* IRP */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f39),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_IRP, 0),
        },
        { /* QPI0 Port 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f32),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_QPI, 0),
        },
        { /* QPI0 Port 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f33),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_QPI, 1),
        },
        { /* QPI1 Port 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f3a),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_QPI, 2),
        },
        { /* R2PCIe */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f34),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_R2PCIE, 0),
        },
        { /* R3QPI0 Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f36),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_R3QPI, 0),
        },
        { /* R3QPI0 Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f37),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_R3QPI, 1),
        },
        { /* R3QPI1 Link 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f3e),
                .driver_data = UNCORE_PCI_DEV_DATA(HSWEP_PCI_UNCORE_R3QPI, 2),
        },
        { /* QPI Port 0 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f86),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT0_FILTER),
        },
        { /* QPI Port 1 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2f96),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT1_FILTER),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver hswep_uncore_pci_driver = {
        .name           = "hswep_uncore",
        .id_table       = hswep_uncore_pci_ids,
};

int hswep_uncore_pci_init(void)
{
        int ret = snbep_pci2phy_map_init(0x2f1e, SNBEP_CPUNODEID, SNBEP_GIDNIDMAP, true);
        if (ret)
                return ret;
        uncore_pci_uncores = hswep_pci_uncores;
        uncore_pci_driver = &hswep_uncore_pci_driver;
        return 0;
}
/* end of Haswell-EP uncore support */

/* BDX uncore support */

static struct intel_uncore_type bdx_uncore_ubox = {
        .name                   = "ubox",
        .num_counters           = 2,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = HSWEP_U_MSR_PMON_CTR0,
        .event_ctl              = HSWEP_U_MSR_PMON_CTL0,
        .event_mask             = SNBEP_U_MSR_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTR,
        .fixed_ctl              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTL,
        .num_shared_regs        = 1,
        .ops                    = &ivbep_uncore_msr_ops,
        .format_group           = &ivbep_uncore_ubox_format_group,
};

static struct event_constraint bdx_uncore_cbox_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x09, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x11, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x36, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x3e, 0x1),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type bdx_uncore_cbox = {
        .name                   = "cbox",
        .num_counters           = 4,
        .num_boxes              = 24,
        .perf_ctr_bits          = 48,
        .event_ctl              = HSWEP_C0_MSR_PMON_CTL0,
        .perf_ctr               = HSWEP_C0_MSR_PMON_CTR0,
        .event_mask             = SNBEP_CBO_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_C0_MSR_PMON_BOX_CTL,
        .msr_offset             = HSWEP_CBO_MSR_OFFSET,
        .num_shared_regs        = 1,
        .constraints            = bdx_uncore_cbox_constraints,
        .ops                    = &hswep_uncore_cbox_ops,
        .format_group           = &hswep_uncore_cbox_format_group,
};

static struct intel_uncore_type bdx_uncore_sbox = {
        .name                   = "sbox",
        .num_counters           = 4,
        .num_boxes              = 4,
        .perf_ctr_bits          = 48,
        .event_ctl              = HSWEP_S0_MSR_PMON_CTL0,
        .perf_ctr               = HSWEP_S0_MSR_PMON_CTR0,
        .event_mask             = HSWEP_S_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_S0_MSR_PMON_BOX_CTL,
        .msr_offset             = HSWEP_SBOX_MSR_OFFSET,
        .ops                    = &hswep_uncore_sbox_msr_ops,
        .format_group           = &hswep_uncore_sbox_format_group,
};

#define BDX_MSR_UNCORE_SBOX     3

static struct intel_uncore_type *bdx_msr_uncores[] = {
        &bdx_uncore_ubox,
        &bdx_uncore_cbox,
        &hswep_uncore_pcu,
        &bdx_uncore_sbox,
        NULL,
};

/* Bit 7 'Use Occupancy' is not available for counter 0 on BDX */
static struct event_constraint bdx_uncore_pcu_constraints[] = {
        EVENT_CONSTRAINT(0x80, 0xe, 0x80),
        EVENT_CONSTRAINT_END
};

#define BDX_PCU_DID                     0x6fc0

void bdx_uncore_cpu_init(void)
{
        if (bdx_uncore_cbox.num_boxes > topology_num_cores_per_package())
                bdx_uncore_cbox.num_boxes = topology_num_cores_per_package();
        uncore_msr_uncores = bdx_msr_uncores;

        /* Detect systems with no SBOXes */
        if (boot_cpu_data.x86_vfm == INTEL_BROADWELL_D || hswep_has_limit_sbox(BDX_PCU_DID))
                uncore_msr_uncores[BDX_MSR_UNCORE_SBOX] = NULL;

        hswep_uncore_pcu.constraints = bdx_uncore_pcu_constraints;
}

static struct intel_uncore_type bdx_uncore_ha = {
        .name           = "ha",
        .num_counters   = 4,
        .num_boxes      = 2,
        .perf_ctr_bits  = 48,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct intel_uncore_type bdx_uncore_imc = {
        .name           = "imc",
        .num_counters   = 4,
        .num_boxes      = 8,
        .perf_ctr_bits  = 48,
        .fixed_ctr_bits = 48,
        .fixed_ctr      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTR,
        .fixed_ctl      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTL,
        .event_descs    = hswep_uncore_imc_events,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct intel_uncore_type bdx_uncore_irp = {
        .name                   = "irp",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCI_PMON_BOX_CTL,
        .ops                    = &hswep_uncore_irp_ops,
        .format_group           = &snbep_uncore_format_group,
};

static struct intel_uncore_type bdx_uncore_qpi = {
        .name                   = "qpi",
        .num_counters           = 4,
        .num_boxes              = 3,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNBEP_PCI_PMON_CTR0,
        .event_ctl              = SNBEP_PCI_PMON_CTL0,
        .event_mask             = SNBEP_QPI_PCI_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNBEP_PCI_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &snbep_uncore_qpi_ops,
        .format_group           = &snbep_uncore_qpi_format_group,
};

static struct event_constraint bdx_uncore_r2pcie_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x10, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x11, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x13, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x23, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x25, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x26, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x28, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x2c, 0x2d, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type bdx_uncore_r2pcie = {
        .name           = "r2pcie",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 48,
        .constraints    = bdx_uncore_r2pcie_constraints,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

static struct event_constraint bdx_uncore_r3qpi_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x01, 0x7),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x07, 0x0a, 0x7),
        UNCORE_EVENT_CONSTRAINT(0x0e, 0x7),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x10, 0x11, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x13, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x14, 0x15, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x1f, 0x23, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x25, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x26, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x28, 0x29, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x2c, 0x2f, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x33, 0x34, 0x3),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x36, 0x39, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type bdx_uncore_r3qpi = {
        .name           = "r3qpi",
        .num_counters   = 3,
        .num_boxes      = 3,
        .perf_ctr_bits  = 48,
        .constraints    = bdx_uncore_r3qpi_constraints,
        SNBEP_UNCORE_PCI_COMMON_INIT(),
};

enum {
        BDX_PCI_UNCORE_HA,
        BDX_PCI_UNCORE_IMC,
        BDX_PCI_UNCORE_IRP,
        BDX_PCI_UNCORE_QPI,
        BDX_PCI_UNCORE_R2PCIE,
        BDX_PCI_UNCORE_R3QPI,
};

static struct intel_uncore_type *bdx_pci_uncores[] = {
        [BDX_PCI_UNCORE_HA]     = &bdx_uncore_ha,
        [BDX_PCI_UNCORE_IMC]    = &bdx_uncore_imc,
        [BDX_PCI_UNCORE_IRP]    = &bdx_uncore_irp,
        [BDX_PCI_UNCORE_QPI]    = &bdx_uncore_qpi,
        [BDX_PCI_UNCORE_R2PCIE] = &bdx_uncore_r2pcie,
        [BDX_PCI_UNCORE_R3QPI]  = &bdx_uncore_r3qpi,
        NULL,
};

static const struct pci_device_id bdx_uncore_pci_ids[] = {
        { /* Home Agent 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f30),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_HA, 0),
        },
        { /* Home Agent 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f38),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_HA, 1),
        },
        { /* MC0 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fb0),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 0),
        },
        { /* MC0 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fb1),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 1),
        },
        { /* MC0 Channel 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fb4),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 2),
        },
        { /* MC0 Channel 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fb5),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 3),
        },
        { /* MC1 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fd0),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 4),
        },
        { /* MC1 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fd1),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 5),
        },
        { /* MC1 Channel 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fd4),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 6),
        },
        { /* MC1 Channel 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6fd5),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IMC, 7),
        },
        { /* IRP */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f39),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_IRP, 0),
        },
        { /* QPI0 Port 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f32),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_QPI, 0),
        },
        { /* QPI0 Port 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f33),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_QPI, 1),
        },
        { /* QPI1 Port 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f3a),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_QPI, 2),
        },
        { /* R2PCIe */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f34),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_R2PCIE, 0),
        },
        { /* R3QPI0 Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f36),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_R3QPI, 0),
        },
        { /* R3QPI0 Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f37),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_R3QPI, 1),
        },
        { /* R3QPI1 Link 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f3e),
                .driver_data = UNCORE_PCI_DEV_DATA(BDX_PCI_UNCORE_R3QPI, 2),
        },
        { /* QPI Port 0 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f86),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT0_FILTER),
        },
        { /* QPI Port 1 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f96),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   SNBEP_PCI_QPI_PORT1_FILTER),
        },
        { /* QPI Port 2 filter  */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x6f46),
                .driver_data = UNCORE_PCI_DEV_DATA(UNCORE_EXTRA_PCI_DEV,
                                                   BDX_PCI_QPI_PORT2_FILTER),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver bdx_uncore_pci_driver = {
        .name           = "bdx_uncore",
        .id_table       = bdx_uncore_pci_ids,
};

int bdx_uncore_pci_init(void)
{
        int ret = snbep_pci2phy_map_init(0x6f1e, SNBEP_CPUNODEID, SNBEP_GIDNIDMAP, true);

        if (ret)
                return ret;
        uncore_pci_uncores = bdx_pci_uncores;
        uncore_pci_driver = &bdx_uncore_pci_driver;
        return 0;
}

/* end of BDX uncore support */

/* SKX uncore support */

static struct intel_uncore_type skx_uncore_ubox = {
        .name                   = "ubox",
        .num_counters           = 2,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = HSWEP_U_MSR_PMON_CTR0,
        .event_ctl              = HSWEP_U_MSR_PMON_CTL0,
        .event_mask             = SNBEP_U_MSR_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTR,
        .fixed_ctl              = HSWEP_U_MSR_PMON_UCLK_FIXED_CTL,
        .ops                    = &ivbep_uncore_msr_ops,
        .format_group           = &ivbep_uncore_ubox_format_group,
};

static struct attribute *skx_uncore_cha_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_filter_tid4.attr,
        &format_attr_filter_state5.attr,
        &format_attr_filter_rem.attr,
        &format_attr_filter_loc.attr,
        &format_attr_filter_nm.attr,
        &format_attr_filter_all_op.attr,
        &format_attr_filter_not_nm.attr,
        &format_attr_filter_opc_0.attr,
        &format_attr_filter_opc_1.attr,
        &format_attr_filter_nc.attr,
        &format_attr_filter_isoc.attr,
        NULL,
};

static const struct attribute_group skx_uncore_chabox_format_group = {
        .name = "format",
        .attrs = skx_uncore_cha_formats_attr,
};

static struct event_constraint skx_uncore_chabox_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x11, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x36, 0x1),
        EVENT_CONSTRAINT_END
};

static struct extra_reg skx_uncore_cha_extra_regs[] = {
        SNBEP_CBO_EVENT_EXTRA_REG(0x0334, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0534, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x0934, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x1134, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x3134, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x9134, 0xffff, 0x4),
        SNBEP_CBO_EVENT_EXTRA_REG(0x35, 0xff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x36, 0xff, 0x8),
        SNBEP_CBO_EVENT_EXTRA_REG(0x38, 0xff, 0x3),
        EVENT_EXTRA_END
};

static u64 skx_cha_filter_mask(int fields)
{
        u64 mask = 0;

        if (fields & 0x1)
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_TID;
        if (fields & 0x2)
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_LINK;
        if (fields & 0x4)
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_STATE;
        if (fields & 0x8) {
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_REM;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_LOC;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_ALL_OPC;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_NM;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_NOT_NM;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_OPC0;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_OPC1;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_NC;
                mask |= SKX_CHA_MSR_PMON_BOX_FILTER_ISOC;
        }
        return mask;
}

static struct event_constraint *
skx_cha_get_constraint(struct intel_uncore_box *box, struct perf_event *event)
{
        return __snbep_cbox_get_constraint(box, event, skx_cha_filter_mask);
}

static int skx_cha_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        struct extra_reg *er;
        int idx = 0;
        /* Any of the CHA events may be filtered by Thread/Core-ID.*/
        if (event->hw.config & SNBEP_CBO_PMON_CTL_TID_EN)
                idx = SKX_CHA_MSR_PMON_BOX_FILTER_TID;

        for (er = skx_uncore_cha_extra_regs; er->msr; er++) {
                if (er->event != (event->hw.config & er->config_mask))
                        continue;
                idx |= er->idx;
        }

        if (idx) {
                reg1->reg = HSWEP_C0_MSR_PMON_BOX_FILTER0 +
                            HSWEP_CBO_MSR_OFFSET * box->pmu->pmu_idx;
                reg1->config = event->attr.config1 & skx_cha_filter_mask(idx);
                reg1->idx = idx;
        }
        return 0;
}

static struct intel_uncore_ops skx_uncore_chabox_ops = {
        /* There is no frz_en for chabox ctl */
        .init_box               = ivbep_uncore_msr_init_box,
        .disable_box            = snbep_uncore_msr_disable_box,
        .enable_box             = snbep_uncore_msr_enable_box,
        .disable_event          = snbep_uncore_msr_disable_event,
        .enable_event           = hswep_cbox_enable_event,
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = skx_cha_hw_config,
        .get_constraint         = skx_cha_get_constraint,
        .put_constraint         = snbep_cbox_put_constraint,
};

static struct intel_uncore_type skx_uncore_chabox = {
        .name                   = "cha",
        .num_counters           = 4,
        .perf_ctr_bits          = 48,
        .event_ctl              = HSWEP_C0_MSR_PMON_CTL0,
        .perf_ctr               = HSWEP_C0_MSR_PMON_CTR0,
        .event_mask             = HSWEP_S_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_C0_MSR_PMON_BOX_CTL,
        .msr_offset             = HSWEP_CBO_MSR_OFFSET,
        .num_shared_regs        = 1,
        .constraints            = skx_uncore_chabox_constraints,
        .ops                    = &skx_uncore_chabox_ops,
        .format_group           = &skx_uncore_chabox_format_group,
};

static struct attribute *skx_uncore_iio_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh9.attr,
        &format_attr_ch_mask.attr,
        &format_attr_fc_mask.attr,
        NULL,
};

static const struct attribute_group skx_uncore_iio_format_group = {
        .name = "format",
        .attrs = skx_uncore_iio_formats_attr,
};

static struct event_constraint skx_uncore_iio_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x83, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x88, 0xc),
        UNCORE_EVENT_CONSTRAINT(0x95, 0xc),
        UNCORE_EVENT_CONSTRAINT(0xc0, 0xc),
        UNCORE_EVENT_CONSTRAINT(0xc5, 0xc),
        UNCORE_EVENT_CONSTRAINT_RANGE(0xd4, 0xd5, 0xc),
        EVENT_CONSTRAINT_END
};

static void skx_iio_enable_event(struct intel_uncore_box *box,
                                 struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;

        wrmsrq(hwc->config_base, hwc->config | SNBEP_PMON_CTL_EN);
}

static struct intel_uncore_ops skx_uncore_iio_ops = {
        .init_box               = ivbep_uncore_msr_init_box,
        .disable_box            = snbep_uncore_msr_disable_box,
        .enable_box             = snbep_uncore_msr_enable_box,
        .disable_event          = snbep_uncore_msr_disable_event,
        .enable_event           = skx_iio_enable_event,
        .read_counter           = uncore_msr_read_counter,
};

static struct intel_uncore_topology *pmu_topology(struct intel_uncore_pmu *pmu, int die)
{
        int idx;

        for (idx = 0; idx < pmu->type->num_boxes; idx++) {
                if (pmu->type->topology[die][idx].pmu_idx == pmu->pmu_idx)
                        return &pmu->type->topology[die][idx];
        }

        return NULL;
}

static umode_t
pmu_iio_mapping_visible(struct kobject *kobj, struct attribute *attr,
                         int die, int zero_bus_pmu)
{
        struct intel_uncore_pmu *pmu = dev_to_uncore_pmu(kobj_to_dev(kobj));
        struct intel_uncore_topology *pmut = pmu_topology(pmu, die);

        return (pmut && !pmut->iio->pci_bus_no && pmu->pmu_idx != zero_bus_pmu) ? 0 : attr->mode;
}

static umode_t
skx_iio_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
{
        /* Root bus 0x00 is valid only for pmu_idx = 0. */
        return pmu_iio_mapping_visible(kobj, attr, die, 0);
}

static ssize_t skx_iio_mapping_show(struct device *dev,
                                    struct device_attribute *attr, char *buf)
{
        struct intel_uncore_pmu *pmu = dev_to_uncore_pmu(dev);
        struct dev_ext_attribute *ea = to_dev_ext_attribute(attr);
        long die = (long)ea->var;
        struct intel_uncore_topology *pmut = pmu_topology(pmu, die);

        return sprintf(buf, "%04x:%02x\n", pmut ? pmut->iio->segment : 0,
                                           pmut ? pmut->iio->pci_bus_no : 0);
}

static int skx_msr_cpu_bus_read(int cpu, u64 *topology)
{
        u64 msr_value;

        if (rdmsrq_on_cpu(cpu, SKX_MSR_CPU_BUS_NUMBER, &msr_value) ||
                        !(msr_value & SKX_MSR_CPU_BUS_VALID_BIT))
                return -ENXIO;

        *topology = msr_value;

        return 0;
}

static int die_to_cpu(int die)
{
        int res = 0, cpu, current_die;
        /*
         * Using cpus_read_lock() to ensure cpu is not going down between
         * looking at cpu_online_mask.
         */
        cpus_read_lock();
        for_each_online_cpu(cpu) {
                current_die = topology_logical_die_id(cpu);
                if (current_die == die) {
                        res = cpu;
                        break;
                }
        }
        cpus_read_unlock();
        return res;
}

enum {
        IIO_TOPOLOGY_TYPE,
        UPI_TOPOLOGY_TYPE,
        TOPOLOGY_MAX
};

static const size_t topology_size[TOPOLOGY_MAX] = {
        sizeof(*((struct intel_uncore_topology *)NULL)->iio),
        sizeof(*((struct intel_uncore_topology *)NULL)->upi)
};

static int pmu_alloc_topology(struct intel_uncore_type *type, int topology_type)
{
        int die, idx;
        struct intel_uncore_topology **topology;

        if (!type->num_boxes)
                return -EPERM;

        topology = kzalloc_objs(*topology, uncore_max_dies());
        if (!topology)
                goto err;

        for (die = 0; die < uncore_max_dies(); die++) {
                topology[die] = kzalloc_objs(**topology, type->num_boxes);
                if (!topology[die])
                        goto clear;
                for (idx = 0; idx < type->num_boxes; idx++) {
                        topology[die][idx].untyped = kcalloc(type->num_boxes,
                                                             topology_size[topology_type],
                                                             GFP_KERNEL);
                        if (!topology[die][idx].untyped)
                                goto clear;
                }
        }

        type->topology = topology;

        return 0;
clear:
        for (; die >= 0; die--) {
                for (idx = 0; idx < type->num_boxes; idx++)
                        kfree(topology[die][idx].untyped);
                kfree(topology[die]);
        }
        kfree(topology);
err:
        return -ENOMEM;
}

static void pmu_free_topology(struct intel_uncore_type *type)
{
        int die, idx;

        if (type->topology) {
                for (die = 0; die < uncore_max_dies(); die++) {
                        for (idx = 0; idx < type->num_boxes; idx++)
                                kfree(type->topology[die][idx].untyped);
                        kfree(type->topology[die]);
                }
                kfree(type->topology);
                type->topology = NULL;
        }
}

static int skx_pmu_get_topology(struct intel_uncore_type *type,
                                 int (*topology_cb)(struct intel_uncore_type*, int, int, u64))
{
        int die, ret = -EPERM;
        u64 cpu_bus_msr;

        for (die = 0; die < uncore_max_dies(); die++) {
                ret = skx_msr_cpu_bus_read(die_to_cpu(die), &cpu_bus_msr);
                if (ret)
                        break;

                ret = uncore_die_to_segment(die);
                if (ret < 0)
                        break;

                ret = topology_cb(type, ret, die, cpu_bus_msr);
                if (ret)
                        break;
        }

        return ret;
}

static int skx_iio_topology_cb(struct intel_uncore_type *type, int segment,
                                int die, u64 cpu_bus_msr)
{
        int idx;
        struct intel_uncore_topology *t;

        for (idx = 0; idx < type->num_boxes; idx++) {
                t = &type->topology[die][idx];
                t->pmu_idx = idx;
                t->iio->segment = segment;
                t->iio->pci_bus_no = (cpu_bus_msr >> (idx * BUS_NUM_STRIDE)) & 0xff;
        }

        return 0;
}

static int skx_iio_get_topology(struct intel_uncore_type *type)
{
        return skx_pmu_get_topology(type, skx_iio_topology_cb);
}

static struct attribute_group skx_iio_mapping_group = {
        .is_visible     = skx_iio_mapping_visible,
};

static const struct attribute_group *skx_iio_attr_update[] = {
        &skx_iio_mapping_group,
        NULL,
};

static void pmu_clear_mapping_attr(const struct attribute_group **groups,
                                   struct attribute_group *ag)
{
        int i;

        for (i = 0; groups[i]; i++) {
                if (groups[i] == ag) {
                        for (i++; groups[i]; i++)
                                groups[i - 1] = groups[i];
                        groups[i - 1] = NULL;
                        break;
                }
        }
}

static void
pmu_set_mapping(struct intel_uncore_type *type, struct attribute_group *ag,
                ssize_t (*show)(struct device*, struct device_attribute*, char*),
                int topology_type)
{
        char buf[64];
        int ret;
        long die = -1;
        struct attribute **attrs = NULL;
        struct dev_ext_attribute *eas = NULL;

        ret = pmu_alloc_topology(type, topology_type);
        if (ret < 0)
                goto clear_attr_update;

        ret = type->get_topology(type);
        if (ret < 0)
                goto clear_topology;

        /* One more for NULL. */
        attrs = kzalloc_objs(*attrs, (uncore_max_dies() + 1));
        if (!attrs)
                goto clear_topology;

        eas = kzalloc_objs(*eas, uncore_max_dies());
        if (!eas)
                goto clear_attrs;

        for (die = 0; die < uncore_max_dies(); die++) {
                snprintf(buf, sizeof(buf), "die%ld", die);
                sysfs_attr_init(&eas[die].attr.attr);
                eas[die].attr.attr.name = kstrdup(buf, GFP_KERNEL);
                if (!eas[die].attr.attr.name)
                        goto err;
                eas[die].attr.attr.mode = 0444;
                eas[die].attr.show = show;
                eas[die].attr.store = NULL;
                eas[die].var = (void *)die;
                attrs[die] = &eas[die].attr.attr;
        }
        ag->attrs = attrs;

        return;
err:
        for (; die >= 0; die--)
                kfree(eas[die].attr.attr.name);
        kfree(eas);
clear_attrs:
        kfree(attrs);
clear_topology:
        pmu_free_topology(type);
clear_attr_update:
        pmu_clear_mapping_attr(type->attr_update, ag);
}

static void
pmu_cleanup_mapping(struct intel_uncore_type *type, struct attribute_group *ag)
{
        struct attribute **attr = ag->attrs;

        if (!attr)
                return;

        for (; *attr; attr++)
                kfree((*attr)->name);
        kfree(attr_to_ext_attr(*ag->attrs));
        kfree(ag->attrs);
        ag->attrs = NULL;
        pmu_free_topology(type);
}

static void
pmu_iio_set_mapping(struct intel_uncore_type *type, struct attribute_group *ag)
{
        pmu_set_mapping(type, ag, skx_iio_mapping_show, IIO_TOPOLOGY_TYPE);
}

static void skx_iio_set_mapping(struct intel_uncore_type *type)
{
        pmu_iio_set_mapping(type, &skx_iio_mapping_group);
}

static void skx_iio_cleanup_mapping(struct intel_uncore_type *type)
{
        pmu_cleanup_mapping(type, &skx_iio_mapping_group);
}

static struct intel_uncore_type skx_uncore_iio = {
        .name                   = "iio",
        .num_counters           = 4,
        .num_boxes              = 6,
        .perf_ctr_bits          = 48,
        .event_ctl              = SKX_IIO0_MSR_PMON_CTL0,
        .perf_ctr               = SKX_IIO0_MSR_PMON_CTR0,
        .event_mask             = SKX_IIO_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = SKX_IIO_PMON_RAW_EVENT_MASK_EXT,
        .box_ctl                = SKX_IIO0_MSR_PMON_BOX_CTL,
        .msr_offset             = SKX_IIO_MSR_OFFSET,
        .constraints            = skx_uncore_iio_constraints,
        .ops                    = &skx_uncore_iio_ops,
        .format_group           = &skx_uncore_iio_format_group,
        .attr_update            = skx_iio_attr_update,
        .get_topology           = skx_iio_get_topology,
        .set_mapping            = skx_iio_set_mapping,
        .cleanup_mapping        = skx_iio_cleanup_mapping,
};

enum perf_uncore_iio_freerunning_type_id {
        SKX_IIO_MSR_IOCLK                       = 0,
        SKX_IIO_MSR_BW                          = 1,
        SKX_IIO_MSR_UTIL                        = 2,

        SKX_IIO_FREERUNNING_TYPE_MAX,
};


static struct freerunning_counters skx_iio_freerunning[] = {
        [SKX_IIO_MSR_IOCLK]     = { 0xa45, 0x1, 0x20, 1, 36 },
        [SKX_IIO_MSR_BW]        = { 0xb00, 0x1, 0x10, 8, 36 },
        [SKX_IIO_MSR_UTIL]      = { 0xb08, 0x1, 0x10, 8, 36 },
};

#define INTEL_UNCORE_FR_EVENT_DESC(name, umask, scl)                    \
        INTEL_UNCORE_EVENT_DESC(name,                                   \
                                "event=0xff,umask=" __stringify(umask)),\
        INTEL_UNCORE_EVENT_DESC(name.scale, __stringify(scl)),          \
        INTEL_UNCORE_EVENT_DESC(name.unit, "MiB")

static struct uncore_event_desc skx_uncore_iio_freerunning_events[] = {
        /* Free-Running IO CLOCKS Counter */
        INTEL_UNCORE_EVENT_DESC(ioclk,                  "event=0xff,umask=0x10"),
        /* Free-Running IIO BANDWIDTH Counters */
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port0,  0x20, 3.814697266e-6),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port1,  0x21, 3.814697266e-6),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port2,  0x22, 3.814697266e-6),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port3,  0x23, 3.814697266e-6),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port0, 0x24, 3.814697266e-6),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port1, 0x25, 3.814697266e-6),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port2, 0x26, 3.814697266e-6),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port3, 0x27, 3.814697266e-6),
        /* Free-running IIO UTILIZATION Counters */
        INTEL_UNCORE_EVENT_DESC(util_in_port0,          "event=0xff,umask=0x30"),
        INTEL_UNCORE_EVENT_DESC(util_out_port0,         "event=0xff,umask=0x31"),
        INTEL_UNCORE_EVENT_DESC(util_in_port1,          "event=0xff,umask=0x32"),
        INTEL_UNCORE_EVENT_DESC(util_out_port1,         "event=0xff,umask=0x33"),
        INTEL_UNCORE_EVENT_DESC(util_in_port2,          "event=0xff,umask=0x34"),
        INTEL_UNCORE_EVENT_DESC(util_out_port2,         "event=0xff,umask=0x35"),
        INTEL_UNCORE_EVENT_DESC(util_in_port3,          "event=0xff,umask=0x36"),
        INTEL_UNCORE_EVENT_DESC(util_out_port3,         "event=0xff,umask=0x37"),
        { /* end: all zeroes */ },
};

static struct intel_uncore_ops skx_uncore_iio_freerunning_ops = {
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = uncore_freerunning_hw_config,
};

static struct attribute *skx_uncore_iio_freerunning_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        NULL,
};

static const struct attribute_group skx_uncore_iio_freerunning_format_group = {
        .name = "format",
        .attrs = skx_uncore_iio_freerunning_formats_attr,
};

static struct intel_uncore_type skx_uncore_iio_free_running = {
        .name                   = "iio_free_running",
        .num_counters           = 17,
        .num_boxes              = 6,
        .num_freerunning_types  = SKX_IIO_FREERUNNING_TYPE_MAX,
        .freerunning            = skx_iio_freerunning,
        .ops                    = &skx_uncore_iio_freerunning_ops,
        .event_descs            = skx_uncore_iio_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

static struct attribute *skx_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group skx_uncore_format_group = {
        .name = "format",
        .attrs = skx_uncore_formats_attr,
};

static struct intel_uncore_type skx_uncore_irp = {
        .name                   = "irp",
        .num_counters           = 2,
        .num_boxes              = 6,
        .perf_ctr_bits          = 48,
        .event_ctl              = SKX_IRP0_MSR_PMON_CTL0,
        .perf_ctr               = SKX_IRP0_MSR_PMON_CTR0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl                = SKX_IRP0_MSR_PMON_BOX_CTL,
        .msr_offset             = SKX_IRP_MSR_OFFSET,
        .ops                    = &skx_uncore_iio_ops,
        .format_group           = &skx_uncore_format_group,
};

static struct attribute *skx_uncore_pcu_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_occ_invert.attr,
        &format_attr_occ_edge_det.attr,
        &format_attr_filter_band0.attr,
        &format_attr_filter_band1.attr,
        &format_attr_filter_band2.attr,
        &format_attr_filter_band3.attr,
        NULL,
};

static struct attribute_group skx_uncore_pcu_format_group = {
        .name = "format",
        .attrs = skx_uncore_pcu_formats_attr,
};

static struct intel_uncore_ops skx_uncore_pcu_ops = {
        IVBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .hw_config              = hswep_pcu_hw_config,
        .get_constraint         = snbep_pcu_get_constraint,
        .put_constraint         = snbep_pcu_put_constraint,
};

static struct intel_uncore_type skx_uncore_pcu = {
        .name                   = "pcu",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .perf_ctr               = HSWEP_PCU_MSR_PMON_CTR0,
        .event_ctl              = HSWEP_PCU_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PCU_MSR_PMON_RAW_EVENT_MASK,
        .box_ctl                = HSWEP_PCU_MSR_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &skx_uncore_pcu_ops,
        .format_group           = &skx_uncore_pcu_format_group,
};

static struct intel_uncore_type *skx_msr_uncores[] = {
        &skx_uncore_ubox,
        &skx_uncore_chabox,
        &skx_uncore_iio,
        &skx_uncore_iio_free_running,
        &skx_uncore_irp,
        &skx_uncore_pcu,
        NULL,
};

/*
 * To determine the number of CHAs, it should read bits 27:0 in the CAPID6
 * register which located at Device 30, Function 3, Offset 0x9C. PCI ID 0x2083.
 */
#define SKX_CAPID6              0x9c
#define SKX_CHA_BIT_MASK        GENMASK(27, 0)

static int skx_count_chabox(void)
{
        struct pci_dev *dev = NULL;
        u32 val = 0;

        dev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x2083, dev);
        if (!dev)
                goto out;

        pci_read_config_dword(dev, SKX_CAPID6, &val);
        val &= SKX_CHA_BIT_MASK;
out:
        pci_dev_put(dev);
        return hweight32(val);
}

void skx_uncore_cpu_init(void)
{
        skx_uncore_chabox.num_boxes = skx_count_chabox();
        uncore_msr_uncores = skx_msr_uncores;
}

static struct intel_uncore_type skx_uncore_imc = {
        .name           = "imc",
        .num_counters   = 4,
        .num_boxes      = 6,
        .perf_ctr_bits  = 48,
        .fixed_ctr_bits = 48,
        .fixed_ctr      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTR,
        .fixed_ctl      = SNBEP_MC_CHy_PCI_PMON_FIXED_CTL,
        .event_descs    = hswep_uncore_imc_events,
        .perf_ctr       = SNBEP_PCI_PMON_CTR0,
        .event_ctl      = SNBEP_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl        = SNBEP_PCI_PMON_BOX_CTL,
        .ops            = &ivbep_uncore_pci_ops,
        .format_group   = &skx_uncore_format_group,
};

static struct attribute *skx_upi_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask_ext.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group skx_upi_uncore_format_group = {
        .name = "format",
        .attrs = skx_upi_uncore_formats_attr,
};

static void skx_upi_uncore_pci_init_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;

        __set_bit(UNCORE_BOX_FLAG_CTL_OFFS8, &box->flags);
        pci_write_config_dword(pdev, SKX_UPI_PCI_PMON_BOX_CTL, IVBEP_PMON_BOX_CTL_INT);
}

static struct intel_uncore_ops skx_upi_uncore_pci_ops = {
        .init_box       = skx_upi_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = snbep_uncore_pci_disable_event,
        .enable_event   = snbep_uncore_pci_enable_event,
        .read_counter   = snbep_uncore_pci_read_counter,
};

static umode_t
skx_upi_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
{
        struct intel_uncore_pmu *pmu = dev_to_uncore_pmu(kobj_to_dev(kobj));

        return pmu->type->topology[die][pmu->pmu_idx].upi->enabled ? attr->mode : 0;
}

static ssize_t skx_upi_mapping_show(struct device *dev,
                                    struct device_attribute *attr, char *buf)
{
        struct intel_uncore_pmu *pmu = dev_to_uncore_pmu(dev);
        struct dev_ext_attribute *ea = to_dev_ext_attribute(attr);
        long die = (long)ea->var;
        struct uncore_upi_topology *upi = pmu->type->topology[die][pmu->pmu_idx].upi;

        return sysfs_emit(buf, "upi_%d,die_%d\n", upi->pmu_idx_to, upi->die_to);
}

#define SKX_UPI_REG_DID                 0x2058
#define SKX_UPI_REGS_ADDR_DEVICE_LINK0  0x0e
#define SKX_UPI_REGS_ADDR_FUNCTION      0x00

/*
 * UPI Link Parameter 0
 * |  Bit  |  Default  |  Description
 * | 19:16 |     0h    | base_nodeid - The NodeID of the sending socket.
 * | 12:8  |    00h    | sending_port - The processor die port number of the sending port.
 */
#define SKX_KTILP0_OFFSET       0x94

/*
 * UPI Pcode Status. This register is used by PCode to store the link training status.
 * |  Bit  |  Default  |  Description
 * |   4   |     0h    | ll_status_valid — Bit indicates the valid training status
 *                       logged from PCode to the BIOS.
 */
#define SKX_KTIPCSTS_OFFSET     0x120

static int upi_fill_topology(struct pci_dev *dev, struct intel_uncore_topology *tp,
                             int pmu_idx)
{
        int ret;
        u32 upi_conf;
        struct uncore_upi_topology *upi = tp->upi;

        tp->pmu_idx = pmu_idx;
        ret = pci_read_config_dword(dev, SKX_KTIPCSTS_OFFSET, &upi_conf);
        if (ret) {
                ret = pcibios_err_to_errno(ret);
                goto err;
        }
        upi->enabled = (upi_conf >> 4) & 1;
        if (upi->enabled) {
                ret = pci_read_config_dword(dev, SKX_KTILP0_OFFSET,
                                            &upi_conf);
                if (ret) {
                        ret = pcibios_err_to_errno(ret);
                        goto err;
                }
                upi->die_to = (upi_conf >> 16) & 0xf;
                upi->pmu_idx_to = (upi_conf >> 8) & 0x1f;
        }
err:
        return ret;
}

static int skx_upi_topology_cb(struct intel_uncore_type *type, int segment,
                                int die, u64 cpu_bus_msr)
{
        int idx, ret;
        struct intel_uncore_topology *upi;
        unsigned int devfn;
        struct pci_dev *dev = NULL;
        u8 bus = cpu_bus_msr >> (3 * BUS_NUM_STRIDE);

        for (idx = 0; idx < type->num_boxes; idx++) {
                upi = &type->topology[die][idx];
                devfn = PCI_DEVFN(SKX_UPI_REGS_ADDR_DEVICE_LINK0 + idx,
                                  SKX_UPI_REGS_ADDR_FUNCTION);
                dev = pci_get_domain_bus_and_slot(segment, bus, devfn);
                if (dev) {
                        ret = upi_fill_topology(dev, upi, idx);
                        if (ret)
                                break;
                }
        }

        pci_dev_put(dev);
        return ret;
}

static int skx_upi_get_topology(struct intel_uncore_type *type)
{
        /* CPX case is not supported */
        if (boot_cpu_data.x86_stepping == 11)
                return -EPERM;

        return skx_pmu_get_topology(type, skx_upi_topology_cb);
}

static struct attribute_group skx_upi_mapping_group = {
        .is_visible     = skx_upi_mapping_visible,
};

static const struct attribute_group *skx_upi_attr_update[] = {
        &skx_upi_mapping_group,
        NULL
};

static void
pmu_upi_set_mapping(struct intel_uncore_type *type, struct attribute_group *ag)
{
        pmu_set_mapping(type, ag, skx_upi_mapping_show, UPI_TOPOLOGY_TYPE);
}

static void skx_upi_set_mapping(struct intel_uncore_type *type)
{
        pmu_upi_set_mapping(type, &skx_upi_mapping_group);
}

static void skx_upi_cleanup_mapping(struct intel_uncore_type *type)
{
        pmu_cleanup_mapping(type, &skx_upi_mapping_group);
}

static struct intel_uncore_type skx_uncore_upi = {
        .name           = "upi",
        .num_counters   = 4,
        .num_boxes      = 3,
        .perf_ctr_bits  = 48,
        .perf_ctr       = SKX_UPI_PCI_PMON_CTR0,
        .event_ctl      = SKX_UPI_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext = SKX_UPI_CTL_UMASK_EXT,
        .box_ctl        = SKX_UPI_PCI_PMON_BOX_CTL,
        .ops            = &skx_upi_uncore_pci_ops,
        .format_group   = &skx_upi_uncore_format_group,
        .attr_update    = skx_upi_attr_update,
        .get_topology   = skx_upi_get_topology,
        .set_mapping    = skx_upi_set_mapping,
        .cleanup_mapping = skx_upi_cleanup_mapping,
};

static void skx_m2m_uncore_pci_init_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;

        __set_bit(UNCORE_BOX_FLAG_CTL_OFFS8, &box->flags);
        pci_write_config_dword(pdev, SKX_M2M_PCI_PMON_BOX_CTL, IVBEP_PMON_BOX_CTL_INT);
}

static struct intel_uncore_ops skx_m2m_uncore_pci_ops = {
        .init_box       = skx_m2m_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = snbep_uncore_pci_disable_event,
        .enable_event   = snbep_uncore_pci_enable_event,
        .read_counter   = snbep_uncore_pci_read_counter,
};

static struct intel_uncore_type skx_uncore_m2m = {
        .name           = "m2m",
        .num_counters   = 4,
        .num_boxes      = 2,
        .perf_ctr_bits  = 48,
        .perf_ctr       = SKX_M2M_PCI_PMON_CTR0,
        .event_ctl      = SKX_M2M_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl        = SKX_M2M_PCI_PMON_BOX_CTL,
        .ops            = &skx_m2m_uncore_pci_ops,
        .format_group   = &skx_uncore_format_group,
};

static struct event_constraint skx_uncore_m2pcie_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x23, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type skx_uncore_m2pcie = {
        .name           = "m2pcie",
        .num_counters   = 4,
        .num_boxes      = 4,
        .perf_ctr_bits  = 48,
        .constraints    = skx_uncore_m2pcie_constraints,
        .perf_ctr       = SNBEP_PCI_PMON_CTR0,
        .event_ctl      = SNBEP_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl        = SNBEP_PCI_PMON_BOX_CTL,
        .ops            = &ivbep_uncore_pci_ops,
        .format_group   = &skx_uncore_format_group,
};

static struct event_constraint skx_uncore_m3upi_constraints[] = {
        UNCORE_EVENT_CONSTRAINT_RANGE(0x1d, 0x1e, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x40, 0x7),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x4e, 0x52, 0x7),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type skx_uncore_m3upi = {
        .name           = "m3upi",
        .num_counters   = 3,
        .num_boxes      = 3,
        .perf_ctr_bits  = 48,
        .constraints    = skx_uncore_m3upi_constraints,
        .perf_ctr       = SNBEP_PCI_PMON_CTR0,
        .event_ctl      = SNBEP_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl        = SNBEP_PCI_PMON_BOX_CTL,
        .ops            = &ivbep_uncore_pci_ops,
        .format_group   = &skx_uncore_format_group,
};

enum {
        SKX_PCI_UNCORE_IMC,
        SKX_PCI_UNCORE_M2M,
        SKX_PCI_UNCORE_UPI,
        SKX_PCI_UNCORE_M2PCIE,
        SKX_PCI_UNCORE_M3UPI,
};

static struct intel_uncore_type *skx_pci_uncores[] = {
        [SKX_PCI_UNCORE_IMC]    = &skx_uncore_imc,
        [SKX_PCI_UNCORE_M2M]    = &skx_uncore_m2m,
        [SKX_PCI_UNCORE_UPI]    = &skx_uncore_upi,
        [SKX_PCI_UNCORE_M2PCIE] = &skx_uncore_m2pcie,
        [SKX_PCI_UNCORE_M3UPI]  = &skx_uncore_m3upi,
        NULL,
};

static const struct pci_device_id skx_uncore_pci_ids[] = {
        { /* MC0 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2042),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(10, 2, SKX_PCI_UNCORE_IMC, 0),
        },
        { /* MC0 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2046),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(10, 6, SKX_PCI_UNCORE_IMC, 1),
        },
        { /* MC0 Channel 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x204a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(11, 2, SKX_PCI_UNCORE_IMC, 2),
        },
        { /* MC1 Channel 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2042),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(12, 2, SKX_PCI_UNCORE_IMC, 3),
        },
        { /* MC1 Channel 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2046),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(12, 6, SKX_PCI_UNCORE_IMC, 4),
        },
        { /* MC1 Channel 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x204a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(13, 2, SKX_PCI_UNCORE_IMC, 5),
        },
        { /* M2M0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2066),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(8, 0, SKX_PCI_UNCORE_M2M, 0),
        },
        { /* M2M1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2066),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(9, 0, SKX_PCI_UNCORE_M2M, 1),
        },
        { /* UPI0 Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2058),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(14, 0, SKX_PCI_UNCORE_UPI, 0),
        },
        { /* UPI0 Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2058),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(15, 0, SKX_PCI_UNCORE_UPI, 1),
        },
        { /* UPI1 Link 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2058),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(16, 0, SKX_PCI_UNCORE_UPI, 2),
        },
        { /* M2PCIe 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2088),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(21, 1, SKX_PCI_UNCORE_M2PCIE, 0),
        },
        { /* M2PCIe 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2088),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(22, 1, SKX_PCI_UNCORE_M2PCIE, 1),
        },
        { /* M2PCIe 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2088),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(23, 1, SKX_PCI_UNCORE_M2PCIE, 2),
        },
        { /* M2PCIe 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x2088),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(21, 5, SKX_PCI_UNCORE_M2PCIE, 3),
        },
        { /* M3UPI0 Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x204D),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(18, 1, SKX_PCI_UNCORE_M3UPI, 0),
        },
        { /* M3UPI0 Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x204E),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(18, 2, SKX_PCI_UNCORE_M3UPI, 1),
        },
        { /* M3UPI1 Link 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x204D),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(18, 5, SKX_PCI_UNCORE_M3UPI, 2),
        },
        { /* end: all zeroes */ }
};


static struct pci_driver skx_uncore_pci_driver = {
        .name           = "skx_uncore",
        .id_table       = skx_uncore_pci_ids,
};

int skx_uncore_pci_init(void)
{
        /* need to double check pci address */
        int ret = snbep_pci2phy_map_init(0x2014, SKX_CPUNODEID, SKX_GIDNIDMAP, false);

        if (ret)
                return ret;

        uncore_pci_uncores = skx_pci_uncores;
        uncore_pci_driver = &skx_uncore_pci_driver;
        return 0;
}

/* end of SKX uncore support */

/* SNR uncore support */

static struct intel_uncore_type snr_uncore_ubox = {
        .name                   = "ubox",
        .num_counters           = 2,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .fixed_ctr_bits         = 48,
        .perf_ctr               = SNR_U_MSR_PMON_CTR0,
        .event_ctl              = SNR_U_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .fixed_ctr              = SNR_U_MSR_PMON_UCLK_FIXED_CTR,
        .fixed_ctl              = SNR_U_MSR_PMON_UCLK_FIXED_CTL,
        .ops                    = &ivbep_uncore_msr_ops,
        .format_group           = &ivbep_uncore_format_group,
};

static struct attribute *snr_uncore_cha_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask_ext2.attr,
        &format_attr_edge.attr,
        &format_attr_tid_en.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_filter_tid5.attr,
        NULL,
};
static const struct attribute_group snr_uncore_chabox_format_group = {
        .name = "format",
        .attrs = snr_uncore_cha_formats_attr,
};

static int snr_cha_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;

        reg1->reg = SNR_C0_MSR_PMON_BOX_FILTER0 +
                    box->pmu->type->msr_offset * box->pmu->pmu_idx;
        reg1->config = event->attr.config1 & SKX_CHA_MSR_PMON_BOX_FILTER_TID;
        reg1->idx = 0;

        return 0;
}

static void snr_cha_enable_event(struct intel_uncore_box *box,
                                   struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;

        if (reg1->idx != EXTRA_REG_NONE)
                wrmsrq(reg1->reg, reg1->config);

        wrmsrq(hwc->config_base, hwc->config | SNBEP_PMON_CTL_EN);
}

static struct intel_uncore_ops snr_uncore_chabox_ops = {
        .init_box               = ivbep_uncore_msr_init_box,
        .disable_box            = snbep_uncore_msr_disable_box,
        .enable_box             = snbep_uncore_msr_enable_box,
        .disable_event          = snbep_uncore_msr_disable_event,
        .enable_event           = snr_cha_enable_event,
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = snr_cha_hw_config,
};

static struct intel_uncore_type snr_uncore_chabox = {
        .name                   = "cha",
        .num_counters           = 4,
        .num_boxes              = 6,
        .perf_ctr_bits          = 48,
        .event_ctl              = SNR_CHA_MSR_PMON_CTL0,
        .perf_ctr               = SNR_CHA_MSR_PMON_CTR0,
        .box_ctl                = SNR_CHA_MSR_PMON_BOX_CTL,
        .msr_offset             = HSWEP_CBO_MSR_OFFSET,
        .event_mask             = HSWEP_S_MSR_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = SNR_CHA_RAW_EVENT_MASK_EXT,
        .ops                    = &snr_uncore_chabox_ops,
        .format_group           = &snr_uncore_chabox_format_group,
};

static struct attribute *snr_uncore_iio_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh9.attr,
        &format_attr_ch_mask2.attr,
        &format_attr_fc_mask2.attr,
        NULL,
};

static const struct attribute_group snr_uncore_iio_format_group = {
        .name = "format",
        .attrs = snr_uncore_iio_formats_attr,
};

static umode_t
snr_iio_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
{
        /* Root bus 0x00 is valid only for pmu_idx = 1. */
        return pmu_iio_mapping_visible(kobj, attr, die, 1);
}

static struct attribute_group snr_iio_mapping_group = {
        .is_visible     = snr_iio_mapping_visible,
};

static const struct attribute_group *snr_iio_attr_update[] = {
        &snr_iio_mapping_group,
        NULL,
};

static int sad_cfg_iio_topology(struct intel_uncore_type *type, u8 *sad_pmon_mapping)
{
        u32 sad_cfg;
        int die, stack_id, ret = -EPERM;
        struct pci_dev *dev = NULL;

        while ((dev = pci_get_device(PCI_VENDOR_ID_INTEL, SNR_ICX_MESH2IIO_MMAP_DID, dev))) {
                ret = pci_read_config_dword(dev, SNR_ICX_SAD_CONTROL_CFG, &sad_cfg);
                if (ret) {
                        ret = pcibios_err_to_errno(ret);
                        break;
                }

                die = uncore_pcibus_to_dieid(dev->bus);
                stack_id = SAD_CONTROL_STACK_ID(sad_cfg);
                if (die < 0 || stack_id >= type->num_boxes) {
                        ret = -EPERM;
                        break;
                }

                /* Convert stack id from SAD_CONTROL to PMON notation. */
                stack_id = sad_pmon_mapping[stack_id];

                type->topology[die][stack_id].iio->segment = pci_domain_nr(dev->bus);
                type->topology[die][stack_id].pmu_idx = stack_id;
                type->topology[die][stack_id].iio->pci_bus_no = dev->bus->number;
        }

        pci_dev_put(dev);

        return ret;
}

/*
 * SNR has a static mapping of stack IDs from SAD_CONTROL_CFG notation to PMON
 */
enum {
        SNR_QAT_PMON_ID,
        SNR_CBDMA_DMI_PMON_ID,
        SNR_NIS_PMON_ID,
        SNR_DLB_PMON_ID,
        SNR_PCIE_GEN3_PMON_ID
};

static u8 snr_sad_pmon_mapping[] = {
        SNR_CBDMA_DMI_PMON_ID,
        SNR_PCIE_GEN3_PMON_ID,
        SNR_DLB_PMON_ID,
        SNR_NIS_PMON_ID,
        SNR_QAT_PMON_ID
};

static int snr_iio_get_topology(struct intel_uncore_type *type)
{
        return sad_cfg_iio_topology(type, snr_sad_pmon_mapping);
}

static void snr_iio_set_mapping(struct intel_uncore_type *type)
{
        pmu_iio_set_mapping(type, &snr_iio_mapping_group);
}

static void snr_iio_cleanup_mapping(struct intel_uncore_type *type)
{
        pmu_cleanup_mapping(type, &snr_iio_mapping_group);
}

static struct event_constraint snr_uncore_iio_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x83, 0x3),
        UNCORE_EVENT_CONSTRAINT(0xc0, 0xc),
        UNCORE_EVENT_CONSTRAINT(0xd5, 0xc),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type snr_uncore_iio = {
        .name                   = "iio",
        .num_counters           = 4,
        .num_boxes              = 5,
        .perf_ctr_bits          = 48,
        .event_ctl              = SNR_IIO_MSR_PMON_CTL0,
        .perf_ctr               = SNR_IIO_MSR_PMON_CTR0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = SNR_IIO_PMON_RAW_EVENT_MASK_EXT,
        .box_ctl                = SNR_IIO_MSR_PMON_BOX_CTL,
        .msr_offset             = SNR_IIO_MSR_OFFSET,
        .constraints            = snr_uncore_iio_constraints,
        .ops                    = &ivbep_uncore_msr_ops,
        .format_group           = &snr_uncore_iio_format_group,
        .attr_update            = snr_iio_attr_update,
        .get_topology           = snr_iio_get_topology,
        .set_mapping            = snr_iio_set_mapping,
        .cleanup_mapping        = snr_iio_cleanup_mapping,
};

static struct intel_uncore_type snr_uncore_irp = {
        .name                   = "irp",
        .num_counters           = 2,
        .num_boxes              = 5,
        .perf_ctr_bits          = 48,
        .event_ctl              = SNR_IRP0_MSR_PMON_CTL0,
        .perf_ctr               = SNR_IRP0_MSR_PMON_CTR0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNR_IRP0_MSR_PMON_BOX_CTL,
        .msr_offset             = SNR_IRP_MSR_OFFSET,
        .ops                    = &ivbep_uncore_msr_ops,
        .format_group           = &ivbep_uncore_format_group,
};

static struct intel_uncore_type snr_uncore_m2pcie = {
        .name           = "m2pcie",
        .num_counters   = 4,
        .num_boxes      = 5,
        .perf_ctr_bits  = 48,
        .event_ctl      = SNR_M2PCIE_MSR_PMON_CTL0,
        .perf_ctr       = SNR_M2PCIE_MSR_PMON_CTR0,
        .box_ctl        = SNR_M2PCIE_MSR_PMON_BOX_CTL,
        .msr_offset     = SNR_M2PCIE_MSR_OFFSET,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .ops            = &ivbep_uncore_msr_ops,
        .format_group   = &ivbep_uncore_format_group,
};

static int snr_pcu_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;
        int ev_sel = hwc->config & SNBEP_PMON_CTL_EV_SEL_MASK;

        if (ev_sel >= 0xb && ev_sel <= 0xe) {
                reg1->reg = SNR_PCU_MSR_PMON_BOX_FILTER;
                reg1->idx = ev_sel - 0xb;
                reg1->config = event->attr.config1 & (0xff << reg1->idx);
        }
        return 0;
}

static struct intel_uncore_ops snr_uncore_pcu_ops = {
        IVBEP_UNCORE_MSR_OPS_COMMON_INIT(),
        .hw_config              = snr_pcu_hw_config,
        .get_constraint         = snbep_pcu_get_constraint,
        .put_constraint         = snbep_pcu_put_constraint,
};

static struct intel_uncore_type snr_uncore_pcu = {
        .name                   = "pcu",
        .num_counters           = 4,
        .num_boxes              = 1,
        .perf_ctr_bits          = 48,
        .perf_ctr               = SNR_PCU_MSR_PMON_CTR0,
        .event_ctl              = SNR_PCU_MSR_PMON_CTL0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl                = SNR_PCU_MSR_PMON_BOX_CTL,
        .num_shared_regs        = 1,
        .ops                    = &snr_uncore_pcu_ops,
        .format_group           = &skx_uncore_pcu_format_group,
};

enum perf_uncore_snr_iio_freerunning_type_id {
        SNR_IIO_MSR_IOCLK,
        SNR_IIO_MSR_BW_IN,

        SNR_IIO_FREERUNNING_TYPE_MAX,
};

static struct freerunning_counters snr_iio_freerunning[] = {
        [SNR_IIO_MSR_IOCLK]     = { 0x1eac, 0x1, 0x10, 1, 48 },
        [SNR_IIO_MSR_BW_IN]     = { 0x1f00, 0x1, 0x10, 8, 48 },
};

static struct uncore_event_desc snr_uncore_iio_freerunning_events[] = {
        /* Free-Running IIO CLOCKS Counter */
        INTEL_UNCORE_EVENT_DESC(ioclk,                  "event=0xff,umask=0x10"),
        /* Free-Running IIO BANDWIDTH IN Counters */
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port0, 0x20, 3.0517578125e-5),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port1, 0x21, 3.0517578125e-5),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port2, 0x22, 3.0517578125e-5),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port3, 0x23, 3.0517578125e-5),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port4, 0x24, 3.0517578125e-5),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port5, 0x25, 3.0517578125e-5),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port6, 0x26, 3.0517578125e-5),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port7, 0x27, 3.0517578125e-5),
        { /* end: all zeroes */ },
};

static struct intel_uncore_type snr_uncore_iio_free_running = {
        .name                   = "iio_free_running",
        .num_counters           = 9,
        .num_boxes              = 5,
        .num_freerunning_types  = SNR_IIO_FREERUNNING_TYPE_MAX,
        .freerunning            = snr_iio_freerunning,
        .ops                    = &skx_uncore_iio_freerunning_ops,
        .event_descs            = snr_uncore_iio_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

static struct intel_uncore_type *snr_msr_uncores[] = {
        &snr_uncore_ubox,
        &snr_uncore_chabox,
        &snr_uncore_iio,
        &snr_uncore_irp,
        &snr_uncore_m2pcie,
        &snr_uncore_pcu,
        &snr_uncore_iio_free_running,
        NULL,
};

void snr_uncore_cpu_init(void)
{
        uncore_msr_uncores = snr_msr_uncores;
}

static void snr_m2m_uncore_pci_init_box(struct intel_uncore_box *box)
{
        struct pci_dev *pdev = box->pci_dev;
        int box_ctl = uncore_pci_box_ctl(box);

        __set_bit(UNCORE_BOX_FLAG_CTL_OFFS8, &box->flags);
        pci_write_config_dword(pdev, box_ctl, IVBEP_PMON_BOX_CTL_INT);
}

static struct intel_uncore_ops snr_m2m_uncore_pci_ops = {
        .init_box       = snr_m2m_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = snbep_uncore_pci_disable_event,
        .enable_event   = snbep_uncore_pci_enable_event,
        .read_counter   = snbep_uncore_pci_read_counter,
};

static struct attribute *snr_m2m_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask_ext3.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group snr_m2m_uncore_format_group = {
        .name = "format",
        .attrs = snr_m2m_uncore_formats_attr,
};

static struct intel_uncore_type snr_uncore_m2m = {
        .name           = "m2m",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 48,
        .perf_ctr       = SNR_M2M_PCI_PMON_CTR0,
        .event_ctl      = SNR_M2M_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext = SNR_M2M_PCI_PMON_UMASK_EXT,
        .box_ctl        = SNR_M2M_PCI_PMON_BOX_CTL,
        .ops            = &snr_m2m_uncore_pci_ops,
        .format_group   = &snr_m2m_uncore_format_group,
};

static void snr_uncore_pci_enable_event(struct intel_uncore_box *box, struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;

        pci_write_config_dword(pdev, hwc->config_base, (u32)(hwc->config | SNBEP_PMON_CTL_EN));
        pci_write_config_dword(pdev, hwc->config_base + 4, (u32)(hwc->config >> 32));
}

static struct intel_uncore_ops snr_pcie3_uncore_pci_ops = {
        .init_box       = snr_m2m_uncore_pci_init_box,
        .disable_box    = snbep_uncore_pci_disable_box,
        .enable_box     = snbep_uncore_pci_enable_box,
        .disable_event  = snbep_uncore_pci_disable_event,
        .enable_event   = snr_uncore_pci_enable_event,
        .read_counter   = snbep_uncore_pci_read_counter,
};

static struct intel_uncore_type snr_uncore_pcie3 = {
        .name           = "pcie3",
        .num_counters   = 4,
        .num_boxes      = 1,
        .perf_ctr_bits  = 48,
        .perf_ctr       = SNR_PCIE3_PCI_PMON_CTR0,
        .event_ctl      = SNR_PCIE3_PCI_PMON_CTL0,
        .event_mask     = SKX_IIO_PMON_RAW_EVENT_MASK,
        .event_mask_ext = SKX_IIO_PMON_RAW_EVENT_MASK_EXT,
        .box_ctl        = SNR_PCIE3_PCI_PMON_BOX_CTL,
        .ops            = &snr_pcie3_uncore_pci_ops,
        .format_group   = &skx_uncore_iio_format_group,
};

enum {
        SNR_PCI_UNCORE_M2M,
        SNR_PCI_UNCORE_PCIE3,
};

static struct intel_uncore_type *snr_pci_uncores[] = {
        [SNR_PCI_UNCORE_M2M]            = &snr_uncore_m2m,
        [SNR_PCI_UNCORE_PCIE3]          = &snr_uncore_pcie3,
        NULL,
};

static const struct pci_device_id snr_uncore_pci_ids[] = {
        { /* M2M */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x344a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(12, 0, SNR_PCI_UNCORE_M2M, 0),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver snr_uncore_pci_driver = {
        .name           = "snr_uncore",
        .id_table       = snr_uncore_pci_ids,
};

static const struct pci_device_id snr_uncore_pci_sub_ids[] = {
        { /* PCIe3 RP */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x334a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(4, 0, SNR_PCI_UNCORE_PCIE3, 0),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver snr_uncore_pci_sub_driver = {
        .name           = "snr_uncore_sub",
        .id_table       = snr_uncore_pci_sub_ids,
};

int snr_uncore_pci_init(void)
{
        /* SNR UBOX DID */
        int ret = snbep_pci2phy_map_init(0x3460, SKX_CPUNODEID,
                                         SKX_GIDNIDMAP, true);

        if (ret)
                return ret;

        uncore_pci_uncores = snr_pci_uncores;
        uncore_pci_driver = &snr_uncore_pci_driver;
        uncore_pci_sub_driver = &snr_uncore_pci_sub_driver;
        return 0;
}

#define SNR_MC_DEVICE_ID        0x3451

static struct pci_dev *snr_uncore_get_mc_dev(unsigned int device, int id)
{
        struct pci_dev *mc_dev = NULL;
        int pkg;

        while (1) {
                mc_dev = pci_get_device(PCI_VENDOR_ID_INTEL, device, mc_dev);
                if (!mc_dev)
                        break;
                pkg = uncore_pcibus_to_dieid(mc_dev->bus);
                if (pkg == id)
                        break;
        }
        return mc_dev;
}

static int snr_uncore_mmio_map(struct intel_uncore_box *box,
                               unsigned int box_ctl, int mem_offset,
                               unsigned int device)
{
        struct pci_dev *pdev = snr_uncore_get_mc_dev(device, box->dieid);
        struct intel_uncore_type *type = box->pmu->type;
        resource_size_t addr;
        u32 pci_dword;

        if (!pdev)
                return -ENODEV;

        pci_read_config_dword(pdev, SNR_IMC_MMIO_BASE_OFFSET, &pci_dword);
        addr = ((resource_size_t)pci_dword & SNR_IMC_MMIO_BASE_MASK) << 23;

        pci_read_config_dword(pdev, mem_offset, &pci_dword);
        addr |= (pci_dword & SNR_IMC_MMIO_MEM0_MASK) << 12;

        addr += box_ctl;

        pci_dev_put(pdev);

        box->io_addr = ioremap(addr, type->mmio_map_size);
        if (!box->io_addr) {
                pr_warn("perf uncore: Failed to ioremap for %s.\n", type->name);
                return -EINVAL;
        }

        return 0;
}

static void __snr_uncore_mmio_init_box(struct intel_uncore_box *box,
                                       unsigned int box_ctl, int mem_offset,
                                       unsigned int device)
{
        if (!snr_uncore_mmio_map(box, box_ctl, mem_offset, device))
                writel(IVBEP_PMON_BOX_CTL_INT, box->io_addr);
}

static void snr_uncore_mmio_init_box(struct intel_uncore_box *box)
{
        __snr_uncore_mmio_init_box(box, uncore_mmio_box_ctl(box),
                                   SNR_IMC_MMIO_MEM0_OFFSET,
                                   SNR_MC_DEVICE_ID);
}

static void snr_uncore_mmio_disable_box(struct intel_uncore_box *box)
{
        u32 config;

        if (!box->io_addr)
                return;

        config = readl(box->io_addr);
        config |= SNBEP_PMON_BOX_CTL_FRZ;
        writel(config, box->io_addr);
}

static void snr_uncore_mmio_enable_box(struct intel_uncore_box *box)
{
        u32 config;

        if (!box->io_addr)
                return;

        config = readl(box->io_addr);
        config &= ~SNBEP_PMON_BOX_CTL_FRZ;
        writel(config, box->io_addr);
}

static void snr_uncore_mmio_enable_event(struct intel_uncore_box *box,
                                           struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;

        if (!box->io_addr)
                return;

        if (!uncore_mmio_is_valid_offset(box, hwc->config_base))
                return;

        writel(hwc->config | SNBEP_PMON_CTL_EN,
               box->io_addr + hwc->config_base);
}

static void snr_uncore_mmio_disable_event(struct intel_uncore_box *box,
                                            struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;

        if (!box->io_addr)
                return;

        if (!uncore_mmio_is_valid_offset(box, hwc->config_base))
                return;

        writel(hwc->config, box->io_addr + hwc->config_base);
}

static struct intel_uncore_ops snr_uncore_mmio_ops = {
        .init_box       = snr_uncore_mmio_init_box,
        .exit_box       = uncore_mmio_exit_box,
        .disable_box    = snr_uncore_mmio_disable_box,
        .enable_box     = snr_uncore_mmio_enable_box,
        .disable_event  = snr_uncore_mmio_disable_event,
        .enable_event   = snr_uncore_mmio_enable_event,
        .read_counter   = uncore_mmio_read_counter,
};

static struct uncore_event_desc snr_uncore_imc_events[] = {
        INTEL_UNCORE_EVENT_DESC(clockticks,      "event=0x00,umask=0x00"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read,  "event=0x04,umask=0x0f"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.unit, "MiB"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write, "event=0x04,umask=0x30"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.unit, "MiB"),
        { /* end: all zeroes */ },
};

static struct intel_uncore_type snr_uncore_imc = {
        .name           = "imc",
        .num_counters   = 4,
        .num_boxes      = 2,
        .perf_ctr_bits  = 48,
        .fixed_ctr_bits = 48,
        .fixed_ctr      = SNR_IMC_MMIO_PMON_FIXED_CTR,
        .fixed_ctl      = SNR_IMC_MMIO_PMON_FIXED_CTL,
        .event_descs    = snr_uncore_imc_events,
        .perf_ctr       = SNR_IMC_MMIO_PMON_CTR0,
        .event_ctl      = SNR_IMC_MMIO_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl        = SNR_IMC_MMIO_PMON_BOX_CTL,
        .mmio_offset    = SNR_IMC_MMIO_OFFSET,
        .mmio_map_size  = SNR_IMC_MMIO_SIZE,
        .ops            = &snr_uncore_mmio_ops,
        .format_group   = &skx_uncore_format_group,
};

enum perf_uncore_snr_imc_freerunning_type_id {
        SNR_IMC_DCLK,
        SNR_IMC_DDR,

        SNR_IMC_FREERUNNING_TYPE_MAX,
};

static struct freerunning_counters snr_imc_freerunning[] = {
        [SNR_IMC_DCLK]  = { 0x22b0, 0x0, 0, 1, 48 },
        [SNR_IMC_DDR]   = { 0x2290, 0x8, 0, 2, 48 },
};

static struct uncore_event_desc snr_uncore_imc_freerunning_events[] = {
        INTEL_UNCORE_EVENT_DESC(dclk,           "event=0xff,umask=0x10"),

        INTEL_UNCORE_FR_EVENT_DESC(read,  0x20, 6.103515625e-5),
        INTEL_UNCORE_FR_EVENT_DESC(write, 0x21, 6.103515625e-5),
        { /* end: all zeroes */ },
};

static struct intel_uncore_ops snr_uncore_imc_freerunning_ops = {
        .init_box       = snr_uncore_mmio_init_box,
        .exit_box       = uncore_mmio_exit_box,
        .read_counter   = uncore_mmio_read_counter,
        .hw_config      = uncore_freerunning_hw_config,
};

static struct intel_uncore_type snr_uncore_imc_free_running = {
        .name                   = "imc_free_running",
        .num_counters           = 3,
        .num_boxes              = 1,
        .num_freerunning_types  = SNR_IMC_FREERUNNING_TYPE_MAX,
        .mmio_map_size          = SNR_IMC_MMIO_SIZE,
        .freerunning            = snr_imc_freerunning,
        .ops                    = &snr_uncore_imc_freerunning_ops,
        .event_descs            = snr_uncore_imc_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

static struct intel_uncore_type *snr_mmio_uncores[] = {
        &snr_uncore_imc,
        &snr_uncore_imc_free_running,
        NULL,
};

void snr_uncore_mmio_init(void)
{
        uncore_mmio_uncores = snr_mmio_uncores;
}

/* end of SNR uncore support */

/* ICX uncore support */

static u64 icx_cha_msr_offsets[] = {
        0x2a0, 0x2ae, 0x2bc, 0x2ca, 0x2d8, 0x2e6, 0x2f4, 0x302, 0x310,
        0x31e, 0x32c, 0x33a, 0x348, 0x356, 0x364, 0x372, 0x380, 0x38e,
        0x3aa, 0x3b8, 0x3c6, 0x3d4, 0x3e2, 0x3f0, 0x3fe, 0x40c, 0x41a,
        0x428, 0x436, 0x444, 0x452, 0x460, 0x46e, 0x47c, 0x0,   0xe,
        0x1c,  0x2a,  0x38,  0x46,
};

static int icx_cha_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        bool tie_en = !!(event->hw.config & SNBEP_CBO_PMON_CTL_TID_EN);

        if (tie_en) {
                reg1->reg = ICX_C34_MSR_PMON_BOX_FILTER0 +
                            icx_cha_msr_offsets[box->pmu->pmu_idx];
                reg1->config = event->attr.config1 & SKX_CHA_MSR_PMON_BOX_FILTER_TID;
                reg1->idx = 0;
        }

        return 0;
}

static struct intel_uncore_ops icx_uncore_chabox_ops = {
        .init_box               = ivbep_uncore_msr_init_box,
        .disable_box            = snbep_uncore_msr_disable_box,
        .enable_box             = snbep_uncore_msr_enable_box,
        .disable_event          = snbep_uncore_msr_disable_event,
        .enable_event           = snr_cha_enable_event,
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = icx_cha_hw_config,
};

static struct intel_uncore_type icx_uncore_chabox = {
        .name                   = "cha",
        .num_counters           = 4,
        .perf_ctr_bits          = 48,
        .event_ctl              = ICX_C34_MSR_PMON_CTL0,
        .perf_ctr               = ICX_C34_MSR_PMON_CTR0,
        .box_ctl                = ICX_C34_MSR_PMON_BOX_CTL,
        .msr_offsets            = icx_cha_msr_offsets,
        .event_mask             = HSWEP_S_MSR_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = SNR_CHA_RAW_EVENT_MASK_EXT,
        .constraints            = skx_uncore_chabox_constraints,
        .ops                    = &icx_uncore_chabox_ops,
        .format_group           = &snr_uncore_chabox_format_group,
};

static u64 icx_msr_offsets[] = {
        0x0, 0x20, 0x40, 0x90, 0xb0, 0xd0,
};

static struct event_constraint icx_uncore_iio_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x02, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x03, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x83, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x88, 0xc),
        UNCORE_EVENT_CONSTRAINT(0xc0, 0xc),
        UNCORE_EVENT_CONSTRAINT(0xc5, 0xc),
        UNCORE_EVENT_CONSTRAINT(0xd5, 0xc),
        EVENT_CONSTRAINT_END
};

static umode_t
icx_iio_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
{
        /* Root bus 0x00 is valid only for pmu_idx = 5. */
        return pmu_iio_mapping_visible(kobj, attr, die, 5);
}

static struct attribute_group icx_iio_mapping_group = {
        .is_visible     = icx_iio_mapping_visible,
};

static const struct attribute_group *icx_iio_attr_update[] = {
        &icx_iio_mapping_group,
        NULL,
};

/*
 * ICX has a static mapping of stack IDs from SAD_CONTROL_CFG notation to PMON
 */
enum {
        ICX_PCIE1_PMON_ID,
        ICX_PCIE2_PMON_ID,
        ICX_PCIE3_PMON_ID,
        ICX_PCIE4_PMON_ID,
        ICX_PCIE5_PMON_ID,
        ICX_CBDMA_DMI_PMON_ID
};

static u8 icx_sad_pmon_mapping[] = {
        ICX_CBDMA_DMI_PMON_ID,
        ICX_PCIE1_PMON_ID,
        ICX_PCIE2_PMON_ID,
        ICX_PCIE3_PMON_ID,
        ICX_PCIE4_PMON_ID,
        ICX_PCIE5_PMON_ID,
};

static int icx_iio_get_topology(struct intel_uncore_type *type)
{
        return sad_cfg_iio_topology(type, icx_sad_pmon_mapping);
}

static void icx_iio_set_mapping(struct intel_uncore_type *type)
{
        /* Detect ICX-D system. This case is not supported */
        if (boot_cpu_data.x86_vfm == INTEL_ICELAKE_D) {
                pmu_clear_mapping_attr(type->attr_update, &icx_iio_mapping_group);
                return;
        }
        pmu_iio_set_mapping(type, &icx_iio_mapping_group);
}

static void icx_iio_cleanup_mapping(struct intel_uncore_type *type)
{
        pmu_cleanup_mapping(type, &icx_iio_mapping_group);
}

static struct intel_uncore_type icx_uncore_iio = {
        .name                   = "iio",
        .num_counters           = 4,
        .num_boxes              = 6,
        .perf_ctr_bits          = 48,
        .event_ctl              = ICX_IIO_MSR_PMON_CTL0,
        .perf_ctr               = ICX_IIO_MSR_PMON_CTR0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = SNR_IIO_PMON_RAW_EVENT_MASK_EXT,
        .box_ctl                = ICX_IIO_MSR_PMON_BOX_CTL,
        .msr_offsets            = icx_msr_offsets,
        .constraints            = icx_uncore_iio_constraints,
        .ops                    = &skx_uncore_iio_ops,
        .format_group           = &snr_uncore_iio_format_group,
        .attr_update            = icx_iio_attr_update,
        .get_topology           = icx_iio_get_topology,
        .set_mapping            = icx_iio_set_mapping,
        .cleanup_mapping        = icx_iio_cleanup_mapping,
};

static struct intel_uncore_type icx_uncore_irp = {
        .name                   = "irp",
        .num_counters           = 2,
        .num_boxes              = 6,
        .perf_ctr_bits          = 48,
        .event_ctl              = ICX_IRP0_MSR_PMON_CTL0,
        .perf_ctr               = ICX_IRP0_MSR_PMON_CTR0,
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl                = ICX_IRP0_MSR_PMON_BOX_CTL,
        .msr_offsets            = icx_msr_offsets,
        .ops                    = &ivbep_uncore_msr_ops,
        .format_group           = &ivbep_uncore_format_group,
};

static struct event_constraint icx_uncore_m2pcie_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x14, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x23, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x2d, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type icx_uncore_m2pcie = {
        .name           = "m2pcie",
        .num_counters   = 4,
        .num_boxes      = 6,
        .perf_ctr_bits  = 48,
        .event_ctl      = ICX_M2PCIE_MSR_PMON_CTL0,
        .perf_ctr       = ICX_M2PCIE_MSR_PMON_CTR0,
        .box_ctl        = ICX_M2PCIE_MSR_PMON_BOX_CTL,
        .msr_offsets    = icx_msr_offsets,
        .constraints    = icx_uncore_m2pcie_constraints,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .ops            = &ivbep_uncore_msr_ops,
        .format_group   = &ivbep_uncore_format_group,
};

enum perf_uncore_icx_iio_freerunning_type_id {
        ICX_IIO_MSR_IOCLK,
        ICX_IIO_MSR_BW_IN,

        ICX_IIO_FREERUNNING_TYPE_MAX,
};

static unsigned icx_iio_clk_freerunning_box_offsets[] = {
        0x0, 0x20, 0x40, 0x90, 0xb0, 0xd0,
};

static unsigned icx_iio_bw_freerunning_box_offsets[] = {
        0x0, 0x10, 0x20, 0x90, 0xa0, 0xb0,
};

static struct freerunning_counters icx_iio_freerunning[] = {
        [ICX_IIO_MSR_IOCLK]     = { 0xa55, 0x1, 0x20, 1, 48, icx_iio_clk_freerunning_box_offsets },
        [ICX_IIO_MSR_BW_IN]     = { 0xaa0, 0x1, 0x10, 8, 48, icx_iio_bw_freerunning_box_offsets },
};

static struct intel_uncore_type icx_uncore_iio_free_running = {
        .name                   = "iio_free_running",
        .num_counters           = 9,
        .num_boxes              = 6,
        .num_freerunning_types  = ICX_IIO_FREERUNNING_TYPE_MAX,
        .freerunning            = icx_iio_freerunning,
        .ops                    = &skx_uncore_iio_freerunning_ops,
        .event_descs            = snr_uncore_iio_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

static struct intel_uncore_type *icx_msr_uncores[] = {
        &skx_uncore_ubox,
        &icx_uncore_chabox,
        &icx_uncore_iio,
        &icx_uncore_irp,
        &icx_uncore_m2pcie,
        &skx_uncore_pcu,
        &icx_uncore_iio_free_running,
        NULL,
};

/*
 * To determine the number of CHAs, it should read CAPID6(Low) and CAPID7 (High)
 * registers which located at Device 30, Function 3
 */
#define ICX_CAPID6              0x9c
#define ICX_CAPID7              0xa0

static u64 icx_count_chabox(void)
{
        struct pci_dev *dev = NULL;
        u64 caps = 0;

        dev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x345b, dev);
        if (!dev)
                goto out;

        pci_read_config_dword(dev, ICX_CAPID6, (u32 *)&caps);
        pci_read_config_dword(dev, ICX_CAPID7, (u32 *)&caps + 1);
out:
        pci_dev_put(dev);
        return hweight64(caps);
}

void icx_uncore_cpu_init(void)
{
        u64 num_boxes = icx_count_chabox();

        if (WARN_ON(num_boxes > ARRAY_SIZE(icx_cha_msr_offsets)))
                return;
        icx_uncore_chabox.num_boxes = num_boxes;
        uncore_msr_uncores = icx_msr_uncores;
}

static struct intel_uncore_type icx_uncore_m2m = {
        .name           = "m2m",
        .num_counters   = 4,
        .num_boxes      = 4,
        .perf_ctr_bits  = 48,
        .perf_ctr       = SNR_M2M_PCI_PMON_CTR0,
        .event_ctl      = SNR_M2M_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext = SNR_M2M_PCI_PMON_UMASK_EXT,
        .box_ctl        = SNR_M2M_PCI_PMON_BOX_CTL,
        .ops            = &snr_m2m_uncore_pci_ops,
        .format_group   = &snr_m2m_uncore_format_group,
};

static struct attribute *icx_upi_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask_ext4.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group icx_upi_uncore_format_group = {
        .name = "format",
        .attrs = icx_upi_uncore_formats_attr,
};

#define ICX_UPI_REGS_ADDR_DEVICE_LINK0  0x02
#define ICX_UPI_REGS_ADDR_FUNCTION      0x01

static int discover_upi_topology(struct intel_uncore_type *type, int ubox_did, int dev_link0)
{
        struct pci_dev *ubox = NULL;
        struct pci_dev *dev = NULL;
        u32 nid, gid;
        int idx, lgc_pkg, ret = -EPERM;
        struct intel_uncore_topology *upi;
        unsigned int devfn;

        /* GIDNIDMAP method supports machines which have less than 8 sockets. */
        if (uncore_max_dies() > 8)
                goto err;

        while ((ubox = pci_get_device(PCI_VENDOR_ID_INTEL, ubox_did, ubox))) {
                ret = upi_nodeid_groupid(ubox, SKX_CPUNODEID, SKX_GIDNIDMAP, &nid, &gid);
                if (ret) {
                        ret = pcibios_err_to_errno(ret);
                        break;
                }

                lgc_pkg = topology_gidnid_map(nid, gid);
                if (lgc_pkg < 0) {
                        ret = -EPERM;
                        goto err;
                }
                for (idx = 0; idx < type->num_boxes; idx++) {
                        upi = &type->topology[lgc_pkg][idx];
                        devfn = PCI_DEVFN(dev_link0 + idx, ICX_UPI_REGS_ADDR_FUNCTION);
                        dev = pci_get_domain_bus_and_slot(pci_domain_nr(ubox->bus),
                                                          ubox->bus->number,
                                                          devfn);
                        if (dev) {
                                ret = upi_fill_topology(dev, upi, idx);
                                if (ret)
                                        goto err;
                        }
                }
        }
err:
        pci_dev_put(ubox);
        pci_dev_put(dev);
        return ret;
}

static int icx_upi_get_topology(struct intel_uncore_type *type)
{
        return discover_upi_topology(type, ICX_UBOX_DID, ICX_UPI_REGS_ADDR_DEVICE_LINK0);
}

static struct attribute_group icx_upi_mapping_group = {
        .is_visible     = skx_upi_mapping_visible,
};

static const struct attribute_group *icx_upi_attr_update[] = {
        &icx_upi_mapping_group,
        NULL
};

static void icx_upi_set_mapping(struct intel_uncore_type *type)
{
        pmu_upi_set_mapping(type, &icx_upi_mapping_group);
}

static void icx_upi_cleanup_mapping(struct intel_uncore_type *type)
{
        pmu_cleanup_mapping(type, &icx_upi_mapping_group);
}

static struct intel_uncore_type icx_uncore_upi = {
        .name           = "upi",
        .num_counters   = 4,
        .num_boxes      = 3,
        .perf_ctr_bits  = 48,
        .perf_ctr       = ICX_UPI_PCI_PMON_CTR0,
        .event_ctl      = ICX_UPI_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext = ICX_UPI_CTL_UMASK_EXT,
        .box_ctl        = ICX_UPI_PCI_PMON_BOX_CTL,
        .ops            = &skx_upi_uncore_pci_ops,
        .format_group   = &icx_upi_uncore_format_group,
        .attr_update    = icx_upi_attr_update,
        .get_topology   = icx_upi_get_topology,
        .set_mapping    = icx_upi_set_mapping,
        .cleanup_mapping = icx_upi_cleanup_mapping,
};

static struct event_constraint icx_uncore_m3upi_constraints[] = {
        UNCORE_EVENT_CONSTRAINT_RANGE(0x1c, 0x1f, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x40, 0x7),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x4e, 0x50, 0x7),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type icx_uncore_m3upi = {
        .name           = "m3upi",
        .num_counters   = 4,
        .num_boxes      = 3,
        .perf_ctr_bits  = 48,
        .perf_ctr       = ICX_M3UPI_PCI_PMON_CTR0,
        .event_ctl      = ICX_M3UPI_PCI_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl        = ICX_M3UPI_PCI_PMON_BOX_CTL,
        .constraints    = icx_uncore_m3upi_constraints,
        .ops            = &ivbep_uncore_pci_ops,
        .format_group   = &skx_uncore_format_group,
};

enum {
        ICX_PCI_UNCORE_M2M,
        ICX_PCI_UNCORE_UPI,
        ICX_PCI_UNCORE_M3UPI,
};

static struct intel_uncore_type *icx_pci_uncores[] = {
        [ICX_PCI_UNCORE_M2M]            = &icx_uncore_m2m,
        [ICX_PCI_UNCORE_UPI]            = &icx_uncore_upi,
        [ICX_PCI_UNCORE_M3UPI]          = &icx_uncore_m3upi,
        NULL,
};

static const struct pci_device_id icx_uncore_pci_ids[] = {
        { /* M2M 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x344a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(12, 0, ICX_PCI_UNCORE_M2M, 0),
        },
        { /* M2M 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x344a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(13, 0, ICX_PCI_UNCORE_M2M, 1),
        },
        { /* M2M 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x344a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(14, 0, ICX_PCI_UNCORE_M2M, 2),
        },
        { /* M2M 3 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x344a),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(15, 0, ICX_PCI_UNCORE_M2M, 3),
        },
        { /* UPI Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3441),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(2, 1, ICX_PCI_UNCORE_UPI, 0),
        },
        { /* UPI Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3441),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(3, 1, ICX_PCI_UNCORE_UPI, 1),
        },
        { /* UPI Link 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3441),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(4, 1, ICX_PCI_UNCORE_UPI, 2),
        },
        { /* M3UPI Link 0 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3446),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(5, 1, ICX_PCI_UNCORE_M3UPI, 0),
        },
        { /* M3UPI Link 1 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3446),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(6, 1, ICX_PCI_UNCORE_M3UPI, 1),
        },
        { /* M3UPI Link 2 */
                PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x3446),
                .driver_data = UNCORE_PCI_DEV_FULL_DATA(7, 1, ICX_PCI_UNCORE_M3UPI, 2),
        },
        { /* end: all zeroes */ }
};

static struct pci_driver icx_uncore_pci_driver = {
        .name           = "icx_uncore",
        .id_table       = icx_uncore_pci_ids,
};

int icx_uncore_pci_init(void)
{
        /* ICX UBOX DID */
        int ret = snbep_pci2phy_map_init(0x3450, SKX_CPUNODEID,
                                         SKX_GIDNIDMAP, true);

        if (ret)
                return ret;

        uncore_pci_uncores = icx_pci_uncores;
        uncore_pci_driver = &icx_uncore_pci_driver;
        return 0;
}

static void icx_uncore_imc_init_box(struct intel_uncore_box *box)
{
        unsigned int box_ctl = box->pmu->type->box_ctl +
                               box->pmu->type->mmio_offset * (box->pmu->pmu_idx % ICX_NUMBER_IMC_CHN);
        int mem_offset = (box->pmu->pmu_idx / ICX_NUMBER_IMC_CHN) * ICX_IMC_MEM_STRIDE +
                         SNR_IMC_MMIO_MEM0_OFFSET;

        __snr_uncore_mmio_init_box(box, box_ctl, mem_offset,
                                   SNR_MC_DEVICE_ID);
}

static struct intel_uncore_ops icx_uncore_mmio_ops = {
        .init_box       = icx_uncore_imc_init_box,
        .exit_box       = uncore_mmio_exit_box,
        .disable_box    = snr_uncore_mmio_disable_box,
        .enable_box     = snr_uncore_mmio_enable_box,
        .disable_event  = snr_uncore_mmio_disable_event,
        .enable_event   = snr_uncore_mmio_enable_event,
        .read_counter   = uncore_mmio_read_counter,
};

static struct intel_uncore_type icx_uncore_imc = {
        .name           = "imc",
        .num_counters   = 4,
        .num_boxes      = 12,
        .perf_ctr_bits  = 48,
        .fixed_ctr_bits = 48,
        .fixed_ctr      = SNR_IMC_MMIO_PMON_FIXED_CTR,
        .fixed_ctl      = SNR_IMC_MMIO_PMON_FIXED_CTL,
        .event_descs    = snr_uncore_imc_events,
        .perf_ctr       = SNR_IMC_MMIO_PMON_CTR0,
        .event_ctl      = SNR_IMC_MMIO_PMON_CTL0,
        .event_mask     = SNBEP_PMON_RAW_EVENT_MASK,
        .box_ctl        = SNR_IMC_MMIO_PMON_BOX_CTL,
        .mmio_offset    = SNR_IMC_MMIO_OFFSET,
        .mmio_map_size  = SNR_IMC_MMIO_SIZE,
        .ops            = &icx_uncore_mmio_ops,
        .format_group   = &skx_uncore_format_group,
};

enum perf_uncore_icx_imc_freerunning_type_id {
        ICX_IMC_DCLK,
        ICX_IMC_DDR,
        ICX_IMC_DDRT,

        ICX_IMC_FREERUNNING_TYPE_MAX,
};

static struct freerunning_counters icx_imc_freerunning[] = {
        [ICX_IMC_DCLK]  = { 0x22b0, 0x0, 0, 1, 48 },
        [ICX_IMC_DDR]   = { 0x2290, 0x8, 0, 2, 48 },
        [ICX_IMC_DDRT]  = { 0x22a0, 0x8, 0, 2, 48 },
};

static struct uncore_event_desc icx_uncore_imc_freerunning_events[] = {
        INTEL_UNCORE_EVENT_DESC(dclk,                   "event=0xff,umask=0x10"),

        INTEL_UNCORE_FR_EVENT_DESC(read,       0x20, 6.103515625e-5),
        INTEL_UNCORE_FR_EVENT_DESC(write,      0x21, 6.103515625e-5),
        INTEL_UNCORE_FR_EVENT_DESC(ddrt_read,  0x30, 6.103515625e-5),
        INTEL_UNCORE_FR_EVENT_DESC(ddrt_write, 0x31, 6.103515625e-5),
        { /* end: all zeroes */ },
};

static void icx_uncore_imc_freerunning_init_box(struct intel_uncore_box *box)
{
        int mem_offset = box->pmu->pmu_idx * ICX_IMC_MEM_STRIDE +
                         SNR_IMC_MMIO_MEM0_OFFSET;

        snr_uncore_mmio_map(box, uncore_mmio_box_ctl(box),
                            mem_offset, SNR_MC_DEVICE_ID);
}

static struct intel_uncore_ops icx_uncore_imc_freerunning_ops = {
        .init_box       = icx_uncore_imc_freerunning_init_box,
        .exit_box       = uncore_mmio_exit_box,
        .read_counter   = uncore_mmio_read_counter,
        .hw_config      = uncore_freerunning_hw_config,
};

static struct intel_uncore_type icx_uncore_imc_free_running = {
        .name                   = "imc_free_running",
        .num_counters           = 5,
        .num_boxes              = 4,
        .num_freerunning_types  = ICX_IMC_FREERUNNING_TYPE_MAX,
        .mmio_map_size          = SNR_IMC_MMIO_SIZE,
        .freerunning            = icx_imc_freerunning,
        .ops                    = &icx_uncore_imc_freerunning_ops,
        .event_descs            = icx_uncore_imc_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

static struct intel_uncore_type *icx_mmio_uncores[] = {
        &icx_uncore_imc,
        &icx_uncore_imc_free_running,
        NULL,
};

void icx_uncore_mmio_init(void)
{
        uncore_mmio_uncores = icx_mmio_uncores;
}

/* end of ICX uncore support */

/* SPR uncore support */

static void spr_uncore_msr_enable_event(struct intel_uncore_box *box,
                                        struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;

        if (reg1->idx != EXTRA_REG_NONE)
                wrmsrq(reg1->reg, reg1->config);

        wrmsrq(hwc->config_base, hwc->config);
}

static void spr_uncore_msr_disable_event(struct intel_uncore_box *box,
                                         struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;
        struct hw_perf_event_extra *reg1 = &hwc->extra_reg;

        if (reg1->idx != EXTRA_REG_NONE)
                wrmsrq(reg1->reg, 0);

        wrmsrq(hwc->config_base, 0);
}

static int spr_cha_hw_config(struct intel_uncore_box *box, struct perf_event *event)
{
        struct hw_perf_event_extra *reg1 = &event->hw.extra_reg;
        bool tie_en = !!(event->hw.config & SPR_CHA_PMON_CTL_TID_EN);
        struct intel_uncore_type *type = box->pmu->type;
        int id = intel_uncore_find_discovery_unit_id(type->boxes, -1, box->pmu->pmu_idx);

        if (tie_en) {
                reg1->reg = SPR_C0_MSR_PMON_BOX_FILTER0 +
                            HSWEP_CBO_MSR_OFFSET * id;
                reg1->config = event->attr.config1 & SPR_CHA_PMON_BOX_FILTER_TID;
                reg1->idx = 0;
        }

        return 0;
}

static struct intel_uncore_ops spr_uncore_chabox_ops = {
        .init_box               = intel_generic_uncore_msr_init_box,
        .disable_box            = intel_generic_uncore_msr_disable_box,
        .enable_box             = intel_generic_uncore_msr_enable_box,
        .disable_event          = spr_uncore_msr_disable_event,
        .enable_event           = spr_uncore_msr_enable_event,
        .read_counter           = uncore_msr_read_counter,
        .hw_config              = spr_cha_hw_config,
        .get_constraint         = uncore_get_constraint,
        .put_constraint         = uncore_put_constraint,
};

static struct attribute *spr_uncore_cha_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask_ext5.attr,
        &format_attr_tid_en2.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_filter_tid5.attr,
        NULL,
};
static const struct attribute_group spr_uncore_chabox_format_group = {
        .name = "format",
        .attrs = spr_uncore_cha_formats_attr,
};

static ssize_t alias_show(struct device *dev,
                          struct device_attribute *attr,
                          char *buf)
{
        struct intel_uncore_pmu *pmu = dev_to_uncore_pmu(dev);
        char pmu_name[UNCORE_PMU_NAME_LEN];

        uncore_get_alias_name(pmu_name, pmu);
        return sysfs_emit(buf, "%s\n", pmu_name);
}

static DEVICE_ATTR_RO(alias);

static struct attribute *uncore_alias_attrs[] = {
        &dev_attr_alias.attr,
        NULL
};

ATTRIBUTE_GROUPS(uncore_alias);

static struct intel_uncore_type spr_uncore_chabox = {
        .name                   = "cha",
        .event_mask             = SPR_CHA_PMON_EVENT_MASK,
        .event_mask_ext         = SPR_CHA_EVENT_MASK_EXT,
        .num_shared_regs        = 1,
        .constraints            = skx_uncore_chabox_constraints,
        .ops                    = &spr_uncore_chabox_ops,
        .format_group           = &spr_uncore_chabox_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type spr_uncore_iio = {
        .name                   = "iio",
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = SNR_IIO_PMON_RAW_EVENT_MASK_EXT,
        .format_group           = &snr_uncore_iio_format_group,
        .attr_update            = uncore_alias_groups,
        .constraints            = icx_uncore_iio_constraints,
};

static struct attribute *spr_uncore_raw_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask_ext4.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group spr_uncore_raw_format_group = {
        .name                   = "format",
        .attrs                  = spr_uncore_raw_formats_attr,
};

#define SPR_UNCORE_COMMON_FORMAT()                              \
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,    \
        .event_mask_ext         = SPR_RAW_EVENT_MASK_EXT,       \
        .format_group           = &spr_uncore_raw_format_group, \
        .attr_update            = uncore_alias_groups

static struct intel_uncore_type spr_uncore_irp = {
        SPR_UNCORE_COMMON_FORMAT(),
        .name                   = "irp",

};

static struct event_constraint spr_uncore_m2pcie_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x14, 0x3),
        UNCORE_EVENT_CONSTRAINT(0x2d, 0x3),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type spr_uncore_m2pcie = {
        SPR_UNCORE_COMMON_FORMAT(),
        .name                   = "m2pcie",
        .constraints            = spr_uncore_m2pcie_constraints,
};

static struct intel_uncore_type spr_uncore_pcu = {
        .name                   = "pcu",
        .attr_update            = uncore_alias_groups,
};

static void spr_uncore_mmio_enable_event(struct intel_uncore_box *box,
                                         struct perf_event *event)
{
        struct hw_perf_event *hwc = &event->hw;

        if (!box->io_addr)
                return;

        if (uncore_pmc_fixed(hwc->idx))
                writel(SNBEP_PMON_CTL_EN, box->io_addr + hwc->config_base);
        else
                writel(hwc->config, box->io_addr + hwc->config_base);
}

static struct intel_uncore_ops spr_uncore_mmio_ops = {
        .init_box               = intel_generic_uncore_mmio_init_box,
        .exit_box               = uncore_mmio_exit_box,
        .disable_box            = intel_generic_uncore_mmio_disable_box,
        .enable_box             = intel_generic_uncore_mmio_enable_box,
        .disable_event          = intel_generic_uncore_mmio_disable_event,
        .enable_event           = spr_uncore_mmio_enable_event,
        .read_counter           = uncore_mmio_read_counter,
};

static struct uncore_event_desc spr_uncore_imc_events[] = {
        INTEL_UNCORE_EVENT_DESC(clockticks,      "event=0x01,umask=0x00"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read,  "event=0x05,umask=0xcf"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read.unit, "MiB"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write, "event=0x05,umask=0xf0"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write.unit, "MiB"),
        { /* end: all zeroes */ },
};

#define SPR_UNCORE_MMIO_COMMON_FORMAT()                         \
        SPR_UNCORE_COMMON_FORMAT(),                             \
        .ops                    = &spr_uncore_mmio_ops

static struct intel_uncore_type spr_uncore_imc = {
        SPR_UNCORE_MMIO_COMMON_FORMAT(),
        .name                   = "imc",
        .fixed_ctr_bits         = 48,
        .fixed_ctr              = SNR_IMC_MMIO_PMON_FIXED_CTR,
        .fixed_ctl              = SNR_IMC_MMIO_PMON_FIXED_CTL,
        .event_descs            = spr_uncore_imc_events,
};

static void spr_uncore_pci_enable_event(struct intel_uncore_box *box,
                                        struct perf_event *event)
{
        struct pci_dev *pdev = box->pci_dev;
        struct hw_perf_event *hwc = &event->hw;

        pci_write_config_dword(pdev, hwc->config_base + 4, (u32)(hwc->config >> 32));
        pci_write_config_dword(pdev, hwc->config_base, (u32)hwc->config);
}

static struct intel_uncore_ops spr_uncore_pci_ops = {
        .init_box               = intel_generic_uncore_pci_init_box,
        .disable_box            = intel_generic_uncore_pci_disable_box,
        .enable_box             = intel_generic_uncore_pci_enable_box,
        .disable_event          = intel_generic_uncore_pci_disable_event,
        .enable_event           = spr_uncore_pci_enable_event,
        .read_counter           = intel_generic_uncore_pci_read_counter,
};

#define SPR_UNCORE_PCI_COMMON_FORMAT()                  \
        SPR_UNCORE_COMMON_FORMAT(),                     \
        .ops                    = &spr_uncore_pci_ops

static struct intel_uncore_type spr_uncore_m2m = {
        SPR_UNCORE_PCI_COMMON_FORMAT(),
        .name                   = "m2m",
};

static struct attribute_group spr_upi_mapping_group = {
        .is_visible     = skx_upi_mapping_visible,
};

static const struct attribute_group *spr_upi_attr_update[] = {
        &uncore_alias_group,
        &spr_upi_mapping_group,
        NULL
};

#define SPR_UPI_REGS_ADDR_DEVICE_LINK0  0x01

static void spr_upi_set_mapping(struct intel_uncore_type *type)
{
        pmu_upi_set_mapping(type, &spr_upi_mapping_group);
}

static void spr_upi_cleanup_mapping(struct intel_uncore_type *type)
{
        pmu_cleanup_mapping(type, &spr_upi_mapping_group);
}

static int spr_upi_get_topology(struct intel_uncore_type *type)
{
        return discover_upi_topology(type, SPR_UBOX_DID, SPR_UPI_REGS_ADDR_DEVICE_LINK0);
}

static struct intel_uncore_type spr_uncore_mdf = {
        SPR_UNCORE_COMMON_FORMAT(),
        .name                   = "mdf",
};

static void spr_uncore_mmio_offs8_init_box(struct intel_uncore_box *box)
{
        __set_bit(UNCORE_BOX_FLAG_CTL_OFFS8, &box->flags);
        intel_generic_uncore_mmio_init_box(box);
}

static struct intel_uncore_ops spr_uncore_mmio_offs8_ops = {
        .init_box               = spr_uncore_mmio_offs8_init_box,
        .exit_box               = uncore_mmio_exit_box,
        .disable_box            = intel_generic_uncore_mmio_disable_box,
        .enable_box             = intel_generic_uncore_mmio_enable_box,
        .disable_event          = intel_generic_uncore_mmio_disable_event,
        .enable_event           = spr_uncore_mmio_enable_event,
        .read_counter           = uncore_mmio_read_counter,
};

#define SPR_UNCORE_MMIO_OFFS8_COMMON_FORMAT()                   \
        SPR_UNCORE_COMMON_FORMAT(),                             \
        .ops                    = &spr_uncore_mmio_offs8_ops

static struct event_constraint spr_uncore_cxlcm_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x02, 0x0f),
        UNCORE_EVENT_CONSTRAINT(0x05, 0x0f),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x40, 0x43, 0xf0),
        UNCORE_EVENT_CONSTRAINT(0x4b, 0xf0),
        UNCORE_EVENT_CONSTRAINT(0x52, 0xf0),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type spr_uncore_cxlcm = {
        SPR_UNCORE_MMIO_OFFS8_COMMON_FORMAT(),
        .name                   = "cxlcm",
        .constraints            = spr_uncore_cxlcm_constraints,
};

static struct intel_uncore_type spr_uncore_cxldp = {
        SPR_UNCORE_MMIO_OFFS8_COMMON_FORMAT(),
        .name                   = "cxldp",
};

static struct intel_uncore_type spr_uncore_hbm = {
        SPR_UNCORE_COMMON_FORMAT(),
        .name                   = "hbm",
};

#define UNCORE_SPR_NUM_UNCORE_TYPES             15
#define UNCORE_SPR_CHA                          0
#define UNCORE_SPR_IIO                          1
#define UNCORE_SPR_IMC                          6
#define UNCORE_SPR_UPI                          8
#define UNCORE_SPR_M3UPI                        9

/*
 * The uncore units, which are supported by the discovery table,
 * are defined here.
 */
static struct intel_uncore_type *spr_uncores[UNCORE_SPR_NUM_UNCORE_TYPES] = {
        &spr_uncore_chabox,
        &spr_uncore_iio,
        &spr_uncore_irp,
        &spr_uncore_m2pcie,
        &spr_uncore_pcu,
        NULL,
        &spr_uncore_imc,
        &spr_uncore_m2m,
        NULL,
        NULL,
        NULL,
        &spr_uncore_mdf,
        &spr_uncore_cxlcm,
        &spr_uncore_cxldp,
        &spr_uncore_hbm,
};

/*
 * The uncore units, which are not supported by the discovery table,
 * are implemented from here.
 */
#define SPR_UNCORE_UPI_NUM_BOXES        4

static u64 spr_upi_pci_offsets[SPR_UNCORE_UPI_NUM_BOXES] = {
        0, 0x8000, 0x10000, 0x18000
};

static void spr_extra_boxes_cleanup(struct intel_uncore_type *type)
{
        struct intel_uncore_discovery_unit *pos;
        struct rb_node *node;

        if (!type->boxes)
                return;

        while (!RB_EMPTY_ROOT(type->boxes)) {
                node = rb_first(type->boxes);
                pos = rb_entry(node, struct intel_uncore_discovery_unit, node);
                rb_erase(node, type->boxes);
                kfree(pos);
        }
        kfree(type->boxes);
        type->boxes = NULL;
}

static struct intel_uncore_type spr_uncore_upi = {
        .event_mask             = SNBEP_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = SPR_RAW_EVENT_MASK_EXT,
        .format_group           = &spr_uncore_raw_format_group,
        .ops                    = &spr_uncore_pci_ops,
        .name                   = "upi",
        .attr_update            = spr_upi_attr_update,
        .get_topology           = spr_upi_get_topology,
        .set_mapping            = spr_upi_set_mapping,
        .cleanup_mapping        = spr_upi_cleanup_mapping,
        .type_id                = UNCORE_SPR_UPI,
        .num_counters           = 4,
        .num_boxes              = SPR_UNCORE_UPI_NUM_BOXES,
        .perf_ctr_bits          = 48,
        .perf_ctr               = ICX_UPI_PCI_PMON_CTR0 - ICX_UPI_PCI_PMON_BOX_CTL,
        .event_ctl              = ICX_UPI_PCI_PMON_CTL0 - ICX_UPI_PCI_PMON_BOX_CTL,
        .box_ctl                = ICX_UPI_PCI_PMON_BOX_CTL,
        .pci_offsets            = spr_upi_pci_offsets,
        .cleanup_extra_boxes    = spr_extra_boxes_cleanup,
};

static struct intel_uncore_type spr_uncore_m3upi = {
        SPR_UNCORE_PCI_COMMON_FORMAT(),
        .name                   = "m3upi",
        .type_id                = UNCORE_SPR_M3UPI,
        .num_counters           = 4,
        .num_boxes              = SPR_UNCORE_UPI_NUM_BOXES,
        .perf_ctr_bits          = 48,
        .perf_ctr               = ICX_M3UPI_PCI_PMON_CTR0 - ICX_M3UPI_PCI_PMON_BOX_CTL,
        .event_ctl              = ICX_M3UPI_PCI_PMON_CTL0 - ICX_M3UPI_PCI_PMON_BOX_CTL,
        .box_ctl                = ICX_M3UPI_PCI_PMON_BOX_CTL,
        .pci_offsets            = spr_upi_pci_offsets,
        .constraints            = icx_uncore_m3upi_constraints,
        .cleanup_extra_boxes    = spr_extra_boxes_cleanup,
};

enum perf_uncore_spr_iio_freerunning_type_id {
        SPR_IIO_MSR_IOCLK,
        SPR_IIO_MSR_BW_IN,
        SPR_IIO_MSR_BW_OUT,

        SPR_IIO_FREERUNNING_TYPE_MAX,
};

static struct freerunning_counters spr_iio_freerunning[] = {
        [SPR_IIO_MSR_IOCLK]     = { 0x340e, 0x1, 0x10, 1, 48 },
        [SPR_IIO_MSR_BW_IN]     = { 0x3800, 0x1, 0x10, 8, 48 },
        [SPR_IIO_MSR_BW_OUT]    = { 0x3808, 0x1, 0x10, 8, 48 },
};

static struct intel_uncore_type spr_uncore_iio_free_running = {
        .name                   = "iio_free_running",
        .num_counters           = 17,
        .num_freerunning_types  = SPR_IIO_FREERUNNING_TYPE_MAX,
        .freerunning            = spr_iio_freerunning,
        .ops                    = &skx_uncore_iio_freerunning_ops,
        .event_descs            = snr_uncore_iio_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

enum perf_uncore_spr_imc_freerunning_type_id {
        SPR_IMC_DCLK,
        SPR_IMC_PQ_CYCLES,

        SPR_IMC_FREERUNNING_TYPE_MAX,
};

static struct freerunning_counters spr_imc_freerunning[] = {
        [SPR_IMC_DCLK]          = { 0x22b0, 0x0, 0, 1, 48 },
        [SPR_IMC_PQ_CYCLES]     = { 0x2318, 0x8, 0, 2, 48 },
};

static struct uncore_event_desc spr_uncore_imc_freerunning_events[] = {
        INTEL_UNCORE_EVENT_DESC(dclk,                   "event=0xff,umask=0x10"),

        INTEL_UNCORE_EVENT_DESC(rpq_cycles,             "event=0xff,umask=0x20"),
        INTEL_UNCORE_EVENT_DESC(wpq_cycles,             "event=0xff,umask=0x21"),
        { /* end: all zeroes */ },
};

#define SPR_MC_DEVICE_ID        0x3251

static void spr_uncore_imc_freerunning_init_box(struct intel_uncore_box *box)
{
        int mem_offset = box->pmu->pmu_idx * ICX_IMC_MEM_STRIDE + SNR_IMC_MMIO_MEM0_OFFSET;

        snr_uncore_mmio_map(box, uncore_mmio_box_ctl(box),
                            mem_offset, SPR_MC_DEVICE_ID);
}

static struct intel_uncore_ops spr_uncore_imc_freerunning_ops = {
        .init_box       = spr_uncore_imc_freerunning_init_box,
        .exit_box       = uncore_mmio_exit_box,
        .read_counter   = uncore_mmio_read_counter,
        .hw_config      = uncore_freerunning_hw_config,
};

static struct intel_uncore_type spr_uncore_imc_free_running = {
        .name                   = "imc_free_running",
        .num_counters           = 3,
        .mmio_map_size          = SNR_IMC_MMIO_SIZE,
        .num_freerunning_types  = SPR_IMC_FREERUNNING_TYPE_MAX,
        .freerunning            = spr_imc_freerunning,
        .ops                    = &spr_uncore_imc_freerunning_ops,
        .event_descs            = spr_uncore_imc_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

#define UNCORE_SPR_MSR_EXTRA_UNCORES            1
#define UNCORE_SPR_MMIO_EXTRA_UNCORES           1
#define UNCORE_SPR_PCI_EXTRA_UNCORES            2

static struct intel_uncore_type *spr_msr_uncores[UNCORE_SPR_MSR_EXTRA_UNCORES] = {
        &spr_uncore_iio_free_running,
};

static struct intel_uncore_type *spr_mmio_uncores[UNCORE_SPR_MMIO_EXTRA_UNCORES] = {
        &spr_uncore_imc_free_running,
};

static struct intel_uncore_type *spr_pci_uncores[UNCORE_SPR_PCI_EXTRA_UNCORES] = {
        &spr_uncore_upi,
        &spr_uncore_m3upi
};

int spr_uncore_units_ignore[] = {
        UNCORE_SPR_UPI,
        UNCORE_SPR_M3UPI,
        UNCORE_IGNORE_END
};

static void uncore_type_customized_copy(struct intel_uncore_type *to_type,
                                        struct intel_uncore_type *from_type)
{
        if (!to_type || !from_type)
                return;

        if (from_type->name)
                to_type->name = from_type->name;
        if (from_type->fixed_ctr_bits)
                to_type->fixed_ctr_bits = from_type->fixed_ctr_bits;
        if (from_type->event_mask)
                to_type->event_mask = from_type->event_mask;
        if (from_type->event_mask_ext)
                to_type->event_mask_ext = from_type->event_mask_ext;
        if (from_type->fixed_ctr)
                to_type->fixed_ctr = from_type->fixed_ctr;
        if (from_type->fixed_ctl)
                to_type->fixed_ctl = from_type->fixed_ctl;
        if (from_type->fixed_ctr_bits)
                to_type->fixed_ctr_bits = from_type->fixed_ctr_bits;
        if (from_type->num_shared_regs)
                to_type->num_shared_regs = from_type->num_shared_regs;
        if (from_type->constraints)
                to_type->constraints = from_type->constraints;
        if (from_type->ops)
                to_type->ops = from_type->ops;
        if (from_type->event_descs)
                to_type->event_descs = from_type->event_descs;
        if (from_type->format_group)
                to_type->format_group = from_type->format_group;
        if (from_type->attr_update)
                to_type->attr_update = from_type->attr_update;
        if (from_type->set_mapping)
                to_type->set_mapping = from_type->set_mapping;
        if (from_type->get_topology)
                to_type->get_topology = from_type->get_topology;
        if (from_type->cleanup_mapping)
                to_type->cleanup_mapping = from_type->cleanup_mapping;
        if (from_type->mmio_map_size)
                to_type->mmio_map_size = from_type->mmio_map_size;
}

struct intel_uncore_type **
uncore_get_uncores(enum uncore_access_type type_id, int num_extra,
                   struct intel_uncore_type **extra, int max_num_types,
                   struct intel_uncore_type **uncores)
{
        struct intel_uncore_type **types, **start_types;
        int i;

        start_types = types = intel_uncore_generic_init_uncores(type_id, num_extra);

        /* Only copy the customized features */
        for (; *types; types++) {
                if ((*types)->type_id >= max_num_types)
                        continue;
                uncore_type_customized_copy(*types, uncores[(*types)->type_id]);
        }

        for (i = 0; i < num_extra; i++, types++)
                *types = extra[i];

        return start_types;
}

static struct intel_uncore_type *
uncore_find_type_by_id(struct intel_uncore_type **types, int type_id)
{
        for (; *types; types++) {
                if (type_id == (*types)->type_id)
                        return *types;
        }

        return NULL;
}

static int uncore_type_max_boxes(struct intel_uncore_type **types,
                                 int type_id)
{
        struct intel_uncore_discovery_unit *unit;
        struct intel_uncore_type *type;
        struct rb_node *node;
        int max = 0;

        type = uncore_find_type_by_id(types, type_id);
        if (!type)
                return 0;

        for (node = rb_first(type->boxes); node; node = rb_next(node)) {
                unit = rb_entry(node, struct intel_uncore_discovery_unit, node);

                /*
                 * on DMR IMH2, the unit id starts from 0x8000,
                 * and we don't need to count it.
                 */
                if ((unit->id > max) && (unit->id < 0x8000))
                        max = unit->id;
        }
        return max + 1;
}

#define SPR_MSR_UNC_CBO_CONFIG          0x2FFE

void spr_uncore_cpu_init(void)
{
        struct intel_uncore_type *type;
        u64 num_cbo;

        uncore_msr_uncores = uncore_get_uncores(UNCORE_ACCESS_MSR,
                                                UNCORE_SPR_MSR_EXTRA_UNCORES,
                                                spr_msr_uncores,
                                                UNCORE_SPR_NUM_UNCORE_TYPES,
                                                spr_uncores);

        type = uncore_find_type_by_id(uncore_msr_uncores, UNCORE_SPR_CHA);
        if (type) {
                /*
                 * The value from the discovery table (stored in the type->num_boxes
                 * of UNCORE_SPR_CHA) is incorrect on some SPR variants because of a
                 * firmware bug. Using the value from SPR_MSR_UNC_CBO_CONFIG to replace it.
                 */
                rdmsrq(SPR_MSR_UNC_CBO_CONFIG, num_cbo);
                /*
                 * The MSR doesn't work on the EMR XCC, but the firmware bug doesn't impact
                 * the EMR XCC. Don't let the value from the MSR replace the existing value.
                 */
                if (num_cbo)
                        type->num_boxes = num_cbo;
        }
        spr_uncore_iio_free_running.num_boxes = uncore_type_max_boxes(uncore_msr_uncores, UNCORE_SPR_IIO);
}

#define SPR_UNCORE_UPI_PCIID            0x3241
#define SPR_UNCORE_UPI0_DEVFN           0x9
#define SPR_UNCORE_M3UPI_PCIID          0x3246
#define SPR_UNCORE_M3UPI0_DEVFN         0x29

static void spr_update_device_location(int type_id)
{
        struct intel_uncore_discovery_unit *unit;
        struct intel_uncore_type *type;
        struct pci_dev *dev = NULL;
        struct rb_root *root;
        u32 device, devfn;
        int die;

        if (type_id == UNCORE_SPR_UPI) {
                type = &spr_uncore_upi;
                device = SPR_UNCORE_UPI_PCIID;
                devfn = SPR_UNCORE_UPI0_DEVFN;
        } else if (type_id == UNCORE_SPR_M3UPI) {
                type = &spr_uncore_m3upi;
                device = SPR_UNCORE_M3UPI_PCIID;
                devfn = SPR_UNCORE_M3UPI0_DEVFN;
        } else
                return;

        root = kzalloc_obj(struct rb_root);
        if (!root) {
                type->num_boxes = 0;
                return;
        }
        *root = RB_ROOT;

        while ((dev = pci_get_device(PCI_VENDOR_ID_INTEL, device, dev)) != NULL) {

                die = uncore_device_to_die(dev);
                if (die < 0)
                        continue;

                unit = kzalloc_obj(*unit);
                if (!unit)
                        continue;
                unit->die = die;
                unit->id = PCI_SLOT(dev->devfn) - PCI_SLOT(devfn);
                unit->addr = pci_domain_nr(dev->bus) << UNCORE_DISCOVERY_PCI_DOMAIN_OFFSET |
                             dev->bus->number << UNCORE_DISCOVERY_PCI_BUS_OFFSET |
                             devfn << UNCORE_DISCOVERY_PCI_DEVFN_OFFSET |
                             type->box_ctl;

                unit->pmu_idx = unit->id;

                uncore_find_add_unit(unit, root, NULL);
        }

        type->boxes = root;
}

int spr_uncore_pci_init(void)
{
        /*
         * The discovery table of UPI on some SPR variant is broken,
         * which impacts the detection of both UPI and M3UPI uncore PMON.
         * Use the pre-defined UPI and M3UPI table to replace.
         *
         * The accurate location, e.g., domain and BUS number,
         * can only be retrieved at load time.
         * Update the location of UPI and M3UPI.
         */
        spr_update_device_location(UNCORE_SPR_UPI);
        spr_update_device_location(UNCORE_SPR_M3UPI);
        uncore_pci_uncores = uncore_get_uncores(UNCORE_ACCESS_PCI,
                                                UNCORE_SPR_PCI_EXTRA_UNCORES,
                                                spr_pci_uncores,
                                                UNCORE_SPR_NUM_UNCORE_TYPES,
                                                spr_uncores);
        return 0;
}

void spr_uncore_mmio_init(void)
{
        int ret = snbep_pci2phy_map_init(0x3250, SKX_CPUNODEID, SKX_GIDNIDMAP, true);

        if (ret) {
                uncore_mmio_uncores = uncore_get_uncores(UNCORE_ACCESS_MMIO, 0, NULL,
                                                         UNCORE_SPR_NUM_UNCORE_TYPES,
                                                         spr_uncores);
        } else {
                uncore_mmio_uncores = uncore_get_uncores(UNCORE_ACCESS_MMIO,
                                                         UNCORE_SPR_MMIO_EXTRA_UNCORES,
                                                         spr_mmio_uncores,
                                                         UNCORE_SPR_NUM_UNCORE_TYPES,
                                                         spr_uncores);

                spr_uncore_imc_free_running.num_boxes = uncore_type_max_boxes(uncore_mmio_uncores, UNCORE_SPR_IMC) / 2;
        }
}

/* end of SPR uncore support */

/* GNR uncore support */

#define UNCORE_GNR_NUM_UNCORE_TYPES     23

int gnr_uncore_units_ignore[] = {
        UNCORE_IGNORE_END
};

static struct intel_uncore_type gnr_uncore_ubox = {
        .name                   = "ubox",
        .attr_update            = uncore_alias_groups,
};

static struct uncore_event_desc gnr_uncore_imc_events[] = {
        INTEL_UNCORE_EVENT_DESC(clockticks,      "event=0x01,umask=0x00"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read_sch0,  "event=0x05,umask=0xcf"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read_sch0.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read_sch0.unit, "MiB"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read_sch1,  "event=0x06,umask=0xcf"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read_sch1.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_read_sch1.unit, "MiB"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write_sch0, "event=0x05,umask=0xf0"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write_sch0.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write_sch0.unit, "MiB"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write_sch1, "event=0x06,umask=0xf0"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write_sch1.scale, "6.103515625e-5"),
        INTEL_UNCORE_EVENT_DESC(cas_count_write_sch1.unit, "MiB"),
        { /* end: all zeroes */ },
};

static struct intel_uncore_type gnr_uncore_imc = {
        SPR_UNCORE_MMIO_COMMON_FORMAT(),
        .name                   = "imc",
        .fixed_ctr_bits         = 48,
        .fixed_ctr              = SNR_IMC_MMIO_PMON_FIXED_CTR,
        .fixed_ctl              = SNR_IMC_MMIO_PMON_FIXED_CTL,
        .event_descs            = gnr_uncore_imc_events,
};

static struct intel_uncore_type gnr_uncore_pciex8 = {
        SPR_UNCORE_PCI_COMMON_FORMAT(),
        .name                   = "pciex8",
};

static struct intel_uncore_type gnr_uncore_pciex16 = {
        SPR_UNCORE_PCI_COMMON_FORMAT(),
        .name                   = "pciex16",
};

static struct intel_uncore_type gnr_uncore_upi = {
        SPR_UNCORE_PCI_COMMON_FORMAT(),
        .name                   = "upi",
};

static struct intel_uncore_type gnr_uncore_b2upi = {
        SPR_UNCORE_PCI_COMMON_FORMAT(),
        .name                   = "b2upi",
};

static struct intel_uncore_type gnr_uncore_b2hot = {
        .name                   = "b2hot",
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type gnr_uncore_b2cmi = {
        SPR_UNCORE_PCI_COMMON_FORMAT(),
        .name                   = "b2cmi",
};

static struct intel_uncore_type gnr_uncore_b2cxl = {
        SPR_UNCORE_MMIO_OFFS8_COMMON_FORMAT(),
        .name                   = "b2cxl",
};

static struct intel_uncore_type gnr_uncore_mdf_sbo = {
        .name                   = "mdf_sbo",
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type *gnr_uncores[UNCORE_GNR_NUM_UNCORE_TYPES] = {
        &spr_uncore_chabox,
        &spr_uncore_iio,
        &spr_uncore_irp,
        NULL,
        &spr_uncore_pcu,
        &gnr_uncore_ubox,
        &gnr_uncore_imc,
        NULL,
        &gnr_uncore_upi,
        NULL,
        NULL,
        NULL,
        &spr_uncore_cxlcm,
        &spr_uncore_cxldp,
        NULL,
        &gnr_uncore_b2hot,
        &gnr_uncore_b2cmi,
        &gnr_uncore_b2cxl,
        &gnr_uncore_b2upi,
        NULL,
        &gnr_uncore_mdf_sbo,
        &gnr_uncore_pciex16,
        &gnr_uncore_pciex8,
};

static struct freerunning_counters gnr_iio_freerunning[] = {
        [SPR_IIO_MSR_IOCLK]     = { 0x290e, 0x01, 0x10, 1, 48 },
        [SPR_IIO_MSR_BW_IN]     = { 0x360e, 0x10, 0x80, 8, 48 },
        [SPR_IIO_MSR_BW_OUT]    = { 0x2e0e, 0x10, 0x80, 8, 48 },
};

void gnr_uncore_cpu_init(void)
{
        uncore_msr_uncores = uncore_get_uncores(UNCORE_ACCESS_MSR,
                                                UNCORE_SPR_MSR_EXTRA_UNCORES,
                                                spr_msr_uncores,
                                                UNCORE_GNR_NUM_UNCORE_TYPES,
                                                gnr_uncores);
        spr_uncore_iio_free_running.num_boxes = uncore_type_max_boxes(uncore_msr_uncores, UNCORE_SPR_IIO);
        spr_uncore_iio_free_running.freerunning = gnr_iio_freerunning;
}

int gnr_uncore_pci_init(void)
{
        uncore_pci_uncores = uncore_get_uncores(UNCORE_ACCESS_PCI, 0, NULL,
                                                UNCORE_GNR_NUM_UNCORE_TYPES,
                                                gnr_uncores);
        return 0;
}

void gnr_uncore_mmio_init(void)
{
        uncore_mmio_uncores = uncore_get_uncores(UNCORE_ACCESS_MMIO, 0, NULL,
                                                 UNCORE_GNR_NUM_UNCORE_TYPES,
                                                 gnr_uncores);
}

/* end of GNR uncore support */

/* DMR uncore support */
#define UNCORE_DMR_NUM_UNCORE_TYPES     52

static struct attribute *dmr_imc_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh10.attr,
        NULL,
};

static const struct attribute_group dmr_imc_uncore_format_group = {
        .name = "format",
        .attrs = dmr_imc_uncore_formats_attr,
};

static struct intel_uncore_type dmr_uncore_imc = {
        .name                   = "imc",
        .fixed_ctr_bits         = 48,
        .fixed_ctr              = DMR_IMC_PMON_FIXED_CTR,
        .fixed_ctl              = DMR_IMC_PMON_FIXED_CTL,
        .ops                    = &spr_uncore_mmio_ops,
        .format_group           = &dmr_imc_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct attribute *dmr_sca_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask_ext5.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        NULL,
};

static const struct attribute_group dmr_sca_uncore_format_group = {
        .name = "format",
        .attrs = dmr_sca_uncore_formats_attr,
};

static struct intel_uncore_type dmr_uncore_sca = {
        .name                   = "sca",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct attribute *dmr_cxlcm_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv2.attr,
        &format_attr_thresh9_2.attr,
        &format_attr_port_en.attr,
        NULL,
};

static const struct attribute_group dmr_cxlcm_uncore_format_group = {
        .name = "format",
        .attrs = dmr_cxlcm_uncore_formats_attr,
};

static struct event_constraint dmr_uncore_cxlcm_constraints[] = {
        UNCORE_EVENT_CONSTRAINT_RANGE(0x1, 0x24, 0x0f),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x41, 0x41, 0xf0),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x50, 0x5e, 0xf0),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x60, 0x61, 0xf0),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type dmr_uncore_cxlcm = {
        .name                   = "cxlcm",
        .event_mask             = GENERIC_PMON_RAW_EVENT_MASK,
        .event_mask_ext         = DMR_CXLCM_EVENT_MASK_EXT,
        .constraints            = dmr_uncore_cxlcm_constraints,
        .format_group           = &dmr_cxlcm_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_hamvf = {
        .name                   = "hamvf",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct event_constraint dmr_uncore_cbo_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x11, 0x1),
        UNCORE_EVENT_CONSTRAINT_RANGE(0x19, 0x1a, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x1f, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x21, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x25, 0x1),
        UNCORE_EVENT_CONSTRAINT(0x36, 0x1),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type dmr_uncore_cbo = {
        .name                   = "cbo",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .constraints            = dmr_uncore_cbo_constraints,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_santa = {
        .name                   = "santa",
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_cncu = {
        .name                   = "cncu",
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_sncu = {
        .name                   = "sncu",
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_ula = {
        .name                   = "ula",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_dda = {
        .name                   = "dda",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct event_constraint dmr_uncore_sbo_constraints[] = {
        UNCORE_EVENT_CONSTRAINT(0x1f, 0x01),
        UNCORE_EVENT_CONSTRAINT(0x25, 0x01),
        EVENT_CONSTRAINT_END
};

static struct intel_uncore_type dmr_uncore_sbo = {
        .name                   = "sbo",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .constraints            = dmr_uncore_sbo_constraints,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_ubr = {
        .name                   = "ubr",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct attribute *dmr_pcie4_uncore_formats_attr[] = {
        &format_attr_event.attr,
        &format_attr_umask.attr,
        &format_attr_edge.attr,
        &format_attr_inv.attr,
        &format_attr_thresh8.attr,
        &format_attr_thresh_ext.attr,
        &format_attr_rs3_sel.attr,
        &format_attr_rx_sel.attr,
        &format_attr_tx_sel.attr,
        &format_attr_iep_sel.attr,
        &format_attr_vc_sel.attr,
        &format_attr_port_sel.attr,
        NULL,
};

static const struct attribute_group dmr_pcie4_uncore_format_group = {
        .name = "format",
        .attrs = dmr_pcie4_uncore_formats_attr,
};

static struct intel_uncore_type dmr_uncore_pcie4 = {
        .name                   = "pcie4",
        .event_mask_ext         = DMR_PCIE4_EVENT_MASK_EXT,
        .format_group           = &dmr_pcie4_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_crs = {
        .name                   = "crs",
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_cpc = {
        .name                   = "cpc",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_itc = {
        .name                   = "itc",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_otc = {
        .name                   = "otc",
        .event_mask_ext         = DMR_HAMVF_EVENT_MASK_EXT,
        .format_group           = &dmr_sca_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_cms = {
        .name                   = "cms",
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type dmr_uncore_pcie6 = {
        .name                   = "pcie6",
        .event_mask_ext         = DMR_PCIE4_EVENT_MASK_EXT,
        .format_group           = &dmr_pcie4_uncore_format_group,
        .attr_update            = uncore_alias_groups,
};

static struct intel_uncore_type *dmr_uncores[UNCORE_DMR_NUM_UNCORE_TYPES] = {
        NULL, NULL, NULL, NULL,
        &spr_uncore_pcu,
        &gnr_uncore_ubox,
        &dmr_uncore_imc,
        NULL,
        NULL, NULL, NULL, NULL,
        NULL, NULL, NULL, NULL,
        NULL, NULL, NULL, NULL,
        NULL, NULL, NULL,
        &dmr_uncore_sca,
        &dmr_uncore_cxlcm,
        NULL, NULL, NULL,
        NULL, NULL,
        &dmr_uncore_hamvf,
        &dmr_uncore_cbo,
        &dmr_uncore_santa,
        &dmr_uncore_cncu,
        &dmr_uncore_sncu,
        &dmr_uncore_ula,
        &dmr_uncore_dda,
        NULL,
        &dmr_uncore_sbo,
        NULL,
        NULL, NULL, NULL,
        &dmr_uncore_ubr,
        NULL,
        &dmr_uncore_pcie4,
        &dmr_uncore_crs,
        &dmr_uncore_cpc,
        &dmr_uncore_itc,
        &dmr_uncore_otc,
        &dmr_uncore_cms,
        &dmr_uncore_pcie6,
};

int dmr_uncore_imh_units_ignore[] = {
        0x13,           /* MSE */
        UNCORE_IGNORE_END
};

int dmr_uncore_cbb_units_ignore[] = {
        0x25,           /* SB2UCIE */
        UNCORE_IGNORE_END
};

static unsigned int dmr_iio_freerunning_box_offsets[] = {
        0x0, 0x8000, 0x18000, 0x20000
};

static void dmr_uncore_freerunning_init_box(struct intel_uncore_box *box)
{
        struct intel_uncore_type *type = box->pmu->type;
        u64 mmio_base;

        if (box->pmu->pmu_idx >= type->num_boxes)
                return;

        mmio_base = DMR_IMH1_HIOP_MMIO_BASE;
        mmio_base += dmr_iio_freerunning_box_offsets[box->pmu->pmu_idx];

        box->io_addr = ioremap(mmio_base, type->mmio_map_size);
        if (!box->io_addr)
                pr_warn("perf uncore: Failed to ioremap for %s.\n", type->name);
}

static struct intel_uncore_ops dmr_uncore_freerunning_ops = {
        .init_box       = dmr_uncore_freerunning_init_box,
        .exit_box       = uncore_mmio_exit_box,
        .read_counter   = uncore_mmio_read_counter,
        .hw_config      = uncore_freerunning_hw_config,
};

enum perf_uncore_dmr_iio_freerunning_type_id {
        DMR_ITC_INB_DATA_BW,
        DMR_ITC_BW_IN,
        DMR_OTC_BW_OUT,
        DMR_OTC_CLOCK_TICKS,

        DMR_IIO_FREERUNNING_TYPE_MAX,
};

static struct freerunning_counters dmr_iio_freerunning[] = {
        [DMR_ITC_INB_DATA_BW]   = { 0x4d40, 0x8, 0, 8, 48},
        [DMR_ITC_BW_IN]         = { 0x6b00, 0x8, 0, 8, 48},
        [DMR_OTC_BW_OUT]        = { 0x6b60, 0x8, 0, 8, 48},
        [DMR_OTC_CLOCK_TICKS]   = { 0x6bb0, 0x8, 0, 1, 48},
};

static struct uncore_event_desc dmr_uncore_iio_freerunning_events[] = {
        /*  ITC Free Running Data BW counter for inbound traffic */
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port0, 0x10, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port1, 0x11, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port2, 0x12, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port3, 0x13, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port4, 0x14, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port5, 0x15, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port6, 0x16, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(inb_data_port7, 0x17, "3.814697266e-6"),

        /*  ITC Free Running BW IN counters */
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port0, 0x20, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port1, 0x21, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port2, 0x22, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port3, 0x23, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port4, 0x24, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port5, 0x25, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port6, 0x26, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_in_port7, 0x27, "3.814697266e-6"),

        /*  ITC Free Running BW OUT counters */
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port0, 0x30, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port1, 0x31, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port2, 0x32, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port3, 0x33, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port4, 0x34, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port5, 0x35, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port6, 0x36, "3.814697266e-6"),
        INTEL_UNCORE_FR_EVENT_DESC(bw_out_port7, 0x37, "3.814697266e-6"),

        /* Free Running Clock Counter */
        INTEL_UNCORE_EVENT_DESC(clockticks, "event=0xff,umask=0x40"),
        { /* end: all zeroes */ },
};

static struct intel_uncore_type dmr_uncore_iio_free_running = {
        .name                   = "iio_free_running",
        .num_counters           = 25,
        .mmio_map_size          = DMR_HIOP_MMIO_SIZE,
        .num_freerunning_types  = DMR_IIO_FREERUNNING_TYPE_MAX,
        .freerunning            = dmr_iio_freerunning,
        .ops                    = &dmr_uncore_freerunning_ops,
        .event_descs            = dmr_uncore_iio_freerunning_events,
        .format_group           = &skx_uncore_iio_freerunning_format_group,
};

#define UNCORE_DMR_MMIO_EXTRA_UNCORES           1
static struct intel_uncore_type *dmr_mmio_uncores[UNCORE_DMR_MMIO_EXTRA_UNCORES] = {
        &dmr_uncore_iio_free_running,
};

int dmr_uncore_pci_init(void)
{
        uncore_pci_uncores = uncore_get_uncores(UNCORE_ACCESS_PCI, 0, NULL,
                                                UNCORE_DMR_NUM_UNCORE_TYPES,
                                                dmr_uncores);
        return 0;
}

void dmr_uncore_mmio_init(void)
{
        uncore_mmio_uncores = uncore_get_uncores(UNCORE_ACCESS_MMIO,
                                                UNCORE_DMR_MMIO_EXTRA_UNCORES,
                                                dmr_mmio_uncores,
                                                UNCORE_DMR_NUM_UNCORE_TYPES,
                                                dmr_uncores);

        dmr_uncore_iio_free_running.num_boxes =
                uncore_type_max_boxes(uncore_mmio_uncores, UNCORE_DMR_ITC);
}
/* end of DMR uncore support */