ocmem
if (a3xx_gpu->ocmem.hdl) {
(unsigned int)(a3xx_gpu->ocmem.base >> 14));
adreno_gpu_ocmem_cleanup(&a3xx_gpu->ocmem);
adreno_gpu, &a3xx_gpu->ocmem);
struct adreno_ocmem ocmem;
(unsigned int)(a4xx_gpu->ocmem.base >> 14));
adreno_gpu_ocmem_cleanup(&a4xx_gpu->ocmem);
&a4xx_gpu->ocmem);
struct adreno_ocmem ocmem;
struct ocmem *ocmem;
ocmem = of_get_ocmem(dev);
if (IS_ERR(ocmem)) {
if (PTR_ERR(ocmem) == -ENODEV) {
return PTR_ERR(ocmem);
ocmem_hdl = ocmem_allocate(ocmem, OCMEM_GRAPHICS, adreno_gpu->info->gmem);
adreno_ocmem->ocmem = ocmem;
ocmem_free(adreno_ocmem->ocmem, OCMEM_GRAPHICS,
struct ocmem *ocmem;
struct adreno_ocmem *ocmem);
void adreno_gpu_ocmem_cleanup(struct adreno_ocmem *ocmem);
static inline void ocmem_write(struct ocmem *ocmem, u32 reg, u32 data)
writel(data, ocmem->mmio + reg);
static inline u32 ocmem_read(struct ocmem *ocmem, u32 reg)
return readl(ocmem->mmio + reg);
static void update_ocmem(struct ocmem *ocmem)
for (i = 0; i < ocmem->config->num_regions; i++) {
struct ocmem_region *region = &ocmem->regions[i];
dev_dbg(ocmem->dev, "ocmem_region_mode_control %x\n",
ocmem_write(ocmem, OCMEM_REG_REGION_MODE_CTL, region_mode_ctrl);
for (i = 0; i < ocmem->config->num_regions; i++) {
struct ocmem_region *region = &ocmem->regions[i];
ocmem_write(ocmem, OCMEM_REG_PSGSC_CTL(i), data);
static unsigned long phys_to_offset(struct ocmem *ocmem,
if (addr < ocmem->memory->start || addr >= ocmem->memory->end)
return addr - ocmem->memory->start;
static unsigned long device_address(struct ocmem *ocmem,
return phys_to_offset(ocmem, addr);
static void update_range(struct ocmem *ocmem, struct ocmem_buf *buf,
for (i = 0; i < ocmem->config->num_regions; i++) {
struct ocmem_region *region = &ocmem->regions[i];
update_ocmem(ocmem);
struct ocmem *of_get_ocmem(struct device *dev)
struct ocmem *ocmem;
ocmem = platform_get_drvdata(pdev);
if (!ocmem) {
return ocmem;
struct ocmem_buf *ocmem_allocate(struct ocmem *ocmem, enum ocmem_client client,
if (test_and_set_bit_lock(BIT(client), &ocmem->active_allocations))
buf->addr = device_address(ocmem, client, buf->offset);
update_range(ocmem, buf, CORE_ON, WIDE_MODE);
dev_err(ocmem->dev, "could not lock: %d\n", ret);
ocmem_write(ocmem, OCMEM_REG_GFX_MPU_START, buf->offset);
ocmem_write(ocmem, OCMEM_REG_GFX_MPU_END,
dev_dbg(ocmem->dev, "using %ldK of OCMEM at 0x%08lx for client %d\n",
clear_bit_unlock(BIT(client), &ocmem->active_allocations);
void ocmem_free(struct ocmem *ocmem, enum ocmem_client client,
update_range(ocmem, buf, CLK_OFF, MODE_DEFAULT);
dev_err(ocmem->dev, "could not unlock: %d\n", ret);
ocmem_write(ocmem, OCMEM_REG_GFX_MPU_START, 0x0);
ocmem_write(ocmem, OCMEM_REG_GFX_MPU_END, 0x0);
clear_bit_unlock(BIT(client), &ocmem->active_allocations);
struct ocmem *ocmem;
ocmem = devm_kzalloc(dev, sizeof(*ocmem), GFP_KERNEL);
if (!ocmem)
ocmem->dev = dev;
ocmem->config = device_get_match_data(dev);
ocmem->core_clk = devm_clk_get(dev, "core");
if (IS_ERR(ocmem->core_clk))
return dev_err_probe(dev, PTR_ERR(ocmem->core_clk),
ocmem->iface_clk = devm_clk_get_optional(dev, "iface");
if (IS_ERR(ocmem->iface_clk))
return dev_err_probe(dev, PTR_ERR(ocmem->iface_clk),
ocmem->mmio = devm_platform_ioremap_resource_byname(pdev, "ctrl");
if (IS_ERR(ocmem->mmio))
return dev_err_probe(&pdev->dev, PTR_ERR(ocmem->mmio),
ocmem->memory = platform_get_resource_byname(pdev, IORESOURCE_MEM,
if (!ocmem->memory) {
WARN_ON(clk_set_rate(ocmem->core_clk, 1000) < 0);
ret = clk_prepare_enable(ocmem->core_clk);
return dev_err_probe(ocmem->dev, ret, "Failed to enable core clock\n");
ret = clk_prepare_enable(ocmem->iface_clk);
clk_disable_unprepare(ocmem->core_clk);
return dev_err_probe(ocmem->dev, ret, "Failed to enable iface clock\n");
reg = ocmem_read(ocmem, OCMEM_REG_HW_VERSION);
reg = ocmem_read(ocmem, OCMEM_REG_HW_PROFILE);
ocmem->num_ports = OCMEM_HW_PROFILE_NUM_PORTS(reg);
ocmem->num_macros = OCMEM_HW_PROFILE_NUM_MACROS(reg);
ocmem->interleaved = !!(reg & OCMEM_HW_PROFILE_INTERLEAVING);
num_banks = ocmem->num_ports / 2;
region_size = ocmem->config->macro_size * num_banks;
ocmem->num_ports, ocmem->config->num_regions,
ocmem->num_macros, ocmem->interleaved ? "" : "not ");
ocmem->regions = devm_kcalloc(dev, ocmem->config->num_regions,
if (!ocmem->regions) {
for (i = 0; i < ocmem->config->num_regions; i++) {
struct ocmem_region *region = &ocmem->regions[i];
if (i == (ocmem->config->num_regions - 1) &&
region->macro_size = ocmem->config->macro_size / 2;
region->macro_size = ocmem->config->macro_size;
platform_set_drvdata(pdev, ocmem);
clk_disable_unprepare(ocmem->core_clk);
clk_disable_unprepare(ocmem->iface_clk);
struct ocmem *ocmem = platform_get_drvdata(pdev);
clk_disable_unprepare(ocmem->core_clk);
clk_disable_unprepare(ocmem->iface_clk);
struct ocmem;
struct ocmem *of_get_ocmem(struct device *dev);
struct ocmem_buf *ocmem_allocate(struct ocmem *ocmem, enum ocmem_client client,
void ocmem_free(struct ocmem *ocmem, enum ocmem_client client,
static inline struct ocmem *of_get_ocmem(struct device *dev)
static inline struct ocmem_buf *ocmem_allocate(struct ocmem *ocmem,
static inline void ocmem_free(struct ocmem *ocmem, enum ocmem_client client,