265 lines
6.1 KiB
C
265 lines
6.1 KiB
C
|
// SPDX-License-Identifier: GPL-2.0-only
|
||
|
/* Copyright (c) 2015-2018, The Linux Foundation. All rights reserved.
|
||
|
*/
|
||
|
|
||
|
#include "dpu_hwio.h"
|
||
|
#include "dpu_hw_catalog.h"
|
||
|
#include "dpu_hw_vbif.h"
|
||
|
|
||
|
#define VBIF_VERSION 0x0000
|
||
|
#define VBIF_CLK_FORCE_CTRL0 0x0008
|
||
|
#define VBIF_CLK_FORCE_CTRL1 0x000C
|
||
|
#define VBIF_QOS_REMAP_00 0x0020
|
||
|
#define VBIF_QOS_REMAP_01 0x0024
|
||
|
#define VBIF_QOS_REMAP_10 0x0028
|
||
|
#define VBIF_QOS_REMAP_11 0x002C
|
||
|
#define VBIF_WRITE_GATHER_EN 0x00AC
|
||
|
#define VBIF_IN_RD_LIM_CONF0 0x00B0
|
||
|
#define VBIF_IN_RD_LIM_CONF1 0x00B4
|
||
|
#define VBIF_IN_RD_LIM_CONF2 0x00B8
|
||
|
#define VBIF_IN_WR_LIM_CONF0 0x00C0
|
||
|
#define VBIF_IN_WR_LIM_CONF1 0x00C4
|
||
|
#define VBIF_IN_WR_LIM_CONF2 0x00C8
|
||
|
#define VBIF_OUT_RD_LIM_CONF0 0x00D0
|
||
|
#define VBIF_OUT_WR_LIM_CONF0 0x00D4
|
||
|
#define VBIF_OUT_AXI_AMEMTYPE_CONF0 0x0160
|
||
|
#define VBIF_OUT_AXI_AMEMTYPE_CONF1 0x0164
|
||
|
#define VBIF_XIN_PND_ERR 0x0190
|
||
|
#define VBIF_XIN_SRC_ERR 0x0194
|
||
|
#define VBIF_XIN_CLR_ERR 0x019C
|
||
|
#define VBIF_XIN_HALT_CTRL0 0x0200
|
||
|
#define VBIF_XIN_HALT_CTRL1 0x0204
|
||
|
#define VBIF_XINL_QOS_RP_REMAP_000 0x0550
|
||
|
#define VBIF_XINL_QOS_LVL_REMAP_000(vbif) (VBIF_XINL_QOS_RP_REMAP_000 + (vbif)->cap->qos_rp_remap_size)
|
||
|
|
||
|
static void dpu_hw_clear_errors(struct dpu_hw_vbif *vbif,
|
||
|
u32 *pnd_errors, u32 *src_errors)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c;
|
||
|
u32 pnd, src;
|
||
|
|
||
|
if (!vbif)
|
||
|
return;
|
||
|
c = &vbif->hw;
|
||
|
pnd = DPU_REG_READ(c, VBIF_XIN_PND_ERR);
|
||
|
src = DPU_REG_READ(c, VBIF_XIN_SRC_ERR);
|
||
|
|
||
|
if (pnd_errors)
|
||
|
*pnd_errors = pnd;
|
||
|
if (src_errors)
|
||
|
*src_errors = src;
|
||
|
|
||
|
DPU_REG_WRITE(c, VBIF_XIN_CLR_ERR, pnd | src);
|
||
|
}
|
||
|
|
||
|
static void dpu_hw_set_mem_type(struct dpu_hw_vbif *vbif,
|
||
|
u32 xin_id, u32 value)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c;
|
||
|
u32 reg_off;
|
||
|
u32 bit_off;
|
||
|
u32 reg_val;
|
||
|
|
||
|
/*
|
||
|
* Assume 4 bits per bit field, 8 fields per 32-bit register so
|
||
|
* 16 bit fields maximum across two registers
|
||
|
*/
|
||
|
if (!vbif || xin_id >= MAX_XIN_COUNT || xin_id >= 16)
|
||
|
return;
|
||
|
|
||
|
c = &vbif->hw;
|
||
|
|
||
|
if (xin_id >= 8) {
|
||
|
xin_id -= 8;
|
||
|
reg_off = VBIF_OUT_AXI_AMEMTYPE_CONF1;
|
||
|
} else {
|
||
|
reg_off = VBIF_OUT_AXI_AMEMTYPE_CONF0;
|
||
|
}
|
||
|
bit_off = (xin_id & 0x7) * 4;
|
||
|
reg_val = DPU_REG_READ(c, reg_off);
|
||
|
reg_val &= ~(0x7 << bit_off);
|
||
|
reg_val |= (value & 0x7) << bit_off;
|
||
|
DPU_REG_WRITE(c, reg_off, reg_val);
|
||
|
}
|
||
|
|
||
|
static void dpu_hw_set_limit_conf(struct dpu_hw_vbif *vbif,
|
||
|
u32 xin_id, bool rd, u32 limit)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c = &vbif->hw;
|
||
|
u32 reg_val;
|
||
|
u32 reg_off;
|
||
|
u32 bit_off;
|
||
|
|
||
|
if (rd)
|
||
|
reg_off = VBIF_IN_RD_LIM_CONF0;
|
||
|
else
|
||
|
reg_off = VBIF_IN_WR_LIM_CONF0;
|
||
|
|
||
|
reg_off += (xin_id / 4) * 4;
|
||
|
bit_off = (xin_id % 4) * 8;
|
||
|
reg_val = DPU_REG_READ(c, reg_off);
|
||
|
reg_val &= ~(0xFF << bit_off);
|
||
|
reg_val |= (limit) << bit_off;
|
||
|
DPU_REG_WRITE(c, reg_off, reg_val);
|
||
|
}
|
||
|
|
||
|
static u32 dpu_hw_get_limit_conf(struct dpu_hw_vbif *vbif,
|
||
|
u32 xin_id, bool rd)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c = &vbif->hw;
|
||
|
u32 reg_val;
|
||
|
u32 reg_off;
|
||
|
u32 bit_off;
|
||
|
u32 limit;
|
||
|
|
||
|
if (rd)
|
||
|
reg_off = VBIF_IN_RD_LIM_CONF0;
|
||
|
else
|
||
|
reg_off = VBIF_IN_WR_LIM_CONF0;
|
||
|
|
||
|
reg_off += (xin_id / 4) * 4;
|
||
|
bit_off = (xin_id % 4) * 8;
|
||
|
reg_val = DPU_REG_READ(c, reg_off);
|
||
|
limit = (reg_val >> bit_off) & 0xFF;
|
||
|
|
||
|
return limit;
|
||
|
}
|
||
|
|
||
|
static void dpu_hw_set_halt_ctrl(struct dpu_hw_vbif *vbif,
|
||
|
u32 xin_id, bool enable)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c = &vbif->hw;
|
||
|
u32 reg_val;
|
||
|
|
||
|
reg_val = DPU_REG_READ(c, VBIF_XIN_HALT_CTRL0);
|
||
|
|
||
|
if (enable)
|
||
|
reg_val |= BIT(xin_id);
|
||
|
else
|
||
|
reg_val &= ~BIT(xin_id);
|
||
|
|
||
|
DPU_REG_WRITE(c, VBIF_XIN_HALT_CTRL0, reg_val);
|
||
|
}
|
||
|
|
||
|
static bool dpu_hw_get_halt_ctrl(struct dpu_hw_vbif *vbif,
|
||
|
u32 xin_id)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c = &vbif->hw;
|
||
|
u32 reg_val;
|
||
|
|
||
|
reg_val = DPU_REG_READ(c, VBIF_XIN_HALT_CTRL1);
|
||
|
|
||
|
return (reg_val & BIT(xin_id)) ? true : false;
|
||
|
}
|
||
|
|
||
|
static void dpu_hw_set_qos_remap(struct dpu_hw_vbif *vbif,
|
||
|
u32 xin_id, u32 level, u32 remap_level)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c;
|
||
|
u32 reg_lvl, reg_val, reg_val_lvl, mask, reg_high, reg_shift;
|
||
|
|
||
|
if (!vbif)
|
||
|
return;
|
||
|
|
||
|
c = &vbif->hw;
|
||
|
|
||
|
reg_lvl = VBIF_XINL_QOS_LVL_REMAP_000(vbif);
|
||
|
reg_high = ((xin_id & 0x8) >> 3) * 4 + (level * 8);
|
||
|
reg_shift = (xin_id & 0x7) * 4;
|
||
|
|
||
|
reg_val = DPU_REG_READ(c, VBIF_XINL_QOS_RP_REMAP_000 + reg_high);
|
||
|
reg_val_lvl = DPU_REG_READ(c, reg_lvl + reg_high);
|
||
|
|
||
|
mask = 0x7 << reg_shift;
|
||
|
|
||
|
reg_val &= ~mask;
|
||
|
reg_val |= (remap_level << reg_shift) & mask;
|
||
|
|
||
|
reg_val_lvl &= ~mask;
|
||
|
reg_val_lvl |= (remap_level << reg_shift) & mask;
|
||
|
|
||
|
DPU_REG_WRITE(c, VBIF_XINL_QOS_RP_REMAP_000 + reg_high, reg_val);
|
||
|
DPU_REG_WRITE(c, reg_lvl + reg_high, reg_val_lvl);
|
||
|
}
|
||
|
|
||
|
static void dpu_hw_set_write_gather_en(struct dpu_hw_vbif *vbif, u32 xin_id)
|
||
|
{
|
||
|
struct dpu_hw_blk_reg_map *c;
|
||
|
u32 reg_val;
|
||
|
|
||
|
if (!vbif || xin_id >= MAX_XIN_COUNT)
|
||
|
return;
|
||
|
|
||
|
c = &vbif->hw;
|
||
|
|
||
|
reg_val = DPU_REG_READ(c, VBIF_WRITE_GATHER_EN);
|
||
|
reg_val |= BIT(xin_id);
|
||
|
DPU_REG_WRITE(c, VBIF_WRITE_GATHER_EN, reg_val);
|
||
|
}
|
||
|
|
||
|
static void _setup_vbif_ops(struct dpu_hw_vbif_ops *ops,
|
||
|
unsigned long cap)
|
||
|
{
|
||
|
ops->set_limit_conf = dpu_hw_set_limit_conf;
|
||
|
ops->get_limit_conf = dpu_hw_get_limit_conf;
|
||
|
ops->set_halt_ctrl = dpu_hw_set_halt_ctrl;
|
||
|
ops->get_halt_ctrl = dpu_hw_get_halt_ctrl;
|
||
|
if (test_bit(DPU_VBIF_QOS_REMAP, &cap))
|
||
|
ops->set_qos_remap = dpu_hw_set_qos_remap;
|
||
|
ops->set_mem_type = dpu_hw_set_mem_type;
|
||
|
ops->clear_errors = dpu_hw_clear_errors;
|
||
|
ops->set_write_gather_en = dpu_hw_set_write_gather_en;
|
||
|
}
|
||
|
|
||
|
static const struct dpu_vbif_cfg *_top_offset(enum dpu_vbif vbif,
|
||
|
const struct dpu_mdss_cfg *m,
|
||
|
void __iomem *addr,
|
||
|
struct dpu_hw_blk_reg_map *b)
|
||
|
{
|
||
|
int i;
|
||
|
|
||
|
for (i = 0; i < m->vbif_count; i++) {
|
||
|
if (vbif == m->vbif[i].id) {
|
||
|
b->blk_addr = addr + m->vbif[i].base;
|
||
|
b->log_mask = DPU_DBG_MASK_VBIF;
|
||
|
return &m->vbif[i];
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return ERR_PTR(-EINVAL);
|
||
|
}
|
||
|
|
||
|
struct dpu_hw_vbif *dpu_hw_vbif_init(enum dpu_vbif idx,
|
||
|
void __iomem *addr,
|
||
|
const struct dpu_mdss_cfg *m)
|
||
|
{
|
||
|
struct dpu_hw_vbif *c;
|
||
|
const struct dpu_vbif_cfg *cfg;
|
||
|
|
||
|
c = kzalloc(sizeof(*c), GFP_KERNEL);
|
||
|
if (!c)
|
||
|
return ERR_PTR(-ENOMEM);
|
||
|
|
||
|
cfg = _top_offset(idx, m, addr, &c->hw);
|
||
|
if (IS_ERR_OR_NULL(cfg)) {
|
||
|
kfree(c);
|
||
|
return ERR_PTR(-EINVAL);
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* Assign ops
|
||
|
*/
|
||
|
c->idx = idx;
|
||
|
c->cap = cfg;
|
||
|
_setup_vbif_ops(&c->ops, c->cap->features);
|
||
|
|
||
|
/* no need to register sub-range in dpu dbg, dump entire vbif io base */
|
||
|
|
||
|
return c;
|
||
|
}
|
||
|
|
||
|
void dpu_hw_vbif_destroy(struct dpu_hw_vbif *vbif)
|
||
|
{
|
||
|
kfree(vbif);
|
||
|
}
|