401 lines
10 KiB
C
401 lines
10 KiB
C
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
||
|
/* backing_ops.c - query/set operations on saved SPU context.
|
||
|
*
|
||
|
* Copyright (C) IBM 2005
|
||
|
* Author: Mark Nutter <mnutter@us.ibm.com>
|
||
|
*
|
||
|
* These register operations allow SPUFS to operate on saved
|
||
|
* SPU contexts rather than hardware.
|
||
|
*/
|
||
|
|
||
|
#include <linux/errno.h>
|
||
|
#include <linux/sched.h>
|
||
|
#include <linux/kernel.h>
|
||
|
#include <linux/mm.h>
|
||
|
#include <linux/vmalloc.h>
|
||
|
#include <linux/smp.h>
|
||
|
#include <linux/stddef.h>
|
||
|
#include <linux/unistd.h>
|
||
|
#include <linux/poll.h>
|
||
|
|
||
|
#include <asm/io.h>
|
||
|
#include <asm/spu.h>
|
||
|
#include <asm/spu_csa.h>
|
||
|
#include <asm/spu_info.h>
|
||
|
#include <asm/mmu_context.h>
|
||
|
#include "spufs.h"
|
||
|
|
||
|
/*
|
||
|
* Reads/writes to various problem and priv2 registers require
|
||
|
* state changes, i.e. generate SPU events, modify channel
|
||
|
* counts, etc.
|
||
|
*/
|
||
|
|
||
|
static void gen_spu_event(struct spu_context *ctx, u32 event)
|
||
|
{
|
||
|
u64 ch0_cnt;
|
||
|
u64 ch0_data;
|
||
|
u64 ch1_data;
|
||
|
|
||
|
ch0_cnt = ctx->csa.spu_chnlcnt_RW[0];
|
||
|
ch0_data = ctx->csa.spu_chnldata_RW[0];
|
||
|
ch1_data = ctx->csa.spu_chnldata_RW[1];
|
||
|
ctx->csa.spu_chnldata_RW[0] |= event;
|
||
|
if ((ch0_cnt == 0) && !(ch0_data & event) && (ch1_data & event)) {
|
||
|
ctx->csa.spu_chnlcnt_RW[0] = 1;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
static int spu_backing_mbox_read(struct spu_context *ctx, u32 * data)
|
||
|
{
|
||
|
u32 mbox_stat;
|
||
|
int ret = 0;
|
||
|
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
mbox_stat = ctx->csa.prob.mb_stat_R;
|
||
|
if (mbox_stat & 0x0000ff) {
|
||
|
/* Read the first available word.
|
||
|
* Implementation note: the depth
|
||
|
* of pu_mb_R is currently 1.
|
||
|
*/
|
||
|
*data = ctx->csa.prob.pu_mb_R;
|
||
|
ctx->csa.prob.mb_stat_R &= ~(0x0000ff);
|
||
|
ctx->csa.spu_chnlcnt_RW[28] = 1;
|
||
|
gen_spu_event(ctx, MFC_PU_MAILBOX_AVAILABLE_EVENT);
|
||
|
ret = 4;
|
||
|
}
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_mbox_stat_read(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.prob.mb_stat_R;
|
||
|
}
|
||
|
|
||
|
static __poll_t spu_backing_mbox_stat_poll(struct spu_context *ctx,
|
||
|
__poll_t events)
|
||
|
{
|
||
|
__poll_t ret;
|
||
|
u32 stat;
|
||
|
|
||
|
ret = 0;
|
||
|
spin_lock_irq(&ctx->csa.register_lock);
|
||
|
stat = ctx->csa.prob.mb_stat_R;
|
||
|
|
||
|
/* if the requested event is there, return the poll
|
||
|
mask, otherwise enable the interrupt to get notified,
|
||
|
but first mark any pending interrupts as done so
|
||
|
we don't get woken up unnecessarily */
|
||
|
|
||
|
if (events & (EPOLLIN | EPOLLRDNORM)) {
|
||
|
if (stat & 0xff0000)
|
||
|
ret |= EPOLLIN | EPOLLRDNORM;
|
||
|
else {
|
||
|
ctx->csa.priv1.int_stat_class2_RW &=
|
||
|
~CLASS2_MAILBOX_INTR;
|
||
|
ctx->csa.priv1.int_mask_class2_RW |=
|
||
|
CLASS2_ENABLE_MAILBOX_INTR;
|
||
|
}
|
||
|
}
|
||
|
if (events & (EPOLLOUT | EPOLLWRNORM)) {
|
||
|
if (stat & 0x00ff00)
|
||
|
ret = EPOLLOUT | EPOLLWRNORM;
|
||
|
else {
|
||
|
ctx->csa.priv1.int_stat_class2_RW &=
|
||
|
~CLASS2_MAILBOX_THRESHOLD_INTR;
|
||
|
ctx->csa.priv1.int_mask_class2_RW |=
|
||
|
CLASS2_ENABLE_MAILBOX_THRESHOLD_INTR;
|
||
|
}
|
||
|
}
|
||
|
spin_unlock_irq(&ctx->csa.register_lock);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static int spu_backing_ibox_read(struct spu_context *ctx, u32 * data)
|
||
|
{
|
||
|
int ret;
|
||
|
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
if (ctx->csa.prob.mb_stat_R & 0xff0000) {
|
||
|
/* Read the first available word.
|
||
|
* Implementation note: the depth
|
||
|
* of puint_mb_R is currently 1.
|
||
|
*/
|
||
|
*data = ctx->csa.priv2.puint_mb_R;
|
||
|
ctx->csa.prob.mb_stat_R &= ~(0xff0000);
|
||
|
ctx->csa.spu_chnlcnt_RW[30] = 1;
|
||
|
gen_spu_event(ctx, MFC_PU_INT_MAILBOX_AVAILABLE_EVENT);
|
||
|
ret = 4;
|
||
|
} else {
|
||
|
/* make sure we get woken up by the interrupt */
|
||
|
ctx->csa.priv1.int_mask_class2_RW |= CLASS2_ENABLE_MAILBOX_INTR;
|
||
|
ret = 0;
|
||
|
}
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static int spu_backing_wbox_write(struct spu_context *ctx, u32 data)
|
||
|
{
|
||
|
int ret;
|
||
|
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
if ((ctx->csa.prob.mb_stat_R) & 0x00ff00) {
|
||
|
int slot = ctx->csa.spu_chnlcnt_RW[29];
|
||
|
int avail = (ctx->csa.prob.mb_stat_R & 0x00ff00) >> 8;
|
||
|
|
||
|
/* We have space to write wbox_data.
|
||
|
* Implementation note: the depth
|
||
|
* of spu_mb_W is currently 4.
|
||
|
*/
|
||
|
BUG_ON(avail != (4 - slot));
|
||
|
ctx->csa.spu_mailbox_data[slot] = data;
|
||
|
ctx->csa.spu_chnlcnt_RW[29] = ++slot;
|
||
|
ctx->csa.prob.mb_stat_R &= ~(0x00ff00);
|
||
|
ctx->csa.prob.mb_stat_R |= (((4 - slot) & 0xff) << 8);
|
||
|
gen_spu_event(ctx, MFC_SPU_MAILBOX_WRITTEN_EVENT);
|
||
|
ret = 4;
|
||
|
} else {
|
||
|
/* make sure we get woken up by the interrupt when space
|
||
|
becomes available */
|
||
|
ctx->csa.priv1.int_mask_class2_RW |=
|
||
|
CLASS2_ENABLE_MAILBOX_THRESHOLD_INTR;
|
||
|
ret = 0;
|
||
|
}
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_signal1_read(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.spu_chnldata_RW[3];
|
||
|
}
|
||
|
|
||
|
static void spu_backing_signal1_write(struct spu_context *ctx, u32 data)
|
||
|
{
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
if (ctx->csa.priv2.spu_cfg_RW & 0x1)
|
||
|
ctx->csa.spu_chnldata_RW[3] |= data;
|
||
|
else
|
||
|
ctx->csa.spu_chnldata_RW[3] = data;
|
||
|
ctx->csa.spu_chnlcnt_RW[3] = 1;
|
||
|
gen_spu_event(ctx, MFC_SIGNAL_1_EVENT);
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_signal2_read(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.spu_chnldata_RW[4];
|
||
|
}
|
||
|
|
||
|
static void spu_backing_signal2_write(struct spu_context *ctx, u32 data)
|
||
|
{
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
if (ctx->csa.priv2.spu_cfg_RW & 0x2)
|
||
|
ctx->csa.spu_chnldata_RW[4] |= data;
|
||
|
else
|
||
|
ctx->csa.spu_chnldata_RW[4] = data;
|
||
|
ctx->csa.spu_chnlcnt_RW[4] = 1;
|
||
|
gen_spu_event(ctx, MFC_SIGNAL_2_EVENT);
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
}
|
||
|
|
||
|
static void spu_backing_signal1_type_set(struct spu_context *ctx, u64 val)
|
||
|
{
|
||
|
u64 tmp;
|
||
|
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
tmp = ctx->csa.priv2.spu_cfg_RW;
|
||
|
if (val)
|
||
|
tmp |= 1;
|
||
|
else
|
||
|
tmp &= ~1;
|
||
|
ctx->csa.priv2.spu_cfg_RW = tmp;
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
}
|
||
|
|
||
|
static u64 spu_backing_signal1_type_get(struct spu_context *ctx)
|
||
|
{
|
||
|
return ((ctx->csa.priv2.spu_cfg_RW & 1) != 0);
|
||
|
}
|
||
|
|
||
|
static void spu_backing_signal2_type_set(struct spu_context *ctx, u64 val)
|
||
|
{
|
||
|
u64 tmp;
|
||
|
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
tmp = ctx->csa.priv2.spu_cfg_RW;
|
||
|
if (val)
|
||
|
tmp |= 2;
|
||
|
else
|
||
|
tmp &= ~2;
|
||
|
ctx->csa.priv2.spu_cfg_RW = tmp;
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
}
|
||
|
|
||
|
static u64 spu_backing_signal2_type_get(struct spu_context *ctx)
|
||
|
{
|
||
|
return ((ctx->csa.priv2.spu_cfg_RW & 2) != 0);
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_npc_read(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.prob.spu_npc_RW;
|
||
|
}
|
||
|
|
||
|
static void spu_backing_npc_write(struct spu_context *ctx, u32 val)
|
||
|
{
|
||
|
ctx->csa.prob.spu_npc_RW = val;
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_status_read(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.prob.spu_status_R;
|
||
|
}
|
||
|
|
||
|
static char *spu_backing_get_ls(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.lscsa->ls;
|
||
|
}
|
||
|
|
||
|
static void spu_backing_privcntl_write(struct spu_context *ctx, u64 val)
|
||
|
{
|
||
|
ctx->csa.priv2.spu_privcntl_RW = val;
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_runcntl_read(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.prob.spu_runcntl_RW;
|
||
|
}
|
||
|
|
||
|
static void spu_backing_runcntl_write(struct spu_context *ctx, u32 val)
|
||
|
{
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
ctx->csa.prob.spu_runcntl_RW = val;
|
||
|
if (val & SPU_RUNCNTL_RUNNABLE) {
|
||
|
ctx->csa.prob.spu_status_R &=
|
||
|
~SPU_STATUS_STOPPED_BY_STOP &
|
||
|
~SPU_STATUS_STOPPED_BY_HALT &
|
||
|
~SPU_STATUS_SINGLE_STEP &
|
||
|
~SPU_STATUS_INVALID_INSTR &
|
||
|
~SPU_STATUS_INVALID_CH;
|
||
|
ctx->csa.prob.spu_status_R |= SPU_STATUS_RUNNING;
|
||
|
} else {
|
||
|
ctx->csa.prob.spu_status_R &= ~SPU_STATUS_RUNNING;
|
||
|
}
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
}
|
||
|
|
||
|
static void spu_backing_runcntl_stop(struct spu_context *ctx)
|
||
|
{
|
||
|
spu_backing_runcntl_write(ctx, SPU_RUNCNTL_STOP);
|
||
|
}
|
||
|
|
||
|
static void spu_backing_master_start(struct spu_context *ctx)
|
||
|
{
|
||
|
struct spu_state *csa = &ctx->csa;
|
||
|
u64 sr1;
|
||
|
|
||
|
spin_lock(&csa->register_lock);
|
||
|
sr1 = csa->priv1.mfc_sr1_RW | MFC_STATE1_MASTER_RUN_CONTROL_MASK;
|
||
|
csa->priv1.mfc_sr1_RW = sr1;
|
||
|
spin_unlock(&csa->register_lock);
|
||
|
}
|
||
|
|
||
|
static void spu_backing_master_stop(struct spu_context *ctx)
|
||
|
{
|
||
|
struct spu_state *csa = &ctx->csa;
|
||
|
u64 sr1;
|
||
|
|
||
|
spin_lock(&csa->register_lock);
|
||
|
sr1 = csa->priv1.mfc_sr1_RW & ~MFC_STATE1_MASTER_RUN_CONTROL_MASK;
|
||
|
csa->priv1.mfc_sr1_RW = sr1;
|
||
|
spin_unlock(&csa->register_lock);
|
||
|
}
|
||
|
|
||
|
static int spu_backing_set_mfc_query(struct spu_context * ctx, u32 mask,
|
||
|
u32 mode)
|
||
|
{
|
||
|
struct spu_problem_collapsed *prob = &ctx->csa.prob;
|
||
|
int ret;
|
||
|
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
ret = -EAGAIN;
|
||
|
if (prob->dma_querytype_RW)
|
||
|
goto out;
|
||
|
ret = 0;
|
||
|
/* FIXME: what are the side-effects of this? */
|
||
|
prob->dma_querymask_RW = mask;
|
||
|
prob->dma_querytype_RW = mode;
|
||
|
/* In the current implementation, the SPU context is always
|
||
|
* acquired in runnable state when new bits are added to the
|
||
|
* mask (tagwait), so it's sufficient just to mask
|
||
|
* dma_tagstatus_R with the 'mask' parameter here.
|
||
|
*/
|
||
|
ctx->csa.prob.dma_tagstatus_R &= mask;
|
||
|
out:
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_read_mfc_tagstatus(struct spu_context * ctx)
|
||
|
{
|
||
|
return ctx->csa.prob.dma_tagstatus_R;
|
||
|
}
|
||
|
|
||
|
static u32 spu_backing_get_mfc_free_elements(struct spu_context *ctx)
|
||
|
{
|
||
|
return ctx->csa.prob.dma_qstatus_R;
|
||
|
}
|
||
|
|
||
|
static int spu_backing_send_mfc_command(struct spu_context *ctx,
|
||
|
struct mfc_dma_command *cmd)
|
||
|
{
|
||
|
int ret;
|
||
|
|
||
|
spin_lock(&ctx->csa.register_lock);
|
||
|
ret = -EAGAIN;
|
||
|
/* FIXME: set up priv2->puq */
|
||
|
spin_unlock(&ctx->csa.register_lock);
|
||
|
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static void spu_backing_restart_dma(struct spu_context *ctx)
|
||
|
{
|
||
|
ctx->csa.priv2.mfc_control_RW |= MFC_CNTL_RESTART_DMA_COMMAND;
|
||
|
}
|
||
|
|
||
|
struct spu_context_ops spu_backing_ops = {
|
||
|
.mbox_read = spu_backing_mbox_read,
|
||
|
.mbox_stat_read = spu_backing_mbox_stat_read,
|
||
|
.mbox_stat_poll = spu_backing_mbox_stat_poll,
|
||
|
.ibox_read = spu_backing_ibox_read,
|
||
|
.wbox_write = spu_backing_wbox_write,
|
||
|
.signal1_read = spu_backing_signal1_read,
|
||
|
.signal1_write = spu_backing_signal1_write,
|
||
|
.signal2_read = spu_backing_signal2_read,
|
||
|
.signal2_write = spu_backing_signal2_write,
|
||
|
.signal1_type_set = spu_backing_signal1_type_set,
|
||
|
.signal1_type_get = spu_backing_signal1_type_get,
|
||
|
.signal2_type_set = spu_backing_signal2_type_set,
|
||
|
.signal2_type_get = spu_backing_signal2_type_get,
|
||
|
.npc_read = spu_backing_npc_read,
|
||
|
.npc_write = spu_backing_npc_write,
|
||
|
.status_read = spu_backing_status_read,
|
||
|
.get_ls = spu_backing_get_ls,
|
||
|
.privcntl_write = spu_backing_privcntl_write,
|
||
|
.runcntl_read = spu_backing_runcntl_read,
|
||
|
.runcntl_write = spu_backing_runcntl_write,
|
||
|
.runcntl_stop = spu_backing_runcntl_stop,
|
||
|
.master_start = spu_backing_master_start,
|
||
|
.master_stop = spu_backing_master_stop,
|
||
|
.set_mfc_query = spu_backing_set_mfc_query,
|
||
|
.read_mfc_tagstatus = spu_backing_read_mfc_tagstatus,
|
||
|
.get_mfc_free_elements = spu_backing_get_mfc_free_elements,
|
||
|
.send_mfc_command = spu_backing_send_mfc_command,
|
||
|
.restart_dma = spu_backing_restart_dma,
|
||
|
};
|