595 lines
16 KiB
C
595 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* LocalPlus Bus FIFO driver for the Freescale MPC52xx.
|
|
*
|
|
* Copyright (C) 2009 Secret Lab Technologies Ltd.
|
|
*
|
|
* Todo:
|
|
* - Add support for multiple requests to be queued.
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/of_irq.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/module.h>
|
|
#include <asm/io.h>
|
|
#include <asm/mpc52xx.h>
|
|
#include <asm/time.h>
|
|
|
|
#include <linux/fsl/bestcomm/bestcomm.h>
|
|
#include <linux/fsl/bestcomm/bestcomm_priv.h>
|
|
#include <linux/fsl/bestcomm/gen_bd.h>
|
|
|
|
MODULE_AUTHOR("Grant Likely <grant.likely@secretlab.ca>");
|
|
MODULE_DESCRIPTION("MPC5200 LocalPlus FIFO device driver");
|
|
MODULE_LICENSE("GPL");
|
|
|
|
#define LPBFIFO_REG_PACKET_SIZE (0x00)
|
|
#define LPBFIFO_REG_START_ADDRESS (0x04)
|
|
#define LPBFIFO_REG_CONTROL (0x08)
|
|
#define LPBFIFO_REG_ENABLE (0x0C)
|
|
#define LPBFIFO_REG_BYTES_DONE_STATUS (0x14)
|
|
#define LPBFIFO_REG_FIFO_DATA (0x40)
|
|
#define LPBFIFO_REG_FIFO_STATUS (0x44)
|
|
#define LPBFIFO_REG_FIFO_CONTROL (0x48)
|
|
#define LPBFIFO_REG_FIFO_ALARM (0x4C)
|
|
|
|
struct mpc52xx_lpbfifo {
|
|
struct device *dev;
|
|
phys_addr_t regs_phys;
|
|
void __iomem *regs;
|
|
int irq;
|
|
spinlock_t lock;
|
|
|
|
struct bcom_task *bcom_tx_task;
|
|
struct bcom_task *bcom_rx_task;
|
|
struct bcom_task *bcom_cur_task;
|
|
|
|
/* Current state data */
|
|
struct mpc52xx_lpbfifo_request *req;
|
|
int dma_irqs_enabled;
|
|
};
|
|
|
|
/* The MPC5200 has only one fifo, so only need one instance structure */
|
|
static struct mpc52xx_lpbfifo lpbfifo;
|
|
|
|
/**
|
|
* mpc52xx_lpbfifo_kick - Trigger the next block of data to be transferred
|
|
*
|
|
* @req: Pointer to request structure
|
|
*/
|
|
static void mpc52xx_lpbfifo_kick(struct mpc52xx_lpbfifo_request *req)
|
|
{
|
|
size_t transfer_size = req->size - req->pos;
|
|
struct bcom_bd *bd;
|
|
void __iomem *reg;
|
|
u32 *data;
|
|
int i;
|
|
int bit_fields;
|
|
int dma = !(req->flags & MPC52XX_LPBFIFO_FLAG_NO_DMA);
|
|
int write = req->flags & MPC52XX_LPBFIFO_FLAG_WRITE;
|
|
int poll_dma = req->flags & MPC52XX_LPBFIFO_FLAG_POLL_DMA;
|
|
|
|
/* Set and clear the reset bits; is good practice in User Manual */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, 0x01010000);
|
|
|
|
/* set master enable bit */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, 0x00000001);
|
|
if (!dma) {
|
|
/* While the FIFO can be setup for transfer sizes as large as
|
|
* 16M-1, the FIFO itself is only 512 bytes deep and it does
|
|
* not generate interrupts for FIFO full events (only transfer
|
|
* complete will raise an IRQ). Therefore when not using
|
|
* Bestcomm to drive the FIFO it needs to either be polled, or
|
|
* transfers need to constrained to the size of the fifo.
|
|
*
|
|
* This driver restricts the size of the transfer
|
|
*/
|
|
if (transfer_size > 512)
|
|
transfer_size = 512;
|
|
|
|
/* Load the FIFO with data */
|
|
if (write) {
|
|
reg = lpbfifo.regs + LPBFIFO_REG_FIFO_DATA;
|
|
data = req->data + req->pos;
|
|
for (i = 0; i < transfer_size; i += 4)
|
|
out_be32(reg, *data++);
|
|
}
|
|
|
|
/* Unmask both error and completion irqs */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, 0x00000301);
|
|
} else {
|
|
/* Choose the correct direction
|
|
*
|
|
* Configure the watermarks so DMA will always complete correctly.
|
|
* It may be worth experimenting with the ALARM value to see if
|
|
* there is a performance impact. However, if it is wrong there
|
|
* is a risk of DMA not transferring the last chunk of data
|
|
*/
|
|
if (write) {
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_FIFO_ALARM, 0x1e4);
|
|
out_8(lpbfifo.regs + LPBFIFO_REG_FIFO_CONTROL, 7);
|
|
lpbfifo.bcom_cur_task = lpbfifo.bcom_tx_task;
|
|
} else {
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_FIFO_ALARM, 0x1ff);
|
|
out_8(lpbfifo.regs + LPBFIFO_REG_FIFO_CONTROL, 0);
|
|
lpbfifo.bcom_cur_task = lpbfifo.bcom_rx_task;
|
|
|
|
if (poll_dma) {
|
|
if (lpbfifo.dma_irqs_enabled) {
|
|
disable_irq(bcom_get_task_irq(lpbfifo.bcom_rx_task));
|
|
lpbfifo.dma_irqs_enabled = 0;
|
|
}
|
|
} else {
|
|
if (!lpbfifo.dma_irqs_enabled) {
|
|
enable_irq(bcom_get_task_irq(lpbfifo.bcom_rx_task));
|
|
lpbfifo.dma_irqs_enabled = 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
bd = bcom_prepare_next_buffer(lpbfifo.bcom_cur_task);
|
|
bd->status = transfer_size;
|
|
if (!write) {
|
|
/*
|
|
* In the DMA read case, the DMA doesn't complete,
|
|
* possibly due to incorrect watermarks in the ALARM
|
|
* and CONTROL regs. For now instead of trying to
|
|
* determine the right watermarks that will make this
|
|
* work, just increase the number of bytes the FIFO is
|
|
* expecting.
|
|
*
|
|
* When submitting another operation, the FIFO will get
|
|
* reset, so the condition of the FIFO waiting for a
|
|
* non-existent 4 bytes will get cleared.
|
|
*/
|
|
transfer_size += 4; /* BLECH! */
|
|
}
|
|
bd->data[0] = req->data_phys + req->pos;
|
|
bcom_submit_next_buffer(lpbfifo.bcom_cur_task, NULL);
|
|
|
|
/* error irq & master enabled bit */
|
|
bit_fields = 0x00000201;
|
|
|
|
/* Unmask irqs */
|
|
if (write && (!poll_dma))
|
|
bit_fields |= 0x00000100; /* completion irq too */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, bit_fields);
|
|
}
|
|
|
|
/* Set transfer size, width, chip select and READ mode */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_START_ADDRESS,
|
|
req->offset + req->pos);
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_PACKET_SIZE, transfer_size);
|
|
|
|
bit_fields = req->cs << 24 | 0x000008;
|
|
if (!write)
|
|
bit_fields |= 0x010000; /* read mode */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_CONTROL, bit_fields);
|
|
|
|
/* Kick it off */
|
|
if (!lpbfifo.req->defer_xfer_start)
|
|
out_8(lpbfifo.regs + LPBFIFO_REG_PACKET_SIZE, 0x01);
|
|
if (dma)
|
|
bcom_enable(lpbfifo.bcom_cur_task);
|
|
}
|
|
|
|
/**
|
|
* mpc52xx_lpbfifo_irq - IRQ handler for LPB FIFO
|
|
* @irq: IRQ number to be handled
|
|
* @dev_id: device ID cookie
|
|
*
|
|
* On transmit, the dma completion irq triggers before the fifo completion
|
|
* triggers. Handle the dma completion here instead of the LPB FIFO Bestcomm
|
|
* task completion irq because everything is not really done until the LPB FIFO
|
|
* completion irq triggers.
|
|
*
|
|
* In other words:
|
|
* For DMA, on receive, the "Fat Lady" is the bestcom completion irq. on
|
|
* transmit, the fifo completion irq is the "Fat Lady". The opera (or in this
|
|
* case the DMA/FIFO operation) is not finished until the "Fat Lady" sings.
|
|
*
|
|
* Reasons for entering this routine:
|
|
* 1) PIO mode rx and tx completion irq
|
|
* 2) DMA interrupt mode tx completion irq
|
|
* 3) DMA polled mode tx
|
|
*
|
|
* Exit conditions:
|
|
* 1) Transfer aborted
|
|
* 2) FIFO complete without DMA; more data to do
|
|
* 3) FIFO complete without DMA; all data transferred
|
|
* 4) FIFO complete using DMA
|
|
*
|
|
* Condition 1 can occur regardless of whether or not DMA is used.
|
|
* It requires executing the callback to report the error and exiting
|
|
* immediately.
|
|
*
|
|
* Condition 2 requires programming the FIFO with the next block of data
|
|
*
|
|
* Condition 3 requires executing the callback to report completion
|
|
*
|
|
* Condition 4 means the same as 3, except that we also retrieve the bcom
|
|
* buffer so DMA doesn't get clogged up.
|
|
*
|
|
* To make things trickier, the spinlock must be dropped before
|
|
* executing the callback, otherwise we could end up with a deadlock
|
|
* or nested spinlock condition. The out path is non-trivial, so
|
|
* extra fiddling is done to make sure all paths lead to the same
|
|
* outbound code.
|
|
*
|
|
* Return: irqreturn code (%IRQ_HANDLED)
|
|
*/
|
|
static irqreturn_t mpc52xx_lpbfifo_irq(int irq, void *dev_id)
|
|
{
|
|
struct mpc52xx_lpbfifo_request *req;
|
|
u32 status = in_8(lpbfifo.regs + LPBFIFO_REG_BYTES_DONE_STATUS);
|
|
void __iomem *reg;
|
|
u32 *data;
|
|
int count, i;
|
|
int do_callback = 0;
|
|
u32 ts;
|
|
unsigned long flags;
|
|
int dma, write, poll_dma;
|
|
|
|
spin_lock_irqsave(&lpbfifo.lock, flags);
|
|
ts = mftb();
|
|
|
|
req = lpbfifo.req;
|
|
if (!req) {
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
pr_err("bogus LPBFIFO IRQ\n");
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
dma = !(req->flags & MPC52XX_LPBFIFO_FLAG_NO_DMA);
|
|
write = req->flags & MPC52XX_LPBFIFO_FLAG_WRITE;
|
|
poll_dma = req->flags & MPC52XX_LPBFIFO_FLAG_POLL_DMA;
|
|
|
|
if (dma && !write) {
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
pr_err("bogus LPBFIFO IRQ (dma and not writing)\n");
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
if ((status & 0x01) == 0) {
|
|
goto out;
|
|
}
|
|
|
|
/* check abort bit */
|
|
if (status & 0x10) {
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, 0x01010000);
|
|
do_callback = 1;
|
|
goto out;
|
|
}
|
|
|
|
/* Read result from hardware */
|
|
count = in_be32(lpbfifo.regs + LPBFIFO_REG_BYTES_DONE_STATUS);
|
|
count &= 0x00ffffff;
|
|
|
|
if (!dma && !write) {
|
|
/* copy the data out of the FIFO */
|
|
reg = lpbfifo.regs + LPBFIFO_REG_FIFO_DATA;
|
|
data = req->data + req->pos;
|
|
for (i = 0; i < count; i += 4)
|
|
*data++ = in_be32(reg);
|
|
}
|
|
|
|
/* Update transfer position and count */
|
|
req->pos += count;
|
|
|
|
/* Decide what to do next */
|
|
if (req->size - req->pos)
|
|
mpc52xx_lpbfifo_kick(req); /* more work to do */
|
|
else
|
|
do_callback = 1;
|
|
|
|
out:
|
|
/* Clear the IRQ */
|
|
out_8(lpbfifo.regs + LPBFIFO_REG_BYTES_DONE_STATUS, 0x01);
|
|
|
|
if (dma && (status & 0x11)) {
|
|
/*
|
|
* Count the DMA as complete only when the FIFO completion
|
|
* status or abort bits are set.
|
|
*
|
|
* (status & 0x01) should always be the case except sometimes
|
|
* when using polled DMA.
|
|
*
|
|
* (status & 0x10) {transfer aborted}: This case needs more
|
|
* testing.
|
|
*/
|
|
bcom_retrieve_buffer(lpbfifo.bcom_cur_task, &status, NULL);
|
|
}
|
|
req->last_byte = ((u8 *)req->data)[req->size - 1];
|
|
|
|
/* When the do_callback flag is set; it means the transfer is finished
|
|
* so set the FIFO as idle */
|
|
if (do_callback)
|
|
lpbfifo.req = NULL;
|
|
|
|
if (irq != 0) /* don't increment on polled case */
|
|
req->irq_count++;
|
|
|
|
req->irq_ticks += mftb() - ts;
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
|
|
/* Spinlock is released; it is now safe to call the callback */
|
|
if (do_callback && req->callback)
|
|
req->callback(req);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/**
|
|
* mpc52xx_lpbfifo_bcom_irq - IRQ handler for LPB FIFO Bestcomm task
|
|
* @irq: IRQ number to be handled
|
|
* @dev_id: device ID cookie
|
|
*
|
|
* Only used when receiving data.
|
|
*
|
|
* Return: irqreturn code (%IRQ_HANDLED)
|
|
*/
|
|
static irqreturn_t mpc52xx_lpbfifo_bcom_irq(int irq, void *dev_id)
|
|
{
|
|
struct mpc52xx_lpbfifo_request *req;
|
|
unsigned long flags;
|
|
u32 status;
|
|
u32 ts;
|
|
|
|
spin_lock_irqsave(&lpbfifo.lock, flags);
|
|
ts = mftb();
|
|
|
|
req = lpbfifo.req;
|
|
if (!req || (req->flags & MPC52XX_LPBFIFO_FLAG_NO_DMA)) {
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
if (irq != 0) /* don't increment on polled case */
|
|
req->irq_count++;
|
|
|
|
if (!bcom_buffer_done(lpbfifo.bcom_cur_task)) {
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
|
|
req->buffer_not_done_cnt++;
|
|
if ((req->buffer_not_done_cnt % 1000) == 0)
|
|
pr_err("transfer stalled\n");
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
bcom_retrieve_buffer(lpbfifo.bcom_cur_task, &status, NULL);
|
|
|
|
req->last_byte = ((u8 *)req->data)[req->size - 1];
|
|
|
|
req->pos = status & 0x00ffffff;
|
|
|
|
/* Mark the FIFO as idle */
|
|
lpbfifo.req = NULL;
|
|
|
|
/* Release the lock before calling out to the callback. */
|
|
req->irq_ticks += mftb() - ts;
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
|
|
if (req->callback)
|
|
req->callback(req);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/**
|
|
* mpc52xx_lpbfifo_poll - Poll for DMA completion
|
|
*/
|
|
void mpc52xx_lpbfifo_poll(void)
|
|
{
|
|
struct mpc52xx_lpbfifo_request *req = lpbfifo.req;
|
|
int dma = !(req->flags & MPC52XX_LPBFIFO_FLAG_NO_DMA);
|
|
int write = req->flags & MPC52XX_LPBFIFO_FLAG_WRITE;
|
|
|
|
/*
|
|
* For more information, see comments on the "Fat Lady"
|
|
*/
|
|
if (dma && write)
|
|
mpc52xx_lpbfifo_irq(0, NULL);
|
|
else
|
|
mpc52xx_lpbfifo_bcom_irq(0, NULL);
|
|
}
|
|
EXPORT_SYMBOL(mpc52xx_lpbfifo_poll);
|
|
|
|
/**
|
|
* mpc52xx_lpbfifo_submit - Submit an LPB FIFO transfer request.
|
|
* @req: Pointer to request structure
|
|
*
|
|
* Return: %0 on success, -errno code on error
|
|
*/
|
|
int mpc52xx_lpbfifo_submit(struct mpc52xx_lpbfifo_request *req)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if (!lpbfifo.regs)
|
|
return -ENODEV;
|
|
|
|
spin_lock_irqsave(&lpbfifo.lock, flags);
|
|
|
|
/* If the req pointer is already set, then a transfer is in progress */
|
|
if (lpbfifo.req) {
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
return -EBUSY;
|
|
}
|
|
|
|
/* Setup the transfer */
|
|
lpbfifo.req = req;
|
|
req->irq_count = 0;
|
|
req->irq_ticks = 0;
|
|
req->buffer_not_done_cnt = 0;
|
|
req->pos = 0;
|
|
|
|
mpc52xx_lpbfifo_kick(req);
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(mpc52xx_lpbfifo_submit);
|
|
|
|
int mpc52xx_lpbfifo_start_xfer(struct mpc52xx_lpbfifo_request *req)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if (!lpbfifo.regs)
|
|
return -ENODEV;
|
|
|
|
spin_lock_irqsave(&lpbfifo.lock, flags);
|
|
|
|
/*
|
|
* If the req pointer is already set and a transfer was
|
|
* started on submit, then this transfer is in progress
|
|
*/
|
|
if (lpbfifo.req && !lpbfifo.req->defer_xfer_start) {
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
return -EBUSY;
|
|
}
|
|
|
|
/*
|
|
* If the req was previously submitted but not
|
|
* started, start it now
|
|
*/
|
|
if (lpbfifo.req && lpbfifo.req == req &&
|
|
lpbfifo.req->defer_xfer_start) {
|
|
out_8(lpbfifo.regs + LPBFIFO_REG_PACKET_SIZE, 0x01);
|
|
}
|
|
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(mpc52xx_lpbfifo_start_xfer);
|
|
|
|
void mpc52xx_lpbfifo_abort(struct mpc52xx_lpbfifo_request *req)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&lpbfifo.lock, flags);
|
|
if (lpbfifo.req == req) {
|
|
/* Put it into reset and clear the state */
|
|
bcom_gen_bd_rx_reset(lpbfifo.bcom_rx_task);
|
|
bcom_gen_bd_tx_reset(lpbfifo.bcom_tx_task);
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, 0x01010000);
|
|
lpbfifo.req = NULL;
|
|
}
|
|
spin_unlock_irqrestore(&lpbfifo.lock, flags);
|
|
}
|
|
EXPORT_SYMBOL(mpc52xx_lpbfifo_abort);
|
|
|
|
static int mpc52xx_lpbfifo_probe(struct platform_device *op)
|
|
{
|
|
struct resource res;
|
|
int rc = -ENOMEM;
|
|
|
|
if (lpbfifo.dev != NULL)
|
|
return -ENOSPC;
|
|
|
|
lpbfifo.irq = irq_of_parse_and_map(op->dev.of_node, 0);
|
|
if (!lpbfifo.irq)
|
|
return -ENODEV;
|
|
|
|
if (of_address_to_resource(op->dev.of_node, 0, &res))
|
|
return -ENODEV;
|
|
lpbfifo.regs_phys = res.start;
|
|
lpbfifo.regs = of_iomap(op->dev.of_node, 0);
|
|
if (!lpbfifo.regs)
|
|
return -ENOMEM;
|
|
|
|
spin_lock_init(&lpbfifo.lock);
|
|
|
|
/* Put FIFO into reset */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, 0x01010000);
|
|
|
|
/* Register the interrupt handler */
|
|
rc = request_irq(lpbfifo.irq, mpc52xx_lpbfifo_irq, 0,
|
|
"mpc52xx-lpbfifo", &lpbfifo);
|
|
if (rc)
|
|
goto err_irq;
|
|
|
|
/* Request the Bestcomm receive (fifo --> memory) task and IRQ */
|
|
lpbfifo.bcom_rx_task =
|
|
bcom_gen_bd_rx_init(2, res.start + LPBFIFO_REG_FIFO_DATA,
|
|
BCOM_INITIATOR_SCLPC, BCOM_IPR_SCLPC,
|
|
16*1024*1024);
|
|
if (!lpbfifo.bcom_rx_task)
|
|
goto err_bcom_rx;
|
|
|
|
rc = request_irq(bcom_get_task_irq(lpbfifo.bcom_rx_task),
|
|
mpc52xx_lpbfifo_bcom_irq, 0,
|
|
"mpc52xx-lpbfifo-rx", &lpbfifo);
|
|
if (rc)
|
|
goto err_bcom_rx_irq;
|
|
|
|
lpbfifo.dma_irqs_enabled = 1;
|
|
|
|
/* Request the Bestcomm transmit (memory --> fifo) task and IRQ */
|
|
lpbfifo.bcom_tx_task =
|
|
bcom_gen_bd_tx_init(2, res.start + LPBFIFO_REG_FIFO_DATA,
|
|
BCOM_INITIATOR_SCLPC, BCOM_IPR_SCLPC);
|
|
if (!lpbfifo.bcom_tx_task)
|
|
goto err_bcom_tx;
|
|
|
|
lpbfifo.dev = &op->dev;
|
|
return 0;
|
|
|
|
err_bcom_tx:
|
|
free_irq(bcom_get_task_irq(lpbfifo.bcom_rx_task), &lpbfifo);
|
|
err_bcom_rx_irq:
|
|
bcom_gen_bd_rx_release(lpbfifo.bcom_rx_task);
|
|
err_bcom_rx:
|
|
free_irq(lpbfifo.irq, &lpbfifo);
|
|
err_irq:
|
|
iounmap(lpbfifo.regs);
|
|
lpbfifo.regs = NULL;
|
|
|
|
dev_err(&op->dev, "mpc52xx_lpbfifo_probe() failed\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
|
|
static int mpc52xx_lpbfifo_remove(struct platform_device *op)
|
|
{
|
|
if (lpbfifo.dev != &op->dev)
|
|
return 0;
|
|
|
|
/* Put FIFO in reset */
|
|
out_be32(lpbfifo.regs + LPBFIFO_REG_ENABLE, 0x01010000);
|
|
|
|
/* Release the bestcomm transmit task */
|
|
free_irq(bcom_get_task_irq(lpbfifo.bcom_tx_task), &lpbfifo);
|
|
bcom_gen_bd_tx_release(lpbfifo.bcom_tx_task);
|
|
|
|
/* Release the bestcomm receive task */
|
|
free_irq(bcom_get_task_irq(lpbfifo.bcom_rx_task), &lpbfifo);
|
|
bcom_gen_bd_rx_release(lpbfifo.bcom_rx_task);
|
|
|
|
free_irq(lpbfifo.irq, &lpbfifo);
|
|
iounmap(lpbfifo.regs);
|
|
lpbfifo.regs = NULL;
|
|
lpbfifo.dev = NULL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct of_device_id mpc52xx_lpbfifo_match[] = {
|
|
{ .compatible = "fsl,mpc5200-lpbfifo", },
|
|
{},
|
|
};
|
|
MODULE_DEVICE_TABLE(of, mpc52xx_lpbfifo_match);
|
|
|
|
static struct platform_driver mpc52xx_lpbfifo_driver = {
|
|
.driver = {
|
|
.name = "mpc52xx-lpbfifo",
|
|
.of_match_table = mpc52xx_lpbfifo_match,
|
|
},
|
|
.probe = mpc52xx_lpbfifo_probe,
|
|
.remove = mpc52xx_lpbfifo_remove,
|
|
};
|
|
module_platform_driver(mpc52xx_lpbfifo_driver);
|