297 lines
8.0 KiB
C
297 lines
8.0 KiB
C
|
// SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
|
||
|
/*
|
||
|
* Copyright (C) 2012-2014, 2018-2019, 2021 Intel Corporation
|
||
|
* Copyright (C) 2013-2015 Intel Mobile Communications GmbH
|
||
|
* Copyright (C) 2016-2017 Intel Deutschland GmbH
|
||
|
*/
|
||
|
#include "iwl-drv.h"
|
||
|
#include "runtime.h"
|
||
|
#include "fw/api/commands.h"
|
||
|
|
||
|
void iwl_free_fw_paging(struct iwl_fw_runtime *fwrt)
|
||
|
{
|
||
|
int i;
|
||
|
|
||
|
if (!fwrt->fw_paging_db[0].fw_paging_block)
|
||
|
return;
|
||
|
|
||
|
for (i = 0; i < NUM_OF_FW_PAGING_BLOCKS; i++) {
|
||
|
struct iwl_fw_paging *paging = &fwrt->fw_paging_db[i];
|
||
|
|
||
|
if (!paging->fw_paging_block) {
|
||
|
IWL_DEBUG_FW(fwrt,
|
||
|
"Paging: block %d already freed, continue to next page\n",
|
||
|
i);
|
||
|
|
||
|
continue;
|
||
|
}
|
||
|
dma_unmap_page(fwrt->trans->dev, paging->fw_paging_phys,
|
||
|
paging->fw_paging_size, DMA_BIDIRECTIONAL);
|
||
|
|
||
|
__free_pages(paging->fw_paging_block,
|
||
|
get_order(paging->fw_paging_size));
|
||
|
paging->fw_paging_block = NULL;
|
||
|
}
|
||
|
|
||
|
memset(fwrt->fw_paging_db, 0, sizeof(fwrt->fw_paging_db));
|
||
|
}
|
||
|
IWL_EXPORT_SYMBOL(iwl_free_fw_paging);
|
||
|
|
||
|
static int iwl_alloc_fw_paging_mem(struct iwl_fw_runtime *fwrt,
|
||
|
const struct fw_img *image)
|
||
|
{
|
||
|
struct page *block;
|
||
|
dma_addr_t phys = 0;
|
||
|
int blk_idx, order, num_of_pages, size;
|
||
|
|
||
|
if (fwrt->fw_paging_db[0].fw_paging_block)
|
||
|
return 0;
|
||
|
|
||
|
/* ensure BLOCK_2_EXP_SIZE is power of 2 of PAGING_BLOCK_SIZE */
|
||
|
BUILD_BUG_ON(BIT(BLOCK_2_EXP_SIZE) != PAGING_BLOCK_SIZE);
|
||
|
|
||
|
num_of_pages = image->paging_mem_size / FW_PAGING_SIZE;
|
||
|
fwrt->num_of_paging_blk =
|
||
|
DIV_ROUND_UP(num_of_pages, NUM_OF_PAGE_PER_GROUP);
|
||
|
fwrt->num_of_pages_in_last_blk =
|
||
|
num_of_pages -
|
||
|
NUM_OF_PAGE_PER_GROUP * (fwrt->num_of_paging_blk - 1);
|
||
|
|
||
|
IWL_DEBUG_FW(fwrt,
|
||
|
"Paging: allocating mem for %d paging blocks, each block holds 8 pages, last block holds %d pages\n",
|
||
|
fwrt->num_of_paging_blk,
|
||
|
fwrt->num_of_pages_in_last_blk);
|
||
|
|
||
|
/*
|
||
|
* Allocate CSS and paging blocks in dram.
|
||
|
*/
|
||
|
for (blk_idx = 0; blk_idx < fwrt->num_of_paging_blk + 1; blk_idx++) {
|
||
|
/* For CSS allocate 4KB, for others PAGING_BLOCK_SIZE (32K) */
|
||
|
size = blk_idx ? PAGING_BLOCK_SIZE : FW_PAGING_SIZE;
|
||
|
order = get_order(size);
|
||
|
block = alloc_pages(GFP_KERNEL, order);
|
||
|
if (!block) {
|
||
|
/* free all the previous pages since we failed */
|
||
|
iwl_free_fw_paging(fwrt);
|
||
|
return -ENOMEM;
|
||
|
}
|
||
|
|
||
|
fwrt->fw_paging_db[blk_idx].fw_paging_block = block;
|
||
|
fwrt->fw_paging_db[blk_idx].fw_paging_size = size;
|
||
|
|
||
|
phys = dma_map_page(fwrt->trans->dev, block, 0,
|
||
|
PAGE_SIZE << order,
|
||
|
DMA_BIDIRECTIONAL);
|
||
|
if (dma_mapping_error(fwrt->trans->dev, phys)) {
|
||
|
/*
|
||
|
* free the previous pages and the current one
|
||
|
* since we failed to map_page.
|
||
|
*/
|
||
|
iwl_free_fw_paging(fwrt);
|
||
|
return -ENOMEM;
|
||
|
}
|
||
|
fwrt->fw_paging_db[blk_idx].fw_paging_phys = phys;
|
||
|
|
||
|
if (!blk_idx)
|
||
|
IWL_DEBUG_FW(fwrt,
|
||
|
"Paging: allocated 4K(CSS) bytes (order %d) for firmware paging.\n",
|
||
|
order);
|
||
|
else
|
||
|
IWL_DEBUG_FW(fwrt,
|
||
|
"Paging: allocated 32K bytes (order %d) for firmware paging.\n",
|
||
|
order);
|
||
|
}
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int iwl_fill_paging_mem(struct iwl_fw_runtime *fwrt,
|
||
|
const struct fw_img *image)
|
||
|
{
|
||
|
int sec_idx, idx, ret;
|
||
|
u32 offset = 0;
|
||
|
|
||
|
/*
|
||
|
* find where is the paging image start point:
|
||
|
* if CPU2 exist and it's in paging format, then the image looks like:
|
||
|
* CPU1 sections (2 or more)
|
||
|
* CPU1_CPU2_SEPARATOR_SECTION delimiter - separate between CPU1 to CPU2
|
||
|
* CPU2 sections (not paged)
|
||
|
* PAGING_SEPARATOR_SECTION delimiter - separate between CPU2
|
||
|
* non paged to CPU2 paging sec
|
||
|
* CPU2 paging CSS
|
||
|
* CPU2 paging image (including instruction and data)
|
||
|
*/
|
||
|
for (sec_idx = 0; sec_idx < image->num_sec; sec_idx++) {
|
||
|
if (image->sec[sec_idx].offset == PAGING_SEPARATOR_SECTION) {
|
||
|
sec_idx++;
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* If paging is enabled there should be at least 2 more sections left
|
||
|
* (one for CSS and one for Paging data)
|
||
|
*/
|
||
|
if (sec_idx >= image->num_sec - 1) {
|
||
|
IWL_ERR(fwrt, "Paging: Missing CSS and/or paging sections\n");
|
||
|
ret = -EINVAL;
|
||
|
goto err;
|
||
|
}
|
||
|
|
||
|
/* copy the CSS block to the dram */
|
||
|
IWL_DEBUG_FW(fwrt, "Paging: load paging CSS to FW, sec = %d\n",
|
||
|
sec_idx);
|
||
|
|
||
|
if (image->sec[sec_idx].len > fwrt->fw_paging_db[0].fw_paging_size) {
|
||
|
IWL_ERR(fwrt, "CSS block is larger than paging size\n");
|
||
|
ret = -EINVAL;
|
||
|
goto err;
|
||
|
}
|
||
|
|
||
|
memcpy(page_address(fwrt->fw_paging_db[0].fw_paging_block),
|
||
|
image->sec[sec_idx].data,
|
||
|
image->sec[sec_idx].len);
|
||
|
fwrt->fw_paging_db[0].fw_offs = image->sec[sec_idx].offset;
|
||
|
dma_sync_single_for_device(fwrt->trans->dev,
|
||
|
fwrt->fw_paging_db[0].fw_paging_phys,
|
||
|
fwrt->fw_paging_db[0].fw_paging_size,
|
||
|
DMA_BIDIRECTIONAL);
|
||
|
|
||
|
IWL_DEBUG_FW(fwrt,
|
||
|
"Paging: copied %d CSS bytes to first block\n",
|
||
|
fwrt->fw_paging_db[0].fw_paging_size);
|
||
|
|
||
|
sec_idx++;
|
||
|
|
||
|
/*
|
||
|
* Copy the paging blocks to the dram. The loop index starts
|
||
|
* from 1 since the CSS block (index 0) was already copied to
|
||
|
* dram. We use num_of_paging_blk + 1 to account for that.
|
||
|
*/
|
||
|
for (idx = 1; idx < fwrt->num_of_paging_blk + 1; idx++) {
|
||
|
struct iwl_fw_paging *block = &fwrt->fw_paging_db[idx];
|
||
|
int remaining = image->sec[sec_idx].len - offset;
|
||
|
int len = block->fw_paging_size;
|
||
|
|
||
|
/*
|
||
|
* For the last block, we copy all that is remaining,
|
||
|
* for all other blocks, we copy fw_paging_size at a
|
||
|
* time. */
|
||
|
if (idx == fwrt->num_of_paging_blk) {
|
||
|
len = remaining;
|
||
|
if (remaining !=
|
||
|
fwrt->num_of_pages_in_last_blk * FW_PAGING_SIZE) {
|
||
|
IWL_ERR(fwrt,
|
||
|
"Paging: last block contains more data than expected %d\n",
|
||
|
remaining);
|
||
|
ret = -EINVAL;
|
||
|
goto err;
|
||
|
}
|
||
|
} else if (block->fw_paging_size > remaining) {
|
||
|
IWL_ERR(fwrt,
|
||
|
"Paging: not enough data in other in block %d (%d)\n",
|
||
|
idx, remaining);
|
||
|
ret = -EINVAL;
|
||
|
goto err;
|
||
|
}
|
||
|
|
||
|
memcpy(page_address(block->fw_paging_block),
|
||
|
(const u8 *)image->sec[sec_idx].data + offset, len);
|
||
|
block->fw_offs = image->sec[sec_idx].offset + offset;
|
||
|
dma_sync_single_for_device(fwrt->trans->dev,
|
||
|
block->fw_paging_phys,
|
||
|
block->fw_paging_size,
|
||
|
DMA_BIDIRECTIONAL);
|
||
|
|
||
|
IWL_DEBUG_FW(fwrt,
|
||
|
"Paging: copied %d paging bytes to block %d\n",
|
||
|
len, idx);
|
||
|
|
||
|
offset += block->fw_paging_size;
|
||
|
}
|
||
|
|
||
|
return 0;
|
||
|
|
||
|
err:
|
||
|
iwl_free_fw_paging(fwrt);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static int iwl_save_fw_paging(struct iwl_fw_runtime *fwrt,
|
||
|
const struct fw_img *fw)
|
||
|
{
|
||
|
int ret;
|
||
|
|
||
|
ret = iwl_alloc_fw_paging_mem(fwrt, fw);
|
||
|
if (ret)
|
||
|
return ret;
|
||
|
|
||
|
return iwl_fill_paging_mem(fwrt, fw);
|
||
|
}
|
||
|
|
||
|
/* send paging cmd to FW in case CPU2 has paging image */
|
||
|
static int iwl_send_paging_cmd(struct iwl_fw_runtime *fwrt,
|
||
|
const struct fw_img *fw)
|
||
|
{
|
||
|
struct iwl_fw_paging_cmd paging_cmd = {
|
||
|
.flags = cpu_to_le32(PAGING_CMD_IS_SECURED |
|
||
|
PAGING_CMD_IS_ENABLED |
|
||
|
(fwrt->num_of_pages_in_last_blk <<
|
||
|
PAGING_CMD_NUM_OF_PAGES_IN_LAST_GRP_POS)),
|
||
|
.block_size = cpu_to_le32(BLOCK_2_EXP_SIZE),
|
||
|
.block_num = cpu_to_le32(fwrt->num_of_paging_blk),
|
||
|
};
|
||
|
struct iwl_host_cmd hcmd = {
|
||
|
.id = WIDE_ID(IWL_ALWAYS_LONG_GROUP, FW_PAGING_BLOCK_CMD),
|
||
|
.len = { sizeof(paging_cmd), },
|
||
|
.data = { &paging_cmd, },
|
||
|
};
|
||
|
int blk_idx;
|
||
|
|
||
|
/* loop for for all paging blocks + CSS block */
|
||
|
for (blk_idx = 0; blk_idx < fwrt->num_of_paging_blk + 1; blk_idx++) {
|
||
|
dma_addr_t addr = fwrt->fw_paging_db[blk_idx].fw_paging_phys;
|
||
|
__le32 phy_addr;
|
||
|
|
||
|
addr = addr >> PAGE_2_EXP_SIZE;
|
||
|
phy_addr = cpu_to_le32(addr);
|
||
|
paging_cmd.device_phy_addr[blk_idx] = phy_addr;
|
||
|
}
|
||
|
|
||
|
return iwl_trans_send_cmd(fwrt->trans, &hcmd);
|
||
|
}
|
||
|
|
||
|
int iwl_init_paging(struct iwl_fw_runtime *fwrt, enum iwl_ucode_type type)
|
||
|
{
|
||
|
const struct fw_img *fw = &fwrt->fw->img[type];
|
||
|
int ret;
|
||
|
|
||
|
if (fwrt->trans->trans_cfg->gen2)
|
||
|
return 0;
|
||
|
|
||
|
/*
|
||
|
* Configure and operate fw paging mechanism.
|
||
|
* The driver configures the paging flow only once.
|
||
|
* The CPU2 paging image is included in the IWL_UCODE_INIT image.
|
||
|
*/
|
||
|
if (!fw->paging_mem_size)
|
||
|
return 0;
|
||
|
|
||
|
ret = iwl_save_fw_paging(fwrt, fw);
|
||
|
if (ret) {
|
||
|
IWL_ERR(fwrt, "failed to save the FW paging image\n");
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
ret = iwl_send_paging_cmd(fwrt, fw);
|
||
|
if (ret) {
|
||
|
IWL_ERR(fwrt, "failed to send the paging cmd\n");
|
||
|
iwl_free_fw_paging(fwrt);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
IWL_EXPORT_SYMBOL(iwl_init_paging);
|