210 lines
5.7 KiB
C
210 lines
5.7 KiB
C
|
// SPDX-License-Identifier: GPL-2.0-only
|
|||
|
|
|||
|
#define pr_fmt(fmt) "rtas-work-area: " fmt
|
|||
|
|
|||
|
#include <linux/genalloc.h>
|
|||
|
#include <linux/log2.h>
|
|||
|
#include <linux/kernel.h>
|
|||
|
#include <linux/memblock.h>
|
|||
|
#include <linux/mempool.h>
|
|||
|
#include <linux/minmax.h>
|
|||
|
#include <linux/mutex.h>
|
|||
|
#include <linux/numa.h>
|
|||
|
#include <linux/sizes.h>
|
|||
|
#include <linux/wait.h>
|
|||
|
|
|||
|
#include <asm/machdep.h>
|
|||
|
#include <asm/rtas-work-area.h>
|
|||
|
#include <asm/rtas.h>
|
|||
|
|
|||
|
enum {
|
|||
|
/*
|
|||
|
* Ensure the pool is page-aligned.
|
|||
|
*/
|
|||
|
RTAS_WORK_AREA_ARENA_ALIGN = PAGE_SIZE,
|
|||
|
/*
|
|||
|
* Don't let a single allocation claim the whole arena.
|
|||
|
*/
|
|||
|
RTAS_WORK_AREA_ARENA_SZ = RTAS_WORK_AREA_MAX_ALLOC_SZ * 2,
|
|||
|
/*
|
|||
|
* The smallest known work area size is for ibm,get-vpd's
|
|||
|
* location code argument, which is limited to 79 characters
|
|||
|
* plus 1 nul terminator.
|
|||
|
*
|
|||
|
* PAPR+ 7.3.20 ibm,get-vpd RTAS Call
|
|||
|
* PAPR+ 12.3.2.4 Converged Location Code Rules - Length Restrictions
|
|||
|
*/
|
|||
|
RTAS_WORK_AREA_MIN_ALLOC_SZ = roundup_pow_of_two(80),
|
|||
|
};
|
|||
|
|
|||
|
static struct {
|
|||
|
struct gen_pool *gen_pool;
|
|||
|
char *arena;
|
|||
|
struct mutex mutex; /* serializes allocations */
|
|||
|
struct wait_queue_head wqh;
|
|||
|
mempool_t descriptor_pool;
|
|||
|
bool available;
|
|||
|
} rwa_state = {
|
|||
|
.mutex = __MUTEX_INITIALIZER(rwa_state.mutex),
|
|||
|
.wqh = __WAIT_QUEUE_HEAD_INITIALIZER(rwa_state.wqh),
|
|||
|
};
|
|||
|
|
|||
|
/*
|
|||
|
* A single work area buffer and descriptor to serve requests early in
|
|||
|
* boot before the allocator is fully initialized. We know 4KB is the
|
|||
|
* most any boot time user needs (they all call ibm,get-system-parameter).
|
|||
|
*/
|
|||
|
static bool early_work_area_in_use __initdata;
|
|||
|
static char early_work_area_buf[SZ_4K] __initdata __aligned(SZ_4K);
|
|||
|
static struct rtas_work_area early_work_area __initdata = {
|
|||
|
.buf = early_work_area_buf,
|
|||
|
.size = sizeof(early_work_area_buf),
|
|||
|
};
|
|||
|
|
|||
|
|
|||
|
static struct rtas_work_area * __init rtas_work_area_alloc_early(size_t size)
|
|||
|
{
|
|||
|
WARN_ON(size > early_work_area.size);
|
|||
|
WARN_ON(early_work_area_in_use);
|
|||
|
early_work_area_in_use = true;
|
|||
|
memset(early_work_area.buf, 0, early_work_area.size);
|
|||
|
return &early_work_area;
|
|||
|
}
|
|||
|
|
|||
|
static void __init rtas_work_area_free_early(struct rtas_work_area *work_area)
|
|||
|
{
|
|||
|
WARN_ON(work_area != &early_work_area);
|
|||
|
WARN_ON(!early_work_area_in_use);
|
|||
|
early_work_area_in_use = false;
|
|||
|
}
|
|||
|
|
|||
|
struct rtas_work_area * __ref __rtas_work_area_alloc(size_t size)
|
|||
|
{
|
|||
|
struct rtas_work_area *area;
|
|||
|
unsigned long addr;
|
|||
|
|
|||
|
might_sleep();
|
|||
|
|
|||
|
/*
|
|||
|
* The rtas_work_area_alloc() wrapper enforces this at build
|
|||
|
* time. Requests that exceed the arena size will block
|
|||
|
* indefinitely.
|
|||
|
*/
|
|||
|
WARN_ON(size > RTAS_WORK_AREA_MAX_ALLOC_SZ);
|
|||
|
|
|||
|
if (!rwa_state.available)
|
|||
|
return rtas_work_area_alloc_early(size);
|
|||
|
/*
|
|||
|
* To ensure FCFS behavior and prevent a high rate of smaller
|
|||
|
* requests from starving larger ones, use the mutex to queue
|
|||
|
* allocations.
|
|||
|
*/
|
|||
|
mutex_lock(&rwa_state.mutex);
|
|||
|
wait_event(rwa_state.wqh,
|
|||
|
(addr = gen_pool_alloc(rwa_state.gen_pool, size)) != 0);
|
|||
|
mutex_unlock(&rwa_state.mutex);
|
|||
|
|
|||
|
area = mempool_alloc(&rwa_state.descriptor_pool, GFP_KERNEL);
|
|||
|
area->buf = (char *)addr;
|
|||
|
area->size = size;
|
|||
|
|
|||
|
return area;
|
|||
|
}
|
|||
|
|
|||
|
void __ref rtas_work_area_free(struct rtas_work_area *area)
|
|||
|
{
|
|||
|
if (!rwa_state.available) {
|
|||
|
rtas_work_area_free_early(area);
|
|||
|
return;
|
|||
|
}
|
|||
|
|
|||
|
gen_pool_free(rwa_state.gen_pool, (unsigned long)area->buf, area->size);
|
|||
|
mempool_free(area, &rwa_state.descriptor_pool);
|
|||
|
wake_up(&rwa_state.wqh);
|
|||
|
}
|
|||
|
|
|||
|
/*
|
|||
|
* Initialization of the work area allocator happens in two parts. To
|
|||
|
* reliably reserve an arena that satisfies RTAS addressing
|
|||
|
* requirements, we must perform a memblock allocation early,
|
|||
|
* immmediately after RTAS instantiation. Then we have to wait until
|
|||
|
* the slab allocator is up before setting up the descriptor mempool
|
|||
|
* and adding the arena to a gen_pool.
|
|||
|
*/
|
|||
|
static __init int rtas_work_area_allocator_init(void)
|
|||
|
{
|
|||
|
const unsigned int order = ilog2(RTAS_WORK_AREA_MIN_ALLOC_SZ);
|
|||
|
const phys_addr_t pa_start = __pa(rwa_state.arena);
|
|||
|
const phys_addr_t pa_end = pa_start + RTAS_WORK_AREA_ARENA_SZ - 1;
|
|||
|
struct gen_pool *pool;
|
|||
|
const int nid = NUMA_NO_NODE;
|
|||
|
int err;
|
|||
|
|
|||
|
err = -ENOMEM;
|
|||
|
if (!rwa_state.arena)
|
|||
|
goto err_out;
|
|||
|
|
|||
|
pool = gen_pool_create(order, nid);
|
|||
|
if (!pool)
|
|||
|
goto err_out;
|
|||
|
/*
|
|||
|
* All RTAS functions that consume work areas are OK with
|
|||
|
* natural alignment, when they have alignment requirements at
|
|||
|
* all.
|
|||
|
*/
|
|||
|
gen_pool_set_algo(pool, gen_pool_first_fit_order_align, NULL);
|
|||
|
|
|||
|
err = gen_pool_add(pool, (unsigned long)rwa_state.arena,
|
|||
|
RTAS_WORK_AREA_ARENA_SZ, nid);
|
|||
|
if (err)
|
|||
|
goto err_destroy;
|
|||
|
|
|||
|
err = mempool_init_kmalloc_pool(&rwa_state.descriptor_pool, 1,
|
|||
|
sizeof(struct rtas_work_area));
|
|||
|
if (err)
|
|||
|
goto err_destroy;
|
|||
|
|
|||
|
rwa_state.gen_pool = pool;
|
|||
|
rwa_state.available = true;
|
|||
|
|
|||
|
pr_debug("arena [%pa-%pa] (%uK), min/max alloc sizes %u/%u\n",
|
|||
|
&pa_start, &pa_end,
|
|||
|
RTAS_WORK_AREA_ARENA_SZ / SZ_1K,
|
|||
|
RTAS_WORK_AREA_MIN_ALLOC_SZ,
|
|||
|
RTAS_WORK_AREA_MAX_ALLOC_SZ);
|
|||
|
|
|||
|
return 0;
|
|||
|
|
|||
|
err_destroy:
|
|||
|
gen_pool_destroy(pool);
|
|||
|
err_out:
|
|||
|
return err;
|
|||
|
}
|
|||
|
machine_arch_initcall(pseries, rtas_work_area_allocator_init);
|
|||
|
|
|||
|
/**
|
|||
|
* rtas_work_area_reserve_arena() - Reserve memory suitable for RTAS work areas.
|
|||
|
*/
|
|||
|
void __init rtas_work_area_reserve_arena(const phys_addr_t limit)
|
|||
|
{
|
|||
|
const phys_addr_t align = RTAS_WORK_AREA_ARENA_ALIGN;
|
|||
|
const phys_addr_t size = RTAS_WORK_AREA_ARENA_SZ;
|
|||
|
const phys_addr_t min = MEMBLOCK_LOW_LIMIT;
|
|||
|
const int nid = NUMA_NO_NODE;
|
|||
|
|
|||
|
/*
|
|||
|
* Too early for a machine_is(pseries) check. But PAPR
|
|||
|
* effectively mandates that ibm,get-system-parameter is
|
|||
|
* present:
|
|||
|
*
|
|||
|
* R1–7.3.16–1. All platforms must support the System
|
|||
|
* Parameters option.
|
|||
|
*
|
|||
|
* So set up the arena if we find that, with a fallback to
|
|||
|
* ibm,configure-connector, just in case.
|
|||
|
*/
|
|||
|
if (rtas_function_implemented(RTAS_FN_IBM_GET_SYSTEM_PARAMETER) ||
|
|||
|
rtas_function_implemented(RTAS_FN_IBM_CONFIGURE_CONNECTOR))
|
|||
|
rwa_state.arena = memblock_alloc_try_nid(size, align, min, limit, nid);
|
|||
|
}
|