2023-08-30 17:31:07 +02:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 OR MIT */
|
|
|
|
/*
|
|
|
|
* Copyright 2014-2022 Advanced Micro Devices, Inc.
|
|
|
|
*
|
|
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
|
|
* to deal in the Software without restriction, including without limitation
|
|
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
|
|
*
|
|
|
|
* The above copyright notice and this permission notice shall be included in
|
|
|
|
* all copies or substantial portions of the Software.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
|
|
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef KFD_DEVICE_QUEUE_MANAGER_H_
|
|
|
|
#define KFD_DEVICE_QUEUE_MANAGER_H_
|
|
|
|
|
|
|
|
#include <linux/rwsem.h>
|
|
|
|
#include <linux/list.h>
|
|
|
|
#include <linux/mutex.h>
|
|
|
|
#include <linux/sched/mm.h>
|
|
|
|
#include "kfd_priv.h"
|
|
|
|
#include "kfd_mqd_manager.h"
|
|
|
|
|
|
|
|
|
|
|
|
#define VMID_NUM 16
|
|
|
|
|
|
|
|
#define KFD_MES_PROCESS_QUANTUM 100000
|
|
|
|
#define KFD_MES_GANG_QUANTUM 10000
|
2023-10-24 12:59:35 +02:00
|
|
|
#define USE_DEFAULT_GRACE_PERIOD 0xffffffff
|
2023-08-30 17:31:07 +02:00
|
|
|
|
|
|
|
struct device_process_node {
|
|
|
|
struct qcm_process_device *qpd;
|
|
|
|
struct list_head list;
|
|
|
|
};
|
|
|
|
|
|
|
|
union SQ_CMD_BITS {
|
|
|
|
struct {
|
|
|
|
uint32_t cmd:3;
|
|
|
|
uint32_t:1;
|
|
|
|
uint32_t mode:3;
|
|
|
|
uint32_t check_vmid:1;
|
|
|
|
uint32_t trap_id:3;
|
|
|
|
uint32_t:5;
|
|
|
|
uint32_t wave_id:4;
|
|
|
|
uint32_t simd_id:2;
|
|
|
|
uint32_t:2;
|
|
|
|
uint32_t queue_id:3;
|
|
|
|
uint32_t:1;
|
|
|
|
uint32_t vm_id:4;
|
|
|
|
} bitfields, bits;
|
|
|
|
uint32_t u32All;
|
|
|
|
signed int i32All;
|
|
|
|
float f32All;
|
|
|
|
};
|
|
|
|
|
|
|
|
union GRBM_GFX_INDEX_BITS {
|
|
|
|
struct {
|
|
|
|
uint32_t instance_index:8;
|
|
|
|
uint32_t sh_index:8;
|
|
|
|
uint32_t se_index:8;
|
|
|
|
uint32_t:5;
|
|
|
|
uint32_t sh_broadcast_writes:1;
|
|
|
|
uint32_t instance_broadcast_writes:1;
|
|
|
|
uint32_t se_broadcast_writes:1;
|
|
|
|
} bitfields, bits;
|
|
|
|
uint32_t u32All;
|
|
|
|
signed int i32All;
|
|
|
|
float f32All;
|
|
|
|
};
|
|
|
|
|
|
|
|
/**
|
|
|
|
* struct device_queue_manager_ops
|
|
|
|
*
|
|
|
|
* @create_queue: Queue creation routine.
|
|
|
|
*
|
|
|
|
* @destroy_queue: Queue destruction routine.
|
|
|
|
*
|
|
|
|
* @update_queue: Queue update routine.
|
|
|
|
*
|
|
|
|
* @exeute_queues: Dispatches the queues list to the H/W.
|
|
|
|
*
|
|
|
|
* @register_process: This routine associates a specific process with device.
|
|
|
|
*
|
|
|
|
* @unregister_process: destroys the associations between process to device.
|
|
|
|
*
|
|
|
|
* @initialize: Initializes the pipelines and memory module for that device.
|
|
|
|
*
|
|
|
|
* @start: Initializes the resources/modules the device needs for queues
|
|
|
|
* execution. This function is called on device initialization and after the
|
|
|
|
* system woke up after suspension.
|
|
|
|
*
|
|
|
|
* @stop: This routine stops execution of all the active queue running on the
|
|
|
|
* H/W and basically this function called on system suspend.
|
|
|
|
*
|
|
|
|
* @uninitialize: Destroys all the device queue manager resources allocated in
|
|
|
|
* initialize routine.
|
|
|
|
*
|
|
|
|
* @create_kernel_queue: Creates kernel queue. Used for debug queue.
|
|
|
|
*
|
|
|
|
* @destroy_kernel_queue: Destroys kernel queue. Used for debug queue.
|
|
|
|
*
|
|
|
|
* @set_cache_memory_policy: Sets memory policy (cached/ non cached) for the
|
|
|
|
* memory apertures.
|
|
|
|
*
|
|
|
|
* @process_termination: Clears all process queues belongs to that device.
|
|
|
|
*
|
|
|
|
* @evict_process_queues: Evict all active queues of a process
|
|
|
|
*
|
|
|
|
* @restore_process_queues: Restore all evicted queues of a process
|
|
|
|
*
|
|
|
|
* @get_wave_state: Retrieves context save state and optionally copies the
|
|
|
|
* control stack, if kept in the MQD, to the given userspace address.
|
|
|
|
*
|
|
|
|
* @reset_queues: reset queues which consume RAS poison
|
|
|
|
* @get_queue_checkpoint_info: Retrieves queue size information for CRIU checkpoint.
|
|
|
|
*
|
|
|
|
* @checkpoint_mqd: checkpoint queue MQD contents for CRIU.
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct device_queue_manager_ops {
|
|
|
|
int (*create_queue)(struct device_queue_manager *dqm,
|
|
|
|
struct queue *q,
|
|
|
|
struct qcm_process_device *qpd,
|
|
|
|
const struct kfd_criu_queue_priv_data *qd,
|
|
|
|
const void *restore_mqd,
|
|
|
|
const void *restore_ctl_stack);
|
|
|
|
|
|
|
|
int (*destroy_queue)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd,
|
|
|
|
struct queue *q);
|
|
|
|
|
|
|
|
int (*update_queue)(struct device_queue_manager *dqm,
|
|
|
|
struct queue *q, struct mqd_update_info *minfo);
|
|
|
|
|
|
|
|
int (*register_process)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
|
|
|
|
int (*unregister_process)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
|
|
|
|
int (*initialize)(struct device_queue_manager *dqm);
|
|
|
|
int (*start)(struct device_queue_manager *dqm);
|
|
|
|
int (*stop)(struct device_queue_manager *dqm);
|
|
|
|
void (*pre_reset)(struct device_queue_manager *dqm);
|
|
|
|
void (*uninitialize)(struct device_queue_manager *dqm);
|
|
|
|
int (*create_kernel_queue)(struct device_queue_manager *dqm,
|
|
|
|
struct kernel_queue *kq,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
|
|
|
|
void (*destroy_kernel_queue)(struct device_queue_manager *dqm,
|
|
|
|
struct kernel_queue *kq,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
|
|
|
|
bool (*set_cache_memory_policy)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd,
|
|
|
|
enum cache_policy default_policy,
|
|
|
|
enum cache_policy alternate_policy,
|
|
|
|
void __user *alternate_aperture_base,
|
|
|
|
uint64_t alternate_aperture_size);
|
|
|
|
|
|
|
|
int (*process_termination)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
|
|
|
|
int (*evict_process_queues)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
int (*restore_process_queues)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
|
|
|
|
int (*get_wave_state)(struct device_queue_manager *dqm,
|
|
|
|
struct queue *q,
|
|
|
|
void __user *ctl_stack,
|
|
|
|
u32 *ctl_stack_used_size,
|
|
|
|
u32 *save_area_used_size);
|
|
|
|
|
|
|
|
int (*reset_queues)(struct device_queue_manager *dqm,
|
|
|
|
uint16_t pasid);
|
|
|
|
void (*get_queue_checkpoint_info)(struct device_queue_manager *dqm,
|
|
|
|
const struct queue *q, u32 *mqd_size,
|
|
|
|
u32 *ctl_stack_size);
|
|
|
|
|
|
|
|
int (*checkpoint_mqd)(struct device_queue_manager *dqm,
|
|
|
|
const struct queue *q,
|
|
|
|
void *mqd,
|
|
|
|
void *ctl_stack);
|
|
|
|
};
|
|
|
|
|
|
|
|
struct device_queue_manager_asic_ops {
|
|
|
|
int (*update_qpd)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
bool (*set_cache_memory_policy)(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd,
|
|
|
|
enum cache_policy default_policy,
|
|
|
|
enum cache_policy alternate_policy,
|
|
|
|
void __user *alternate_aperture_base,
|
|
|
|
uint64_t alternate_aperture_size);
|
|
|
|
void (*init_sdma_vm)(struct device_queue_manager *dqm,
|
|
|
|
struct queue *q,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
struct mqd_manager * (*mqd_manager_init)(enum KFD_MQD_TYPE type,
|
2023-10-24 12:59:35 +02:00
|
|
|
struct kfd_node *dev);
|
2023-08-30 17:31:07 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
/**
|
|
|
|
* struct device_queue_manager
|
|
|
|
*
|
|
|
|
* This struct is a base class for the kfd queues scheduler in the
|
|
|
|
* device level. The device base class should expose the basic operations
|
|
|
|
* for queue creation and queue destruction. This base class hides the
|
|
|
|
* scheduling mode of the driver and the specific implementation of the
|
|
|
|
* concrete device. This class is the only class in the queues scheduler
|
|
|
|
* that configures the H/W.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct device_queue_manager {
|
|
|
|
struct device_queue_manager_ops ops;
|
|
|
|
struct device_queue_manager_asic_ops asic_ops;
|
|
|
|
|
|
|
|
struct mqd_manager *mqd_mgrs[KFD_MQD_TYPE_MAX];
|
|
|
|
struct packet_manager packet_mgr;
|
2023-10-24 12:59:35 +02:00
|
|
|
struct kfd_node *dev;
|
2023-08-30 17:31:07 +02:00
|
|
|
struct mutex lock_hidden; /* use dqm_lock/unlock(dqm) */
|
|
|
|
struct list_head queues;
|
|
|
|
unsigned int saved_flags;
|
|
|
|
unsigned int processes_count;
|
|
|
|
unsigned int active_queue_count;
|
|
|
|
unsigned int active_cp_queue_count;
|
|
|
|
unsigned int gws_queue_count;
|
|
|
|
unsigned int total_queue_count;
|
|
|
|
unsigned int next_pipe_to_allocate;
|
|
|
|
unsigned int *allocated_queues;
|
2023-10-24 12:59:35 +02:00
|
|
|
DECLARE_BITMAP(sdma_bitmap, KFD_MAX_SDMA_QUEUES);
|
|
|
|
DECLARE_BITMAP(xgmi_sdma_bitmap, KFD_MAX_SDMA_QUEUES);
|
2023-08-30 17:31:07 +02:00
|
|
|
/* the pasid mapping for each kfd vmid */
|
|
|
|
uint16_t vmid_pasid[VMID_NUM];
|
|
|
|
uint64_t pipelines_addr;
|
|
|
|
uint64_t fence_gpu_addr;
|
|
|
|
uint64_t *fence_addr;
|
|
|
|
struct kfd_mem_obj *fence_mem;
|
|
|
|
bool active_runlist;
|
|
|
|
int sched_policy;
|
2023-10-24 12:59:35 +02:00
|
|
|
uint32_t trap_debug_vmid;
|
2023-08-30 17:31:07 +02:00
|
|
|
|
|
|
|
/* hw exception */
|
|
|
|
bool is_hws_hang;
|
|
|
|
bool is_resetting;
|
|
|
|
struct work_struct hw_exception_work;
|
|
|
|
struct kfd_mem_obj hiq_sdma_mqd;
|
|
|
|
bool sched_running;
|
2023-10-24 12:59:35 +02:00
|
|
|
|
|
|
|
/* used for GFX 9.4.3 only */
|
|
|
|
uint32_t current_logical_xcc_start;
|
|
|
|
|
|
|
|
uint32_t wait_times;
|
|
|
|
|
|
|
|
wait_queue_head_t destroy_wait;
|
2023-08-30 17:31:07 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
void device_queue_manager_init_cik(
|
|
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
|
|
void device_queue_manager_init_cik_hawaii(
|
|
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
|
|
void device_queue_manager_init_vi(
|
|
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
|
|
void device_queue_manager_init_vi_tonga(
|
|
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
|
|
void device_queue_manager_init_v9(
|
|
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
|
|
void device_queue_manager_init_v10_navi10(
|
|
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
|
|
void device_queue_manager_init_v11(
|
|
|
|
struct device_queue_manager_asic_ops *asic_ops);
|
|
|
|
void program_sh_mem_settings(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
unsigned int get_cp_queues_num(struct device_queue_manager *dqm);
|
|
|
|
unsigned int get_queues_per_pipe(struct device_queue_manager *dqm);
|
|
|
|
unsigned int get_pipes_per_mec(struct device_queue_manager *dqm);
|
|
|
|
unsigned int get_num_sdma_queues(struct device_queue_manager *dqm);
|
|
|
|
unsigned int get_num_xgmi_sdma_queues(struct device_queue_manager *dqm);
|
2023-10-24 12:59:35 +02:00
|
|
|
int reserve_debug_trap_vmid(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
int release_debug_trap_vmid(struct device_queue_manager *dqm,
|
|
|
|
struct qcm_process_device *qpd);
|
|
|
|
int suspend_queues(struct kfd_process *p,
|
|
|
|
uint32_t num_queues,
|
|
|
|
uint32_t grace_period,
|
|
|
|
uint64_t exception_clear_mask,
|
|
|
|
uint32_t *usr_queue_id_array);
|
|
|
|
int resume_queues(struct kfd_process *p,
|
|
|
|
uint32_t num_queues,
|
|
|
|
uint32_t *usr_queue_id_array);
|
|
|
|
void set_queue_snapshot_entry(struct queue *q,
|
|
|
|
uint64_t exception_clear_mask,
|
|
|
|
struct kfd_queue_snapshot_entry *qss_entry);
|
|
|
|
int debug_lock_and_unmap(struct device_queue_manager *dqm);
|
|
|
|
int debug_map_and_unlock(struct device_queue_manager *dqm);
|
|
|
|
int debug_refresh_runlist(struct device_queue_manager *dqm);
|
2023-08-30 17:31:07 +02:00
|
|
|
|
|
|
|
static inline unsigned int get_sh_mem_bases_32(struct kfd_process_device *pdd)
|
|
|
|
{
|
|
|
|
return (pdd->lds_base >> 16) & 0xFF;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline unsigned int
|
|
|
|
get_sh_mem_bases_nybble_64(struct kfd_process_device *pdd)
|
|
|
|
{
|
|
|
|
return (pdd->lds_base >> 60) & 0x0E;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* The DQM lock can be taken in MMU notifiers. Make sure no reclaim-FS
|
|
|
|
* happens while holding this lock anywhere to prevent deadlocks when
|
|
|
|
* an MMU notifier runs in reclaim-FS context.
|
|
|
|
*/
|
|
|
|
static inline void dqm_lock(struct device_queue_manager *dqm)
|
|
|
|
{
|
|
|
|
mutex_lock(&dqm->lock_hidden);
|
|
|
|
dqm->saved_flags = memalloc_noreclaim_save();
|
|
|
|
}
|
|
|
|
static inline void dqm_unlock(struct device_queue_manager *dqm)
|
|
|
|
{
|
|
|
|
memalloc_noreclaim_restore(dqm->saved_flags);
|
|
|
|
mutex_unlock(&dqm->lock_hidden);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int read_sdma_queue_counter(uint64_t __user *q_rptr, uint64_t *val)
|
|
|
|
{
|
|
|
|
/* SDMA activity counter is stored at queue's RPTR + 0x8 location. */
|
|
|
|
return get_user(*val, q_rptr + 1);
|
|
|
|
}
|
|
|
|
#endif /* KFD_DEVICE_QUEUE_MANAGER_H_ */
|