Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v6.19 338 lines 8.4 kB view raw
1/* SPDX-License-Identifier: GPL-2.0-only */ 2/* 3 * AMD Passthru DMA device driver 4 * -- Based on the CCP driver 5 * 6 * Copyright (C) 2016,2021 Advanced Micro Devices, Inc. 7 * 8 * Author: Sanjay R Mehta <sanju.mehta@amd.com> 9 * Author: Tom Lendacky <thomas.lendacky@amd.com> 10 * Author: Gary R Hook <gary.hook@amd.com> 11 */ 12 13#ifndef __PT_DEV_H__ 14#define __PT_DEV_H__ 15 16#include <linux/device.h> 17#include <linux/dmaengine.h> 18#include <linux/pci.h> 19#include <linux/spinlock.h> 20#include <linux/mutex.h> 21#include <linux/list.h> 22#include <linux/wait.h> 23#include <linux/dmapool.h> 24 25#include "../../virt-dma.h" 26 27#define MAX_PT_NAME_LEN 16 28#define MAX_DMAPOOL_NAME_LEN 32 29 30#define MAX_HW_QUEUES 1 31#define MAX_CMD_QLEN 100 32 33#define PT_ENGINE_PASSTHRU 5 34 35/* Register Mappings */ 36#define IRQ_MASK_REG 0x040 37#define IRQ_STATUS_REG 0x200 38 39#define CMD_Q_ERROR(__qs) ((__qs) & 0x0000003f) 40 41#define CMD_QUEUE_PRIO_OFFSET 0x00 42#define CMD_REQID_CONFIG_OFFSET 0x04 43#define CMD_TIMEOUT_OFFSET 0x08 44#define CMD_PT_VERSION 0x10 45 46#define CMD_Q_CONTROL_BASE 0x0000 47#define CMD_Q_TAIL_LO_BASE 0x0004 48#define CMD_Q_HEAD_LO_BASE 0x0008 49#define CMD_Q_INT_ENABLE_BASE 0x000C 50#define CMD_Q_INTERRUPT_STATUS_BASE 0x0010 51 52#define CMD_Q_STATUS_BASE 0x0100 53#define CMD_Q_INT_STATUS_BASE 0x0104 54#define CMD_Q_DMA_STATUS_BASE 0x0108 55#define CMD_Q_DMA_READ_STATUS_BASE 0x010C 56#define CMD_Q_DMA_WRITE_STATUS_BASE 0x0110 57#define CMD_Q_ABORT_BASE 0x0114 58#define CMD_Q_AX_CACHE_BASE 0x0118 59 60#define CMD_CONFIG_OFFSET 0x1120 61#define CMD_CLK_GATE_CTL_OFFSET 0x6004 62 63#define CMD_DESC_DW0_VAL 0x500012 64 65/* Address offset for virtual queue registers */ 66#define CMD_Q_STATUS_INCR 0x1000 67 68/* Bit masks */ 69#define CMD_CONFIG_REQID 0 70#define CMD_TIMEOUT_DISABLE 0 71#define CMD_CLK_DYN_GATING_DIS 0 72#define CMD_CLK_SW_GATE_MODE 0 73#define CMD_CLK_GATE_CTL 0 74#define CMD_QUEUE_PRIO GENMASK(2, 1) 75#define CMD_CONFIG_VHB_EN BIT(0) 76#define CMD_CLK_DYN_GATING_EN BIT(0) 77#define CMD_CLK_HW_GATE_MODE BIT(0) 78#define CMD_CLK_GATE_ON_DELAY BIT(12) 79#define CMD_CLK_GATE_OFF_DELAY BIT(12) 80 81#define CMD_CLK_GATE_CONFIG (CMD_CLK_GATE_CTL | \ 82 CMD_CLK_HW_GATE_MODE | \ 83 CMD_CLK_GATE_ON_DELAY | \ 84 CMD_CLK_DYN_GATING_EN | \ 85 CMD_CLK_GATE_OFF_DELAY) 86 87#define CMD_Q_LEN 32 88#define CMD_Q_RUN BIT(0) 89#define CMD_Q_HALT BIT(1) 90#define CMD_Q_MEM_LOCATION BIT(2) 91#define CMD_Q_SIZE_MASK GENMASK(4, 0) 92#define CMD_Q_SIZE GENMASK(7, 3) 93#define CMD_Q_SHIFT GENMASK(1, 0) 94#define QUEUE_SIZE_VAL ((ffs(CMD_Q_LEN) - 2) & \ 95 CMD_Q_SIZE_MASK) 96#define Q_PTR_MASK (2 << (QUEUE_SIZE_VAL + 5) - 1) 97#define Q_DESC_SIZE sizeof(struct ptdma_desc) 98#define Q_SIZE(n) (CMD_Q_LEN * (n)) 99 100#define INT_COMPLETION BIT(0) 101#define INT_ERROR BIT(1) 102#define INT_QUEUE_STOPPED BIT(2) 103#define INT_EMPTY_QUEUE BIT(3) 104#define SUPPORTED_INTERRUPTS (INT_COMPLETION | INT_ERROR) 105 106/****** Local Storage Block ******/ 107#define LSB_START 0 108#define LSB_END 127 109#define LSB_COUNT (LSB_END - LSB_START + 1) 110 111#define PT_DMAPOOL_MAX_SIZE 64 112#define PT_DMAPOOL_ALIGN BIT(5) 113 114#define PT_PASSTHRU_BLOCKSIZE 512 115 116struct pt_device; 117 118struct pt_tasklet_data { 119 struct completion completion; 120 struct pt_cmd *cmd; 121}; 122 123/* 124 * struct pt_passthru_engine - pass-through operation 125 * without performing DMA mapping 126 * @mask: mask to be applied to data 127 * @mask_len: length in bytes of mask 128 * @src_dma: data to be used for this operation 129 * @dst_dma: data produced by this operation 130 * @src_len: length in bytes of data used for this operation 131 * 132 * Variables required to be set when calling pt_enqueue_cmd(): 133 * - bit_mod, byte_swap, src, dst, src_len 134 * - mask, mask_len if bit_mod is not PT_PASSTHRU_BITWISE_NOOP 135 */ 136struct pt_passthru_engine { 137 dma_addr_t mask; 138 u32 mask_len; /* In bytes */ 139 140 dma_addr_t src_dma, dst_dma; 141 u64 src_len; /* In bytes */ 142}; 143 144/* 145 * struct pt_cmd - PTDMA operation request 146 * @entry: list element 147 * @work: work element used for callbacks 148 * @pt: PT device to be run on 149 * @ret: operation return code 150 * @flags: cmd processing flags 151 * @engine: PTDMA operation to perform (passthru) 152 * @engine_error: PT engine return code 153 * @passthru: engine specific structures, refer to specific engine struct below 154 * @callback: operation completion callback function 155 * @data: parameter value to be supplied to the callback function 156 * 157 * Variables required to be set when calling pt_enqueue_cmd(): 158 * - engine, callback 159 * - See the operation structures below for what is required for each 160 * operation. 161 */ 162struct pt_cmd { 163 struct list_head entry; 164 struct work_struct work; 165 struct pt_device *pt; 166 int ret; 167 u32 engine; 168 u32 engine_error; 169 struct pt_passthru_engine passthru; 170 /* Completion callback support */ 171 void (*pt_cmd_callback)(void *data, int err); 172 void *data; 173}; 174 175struct pt_dma_desc { 176 struct virt_dma_desc vd; 177 struct pt_device *pt; 178 enum dma_status status; 179 size_t len; 180 bool issued_to_hw; 181 struct pt_cmd pt_cmd; 182}; 183 184struct pt_dma_chan { 185 struct virt_dma_chan vc; 186 struct pt_device *pt; 187 u32 id; 188}; 189 190struct pt_cmd_queue { 191 struct pt_device *pt; 192 193 /* Queue dma pool */ 194 struct dma_pool *dma_pool; 195 196 /* Queue base address (not necessarily aligned)*/ 197 struct ptdma_desc *qbase; 198 199 /* Aligned queue start address (per requirement) */ 200 spinlock_t q_lock ____cacheline_aligned; 201 unsigned int qidx; 202 203 unsigned int qsize; 204 dma_addr_t qbase_dma; 205 dma_addr_t qdma_tail; 206 207 unsigned int active; 208 unsigned int suspended; 209 210 /* Interrupt flag */ 211 bool int_en; 212 213 /* Register addresses for queue */ 214 void __iomem *reg_control; 215 u32 qcontrol; /* Cached control register */ 216 217 /* Status values from job */ 218 u32 int_status; 219 u32 q_status; 220 u32 q_int_status; 221 u32 cmd_error; 222 /* Queue Statistics */ 223 unsigned long total_pt_ops; 224} ____cacheline_aligned; 225 226struct pt_device { 227 struct list_head entry; 228 229 unsigned int ord; 230 char name[MAX_PT_NAME_LEN]; 231 232 struct device *dev; 233 234 /* Bus specific device information */ 235 struct pt_msix *pt_msix; 236 237 struct pt_dev_vdata *dev_vdata; 238 239 unsigned int pt_irq; 240 241 /* I/O area used for device communication */ 242 void __iomem *io_regs; 243 244 spinlock_t cmd_lock ____cacheline_aligned; 245 unsigned int cmd_count; 246 struct list_head cmd; 247 248 /* 249 * The command queue. This represent the queue available on the 250 * PTDMA that are available for processing cmds 251 */ 252 struct pt_cmd_queue cmd_q; 253 254 /* Support for the DMA Engine capabilities */ 255 struct dma_device dma_dev; 256 struct pt_dma_chan *pt_dma_chan; 257 struct kmem_cache *dma_desc_cache; 258 259 wait_queue_head_t lsb_queue; 260 261 /* Device Statistics */ 262 unsigned long total_interrupts; 263 264 struct pt_tasklet_data tdata; 265 int ver; 266}; 267 268/* 269 * descriptor for PTDMA commands 270 * 8 32-bit words: 271 * word 0: function; engine; control bits 272 * word 1: length of source data 273 * word 2: low 32 bits of source pointer 274 * word 3: upper 16 bits of source pointer; source memory type 275 * word 4: low 32 bits of destination pointer 276 * word 5: upper 16 bits of destination pointer; destination memory type 277 * word 6: reserved 32 bits 278 * word 7: reserved 32 bits 279 */ 280 281#define DWORD0_SOC BIT(0) 282#define DWORD0_IOC BIT(1) 283 284struct dword3 { 285 unsigned int src_hi:16; 286 unsigned int src_mem:2; 287 unsigned int lsb_cxt_id:8; 288 unsigned int rsvd1:5; 289 unsigned int fixed:1; 290}; 291 292struct dword5 { 293 unsigned int dst_hi:16; 294 unsigned int dst_mem:2; 295 unsigned int rsvd1:13; 296 unsigned int fixed:1; 297}; 298 299struct ptdma_desc { 300 u32 dw0; 301 u32 length; 302 u32 src_lo; 303 struct dword3 dw3; 304 u32 dst_lo; 305 struct dword5 dw5; 306 __le32 rsvd1; 307 __le32 rsvd2; 308}; 309 310/* Structure to hold PT device data */ 311struct pt_dev_vdata { 312 const unsigned int bar; 313}; 314 315int pt_dmaengine_register(struct pt_device *pt); 316void pt_dmaengine_unregister(struct pt_device *pt); 317 318void ptdma_debugfs_setup(struct pt_device *pt); 319int pt_core_init(struct pt_device *pt); 320void pt_core_destroy(struct pt_device *pt); 321 322int pt_core_perform_passthru(struct pt_cmd_queue *cmd_q, 323 struct pt_passthru_engine *pt_engine); 324 325void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q); 326void pt_start_queue(struct pt_cmd_queue *cmd_q); 327void pt_stop_queue(struct pt_cmd_queue *cmd_q); 328 329static inline void pt_core_disable_queue_interrupts(struct pt_device *pt) 330{ 331 iowrite32(0, pt->cmd_q.reg_control + 0x000C); 332} 333 334static inline void pt_core_enable_queue_interrupts(struct pt_device *pt) 335{ 336 iowrite32(SUPPORTED_INTERRUPTS, pt->cmd_q.reg_control + 0x000C); 337} 338#endif