1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3 * AMD Passthru DMA device driver
4 * -- Based on the CCP driver
5 *
6 * Copyright (C) 2016,2021 Advanced Micro Devices, Inc.
7 *
8 * Author: Sanjay R Mehta <[email protected]>
9 * Author: Tom Lendacky <[email protected]>
10 * Author: Gary R Hook <[email protected]>
11 */
12
13 #ifndef __PT_DEV_H__
14 #define __PT_DEV_H__
15
16 #include <linux/device.h>
17 #include <linux/dmaengine.h>
18 #include <linux/pci.h>
19 #include <linux/spinlock.h>
20 #include <linux/mutex.h>
21 #include <linux/list.h>
22 #include <linux/wait.h>
23 #include <linux/dmapool.h>
24
25 #include "../../virt-dma.h"
26
27 #define MAX_PT_NAME_LEN 16
28 #define MAX_DMAPOOL_NAME_LEN 32
29
30 #define MAX_HW_QUEUES 1
31 #define MAX_CMD_QLEN 100
32
33 #define PT_ENGINE_PASSTHRU 5
34
35 /* Register Mappings */
36 #define IRQ_MASK_REG 0x040
37 #define IRQ_STATUS_REG 0x200
38
39 #define CMD_Q_ERROR(__qs) ((__qs) & 0x0000003f)
40
41 #define CMD_QUEUE_PRIO_OFFSET 0x00
42 #define CMD_REQID_CONFIG_OFFSET 0x04
43 #define CMD_TIMEOUT_OFFSET 0x08
44 #define CMD_PT_VERSION 0x10
45
46 #define CMD_Q_CONTROL_BASE 0x0000
47 #define CMD_Q_TAIL_LO_BASE 0x0004
48 #define CMD_Q_HEAD_LO_BASE 0x0008
49 #define CMD_Q_INT_ENABLE_BASE 0x000C
50 #define CMD_Q_INTERRUPT_STATUS_BASE 0x0010
51
52 #define CMD_Q_STATUS_BASE 0x0100
53 #define CMD_Q_INT_STATUS_BASE 0x0104
54 #define CMD_Q_DMA_STATUS_BASE 0x0108
55 #define CMD_Q_DMA_READ_STATUS_BASE 0x010C
56 #define CMD_Q_DMA_WRITE_STATUS_BASE 0x0110
57 #define CMD_Q_ABORT_BASE 0x0114
58 #define CMD_Q_AX_CACHE_BASE 0x0118
59
60 #define CMD_CONFIG_OFFSET 0x1120
61 #define CMD_CLK_GATE_CTL_OFFSET 0x6004
62
63 #define CMD_DESC_DW0_VAL 0x500012
64
65 /* Address offset for virtual queue registers */
66 #define CMD_Q_STATUS_INCR 0x1000
67
68 /* Bit masks */
69 #define CMD_CONFIG_REQID 0
70 #define CMD_TIMEOUT_DISABLE 0
71 #define CMD_CLK_DYN_GATING_DIS 0
72 #define CMD_CLK_SW_GATE_MODE 0
73 #define CMD_CLK_GATE_CTL 0
74 #define CMD_QUEUE_PRIO GENMASK(2, 1)
75 #define CMD_CONFIG_VHB_EN BIT(0)
76 #define CMD_CLK_DYN_GATING_EN BIT(0)
77 #define CMD_CLK_HW_GATE_MODE BIT(0)
78 #define CMD_CLK_GATE_ON_DELAY BIT(12)
79 #define CMD_CLK_GATE_OFF_DELAY BIT(12)
80
81 #define CMD_CLK_GATE_CONFIG (CMD_CLK_GATE_CTL | \
82 CMD_CLK_HW_GATE_MODE | \
83 CMD_CLK_GATE_ON_DELAY | \
84 CMD_CLK_DYN_GATING_EN | \
85 CMD_CLK_GATE_OFF_DELAY)
86
87 #define CMD_Q_LEN 32
88 #define CMD_Q_RUN BIT(0)
89 #define CMD_Q_HALT BIT(1)
90 #define CMD_Q_MEM_LOCATION BIT(2)
91 #define CMD_Q_SIZE_MASK GENMASK(4, 0)
92 #define CMD_Q_SIZE GENMASK(7, 3)
93 #define CMD_Q_SHIFT GENMASK(1, 0)
94 #define QUEUE_SIZE_VAL ((ffs(CMD_Q_LEN) - 2) & \
95 CMD_Q_SIZE_MASK)
96 #define Q_PTR_MASK (2 << (QUEUE_SIZE_VAL + 5) - 1)
97 #define Q_DESC_SIZE sizeof(struct ptdma_desc)
98 #define Q_SIZE(n) (CMD_Q_LEN * (n))
99
100 #define INT_COMPLETION BIT(0)
101 #define INT_ERROR BIT(1)
102 #define INT_QUEUE_STOPPED BIT(2)
103 #define INT_EMPTY_QUEUE BIT(3)
104 #define SUPPORTED_INTERRUPTS (INT_COMPLETION | INT_ERROR)
105
106 /****** Local Storage Block ******/
107 #define LSB_START 0
108 #define LSB_END 127
109 #define LSB_COUNT (LSB_END - LSB_START + 1)
110
111 #define PT_DMAPOOL_MAX_SIZE 64
112 #define PT_DMAPOOL_ALIGN BIT(5)
113
114 #define PT_PASSTHRU_BLOCKSIZE 512
115
116 struct pt_device;
117
118 struct pt_tasklet_data {
119 struct completion completion;
120 struct pt_cmd *cmd;
121 };
122
123 /*
124 * struct pt_passthru_engine - pass-through operation
125 * without performing DMA mapping
126 * @mask: mask to be applied to data
127 * @mask_len: length in bytes of mask
128 * @src_dma: data to be used for this operation
129 * @dst_dma: data produced by this operation
130 * @src_len: length in bytes of data used for this operation
131 *
132 * Variables required to be set when calling pt_enqueue_cmd():
133 * - bit_mod, byte_swap, src, dst, src_len
134 * - mask, mask_len if bit_mod is not PT_PASSTHRU_BITWISE_NOOP
135 */
136 struct pt_passthru_engine {
137 dma_addr_t mask;
138 u32 mask_len; /* In bytes */
139
140 dma_addr_t src_dma, dst_dma;
141 u64 src_len; /* In bytes */
142 };
143
144 /*
145 * struct pt_cmd - PTDMA operation request
146 * @entry: list element
147 * @work: work element used for callbacks
148 * @pt: PT device to be run on
149 * @ret: operation return code
150 * @flags: cmd processing flags
151 * @engine: PTDMA operation to perform (passthru)
152 * @engine_error: PT engine return code
153 * @passthru: engine specific structures, refer to specific engine struct below
154 * @callback: operation completion callback function
155 * @data: parameter value to be supplied to the callback function
156 *
157 * Variables required to be set when calling pt_enqueue_cmd():
158 * - engine, callback
159 * - See the operation structures below for what is required for each
160 * operation.
161 */
162 struct pt_cmd {
163 struct list_head entry;
164 struct work_struct work;
165 struct pt_device *pt;
166 int ret;
167 u32 engine;
168 u32 engine_error;
169 struct pt_passthru_engine passthru;
170 /* Completion callback support */
171 void (*pt_cmd_callback)(void *data, int err);
172 void *data;
173 };
174
175 struct pt_dma_desc {
176 struct virt_dma_desc vd;
177 struct pt_device *pt;
178 enum dma_status status;
179 size_t len;
180 bool issued_to_hw;
181 struct pt_cmd pt_cmd;
182 };
183
184 struct pt_dma_chan {
185 struct virt_dma_chan vc;
186 struct pt_device *pt;
187 u32 id;
188 };
189
190 struct pt_cmd_queue {
191 struct pt_device *pt;
192
193 /* Queue dma pool */
194 struct dma_pool *dma_pool;
195
196 /* Queue base address (not necessarily aligned)*/
197 struct ptdma_desc *qbase;
198
199 /* Aligned queue start address (per requirement) */
200 spinlock_t q_lock ____cacheline_aligned;
201 unsigned int qidx;
202
203 unsigned int qsize;
204 dma_addr_t qbase_dma;
205 dma_addr_t qdma_tail;
206
207 unsigned int active;
208 unsigned int suspended;
209
210 /* Interrupt flag */
211 bool int_en;
212
213 /* Register addresses for queue */
214 void __iomem *reg_control;
215 u32 qcontrol; /* Cached control register */
216
217 /* Status values from job */
218 u32 int_status;
219 u32 q_status;
220 u32 q_int_status;
221 u32 cmd_error;
222 /* Queue Statistics */
223 unsigned long total_pt_ops;
224 } ____cacheline_aligned;
225
226 struct pt_device {
227 struct list_head entry;
228
229 unsigned int ord;
230 char name[MAX_PT_NAME_LEN];
231
232 struct device *dev;
233
234 /* Bus specific device information */
235 struct pt_msix *pt_msix;
236
237 struct pt_dev_vdata *dev_vdata;
238
239 unsigned int pt_irq;
240
241 /* I/O area used for device communication */
242 void __iomem *io_regs;
243
244 spinlock_t cmd_lock ____cacheline_aligned;
245 unsigned int cmd_count;
246 struct list_head cmd;
247
248 /*
249 * The command queue. This represent the queue available on the
250 * PTDMA that are available for processing cmds
251 */
252 struct pt_cmd_queue cmd_q;
253
254 /* Support for the DMA Engine capabilities */
255 struct dma_device dma_dev;
256 struct pt_dma_chan *pt_dma_chan;
257 struct kmem_cache *dma_cmd_cache;
258 struct kmem_cache *dma_desc_cache;
259
260 wait_queue_head_t lsb_queue;
261
262 /* Device Statistics */
263 unsigned long total_interrupts;
264
265 struct pt_tasklet_data tdata;
266 int ver;
267 };
268
269 /*
270 * descriptor for PTDMA commands
271 * 8 32-bit words:
272 * word 0: function; engine; control bits
273 * word 1: length of source data
274 * word 2: low 32 bits of source pointer
275 * word 3: upper 16 bits of source pointer; source memory type
276 * word 4: low 32 bits of destination pointer
277 * word 5: upper 16 bits of destination pointer; destination memory type
278 * word 6: reserved 32 bits
279 * word 7: reserved 32 bits
280 */
281
282 #define DWORD0_SOC BIT(0)
283 #define DWORD0_IOC BIT(1)
284
285 struct dword3 {
286 unsigned int src_hi:16;
287 unsigned int src_mem:2;
288 unsigned int lsb_cxt_id:8;
289 unsigned int rsvd1:5;
290 unsigned int fixed:1;
291 };
292
293 struct dword5 {
294 unsigned int dst_hi:16;
295 unsigned int dst_mem:2;
296 unsigned int rsvd1:13;
297 unsigned int fixed:1;
298 };
299
300 struct ptdma_desc {
301 u32 dw0;
302 u32 length;
303 u32 src_lo;
304 struct dword3 dw3;
305 u32 dst_lo;
306 struct dword5 dw5;
307 __le32 rsvd1;
308 __le32 rsvd2;
309 };
310
311 /* Structure to hold PT device data */
312 struct pt_dev_vdata {
313 const unsigned int bar;
314 };
315
316 int pt_dmaengine_register(struct pt_device *pt);
317 void pt_dmaengine_unregister(struct pt_device *pt);
318
319 void ptdma_debugfs_setup(struct pt_device *pt);
320 int pt_core_init(struct pt_device *pt);
321 void pt_core_destroy(struct pt_device *pt);
322
323 int pt_core_perform_passthru(struct pt_cmd_queue *cmd_q,
324 struct pt_passthru_engine *pt_engine);
325
326 void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q);
327 void pt_start_queue(struct pt_cmd_queue *cmd_q);
328 void pt_stop_queue(struct pt_cmd_queue *cmd_q);
329
pt_core_disable_queue_interrupts(struct pt_device * pt)330 static inline void pt_core_disable_queue_interrupts(struct pt_device *pt)
331 {
332 iowrite32(0, pt->cmd_q.reg_control + 0x000C);
333 }
334
pt_core_enable_queue_interrupts(struct pt_device * pt)335 static inline void pt_core_enable_queue_interrupts(struct pt_device *pt)
336 {
337 iowrite32(SUPPORTED_INTERRUPTS, pt->cmd_q.reg_control + 0x000C);
338 }
339 #endif
340