tests/acceptance: Ignore binary data sent on serial console
[qemu/ar7.git] / hw / dma / sifive_pdma.c
blob9b2ac2017d9d36849e5ba7c5a3b2baee49f1b9e9
1 /*
2 * SiFive Platform DMA emulation
4 * Copyright (c) 2020 Wind River Systems, Inc.
6 * Author:
7 * Bin Meng <bin.meng@windriver.com>
9 * This program is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU General Public License as
11 * published by the Free Software Foundation; either version 2 or
12 * (at your option) version 3 of the License.
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 * GNU General Public License for more details.
19 * You should have received a copy of the GNU General Public License along
20 * with this program; if not, see <http://www.gnu.org/licenses/>.
23 #include "qemu/osdep.h"
24 #include "qemu/bitops.h"
25 #include "qemu/log.h"
26 #include "qapi/error.h"
27 #include "hw/irq.h"
28 #include "hw/qdev-properties.h"
29 #include "hw/sysbus.h"
30 #include "migration/vmstate.h"
31 #include "sysemu/dma.h"
32 #include "hw/dma/sifive_pdma.h"
34 #define DMA_CONTROL 0x000
35 #define CONTROL_CLAIM BIT(0)
36 #define CONTROL_RUN BIT(1)
37 #define CONTROL_DONE_IE BIT(14)
38 #define CONTROL_ERR_IE BIT(15)
39 #define CONTROL_DONE BIT(30)
40 #define CONTROL_ERR BIT(31)
42 #define DMA_NEXT_CONFIG 0x004
43 #define CONFIG_REPEAT BIT(2)
44 #define CONFIG_ORDER BIT(3)
45 #define CONFIG_WRSZ_SHIFT 24
46 #define CONFIG_RDSZ_SHIFT 28
47 #define CONFIG_SZ_MASK 0xf
49 #define DMA_NEXT_BYTES 0x008
50 #define DMA_NEXT_DST 0x010
51 #define DMA_NEXT_SRC 0x018
52 #define DMA_EXEC_CONFIG 0x104
53 #define DMA_EXEC_BYTES 0x108
54 #define DMA_EXEC_DST 0x110
55 #define DMA_EXEC_SRC 0x118
57 enum dma_chan_state {
58 DMA_CHAN_STATE_IDLE,
59 DMA_CHAN_STATE_STARTED,
60 DMA_CHAN_STATE_ERROR,
61 DMA_CHAN_STATE_DONE
64 static void sifive_pdma_run(SiFivePDMAState *s, int ch)
66 uint64_t bytes = s->chan[ch].next_bytes;
67 uint64_t dst = s->chan[ch].next_dst;
68 uint64_t src = s->chan[ch].next_src;
69 uint32_t config = s->chan[ch].next_config;
70 int wsize, rsize, size;
71 uint8_t buf[64];
72 int n;
74 /* do nothing if bytes to transfer is zero */
75 if (!bytes) {
76 goto error;
80 * The manual does not describe how the hardware behaviors when
81 * config.wsize and config.rsize are given different values.
82 * A common case is memory to memory DMA, and in this case they
83 * are normally the same. Abort if this expectation fails.
85 wsize = (config >> CONFIG_WRSZ_SHIFT) & CONFIG_SZ_MASK;
86 rsize = (config >> CONFIG_RDSZ_SHIFT) & CONFIG_SZ_MASK;
87 if (wsize != rsize) {
88 goto error;
92 * Calculate the transaction size
94 * size field is base 2 logarithm of DMA transaction size,
95 * but there is an upper limit of 64 bytes per transaction.
97 size = wsize;
98 if (size > 6) {
99 size = 6;
101 size = 1 << size;
103 /* the bytes to transfer should be multiple of transaction size */
104 if (bytes % size) {
105 goto error;
108 /* indicate a DMA transfer is started */
109 s->chan[ch].state = DMA_CHAN_STATE_STARTED;
110 s->chan[ch].control &= ~CONTROL_DONE;
111 s->chan[ch].control &= ~CONTROL_ERR;
113 /* load the next_ registers into their exec_ counterparts */
114 s->chan[ch].exec_config = config;
115 s->chan[ch].exec_bytes = bytes;
116 s->chan[ch].exec_dst = dst;
117 s->chan[ch].exec_src = src;
119 for (n = 0; n < bytes / size; n++) {
120 cpu_physical_memory_read(s->chan[ch].exec_src, buf, size);
121 cpu_physical_memory_write(s->chan[ch].exec_dst, buf, size);
122 s->chan[ch].exec_src += size;
123 s->chan[ch].exec_dst += size;
124 s->chan[ch].exec_bytes -= size;
127 /* indicate a DMA transfer is done */
128 s->chan[ch].state = DMA_CHAN_STATE_DONE;
129 s->chan[ch].control &= ~CONTROL_RUN;
130 s->chan[ch].control |= CONTROL_DONE;
132 /* reload exec_ registers if repeat is required */
133 if (s->chan[ch].next_config & CONFIG_REPEAT) {
134 s->chan[ch].exec_bytes = bytes;
135 s->chan[ch].exec_dst = dst;
136 s->chan[ch].exec_src = src;
139 return;
141 error:
142 s->chan[ch].state = DMA_CHAN_STATE_ERROR;
143 s->chan[ch].control |= CONTROL_ERR;
144 return;
147 static inline void sifive_pdma_update_irq(SiFivePDMAState *s, int ch)
149 bool done_ie, err_ie;
151 done_ie = !!(s->chan[ch].control & CONTROL_DONE_IE);
152 err_ie = !!(s->chan[ch].control & CONTROL_ERR_IE);
154 if (done_ie && (s->chan[ch].control & CONTROL_DONE)) {
155 qemu_irq_raise(s->irq[ch * 2]);
156 } else {
157 qemu_irq_lower(s->irq[ch * 2]);
160 if (err_ie && (s->chan[ch].control & CONTROL_ERR)) {
161 qemu_irq_raise(s->irq[ch * 2 + 1]);
162 } else {
163 qemu_irq_lower(s->irq[ch * 2 + 1]);
166 s->chan[ch].state = DMA_CHAN_STATE_IDLE;
169 static uint64_t sifive_pdma_read(void *opaque, hwaddr offset, unsigned size)
171 SiFivePDMAState *s = opaque;
172 int ch = SIFIVE_PDMA_CHAN_NO(offset);
173 uint64_t val = 0;
175 if (ch >= SIFIVE_PDMA_CHANS) {
176 qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid channel no %d\n",
177 __func__, ch);
178 return 0;
181 offset &= 0xfff;
182 switch (offset) {
183 case DMA_CONTROL:
184 val = s->chan[ch].control;
185 break;
186 case DMA_NEXT_CONFIG:
187 val = s->chan[ch].next_config;
188 break;
189 case DMA_NEXT_BYTES:
190 val = s->chan[ch].next_bytes;
191 break;
192 case DMA_NEXT_DST:
193 val = s->chan[ch].next_dst;
194 break;
195 case DMA_NEXT_SRC:
196 val = s->chan[ch].next_src;
197 break;
198 case DMA_EXEC_CONFIG:
199 val = s->chan[ch].exec_config;
200 break;
201 case DMA_EXEC_BYTES:
202 val = s->chan[ch].exec_bytes;
203 break;
204 case DMA_EXEC_DST:
205 val = s->chan[ch].exec_dst;
206 break;
207 case DMA_EXEC_SRC:
208 val = s->chan[ch].exec_src;
209 break;
210 default:
211 qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad offset 0x%" HWADDR_PRIX "\n",
212 __func__, offset);
213 break;
216 return val;
219 static void sifive_pdma_write(void *opaque, hwaddr offset,
220 uint64_t value, unsigned size)
222 SiFivePDMAState *s = opaque;
223 int ch = SIFIVE_PDMA_CHAN_NO(offset);
225 if (ch >= SIFIVE_PDMA_CHANS) {
226 qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid channel no %d\n",
227 __func__, ch);
228 return;
231 offset &= 0xfff;
232 switch (offset) {
233 case DMA_CONTROL:
234 s->chan[ch].control = value;
236 if (value & CONTROL_RUN) {
237 sifive_pdma_run(s, ch);
240 sifive_pdma_update_irq(s, ch);
241 break;
242 case DMA_NEXT_CONFIG:
243 s->chan[ch].next_config = value;
244 break;
245 case DMA_NEXT_BYTES:
246 s->chan[ch].next_bytes = value;
247 break;
248 case DMA_NEXT_DST:
249 s->chan[ch].next_dst = value;
250 break;
251 case DMA_NEXT_SRC:
252 s->chan[ch].next_src = value;
253 break;
254 case DMA_EXEC_CONFIG:
255 case DMA_EXEC_BYTES:
256 case DMA_EXEC_DST:
257 case DMA_EXEC_SRC:
258 /* these are read-only registers */
259 break;
260 default:
261 qemu_log_mask(LOG_GUEST_ERROR, "%s: Bad offset 0x%" HWADDR_PRIX "\n",
262 __func__, offset);
263 break;
267 static const MemoryRegionOps sifive_pdma_ops = {
268 .read = sifive_pdma_read,
269 .write = sifive_pdma_write,
270 .endianness = DEVICE_LITTLE_ENDIAN,
271 /* there are 32-bit and 64-bit wide registers */
272 .impl = {
273 .min_access_size = 4,
274 .max_access_size = 8,
278 static void sifive_pdma_realize(DeviceState *dev, Error **errp)
280 SiFivePDMAState *s = SIFIVE_PDMA(dev);
281 int i;
283 memory_region_init_io(&s->iomem, OBJECT(dev), &sifive_pdma_ops, s,
284 TYPE_SIFIVE_PDMA, SIFIVE_PDMA_REG_SIZE);
285 sysbus_init_mmio(SYS_BUS_DEVICE(dev), &s->iomem);
287 for (i = 0; i < SIFIVE_PDMA_IRQS; i++) {
288 sysbus_init_irq(SYS_BUS_DEVICE(dev), &s->irq[i]);
292 static void sifive_pdma_class_init(ObjectClass *klass, void *data)
294 DeviceClass *dc = DEVICE_CLASS(klass);
296 dc->desc = "SiFive Platform DMA controller";
297 dc->realize = sifive_pdma_realize;
300 static const TypeInfo sifive_pdma_info = {
301 .name = TYPE_SIFIVE_PDMA,
302 .parent = TYPE_SYS_BUS_DEVICE,
303 .instance_size = sizeof(SiFivePDMAState),
304 .class_init = sifive_pdma_class_init,
307 static void sifive_pdma_register_types(void)
309 type_register_static(&sifive_pdma_info);
312 type_init(sifive_pdma_register_types)