1 /***************************************************************************
2 * Copyright (C) 2007 by Dominic Rath *
3 * Dominic.Rath@gmx.de *
5 * This program is free software; you can redistribute it and/or modify *
6 * it under the terms of the GNU General Public License as published by *
7 * the Free Software Foundation; either version 2 of the License, or *
8 * (at your option) any later version. *
10 * This program is distributed in the hope that it will be useful, *
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of *
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
13 * GNU General Public License for more details. *
15 * You should have received a copy of the GNU General Public License *
16 * along with this program; if not, write to the *
17 * Free Software Foundation, Inc., *
18 * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. *
19 ***************************************************************************/
30 static char* etb_reg_list
[] =
37 "ETB_ram_read_pointer",
38 "ETB_ram_write_pointer",
39 "ETB_trigger_counter",
43 static int etb_get_reg(struct reg
*reg
);
45 static int etb_set_instr(struct etb
*etb
, uint32_t new_instr
)
53 if (buf_get_u32(tap
->cur_instr
, 0, tap
->ir_length
) != new_instr
)
55 struct scan_field field
;
58 field
.num_bits
= tap
->ir_length
;
59 field
.out_value
= calloc(DIV_ROUND_UP(field
.num_bits
, 8), 1);
60 buf_set_u32(field
.out_value
, 0, field
.num_bits
, new_instr
);
62 field
.in_value
= NULL
;
64 jtag_add_ir_scan(1, &field
, jtag_get_end_state());
66 free(field
.out_value
);
72 static int etb_scann(struct etb
*etb
, uint32_t new_scan_chain
)
74 if (etb
->cur_scan_chain
!= new_scan_chain
)
76 struct scan_field field
;
80 field
.out_value
= calloc(DIV_ROUND_UP(field
.num_bits
, 8), 1);
81 buf_set_u32(field
.out_value
, 0, field
.num_bits
, new_scan_chain
);
83 field
.in_value
= NULL
;
85 /* select INTEST instruction */
86 etb_set_instr(etb
, 0x2);
87 jtag_add_dr_scan(1, &field
, jtag_get_end_state());
89 etb
->cur_scan_chain
= new_scan_chain
;
91 free(field
.out_value
);
97 static int etb_read_reg_w_check(struct reg
*, uint8_t *, uint8_t *);
98 static int etb_set_reg_w_exec(struct reg
*, uint8_t *);
100 static int etb_read_reg(struct reg
*reg
)
102 return etb_read_reg_w_check(reg
, NULL
, NULL
);
105 static int etb_get_reg(struct reg
*reg
)
109 if ((retval
= etb_read_reg(reg
)) != ERROR_OK
)
111 LOG_ERROR("BUG: error scheduling ETB register read");
115 if ((retval
= jtag_execute_queue()) != ERROR_OK
)
117 LOG_ERROR("ETB register read failed");
124 static const struct reg_arch_type etb_reg_type
= {
126 .set
= etb_set_reg_w_exec
,
129 struct reg_cache
* etb_build_reg_cache(struct etb
*etb
)
131 struct reg_cache
*reg_cache
= malloc(sizeof(struct reg_cache
));
132 struct reg
*reg_list
= NULL
;
133 struct etb_reg
*arch_info
= NULL
;
137 /* the actual registers are kept in two arrays */
138 reg_list
= calloc(num_regs
, sizeof(struct reg
));
139 arch_info
= calloc(num_regs
, sizeof(struct etb_reg
));
141 /* fill in values for the reg cache */
142 reg_cache
->name
= "etb registers";
143 reg_cache
->next
= NULL
;
144 reg_cache
->reg_list
= reg_list
;
145 reg_cache
->num_regs
= num_regs
;
147 /* set up registers */
148 for (i
= 0; i
< num_regs
; i
++)
150 reg_list
[i
].name
= etb_reg_list
[i
];
151 reg_list
[i
].size
= 32;
152 reg_list
[i
].dirty
= 0;
153 reg_list
[i
].valid
= 0;
154 reg_list
[i
].value
= calloc(1, 4);
155 reg_list
[i
].arch_info
= &arch_info
[i
];
156 reg_list
[i
].type
= &etb_reg_type
;
157 reg_list
[i
].size
= 32;
158 arch_info
[i
].addr
= i
;
159 arch_info
[i
].etb
= etb
;
165 static void etb_getbuf(jtag_callback_data_t arg
)
167 uint8_t *in
= (uint8_t *)arg
;
169 *((uint32_t *)in
) = buf_get_u32(in
, 0, 32);
173 static int etb_read_ram(struct etb
*etb
, uint32_t *data
, int num_frames
)
175 struct scan_field fields
[3];
178 jtag_set_end_state(TAP_IDLE
);
180 etb_set_instr(etb
, 0xc);
182 fields
[0].tap
= etb
->tap
;
183 fields
[0].num_bits
= 32;
184 fields
[0].out_value
= NULL
;
185 fields
[0].in_value
= NULL
;
187 fields
[1].tap
= etb
->tap
;
188 fields
[1].num_bits
= 7;
189 fields
[1].out_value
= malloc(1);
190 buf_set_u32(fields
[1].out_value
, 0, 7, 4);
191 fields
[1].in_value
= NULL
;
193 fields
[2].tap
= etb
->tap
;
194 fields
[2].num_bits
= 1;
195 fields
[2].out_value
= malloc(1);
196 buf_set_u32(fields
[2].out_value
, 0, 1, 0);
197 fields
[2].in_value
= NULL
;
199 jtag_add_dr_scan(3, fields
, jtag_get_end_state());
201 for (i
= 0; i
< num_frames
; i
++)
203 /* ensure nR/W reamins set to read */
204 buf_set_u32(fields
[2].out_value
, 0, 1, 0);
206 /* address remains set to 0x4 (RAM data) until we read the last frame */
207 if (i
< num_frames
- 1)
208 buf_set_u32(fields
[1].out_value
, 0, 7, 4);
210 buf_set_u32(fields
[1].out_value
, 0, 7, 0);
212 fields
[0].in_value
= (uint8_t *)(data
+ i
);
213 jtag_add_dr_scan(3, fields
, jtag_get_end_state());
215 jtag_add_callback(etb_getbuf
, (jtag_callback_data_t
)(data
+ i
));
218 jtag_execute_queue();
220 free(fields
[1].out_value
);
221 free(fields
[2].out_value
);
226 static int etb_read_reg_w_check(struct reg
*reg
,
227 uint8_t* check_value
, uint8_t* check_mask
)
229 struct etb_reg
*etb_reg
= reg
->arch_info
;
230 uint8_t reg_addr
= etb_reg
->addr
& 0x7f;
231 struct scan_field fields
[3];
233 LOG_DEBUG("%i", (int)(etb_reg
->addr
));
235 jtag_set_end_state(TAP_IDLE
);
236 etb_scann(etb_reg
->etb
, 0x0);
237 etb_set_instr(etb_reg
->etb
, 0xc);
239 fields
[0].tap
= etb_reg
->etb
->tap
;
240 fields
[0].num_bits
= 32;
241 fields
[0].out_value
= reg
->value
;
242 fields
[0].in_value
= NULL
;
243 fields
[0].check_value
= NULL
;
244 fields
[0].check_mask
= NULL
;
246 fields
[1].tap
= etb_reg
->etb
->tap
;
247 fields
[1].num_bits
= 7;
248 fields
[1].out_value
= malloc(1);
249 buf_set_u32(fields
[1].out_value
, 0, 7, reg_addr
);
250 fields
[1].in_value
= NULL
;
251 fields
[1].check_value
= NULL
;
252 fields
[1].check_mask
= NULL
;
254 fields
[2].tap
= etb_reg
->etb
->tap
;
255 fields
[2].num_bits
= 1;
256 fields
[2].out_value
= malloc(1);
257 buf_set_u32(fields
[2].out_value
, 0, 1, 0);
258 fields
[2].in_value
= NULL
;
259 fields
[2].check_value
= NULL
;
260 fields
[2].check_mask
= NULL
;
262 jtag_add_dr_scan(3, fields
, jtag_get_end_state());
264 /* read the identification register in the second run, to make sure we
265 * don't read the ETB data register twice, skipping every second entry
267 buf_set_u32(fields
[1].out_value
, 0, 7, 0x0);
268 fields
[0].in_value
= reg
->value
;
269 fields
[0].check_value
= check_value
;
270 fields
[0].check_mask
= check_mask
;
272 jtag_add_dr_scan_check(3, fields
, jtag_get_end_state());
274 free(fields
[1].out_value
);
275 free(fields
[2].out_value
);
280 static int etb_write_reg(struct reg
*, uint32_t);
282 static int etb_set_reg(struct reg
*reg
, uint32_t value
)
286 if ((retval
= etb_write_reg(reg
, value
)) != ERROR_OK
)
288 LOG_ERROR("BUG: error scheduling ETB register write");
292 buf_set_u32(reg
->value
, 0, reg
->size
, value
);
299 static int etb_set_reg_w_exec(struct reg
*reg
, uint8_t *buf
)
303 etb_set_reg(reg
, buf_get_u32(buf
, 0, reg
->size
));
305 if ((retval
= jtag_execute_queue()) != ERROR_OK
)
307 LOG_ERROR("ETB: register write failed");
313 static int etb_write_reg(struct reg
*reg
, uint32_t value
)
315 struct etb_reg
*etb_reg
= reg
->arch_info
;
316 uint8_t reg_addr
= etb_reg
->addr
& 0x7f;
317 struct scan_field fields
[3];
319 LOG_DEBUG("%i: 0x%8.8" PRIx32
"", (int)(etb_reg
->addr
), value
);
321 jtag_set_end_state(TAP_IDLE
);
322 etb_scann(etb_reg
->etb
, 0x0);
323 etb_set_instr(etb_reg
->etb
, 0xc);
325 fields
[0].tap
= etb_reg
->etb
->tap
;
326 fields
[0].num_bits
= 32;
327 fields
[0].out_value
= malloc(4);
328 buf_set_u32(fields
[0].out_value
, 0, 32, value
);
329 fields
[0].in_value
= NULL
;
331 fields
[1].tap
= etb_reg
->etb
->tap
;
332 fields
[1].num_bits
= 7;
333 fields
[1].out_value
= malloc(1);
334 buf_set_u32(fields
[1].out_value
, 0, 7, reg_addr
);
335 fields
[1].in_value
= NULL
;
337 fields
[2].tap
= etb_reg
->etb
->tap
;
338 fields
[2].num_bits
= 1;
339 fields
[2].out_value
= malloc(1);
340 buf_set_u32(fields
[2].out_value
, 0, 1, 1);
342 fields
[2].in_value
= NULL
;
344 free(fields
[0].out_value
);
345 free(fields
[1].out_value
);
346 free(fields
[2].out_value
);
351 COMMAND_HANDLER(handle_etb_config_command
)
353 struct target
*target
;
354 struct jtag_tap
*tap
;
359 return ERROR_COMMAND_SYNTAX_ERROR
;
362 target
= get_target(CMD_ARGV
[0]);
366 LOG_ERROR("ETB: target '%s' not defined", CMD_ARGV
[0]);
370 arm
= target_to_arm(target
);
373 command_print(CMD_CTX
, "ETB: '%s' isn't an ARM", CMD_ARGV
[0]);
377 tap
= jtag_tap_by_string(CMD_ARGV
[1]);
380 command_print(CMD_CTX
, "ETB: TAP %s does not exist", CMD_ARGV
[1]);
386 struct etb
*etb
= malloc(sizeof(struct etb
));
388 arm
->etm
->capture_driver_priv
= etb
;
391 etb
->cur_scan_chain
= 0xffffffff;
392 etb
->reg_cache
= NULL
;
398 LOG_ERROR("ETM: target has no ETM defined, ETB left unconfigured");
405 static const struct command_registration etb_config_command_handlers
[] = {
408 .handler
= &handle_etb_config_command
,
409 .mode
= COMMAND_CONFIG
,
410 .usage
= "<target> <tap>",
412 COMMAND_REGISTRATION_DONE
414 static const struct command_registration etb_command_handlers
[] = {
418 .help
= "Emebdded Trace Buffer command group",
419 .chain
= etb_config_command_handlers
,
421 COMMAND_REGISTRATION_DONE
424 static int etb_init(struct etm_context
*etm_ctx
)
426 struct etb
*etb
= etm_ctx
->capture_driver_priv
;
428 etb
->etm_ctx
= etm_ctx
;
430 /* identify ETB RAM depth and width */
431 etb_read_reg(&etb
->reg_cache
->reg_list
[ETB_RAM_DEPTH
]);
432 etb_read_reg(&etb
->reg_cache
->reg_list
[ETB_RAM_WIDTH
]);
433 jtag_execute_queue();
435 etb
->ram_depth
= buf_get_u32(etb
->reg_cache
->reg_list
[ETB_RAM_DEPTH
].value
, 0, 32);
436 etb
->ram_width
= buf_get_u32(etb
->reg_cache
->reg_list
[ETB_RAM_WIDTH
].value
, 0, 32);
441 static trace_status_t
etb_status(struct etm_context
*etm_ctx
)
443 struct etb
*etb
= etm_ctx
->capture_driver_priv
;
444 struct reg
*control
= &etb
->reg_cache
->reg_list
[ETB_CTRL
];
445 struct reg
*status
= &etb
->reg_cache
->reg_list
[ETB_STATUS
];
446 trace_status_t retval
= 0;
447 int etb_timeout
= 100;
449 etb
->etm_ctx
= etm_ctx
;
451 /* read control and status registers */
452 etb_read_reg(control
);
453 etb_read_reg(status
);
454 jtag_execute_queue();
456 /* See if it's (still) active */
457 retval
= buf_get_u32(control
->value
, 0, 1) ? TRACE_RUNNING
: TRACE_IDLE
;
459 /* check Full bit to identify wraparound/overflow */
460 if (buf_get_u32(status
->value
, 0, 1) == 1)
461 retval
|= TRACE_OVERFLOWED
;
463 /* check Triggered bit to identify trigger condition */
464 if (buf_get_u32(status
->value
, 1, 1) == 1)
465 retval
|= TRACE_TRIGGERED
;
467 /* check AcqComp to see if trigger counter dropped to zero */
468 if (buf_get_u32(status
->value
, 2, 1) == 1) {
469 /* wait for DFEmpty */
470 while (etb_timeout
-- && buf_get_u32(status
->value
, 3, 1) == 0)
473 if (etb_timeout
== 0)
474 LOG_ERROR("ETB: DFEmpty won't go high, status 0x%02x",
475 (unsigned) buf_get_u32(status
->value
, 0, 4));
477 if (!(etm_ctx
->capture_status
& TRACE_TRIGGERED
))
478 LOG_WARNING("ETB: trace complete without triggering?");
480 retval
|= TRACE_COMPLETED
;
483 /* NOTE: using a trigger is optional; and at least ETB11 has a mode
484 * where it can ignore the trigger counter.
487 /* update recorded state */
488 etm_ctx
->capture_status
= retval
;
493 static int etb_read_trace(struct etm_context
*etm_ctx
)
495 struct etb
*etb
= etm_ctx
->capture_driver_priv
;
497 int num_frames
= etb
->ram_depth
;
498 uint32_t *trace_data
= NULL
;
501 etb_read_reg(&etb
->reg_cache
->reg_list
[ETB_STATUS
]);
502 etb_read_reg(&etb
->reg_cache
->reg_list
[ETB_RAM_WRITE_POINTER
]);
503 jtag_execute_queue();
505 /* check if we overflowed, and adjust first frame of the trace accordingly
506 * if we didn't overflow, read only up to the frame that would be written next,
507 * i.e. don't read invalid entries
509 if (buf_get_u32(etb
->reg_cache
->reg_list
[ETB_STATUS
].value
, 0, 1))
511 first_frame
= buf_get_u32(etb
->reg_cache
->reg_list
[ETB_RAM_WRITE_POINTER
].value
, 0, 32);
515 num_frames
= buf_get_u32(etb
->reg_cache
->reg_list
[ETB_RAM_WRITE_POINTER
].value
, 0, 32);
518 etb_write_reg(&etb
->reg_cache
->reg_list
[ETB_RAM_READ_POINTER
], first_frame
);
520 /* read data into temporary array for unpacking */
521 trace_data
= malloc(sizeof(uint32_t) * num_frames
);
522 etb_read_ram(etb
, trace_data
, num_frames
);
524 if (etm_ctx
->trace_depth
> 0)
526 free(etm_ctx
->trace_data
);
529 if ((etm_ctx
->portmode
& ETM_PORT_WIDTH_MASK
) == ETM_PORT_4BIT
)
530 etm_ctx
->trace_depth
= num_frames
* 3;
531 else if ((etm_ctx
->portmode
& ETM_PORT_WIDTH_MASK
) == ETM_PORT_8BIT
)
532 etm_ctx
->trace_depth
= num_frames
* 2;
534 etm_ctx
->trace_depth
= num_frames
;
536 etm_ctx
->trace_data
= malloc(sizeof(struct etmv1_trace_data
) * etm_ctx
->trace_depth
);
538 for (i
= 0, j
= 0; i
< num_frames
; i
++)
540 if ((etm_ctx
->portmode
& ETM_PORT_WIDTH_MASK
) == ETM_PORT_4BIT
)
543 etm_ctx
->trace_data
[j
].pipestat
= trace_data
[i
] & 0x7;
544 etm_ctx
->trace_data
[j
].packet
= (trace_data
[i
] & 0x78) >> 3;
545 etm_ctx
->trace_data
[j
].flags
= 0;
546 if ((trace_data
[i
] & 0x80) >> 7)
548 etm_ctx
->trace_data
[j
].flags
|= ETMV1_TRACESYNC_CYCLE
;
550 if (etm_ctx
->trace_data
[j
].pipestat
== STAT_TR
)
552 etm_ctx
->trace_data
[j
].pipestat
= etm_ctx
->trace_data
[j
].packet
& 0x7;
553 etm_ctx
->trace_data
[j
].flags
|= ETMV1_TRIGGER_CYCLE
;
556 /* trace word j + 1 */
557 etm_ctx
->trace_data
[j
+ 1].pipestat
= (trace_data
[i
] & 0x100) >> 8;
558 etm_ctx
->trace_data
[j
+ 1].packet
= (trace_data
[i
] & 0x7800) >> 11;
559 etm_ctx
->trace_data
[j
+ 1].flags
= 0;
560 if ((trace_data
[i
] & 0x8000) >> 15)
562 etm_ctx
->trace_data
[j
+ 1].flags
|= ETMV1_TRACESYNC_CYCLE
;
564 if (etm_ctx
->trace_data
[j
+ 1].pipestat
== STAT_TR
)
566 etm_ctx
->trace_data
[j
+ 1].pipestat
= etm_ctx
->trace_data
[j
+ 1].packet
& 0x7;
567 etm_ctx
->trace_data
[j
+ 1].flags
|= ETMV1_TRIGGER_CYCLE
;
570 /* trace word j + 2 */
571 etm_ctx
->trace_data
[j
+ 2].pipestat
= (trace_data
[i
] & 0x10000) >> 16;
572 etm_ctx
->trace_data
[j
+ 2].packet
= (trace_data
[i
] & 0x780000) >> 19;
573 etm_ctx
->trace_data
[j
+ 2].flags
= 0;
574 if ((trace_data
[i
] & 0x800000) >> 23)
576 etm_ctx
->trace_data
[j
+ 2].flags
|= ETMV1_TRACESYNC_CYCLE
;
578 if (etm_ctx
->trace_data
[j
+ 2].pipestat
== STAT_TR
)
580 etm_ctx
->trace_data
[j
+ 2].pipestat
= etm_ctx
->trace_data
[j
+ 2].packet
& 0x7;
581 etm_ctx
->trace_data
[j
+ 2].flags
|= ETMV1_TRIGGER_CYCLE
;
586 else if ((etm_ctx
->portmode
& ETM_PORT_WIDTH_MASK
) == ETM_PORT_8BIT
)
589 etm_ctx
->trace_data
[j
].pipestat
= trace_data
[i
] & 0x7;
590 etm_ctx
->trace_data
[j
].packet
= (trace_data
[i
] & 0x7f8) >> 3;
591 etm_ctx
->trace_data
[j
].flags
= 0;
592 if ((trace_data
[i
] & 0x800) >> 11)
594 etm_ctx
->trace_data
[j
].flags
|= ETMV1_TRACESYNC_CYCLE
;
596 if (etm_ctx
->trace_data
[j
].pipestat
== STAT_TR
)
598 etm_ctx
->trace_data
[j
].pipestat
= etm_ctx
->trace_data
[j
].packet
& 0x7;
599 etm_ctx
->trace_data
[j
].flags
|= ETMV1_TRIGGER_CYCLE
;
602 /* trace word j + 1 */
603 etm_ctx
->trace_data
[j
+ 1].pipestat
= (trace_data
[i
] & 0x7000) >> 12;
604 etm_ctx
->trace_data
[j
+ 1].packet
= (trace_data
[i
] & 0x7f8000) >> 15;
605 etm_ctx
->trace_data
[j
+ 1].flags
= 0;
606 if ((trace_data
[i
] & 0x800000) >> 23)
608 etm_ctx
->trace_data
[j
+ 1].flags
|= ETMV1_TRACESYNC_CYCLE
;
610 if (etm_ctx
->trace_data
[j
+ 1].pipestat
== STAT_TR
)
612 etm_ctx
->trace_data
[j
+ 1].pipestat
= etm_ctx
->trace_data
[j
+ 1].packet
& 0x7;
613 etm_ctx
->trace_data
[j
+ 1].flags
|= ETMV1_TRIGGER_CYCLE
;
621 etm_ctx
->trace_data
[j
].pipestat
= trace_data
[i
] & 0x7;
622 etm_ctx
->trace_data
[j
].packet
= (trace_data
[i
] & 0x7fff8) >> 3;
623 etm_ctx
->trace_data
[j
].flags
= 0;
624 if ((trace_data
[i
] & 0x80000) >> 19)
626 etm_ctx
->trace_data
[j
].flags
|= ETMV1_TRACESYNC_CYCLE
;
628 if (etm_ctx
->trace_data
[j
].pipestat
== STAT_TR
)
630 etm_ctx
->trace_data
[j
].pipestat
= etm_ctx
->trace_data
[j
].packet
& 0x7;
631 etm_ctx
->trace_data
[j
].flags
|= ETMV1_TRIGGER_CYCLE
;
643 static int etb_start_capture(struct etm_context
*etm_ctx
)
645 struct etb
*etb
= etm_ctx
->capture_driver_priv
;
646 uint32_t etb_ctrl_value
= 0x1;
647 uint32_t trigger_count
;
649 if ((etm_ctx
->portmode
& ETM_PORT_MODE_MASK
) == ETM_PORT_DEMUXED
)
651 if ((etm_ctx
->portmode
& ETM_PORT_WIDTH_MASK
) != ETM_PORT_8BIT
)
653 LOG_ERROR("ETB can't run in demultiplexed mode with a 4 or 16 bit port");
654 return ERROR_ETM_PORTMODE_NOT_SUPPORTED
;
656 etb_ctrl_value
|= 0x2;
659 if ((etm_ctx
->portmode
& ETM_PORT_MODE_MASK
) == ETM_PORT_MUXED
) {
660 LOG_ERROR("ETB: can't run in multiplexed mode");
661 return ERROR_ETM_PORTMODE_NOT_SUPPORTED
;
664 trigger_count
= (etb
->ram_depth
* etm_ctx
->trigger_percent
) / 100;
666 etb_write_reg(&etb
->reg_cache
->reg_list
[ETB_TRIGGER_COUNTER
], trigger_count
);
667 etb_write_reg(&etb
->reg_cache
->reg_list
[ETB_RAM_WRITE_POINTER
], 0x0);
668 etb_write_reg(&etb
->reg_cache
->reg_list
[ETB_CTRL
], etb_ctrl_value
);
669 jtag_execute_queue();
671 /* we're starting a new trace, initialize capture status */
672 etm_ctx
->capture_status
= TRACE_RUNNING
;
677 static int etb_stop_capture(struct etm_context
*etm_ctx
)
679 struct etb
*etb
= etm_ctx
->capture_driver_priv
;
680 struct reg
*etb_ctrl_reg
= &etb
->reg_cache
->reg_list
[ETB_CTRL
];
682 etb_write_reg(etb_ctrl_reg
, 0x0);
683 jtag_execute_queue();
685 /* trace stopped, just clear running flag, but preserve others */
686 etm_ctx
->capture_status
&= ~TRACE_RUNNING
;
691 struct etm_capture_driver etb_capture_driver
=
694 .commands
= etb_command_handlers
,
696 .status
= etb_status
,
697 .start_capture
= etb_start_capture
,
698 .stop_capture
= etb_stop_capture
,
699 .read_trace
= etb_read_trace
,