5 #include "util/cache.h"
6 #include "util/symbol.h"
7 #include "util/thread.h"
8 #include "util/header.h"
9 #include "util/session.h"
11 #include "util/parse-options.h"
12 #include "util/trace-event.h"
14 #include "util/debug.h"
16 #include <sys/prctl.h>
18 #include <semaphore.h>
22 static char const *input_name
= "perf.data";
24 static char default_sort_order
[] = "avg, max, switch, runtime";
25 static const char *sort_order
= default_sort_order
;
27 static int profile_cpu
= -1;
29 #define PR_SET_NAME 15 /* Set process name */
32 static u64 run_measurement_overhead
;
33 static u64 sleep_measurement_overhead
;
40 static unsigned long nr_tasks
;
49 unsigned long nr_events
;
50 unsigned long curr_event
;
51 struct sched_atom
**atoms
;
62 enum sched_event_type
{
66 SCHED_EVENT_MIGRATION
,
70 enum sched_event_type type
;
76 struct task_desc
*wakee
;
79 static struct task_desc
*pid_to_task
[MAX_PID
];
81 static struct task_desc
**tasks
;
83 static pthread_mutex_t start_work_mutex
= PTHREAD_MUTEX_INITIALIZER
;
84 static u64 start_time
;
86 static pthread_mutex_t work_done_wait_mutex
= PTHREAD_MUTEX_INITIALIZER
;
88 static unsigned long nr_run_events
;
89 static unsigned long nr_sleep_events
;
90 static unsigned long nr_wakeup_events
;
92 static unsigned long nr_sleep_corrections
;
93 static unsigned long nr_run_events_optimized
;
95 static unsigned long targetless_wakeups
;
96 static unsigned long multitarget_wakeups
;
99 static u64 runavg_cpu_usage
;
100 static u64 parent_cpu_usage
;
101 static u64 runavg_parent_cpu_usage
;
103 static unsigned long nr_runs
;
104 static u64 sum_runtime
;
105 static u64 sum_fluct
;
108 static unsigned int replay_repeat
= 10;
109 static unsigned long nr_timestamps
;
110 static unsigned long nr_unordered_timestamps
;
111 static unsigned long nr_state_machine_bugs
;
112 static unsigned long nr_context_switch_bugs
;
113 static unsigned long nr_events
;
114 static unsigned long nr_lost_chunks
;
115 static unsigned long nr_lost_events
;
117 #define TASK_STATE_TO_CHAR_STR "RSDTtZX"
127 struct list_head list
;
128 enum thread_state state
;
136 struct list_head work_list
;
137 struct thread
*thread
;
146 typedef int (*sort_fn_t
)(struct work_atoms
*, struct work_atoms
*);
148 static struct rb_root atom_root
, sorted_atom_root
;
150 static u64 all_runtime
;
151 static u64 all_count
;
154 static u64
get_nsecs(void)
158 clock_gettime(CLOCK_MONOTONIC
, &ts
);
160 return ts
.tv_sec
* 1000000000ULL + ts
.tv_nsec
;
163 static void burn_nsecs(u64 nsecs
)
165 u64 T0
= get_nsecs(), T1
;
169 } while (T1
+ run_measurement_overhead
< T0
+ nsecs
);
172 static void sleep_nsecs(u64 nsecs
)
176 ts
.tv_nsec
= nsecs
% 999999999;
177 ts
.tv_sec
= nsecs
/ 999999999;
179 nanosleep(&ts
, NULL
);
182 static void calibrate_run_measurement_overhead(void)
184 u64 T0
, T1
, delta
, min_delta
= 1000000000ULL;
187 for (i
= 0; i
< 10; i
++) {
192 min_delta
= min(min_delta
, delta
);
194 run_measurement_overhead
= min_delta
;
196 printf("run measurement overhead: %Ld nsecs\n", min_delta
);
199 static void calibrate_sleep_measurement_overhead(void)
201 u64 T0
, T1
, delta
, min_delta
= 1000000000ULL;
204 for (i
= 0; i
< 10; i
++) {
209 min_delta
= min(min_delta
, delta
);
212 sleep_measurement_overhead
= min_delta
;
214 printf("sleep measurement overhead: %Ld nsecs\n", min_delta
);
217 static struct sched_atom
*
218 get_new_event(struct task_desc
*task
, u64 timestamp
)
220 struct sched_atom
*event
= zalloc(sizeof(*event
));
221 unsigned long idx
= task
->nr_events
;
224 event
->timestamp
= timestamp
;
228 size
= sizeof(struct sched_atom
*) * task
->nr_events
;
229 task
->atoms
= realloc(task
->atoms
, size
);
230 BUG_ON(!task
->atoms
);
232 task
->atoms
[idx
] = event
;
237 static struct sched_atom
*last_event(struct task_desc
*task
)
239 if (!task
->nr_events
)
242 return task
->atoms
[task
->nr_events
- 1];
246 add_sched_event_run(struct task_desc
*task
, u64 timestamp
, u64 duration
)
248 struct sched_atom
*event
, *curr_event
= last_event(task
);
251 * optimize an existing RUN event by merging this one
254 if (curr_event
&& curr_event
->type
== SCHED_EVENT_RUN
) {
255 nr_run_events_optimized
++;
256 curr_event
->duration
+= duration
;
260 event
= get_new_event(task
, timestamp
);
262 event
->type
= SCHED_EVENT_RUN
;
263 event
->duration
= duration
;
269 add_sched_event_wakeup(struct task_desc
*task
, u64 timestamp
,
270 struct task_desc
*wakee
)
272 struct sched_atom
*event
, *wakee_event
;
274 event
= get_new_event(task
, timestamp
);
275 event
->type
= SCHED_EVENT_WAKEUP
;
276 event
->wakee
= wakee
;
278 wakee_event
= last_event(wakee
);
279 if (!wakee_event
|| wakee_event
->type
!= SCHED_EVENT_SLEEP
) {
280 targetless_wakeups
++;
283 if (wakee_event
->wait_sem
) {
284 multitarget_wakeups
++;
288 wakee_event
->wait_sem
= zalloc(sizeof(*wakee_event
->wait_sem
));
289 sem_init(wakee_event
->wait_sem
, 0, 0);
290 wakee_event
->specific_wait
= 1;
291 event
->wait_sem
= wakee_event
->wait_sem
;
297 add_sched_event_sleep(struct task_desc
*task
, u64 timestamp
,
298 u64 task_state __used
)
300 struct sched_atom
*event
= get_new_event(task
, timestamp
);
302 event
->type
= SCHED_EVENT_SLEEP
;
307 static struct task_desc
*register_pid(unsigned long pid
, const char *comm
)
309 struct task_desc
*task
;
311 BUG_ON(pid
>= MAX_PID
);
313 task
= pid_to_task
[pid
];
318 task
= zalloc(sizeof(*task
));
321 strcpy(task
->comm
, comm
);
323 * every task starts in sleeping state - this gets ignored
324 * if there's no wakeup pointing to this sleep state:
326 add_sched_event_sleep(task
, 0, 0);
328 pid_to_task
[pid
] = task
;
330 tasks
= realloc(tasks
, nr_tasks
*sizeof(struct task_task
*));
332 tasks
[task
->nr
] = task
;
335 printf("registered task #%ld, PID %ld (%s)\n", nr_tasks
, pid
, comm
);
341 static void print_task_traces(void)
343 struct task_desc
*task
;
346 for (i
= 0; i
< nr_tasks
; i
++) {
348 printf("task %6ld (%20s:%10ld), nr_events: %ld\n",
349 task
->nr
, task
->comm
, task
->pid
, task
->nr_events
);
353 static void add_cross_task_wakeups(void)
355 struct task_desc
*task1
, *task2
;
358 for (i
= 0; i
< nr_tasks
; i
++) {
364 add_sched_event_wakeup(task1
, 0, task2
);
369 process_sched_event(struct task_desc
*this_task __used
, struct sched_atom
*atom
)
376 delta
= start_time
+ atom
->timestamp
- now
;
378 switch (atom
->type
) {
379 case SCHED_EVENT_RUN
:
380 burn_nsecs(atom
->duration
);
382 case SCHED_EVENT_SLEEP
:
384 ret
= sem_wait(atom
->wait_sem
);
387 case SCHED_EVENT_WAKEUP
:
389 ret
= sem_post(atom
->wait_sem
);
392 case SCHED_EVENT_MIGRATION
:
399 static u64
get_cpu_usage_nsec_parent(void)
405 err
= getrusage(RUSAGE_SELF
, &ru
);
408 sum
= ru
.ru_utime
.tv_sec
*1e9
+ ru
.ru_utime
.tv_usec
*1e3
;
409 sum
+= ru
.ru_stime
.tv_sec
*1e9
+ ru
.ru_stime
.tv_usec
*1e3
;
414 static int self_open_counters(void)
416 struct perf_event_attr attr
;
419 memset(&attr
, 0, sizeof(attr
));
421 attr
.type
= PERF_TYPE_SOFTWARE
;
422 attr
.config
= PERF_COUNT_SW_TASK_CLOCK
;
424 fd
= sys_perf_event_open(&attr
, 0, -1, -1, 0);
427 die("Error: sys_perf_event_open() syscall returned"
428 "with %d (%s)\n", fd
, strerror(errno
));
432 static u64
get_cpu_usage_nsec_self(int fd
)
437 ret
= read(fd
, &runtime
, sizeof(runtime
));
438 BUG_ON(ret
!= sizeof(runtime
));
443 static void *thread_func(void *ctx
)
445 struct task_desc
*this_task
= ctx
;
446 u64 cpu_usage_0
, cpu_usage_1
;
447 unsigned long i
, ret
;
451 sprintf(comm2
, ":%s", this_task
->comm
);
452 prctl(PR_SET_NAME
, comm2
);
453 fd
= self_open_counters();
456 ret
= sem_post(&this_task
->ready_for_work
);
458 ret
= pthread_mutex_lock(&start_work_mutex
);
460 ret
= pthread_mutex_unlock(&start_work_mutex
);
463 cpu_usage_0
= get_cpu_usage_nsec_self(fd
);
465 for (i
= 0; i
< this_task
->nr_events
; i
++) {
466 this_task
->curr_event
= i
;
467 process_sched_event(this_task
, this_task
->atoms
[i
]);
470 cpu_usage_1
= get_cpu_usage_nsec_self(fd
);
471 this_task
->cpu_usage
= cpu_usage_1
- cpu_usage_0
;
472 ret
= sem_post(&this_task
->work_done_sem
);
475 ret
= pthread_mutex_lock(&work_done_wait_mutex
);
477 ret
= pthread_mutex_unlock(&work_done_wait_mutex
);
483 static void create_tasks(void)
485 struct task_desc
*task
;
490 err
= pthread_attr_init(&attr
);
492 err
= pthread_attr_setstacksize(&attr
, (size_t)(16*1024));
494 err
= pthread_mutex_lock(&start_work_mutex
);
496 err
= pthread_mutex_lock(&work_done_wait_mutex
);
498 for (i
= 0; i
< nr_tasks
; i
++) {
500 sem_init(&task
->sleep_sem
, 0, 0);
501 sem_init(&task
->ready_for_work
, 0, 0);
502 sem_init(&task
->work_done_sem
, 0, 0);
503 task
->curr_event
= 0;
504 err
= pthread_create(&task
->thread
, &attr
, thread_func
, task
);
509 static void wait_for_tasks(void)
511 u64 cpu_usage_0
, cpu_usage_1
;
512 struct task_desc
*task
;
513 unsigned long i
, ret
;
515 start_time
= get_nsecs();
517 pthread_mutex_unlock(&work_done_wait_mutex
);
519 for (i
= 0; i
< nr_tasks
; i
++) {
521 ret
= sem_wait(&task
->ready_for_work
);
523 sem_init(&task
->ready_for_work
, 0, 0);
525 ret
= pthread_mutex_lock(&work_done_wait_mutex
);
528 cpu_usage_0
= get_cpu_usage_nsec_parent();
530 pthread_mutex_unlock(&start_work_mutex
);
532 for (i
= 0; i
< nr_tasks
; i
++) {
534 ret
= sem_wait(&task
->work_done_sem
);
536 sem_init(&task
->work_done_sem
, 0, 0);
537 cpu_usage
+= task
->cpu_usage
;
541 cpu_usage_1
= get_cpu_usage_nsec_parent();
542 if (!runavg_cpu_usage
)
543 runavg_cpu_usage
= cpu_usage
;
544 runavg_cpu_usage
= (runavg_cpu_usage
*9 + cpu_usage
)/10;
546 parent_cpu_usage
= cpu_usage_1
- cpu_usage_0
;
547 if (!runavg_parent_cpu_usage
)
548 runavg_parent_cpu_usage
= parent_cpu_usage
;
549 runavg_parent_cpu_usage
= (runavg_parent_cpu_usage
*9 +
550 parent_cpu_usage
)/10;
552 ret
= pthread_mutex_lock(&start_work_mutex
);
555 for (i
= 0; i
< nr_tasks
; i
++) {
557 sem_init(&task
->sleep_sem
, 0, 0);
558 task
->curr_event
= 0;
562 static void run_one_test(void)
564 u64 T0
, T1
, delta
, avg_delta
, fluct
, std_dev
;
571 sum_runtime
+= delta
;
574 avg_delta
= sum_runtime
/ nr_runs
;
575 if (delta
< avg_delta
)
576 fluct
= avg_delta
- delta
;
578 fluct
= delta
- avg_delta
;
580 std_dev
= sum_fluct
/ nr_runs
/ sqrt(nr_runs
);
583 run_avg
= (run_avg
*9 + delta
)/10;
585 printf("#%-3ld: %0.3f, ",
586 nr_runs
, (double)delta
/1000000.0);
588 printf("ravg: %0.2f, ",
589 (double)run_avg
/1e6
);
591 printf("cpu: %0.2f / %0.2f",
592 (double)cpu_usage
/1e6
, (double)runavg_cpu_usage
/1e6
);
596 * rusage statistics done by the parent, these are less
597 * accurate than the sum_exec_runtime based statistics:
599 printf(" [%0.2f / %0.2f]",
600 (double)parent_cpu_usage
/1e6
,
601 (double)runavg_parent_cpu_usage
/1e6
);
606 if (nr_sleep_corrections
)
607 printf(" (%ld sleep corrections)\n", nr_sleep_corrections
);
608 nr_sleep_corrections
= 0;
611 static void test_calibrations(void)
619 printf("the run test took %Ld nsecs\n", T1
-T0
);
625 printf("the sleep test took %Ld nsecs\n", T1
-T0
);
628 #define FILL_FIELD(ptr, field, event, data) \
629 ptr.field = (typeof(ptr.field)) raw_field_value(event, #field, data)
631 #define FILL_ARRAY(ptr, array, event, data) \
633 void *__array = raw_field_ptr(event, #array, data); \
634 memcpy(ptr.array, __array, sizeof(ptr.array)); \
637 #define FILL_COMMON_FIELDS(ptr, event, data) \
639 FILL_FIELD(ptr, common_type, event, data); \
640 FILL_FIELD(ptr, common_flags, event, data); \
641 FILL_FIELD(ptr, common_preempt_count, event, data); \
642 FILL_FIELD(ptr, common_pid, event, data); \
643 FILL_FIELD(ptr, common_tgid, event, data); \
648 struct trace_switch_event
{
653 u8 common_preempt_count
;
666 struct trace_runtime_event
{
671 u8 common_preempt_count
;
681 struct trace_wakeup_event
{
686 u8 common_preempt_count
;
698 struct trace_fork_event
{
703 u8 common_preempt_count
;
707 char parent_comm
[16];
713 struct trace_migrate_task_event
{
718 u8 common_preempt_count
;
729 struct trace_sched_handler
{
730 void (*switch_event
)(struct trace_switch_event
*,
731 struct perf_session
*,
735 struct thread
*thread
);
737 void (*runtime_event
)(struct trace_runtime_event
*,
738 struct perf_session
*,
742 struct thread
*thread
);
744 void (*wakeup_event
)(struct trace_wakeup_event
*,
745 struct perf_session
*,
749 struct thread
*thread
);
751 void (*fork_event
)(struct trace_fork_event
*,
755 struct thread
*thread
);
757 void (*migrate_task_event
)(struct trace_migrate_task_event
*,
758 struct perf_session
*session
,
762 struct thread
*thread
);
767 replay_wakeup_event(struct trace_wakeup_event
*wakeup_event
,
768 struct perf_session
*session __used
,
771 u64 timestamp __used
,
772 struct thread
*thread __used
)
774 struct task_desc
*waker
, *wakee
;
777 printf("sched_wakeup event %p\n", event
);
779 printf(" ... pid %d woke up %s/%d\n",
780 wakeup_event
->common_pid
,
785 waker
= register_pid(wakeup_event
->common_pid
, "<unknown>");
786 wakee
= register_pid(wakeup_event
->pid
, wakeup_event
->comm
);
788 add_sched_event_wakeup(waker
, timestamp
, wakee
);
791 static u64 cpu_last_switched
[MAX_CPUS
];
794 replay_switch_event(struct trace_switch_event
*switch_event
,
795 struct perf_session
*session __used
,
799 struct thread
*thread __used
)
801 struct task_desc
*prev
, *next
;
806 printf("sched_switch event %p\n", event
);
808 if (cpu
>= MAX_CPUS
|| cpu
< 0)
811 timestamp0
= cpu_last_switched
[cpu
];
813 delta
= timestamp
- timestamp0
;
818 die("hm, delta: %Ld < 0 ?\n", delta
);
821 printf(" ... switch from %s/%d to %s/%d [ran %Ld nsecs]\n",
822 switch_event
->prev_comm
, switch_event
->prev_pid
,
823 switch_event
->next_comm
, switch_event
->next_pid
,
827 prev
= register_pid(switch_event
->prev_pid
, switch_event
->prev_comm
);
828 next
= register_pid(switch_event
->next_pid
, switch_event
->next_comm
);
830 cpu_last_switched
[cpu
] = timestamp
;
832 add_sched_event_run(prev
, timestamp
, delta
);
833 add_sched_event_sleep(prev
, timestamp
, switch_event
->prev_state
);
838 replay_fork_event(struct trace_fork_event
*fork_event
,
841 u64 timestamp __used
,
842 struct thread
*thread __used
)
845 printf("sched_fork event %p\n", event
);
846 printf("... parent: %s/%d\n", fork_event
->parent_comm
, fork_event
->parent_pid
);
847 printf("... child: %s/%d\n", fork_event
->child_comm
, fork_event
->child_pid
);
849 register_pid(fork_event
->parent_pid
, fork_event
->parent_comm
);
850 register_pid(fork_event
->child_pid
, fork_event
->child_comm
);
853 static struct trace_sched_handler replay_ops
= {
854 .wakeup_event
= replay_wakeup_event
,
855 .switch_event
= replay_switch_event
,
856 .fork_event
= replay_fork_event
,
859 struct sort_dimension
{
862 struct list_head list
;
865 static LIST_HEAD(cmp_pid
);
868 thread_lat_cmp(struct list_head
*list
, struct work_atoms
*l
, struct work_atoms
*r
)
870 struct sort_dimension
*sort
;
873 BUG_ON(list_empty(list
));
875 list_for_each_entry(sort
, list
, list
) {
876 ret
= sort
->cmp(l
, r
);
884 static struct work_atoms
*
885 thread_atoms_search(struct rb_root
*root
, struct thread
*thread
,
886 struct list_head
*sort_list
)
888 struct rb_node
*node
= root
->rb_node
;
889 struct work_atoms key
= { .thread
= thread
};
892 struct work_atoms
*atoms
;
895 atoms
= container_of(node
, struct work_atoms
, node
);
897 cmp
= thread_lat_cmp(sort_list
, &key
, atoms
);
899 node
= node
->rb_left
;
901 node
= node
->rb_right
;
903 BUG_ON(thread
!= atoms
->thread
);
911 __thread_latency_insert(struct rb_root
*root
, struct work_atoms
*data
,
912 struct list_head
*sort_list
)
914 struct rb_node
**new = &(root
->rb_node
), *parent
= NULL
;
917 struct work_atoms
*this;
920 this = container_of(*new, struct work_atoms
, node
);
923 cmp
= thread_lat_cmp(sort_list
, data
, this);
926 new = &((*new)->rb_left
);
928 new = &((*new)->rb_right
);
931 rb_link_node(&data
->node
, parent
, new);
932 rb_insert_color(&data
->node
, root
);
935 static void thread_atoms_insert(struct thread
*thread
)
937 struct work_atoms
*atoms
= zalloc(sizeof(*atoms
));
941 atoms
->thread
= thread
;
942 INIT_LIST_HEAD(&atoms
->work_list
);
943 __thread_latency_insert(&atom_root
, atoms
, &cmp_pid
);
947 latency_fork_event(struct trace_fork_event
*fork_event __used
,
948 struct event
*event __used
,
950 u64 timestamp __used
,
951 struct thread
*thread __used
)
953 /* should insert the newcomer */
957 static char sched_out_state(struct trace_switch_event
*switch_event
)
959 const char *str
= TASK_STATE_TO_CHAR_STR
;
961 return str
[switch_event
->prev_state
];
965 add_sched_out_event(struct work_atoms
*atoms
,
969 struct work_atom
*atom
= zalloc(sizeof(*atom
));
973 atom
->sched_out_time
= timestamp
;
975 if (run_state
== 'R') {
976 atom
->state
= THREAD_WAIT_CPU
;
977 atom
->wake_up_time
= atom
->sched_out_time
;
980 list_add_tail(&atom
->list
, &atoms
->work_list
);
984 add_runtime_event(struct work_atoms
*atoms
, u64 delta
, u64 timestamp __used
)
986 struct work_atom
*atom
;
988 BUG_ON(list_empty(&atoms
->work_list
));
990 atom
= list_entry(atoms
->work_list
.prev
, struct work_atom
, list
);
992 atom
->runtime
+= delta
;
993 atoms
->total_runtime
+= delta
;
997 add_sched_in_event(struct work_atoms
*atoms
, u64 timestamp
)
999 struct work_atom
*atom
;
1002 if (list_empty(&atoms
->work_list
))
1005 atom
= list_entry(atoms
->work_list
.prev
, struct work_atom
, list
);
1007 if (atom
->state
!= THREAD_WAIT_CPU
)
1010 if (timestamp
< atom
->wake_up_time
) {
1011 atom
->state
= THREAD_IGNORE
;
1015 atom
->state
= THREAD_SCHED_IN
;
1016 atom
->sched_in_time
= timestamp
;
1018 delta
= atom
->sched_in_time
- atom
->wake_up_time
;
1019 atoms
->total_lat
+= delta
;
1020 if (delta
> atoms
->max_lat
) {
1021 atoms
->max_lat
= delta
;
1022 atoms
->max_lat_at
= timestamp
;
1028 latency_switch_event(struct trace_switch_event
*switch_event
,
1029 struct perf_session
*session
,
1030 struct event
*event __used
,
1033 struct thread
*thread __used
)
1035 struct work_atoms
*out_events
, *in_events
;
1036 struct thread
*sched_out
, *sched_in
;
1040 BUG_ON(cpu
>= MAX_CPUS
|| cpu
< 0);
1042 timestamp0
= cpu_last_switched
[cpu
];
1043 cpu_last_switched
[cpu
] = timestamp
;
1045 delta
= timestamp
- timestamp0
;
1050 die("hm, delta: %Ld < 0 ?\n", delta
);
1053 sched_out
= perf_session__findnew(session
, switch_event
->prev_pid
);
1054 sched_in
= perf_session__findnew(session
, switch_event
->next_pid
);
1056 out_events
= thread_atoms_search(&atom_root
, sched_out
, &cmp_pid
);
1058 thread_atoms_insert(sched_out
);
1059 out_events
= thread_atoms_search(&atom_root
, sched_out
, &cmp_pid
);
1061 die("out-event: Internal tree error");
1063 add_sched_out_event(out_events
, sched_out_state(switch_event
), timestamp
);
1065 in_events
= thread_atoms_search(&atom_root
, sched_in
, &cmp_pid
);
1067 thread_atoms_insert(sched_in
);
1068 in_events
= thread_atoms_search(&atom_root
, sched_in
, &cmp_pid
);
1070 die("in-event: Internal tree error");
1072 * Take came in we have not heard about yet,
1073 * add in an initial atom in runnable state:
1075 add_sched_out_event(in_events
, 'R', timestamp
);
1077 add_sched_in_event(in_events
, timestamp
);
1081 latency_runtime_event(struct trace_runtime_event
*runtime_event
,
1082 struct perf_session
*session
,
1083 struct event
*event __used
,
1086 struct thread
*this_thread __used
)
1088 struct thread
*thread
= perf_session__findnew(session
, runtime_event
->pid
);
1089 struct work_atoms
*atoms
= thread_atoms_search(&atom_root
, thread
, &cmp_pid
);
1091 BUG_ON(cpu
>= MAX_CPUS
|| cpu
< 0);
1093 thread_atoms_insert(thread
);
1094 atoms
= thread_atoms_search(&atom_root
, thread
, &cmp_pid
);
1096 die("in-event: Internal tree error");
1097 add_sched_out_event(atoms
, 'R', timestamp
);
1100 add_runtime_event(atoms
, runtime_event
->runtime
, timestamp
);
1104 latency_wakeup_event(struct trace_wakeup_event
*wakeup_event
,
1105 struct perf_session
*session
,
1106 struct event
*__event __used
,
1109 struct thread
*thread __used
)
1111 struct work_atoms
*atoms
;
1112 struct work_atom
*atom
;
1113 struct thread
*wakee
;
1115 /* Note for later, it may be interesting to observe the failing cases */
1116 if (!wakeup_event
->success
)
1119 wakee
= perf_session__findnew(session
, wakeup_event
->pid
);
1120 atoms
= thread_atoms_search(&atom_root
, wakee
, &cmp_pid
);
1122 thread_atoms_insert(wakee
);
1123 atoms
= thread_atoms_search(&atom_root
, wakee
, &cmp_pid
);
1125 die("wakeup-event: Internal tree error");
1126 add_sched_out_event(atoms
, 'S', timestamp
);
1129 BUG_ON(list_empty(&atoms
->work_list
));
1131 atom
= list_entry(atoms
->work_list
.prev
, struct work_atom
, list
);
1134 * You WILL be missing events if you've recorded only
1135 * one CPU, or are only looking at only one, so don't
1136 * make useless noise.
1138 if (profile_cpu
== -1 && atom
->state
!= THREAD_SLEEPING
)
1139 nr_state_machine_bugs
++;
1142 if (atom
->sched_out_time
> timestamp
) {
1143 nr_unordered_timestamps
++;
1147 atom
->state
= THREAD_WAIT_CPU
;
1148 atom
->wake_up_time
= timestamp
;
1152 latency_migrate_task_event(struct trace_migrate_task_event
*migrate_task_event
,
1153 struct perf_session
*session
,
1154 struct event
*__event __used
,
1157 struct thread
*thread __used
)
1159 struct work_atoms
*atoms
;
1160 struct work_atom
*atom
;
1161 struct thread
*migrant
;
1164 * Only need to worry about migration when profiling one CPU.
1166 if (profile_cpu
== -1)
1169 migrant
= perf_session__findnew(session
, migrate_task_event
->pid
);
1170 atoms
= thread_atoms_search(&atom_root
, migrant
, &cmp_pid
);
1172 thread_atoms_insert(migrant
);
1173 register_pid(migrant
->pid
, migrant
->comm
);
1174 atoms
= thread_atoms_search(&atom_root
, migrant
, &cmp_pid
);
1176 die("migration-event: Internal tree error");
1177 add_sched_out_event(atoms
, 'R', timestamp
);
1180 BUG_ON(list_empty(&atoms
->work_list
));
1182 atom
= list_entry(atoms
->work_list
.prev
, struct work_atom
, list
);
1183 atom
->sched_in_time
= atom
->sched_out_time
= atom
->wake_up_time
= timestamp
;
1187 if (atom
->sched_out_time
> timestamp
)
1188 nr_unordered_timestamps
++;
1191 static struct trace_sched_handler lat_ops
= {
1192 .wakeup_event
= latency_wakeup_event
,
1193 .switch_event
= latency_switch_event
,
1194 .runtime_event
= latency_runtime_event
,
1195 .fork_event
= latency_fork_event
,
1196 .migrate_task_event
= latency_migrate_task_event
,
1199 static void output_lat_thread(struct work_atoms
*work_list
)
1205 if (!work_list
->nb_atoms
)
1208 * Ignore idle threads:
1210 if (!strcmp(work_list
->thread
->comm
, "swapper"))
1213 all_runtime
+= work_list
->total_runtime
;
1214 all_count
+= work_list
->nb_atoms
;
1216 ret
= printf(" %s:%d ", work_list
->thread
->comm
, work_list
->thread
->pid
);
1218 for (i
= 0; i
< 24 - ret
; i
++)
1221 avg
= work_list
->total_lat
/ work_list
->nb_atoms
;
1223 printf("|%11.3f ms |%9llu | avg:%9.3f ms | max:%9.3f ms | max at: %9.6f s\n",
1224 (double)work_list
->total_runtime
/ 1e6
,
1225 work_list
->nb_atoms
, (double)avg
/ 1e6
,
1226 (double)work_list
->max_lat
/ 1e6
,
1227 (double)work_list
->max_lat_at
/ 1e9
);
1230 static int pid_cmp(struct work_atoms
*l
, struct work_atoms
*r
)
1232 if (l
->thread
->pid
< r
->thread
->pid
)
1234 if (l
->thread
->pid
> r
->thread
->pid
)
1240 static struct sort_dimension pid_sort_dimension
= {
1245 static int avg_cmp(struct work_atoms
*l
, struct work_atoms
*r
)
1255 avgl
= l
->total_lat
/ l
->nb_atoms
;
1256 avgr
= r
->total_lat
/ r
->nb_atoms
;
1266 static struct sort_dimension avg_sort_dimension
= {
1271 static int max_cmp(struct work_atoms
*l
, struct work_atoms
*r
)
1273 if (l
->max_lat
< r
->max_lat
)
1275 if (l
->max_lat
> r
->max_lat
)
1281 static struct sort_dimension max_sort_dimension
= {
1286 static int switch_cmp(struct work_atoms
*l
, struct work_atoms
*r
)
1288 if (l
->nb_atoms
< r
->nb_atoms
)
1290 if (l
->nb_atoms
> r
->nb_atoms
)
1296 static struct sort_dimension switch_sort_dimension
= {
1301 static int runtime_cmp(struct work_atoms
*l
, struct work_atoms
*r
)
1303 if (l
->total_runtime
< r
->total_runtime
)
1305 if (l
->total_runtime
> r
->total_runtime
)
1311 static struct sort_dimension runtime_sort_dimension
= {
1316 static struct sort_dimension
*available_sorts
[] = {
1317 &pid_sort_dimension
,
1318 &avg_sort_dimension
,
1319 &max_sort_dimension
,
1320 &switch_sort_dimension
,
1321 &runtime_sort_dimension
,
1324 #define NB_AVAILABLE_SORTS (int)(sizeof(available_sorts) / sizeof(struct sort_dimension *))
1326 static LIST_HEAD(sort_list
);
1328 static int sort_dimension__add(const char *tok
, struct list_head
*list
)
1332 for (i
= 0; i
< NB_AVAILABLE_SORTS
; i
++) {
1333 if (!strcmp(available_sorts
[i
]->name
, tok
)) {
1334 list_add_tail(&available_sorts
[i
]->list
, list
);
1343 static void setup_sorting(void);
1345 static void sort_lat(void)
1347 struct rb_node
*node
;
1350 struct work_atoms
*data
;
1351 node
= rb_first(&atom_root
);
1355 rb_erase(node
, &atom_root
);
1356 data
= rb_entry(node
, struct work_atoms
, node
);
1357 __thread_latency_insert(&sorted_atom_root
, data
, &sort_list
);
1361 static struct trace_sched_handler
*trace_handler
;
1364 process_sched_wakeup_event(void *data
, struct perf_session
*session
,
1365 struct event
*event
,
1367 u64 timestamp __used
,
1368 struct thread
*thread __used
)
1370 struct trace_wakeup_event wakeup_event
;
1372 FILL_COMMON_FIELDS(wakeup_event
, event
, data
);
1374 FILL_ARRAY(wakeup_event
, comm
, event
, data
);
1375 FILL_FIELD(wakeup_event
, pid
, event
, data
);
1376 FILL_FIELD(wakeup_event
, prio
, event
, data
);
1377 FILL_FIELD(wakeup_event
, success
, event
, data
);
1378 FILL_FIELD(wakeup_event
, cpu
, event
, data
);
1380 if (trace_handler
->wakeup_event
)
1381 trace_handler
->wakeup_event(&wakeup_event
, session
, event
,
1382 cpu
, timestamp
, thread
);
1386 * Track the current task - that way we can know whether there's any
1387 * weird events, such as a task being switched away that is not current.
1391 static u32 curr_pid
[MAX_CPUS
] = { [0 ... MAX_CPUS
-1] = -1 };
1393 static struct thread
*curr_thread
[MAX_CPUS
];
1395 static char next_shortname1
= 'A';
1396 static char next_shortname2
= '0';
1399 map_switch_event(struct trace_switch_event
*switch_event
,
1400 struct perf_session
*session
,
1401 struct event
*event __used
,
1404 struct thread
*thread __used
)
1406 struct thread
*sched_out
, *sched_in
;
1412 BUG_ON(this_cpu
>= MAX_CPUS
|| this_cpu
< 0);
1414 if (this_cpu
> max_cpu
)
1417 timestamp0
= cpu_last_switched
[this_cpu
];
1418 cpu_last_switched
[this_cpu
] = timestamp
;
1420 delta
= timestamp
- timestamp0
;
1425 die("hm, delta: %Ld < 0 ?\n", delta
);
1428 sched_out
= perf_session__findnew(session
, switch_event
->prev_pid
);
1429 sched_in
= perf_session__findnew(session
, switch_event
->next_pid
);
1431 curr_thread
[this_cpu
] = sched_in
;
1436 if (!sched_in
->shortname
[0]) {
1437 sched_in
->shortname
[0] = next_shortname1
;
1438 sched_in
->shortname
[1] = next_shortname2
;
1440 if (next_shortname1
< 'Z') {
1443 next_shortname1
='A';
1444 if (next_shortname2
< '9') {
1447 next_shortname2
='0';
1453 for (cpu
= 0; cpu
<= max_cpu
; cpu
++) {
1454 if (cpu
!= this_cpu
)
1459 if (curr_thread
[cpu
]) {
1460 if (curr_thread
[cpu
]->pid
)
1461 printf("%2s ", curr_thread
[cpu
]->shortname
);
1468 printf(" %12.6f secs ", (double)timestamp
/1e9
);
1469 if (new_shortname
) {
1470 printf("%s => %s:%d\n",
1471 sched_in
->shortname
, sched_in
->comm
, sched_in
->pid
);
1479 process_sched_switch_event(void *data
, struct perf_session
*session
,
1480 struct event
*event
,
1482 u64 timestamp __used
,
1483 struct thread
*thread __used
)
1485 struct trace_switch_event switch_event
;
1487 FILL_COMMON_FIELDS(switch_event
, event
, data
);
1489 FILL_ARRAY(switch_event
, prev_comm
, event
, data
);
1490 FILL_FIELD(switch_event
, prev_pid
, event
, data
);
1491 FILL_FIELD(switch_event
, prev_prio
, event
, data
);
1492 FILL_FIELD(switch_event
, prev_state
, event
, data
);
1493 FILL_ARRAY(switch_event
, next_comm
, event
, data
);
1494 FILL_FIELD(switch_event
, next_pid
, event
, data
);
1495 FILL_FIELD(switch_event
, next_prio
, event
, data
);
1497 if (curr_pid
[this_cpu
] != (u32
)-1) {
1499 * Are we trying to switch away a PID that is
1502 if (curr_pid
[this_cpu
] != switch_event
.prev_pid
)
1503 nr_context_switch_bugs
++;
1505 if (trace_handler
->switch_event
)
1506 trace_handler
->switch_event(&switch_event
, session
, event
,
1507 this_cpu
, timestamp
, thread
);
1509 curr_pid
[this_cpu
] = switch_event
.next_pid
;
1513 process_sched_runtime_event(void *data
, struct perf_session
*session
,
1514 struct event
*event
,
1516 u64 timestamp __used
,
1517 struct thread
*thread __used
)
1519 struct trace_runtime_event runtime_event
;
1521 FILL_ARRAY(runtime_event
, comm
, event
, data
);
1522 FILL_FIELD(runtime_event
, pid
, event
, data
);
1523 FILL_FIELD(runtime_event
, runtime
, event
, data
);
1524 FILL_FIELD(runtime_event
, vruntime
, event
, data
);
1526 if (trace_handler
->runtime_event
)
1527 trace_handler
->runtime_event(&runtime_event
, session
, event
, cpu
, timestamp
, thread
);
1531 process_sched_fork_event(void *data
,
1532 struct event
*event
,
1534 u64 timestamp __used
,
1535 struct thread
*thread __used
)
1537 struct trace_fork_event fork_event
;
1539 FILL_COMMON_FIELDS(fork_event
, event
, data
);
1541 FILL_ARRAY(fork_event
, parent_comm
, event
, data
);
1542 FILL_FIELD(fork_event
, parent_pid
, event
, data
);
1543 FILL_ARRAY(fork_event
, child_comm
, event
, data
);
1544 FILL_FIELD(fork_event
, child_pid
, event
, data
);
1546 if (trace_handler
->fork_event
)
1547 trace_handler
->fork_event(&fork_event
, event
,
1548 cpu
, timestamp
, thread
);
1552 process_sched_exit_event(struct event
*event
,
1554 u64 timestamp __used
,
1555 struct thread
*thread __used
)
1558 printf("sched_exit event %p\n", event
);
1562 process_sched_migrate_task_event(void *data
, struct perf_session
*session
,
1563 struct event
*event
,
1565 u64 timestamp __used
,
1566 struct thread
*thread __used
)
1568 struct trace_migrate_task_event migrate_task_event
;
1570 FILL_COMMON_FIELDS(migrate_task_event
, event
, data
);
1572 FILL_ARRAY(migrate_task_event
, comm
, event
, data
);
1573 FILL_FIELD(migrate_task_event
, pid
, event
, data
);
1574 FILL_FIELD(migrate_task_event
, prio
, event
, data
);
1575 FILL_FIELD(migrate_task_event
, cpu
, event
, data
);
1577 if (trace_handler
->migrate_task_event
)
1578 trace_handler
->migrate_task_event(&migrate_task_event
, session
,
1579 event
, cpu
, timestamp
, thread
);
1583 process_raw_event(event_t
*raw_event __used
, struct perf_session
*session
,
1584 void *data
, int cpu
, u64 timestamp
, struct thread
*thread
)
1586 struct event
*event
;
1590 type
= trace_parse_common_type(data
);
1591 event
= trace_find_event(type
);
1593 if (!strcmp(event
->name
, "sched_switch"))
1594 process_sched_switch_event(data
, session
, event
, cpu
, timestamp
, thread
);
1595 if (!strcmp(event
->name
, "sched_stat_runtime"))
1596 process_sched_runtime_event(data
, session
, event
, cpu
, timestamp
, thread
);
1597 if (!strcmp(event
->name
, "sched_wakeup"))
1598 process_sched_wakeup_event(data
, session
, event
, cpu
, timestamp
, thread
);
1599 if (!strcmp(event
->name
, "sched_wakeup_new"))
1600 process_sched_wakeup_event(data
, session
, event
, cpu
, timestamp
, thread
);
1601 if (!strcmp(event
->name
, "sched_process_fork"))
1602 process_sched_fork_event(data
, event
, cpu
, timestamp
, thread
);
1603 if (!strcmp(event
->name
, "sched_process_exit"))
1604 process_sched_exit_event(event
, cpu
, timestamp
, thread
);
1605 if (!strcmp(event
->name
, "sched_migrate_task"))
1606 process_sched_migrate_task_event(data
, session
, event
, cpu
, timestamp
, thread
);
1609 static int process_sample_event(event_t
*event
, struct perf_session
*session
)
1611 struct sample_data data
;
1612 struct thread
*thread
;
1614 if (!(session
->sample_type
& PERF_SAMPLE_RAW
))
1617 memset(&data
, 0, sizeof(data
));
1622 event__parse_sample(event
, session
->sample_type
, &data
);
1624 dump_printf("(IP, %d): %d/%d: %#Lx period: %Ld\n", event
->header
.misc
,
1625 data
.pid
, data
.tid
, data
.ip
, data
.period
);
1627 thread
= perf_session__findnew(session
, data
.pid
);
1628 if (thread
== NULL
) {
1629 pr_debug("problem processing %d event, skipping it.\n",
1630 event
->header
.type
);
1634 dump_printf(" ... thread: %s:%d\n", thread
->comm
, thread
->pid
);
1636 if (profile_cpu
!= -1 && profile_cpu
!= (int)data
.cpu
)
1639 process_raw_event(event
, session
, data
.raw_data
, data
.cpu
, data
.time
, thread
);
1644 static struct perf_event_ops event_ops
= {
1645 .sample
= process_sample_event
,
1646 .comm
= event__process_comm
,
1647 .lost
= event__process_lost
,
1648 .fork
= event__process_task
,
1649 .ordered_samples
= true,
1652 static int read_events(void)
1655 struct perf_session
*session
= perf_session__new(input_name
, O_RDONLY
, 0, false);
1656 if (session
== NULL
)
1659 if (perf_session__has_traces(session
, "record -R")) {
1660 err
= perf_session__process_events(session
, &event_ops
);
1661 nr_events
= session
->hists
.stats
.nr_events
[0];
1662 nr_lost_events
= session
->hists
.stats
.total_lost
;
1663 nr_lost_chunks
= session
->hists
.stats
.nr_events
[PERF_RECORD_LOST
];
1666 perf_session__delete(session
);
1670 static void print_bad_events(void)
1672 if (nr_unordered_timestamps
&& nr_timestamps
) {
1673 printf(" INFO: %.3f%% unordered timestamps (%ld out of %ld)\n",
1674 (double)nr_unordered_timestamps
/(double)nr_timestamps
*100.0,
1675 nr_unordered_timestamps
, nr_timestamps
);
1677 if (nr_lost_events
&& nr_events
) {
1678 printf(" INFO: %.3f%% lost events (%ld out of %ld, in %ld chunks)\n",
1679 (double)nr_lost_events
/(double)nr_events
*100.0,
1680 nr_lost_events
, nr_events
, nr_lost_chunks
);
1682 if (nr_state_machine_bugs
&& nr_timestamps
) {
1683 printf(" INFO: %.3f%% state machine bugs (%ld out of %ld)",
1684 (double)nr_state_machine_bugs
/(double)nr_timestamps
*100.0,
1685 nr_state_machine_bugs
, nr_timestamps
);
1687 printf(" (due to lost events?)");
1690 if (nr_context_switch_bugs
&& nr_timestamps
) {
1691 printf(" INFO: %.3f%% context switch bugs (%ld out of %ld)",
1692 (double)nr_context_switch_bugs
/(double)nr_timestamps
*100.0,
1693 nr_context_switch_bugs
, nr_timestamps
);
1695 printf(" (due to lost events?)");
1700 static void __cmd_lat(void)
1702 struct rb_node
*next
;
1708 printf("\n ---------------------------------------------------------------------------------------------------------------\n");
1709 printf(" Task | Runtime ms | Switches | Average delay ms | Maximum delay ms | Maximum delay at |\n");
1710 printf(" ---------------------------------------------------------------------------------------------------------------\n");
1712 next
= rb_first(&sorted_atom_root
);
1715 struct work_atoms
*work_list
;
1717 work_list
= rb_entry(next
, struct work_atoms
, node
);
1718 output_lat_thread(work_list
);
1719 next
= rb_next(next
);
1722 printf(" -----------------------------------------------------------------------------------------\n");
1723 printf(" TOTAL: |%11.3f ms |%9Ld |\n",
1724 (double)all_runtime
/1e6
, all_count
);
1726 printf(" ---------------------------------------------------\n");
1733 static struct trace_sched_handler map_ops
= {
1734 .wakeup_event
= NULL
,
1735 .switch_event
= map_switch_event
,
1736 .runtime_event
= NULL
,
1740 static void __cmd_map(void)
1742 max_cpu
= sysconf(_SC_NPROCESSORS_CONF
);
1749 static void __cmd_replay(void)
1753 calibrate_run_measurement_overhead();
1754 calibrate_sleep_measurement_overhead();
1756 test_calibrations();
1760 printf("nr_run_events: %ld\n", nr_run_events
);
1761 printf("nr_sleep_events: %ld\n", nr_sleep_events
);
1762 printf("nr_wakeup_events: %ld\n", nr_wakeup_events
);
1764 if (targetless_wakeups
)
1765 printf("target-less wakeups: %ld\n", targetless_wakeups
);
1766 if (multitarget_wakeups
)
1767 printf("multi-target wakeups: %ld\n", multitarget_wakeups
);
1768 if (nr_run_events_optimized
)
1769 printf("run atoms optimized: %ld\n",
1770 nr_run_events_optimized
);
1772 print_task_traces();
1773 add_cross_task_wakeups();
1776 printf("------------------------------------------------------------\n");
1777 for (i
= 0; i
< replay_repeat
; i
++)
1782 static const char * const sched_usage
[] = {
1783 "perf sched [<options>] {record|latency|map|replay|trace}",
1787 static const struct option sched_options
[] = {
1788 OPT_STRING('i', "input", &input_name
, "file",
1790 OPT_INCR('v', "verbose", &verbose
,
1791 "be more verbose (show symbol address, etc)"),
1792 OPT_BOOLEAN('D', "dump-raw-trace", &dump_trace
,
1793 "dump raw trace in ASCII"),
1797 static const char * const latency_usage
[] = {
1798 "perf sched latency [<options>]",
1802 static const struct option latency_options
[] = {
1803 OPT_STRING('s', "sort", &sort_order
, "key[,key2...]",
1804 "sort by key(s): runtime, switch, avg, max"),
1805 OPT_INCR('v', "verbose", &verbose
,
1806 "be more verbose (show symbol address, etc)"),
1807 OPT_INTEGER('C', "CPU", &profile_cpu
,
1808 "CPU to profile on"),
1809 OPT_BOOLEAN('D', "dump-raw-trace", &dump_trace
,
1810 "dump raw trace in ASCII"),
1814 static const char * const replay_usage
[] = {
1815 "perf sched replay [<options>]",
1819 static const struct option replay_options
[] = {
1820 OPT_UINTEGER('r', "repeat", &replay_repeat
,
1821 "repeat the workload replay N times (-1: infinite)"),
1822 OPT_INCR('v', "verbose", &verbose
,
1823 "be more verbose (show symbol address, etc)"),
1824 OPT_BOOLEAN('D', "dump-raw-trace", &dump_trace
,
1825 "dump raw trace in ASCII"),
1829 static void setup_sorting(void)
1831 char *tmp
, *tok
, *str
= strdup(sort_order
);
1833 for (tok
= strtok_r(str
, ", ", &tmp
);
1834 tok
; tok
= strtok_r(NULL
, ", ", &tmp
)) {
1835 if (sort_dimension__add(tok
, &sort_list
) < 0) {
1836 error("Unknown --sort key: `%s'", tok
);
1837 usage_with_options(latency_usage
, latency_options
);
1843 sort_dimension__add("pid", &cmp_pid
);
1846 static const char *record_args
[] = {
1853 "-e", "sched:sched_switch:r",
1854 "-e", "sched:sched_stat_wait:r",
1855 "-e", "sched:sched_stat_sleep:r",
1856 "-e", "sched:sched_stat_iowait:r",
1857 "-e", "sched:sched_stat_runtime:r",
1858 "-e", "sched:sched_process_exit:r",
1859 "-e", "sched:sched_process_fork:r",
1860 "-e", "sched:sched_wakeup:r",
1861 "-e", "sched:sched_migrate_task:r",
1864 static int __cmd_record(int argc
, const char **argv
)
1866 unsigned int rec_argc
, i
, j
;
1867 const char **rec_argv
;
1869 rec_argc
= ARRAY_SIZE(record_args
) + argc
- 1;
1870 rec_argv
= calloc(rec_argc
+ 1, sizeof(char *));
1872 for (i
= 0; i
< ARRAY_SIZE(record_args
); i
++)
1873 rec_argv
[i
] = strdup(record_args
[i
]);
1875 for (j
= 1; j
< (unsigned int)argc
; j
++, i
++)
1876 rec_argv
[i
] = argv
[j
];
1878 BUG_ON(i
!= rec_argc
);
1880 return cmd_record(i
, rec_argv
, NULL
);
1883 int cmd_sched(int argc
, const char **argv
, const char *prefix __used
)
1885 argc
= parse_options(argc
, argv
, sched_options
, sched_usage
,
1886 PARSE_OPT_STOP_AT_NON_OPTION
);
1888 usage_with_options(sched_usage
, sched_options
);
1891 * Aliased to 'perf trace' for now:
1893 if (!strcmp(argv
[0], "trace"))
1894 return cmd_trace(argc
, argv
, prefix
);
1897 if (!strncmp(argv
[0], "rec", 3)) {
1898 return __cmd_record(argc
, argv
);
1899 } else if (!strncmp(argv
[0], "lat", 3)) {
1900 trace_handler
= &lat_ops
;
1902 argc
= parse_options(argc
, argv
, latency_options
, latency_usage
, 0);
1904 usage_with_options(latency_usage
, latency_options
);
1908 } else if (!strcmp(argv
[0], "map")) {
1909 trace_handler
= &map_ops
;
1912 } else if (!strncmp(argv
[0], "rep", 3)) {
1913 trace_handler
= &replay_ops
;
1915 argc
= parse_options(argc
, argv
, replay_options
, replay_usage
, 0);
1917 usage_with_options(replay_usage
, replay_options
);
1921 usage_with_options(sched_usage
, sched_options
);