4 * Copyright IBM, Corp. 2011
7 * Stefan Hajnoczi <stefanha@linux.vnet.ibm.com>
9 * This work is licensed under the terms of the GNU LGPL, version 2 or later.
10 * See the COPYING.LIB file in the top-level directory.
14 #include "qemu/osdep.h"
15 #include "qemu/coroutine.h"
16 #include "qemu/coroutine_int.h"
17 #include "qemu/lockable.h"
20 * Check that qemu_in_coroutine() works
23 static void coroutine_fn
verify_in_coroutine(void *opaque
)
25 g_assert(qemu_in_coroutine());
28 static void test_in_coroutine(void)
32 g_assert(!qemu_in_coroutine());
34 coroutine
= qemu_coroutine_create(verify_in_coroutine
, NULL
);
35 qemu_coroutine_enter(coroutine
);
39 * Check that qemu_coroutine_self() works
42 static void coroutine_fn
verify_self(void *opaque
)
44 Coroutine
**p_co
= opaque
;
45 g_assert(qemu_coroutine_self() == *p_co
);
48 static void test_self(void)
52 coroutine
= qemu_coroutine_create(verify_self
, &coroutine
);
53 qemu_coroutine_enter(coroutine
);
57 * Check that qemu_coroutine_entered() works
60 static void coroutine_fn
verify_entered_step_2(void *opaque
)
62 Coroutine
*caller
= (Coroutine
*)opaque
;
64 g_assert(qemu_coroutine_entered(caller
));
65 g_assert(qemu_coroutine_entered(qemu_coroutine_self()));
66 qemu_coroutine_yield();
68 /* Once more to check it still works after yielding */
69 g_assert(qemu_coroutine_entered(caller
));
70 g_assert(qemu_coroutine_entered(qemu_coroutine_self()));
73 static void coroutine_fn
verify_entered_step_1(void *opaque
)
75 Coroutine
*self
= qemu_coroutine_self();
78 g_assert(qemu_coroutine_entered(self
));
80 coroutine
= qemu_coroutine_create(verify_entered_step_2
, self
);
81 g_assert(!qemu_coroutine_entered(coroutine
));
82 qemu_coroutine_enter(coroutine
);
83 g_assert(!qemu_coroutine_entered(coroutine
));
84 qemu_coroutine_enter(coroutine
);
87 static void test_entered(void)
91 coroutine
= qemu_coroutine_create(verify_entered_step_1
, NULL
);
92 g_assert(!qemu_coroutine_entered(coroutine
));
93 qemu_coroutine_enter(coroutine
);
97 * Check that coroutines may nest multiple levels
101 unsigned int n_enter
; /* num coroutines entered */
102 unsigned int n_return
; /* num coroutines returned */
103 unsigned int max
; /* maximum level of nesting */
106 static void coroutine_fn
nest(void *opaque
)
108 NestData
*nd
= opaque
;
112 if (nd
->n_enter
< nd
->max
) {
115 child
= qemu_coroutine_create(nest
, nd
);
116 qemu_coroutine_enter(child
);
122 static void test_nesting(void)
131 root
= qemu_coroutine_create(nest
, &nd
);
132 qemu_coroutine_enter(root
);
134 /* Must enter and return from max nesting level */
135 g_assert_cmpint(nd
.n_enter
, ==, nd
.max
);
136 g_assert_cmpint(nd
.n_return
, ==, nd
.max
);
140 * Check that yield/enter transfer control correctly
143 static void coroutine_fn
yield_5_times(void *opaque
)
148 for (i
= 0; i
< 5; i
++) {
149 qemu_coroutine_yield();
154 static void test_yield(void)
156 Coroutine
*coroutine
;
158 int i
= -1; /* one extra time to return from coroutine */
160 coroutine
= qemu_coroutine_create(yield_5_times
, &done
);
162 qemu_coroutine_enter(coroutine
);
165 g_assert_cmpint(i
, ==, 5); /* coroutine must yield 5 times */
168 static void coroutine_fn
c2_fn(void *opaque
)
170 qemu_coroutine_yield();
173 static void coroutine_fn
c1_fn(void *opaque
)
175 Coroutine
*c2
= opaque
;
176 qemu_coroutine_enter(c2
);
179 static void test_no_dangling_access(void)
185 c2
= qemu_coroutine_create(c2_fn
, NULL
);
186 c1
= qemu_coroutine_create(c1_fn
, c2
);
188 qemu_coroutine_enter(c1
);
190 /* c1 shouldn't be used any more now; make sure we segfault if it is */
192 memset(c1
, 0xff, sizeof(Coroutine
));
193 qemu_coroutine_enter(c2
);
195 /* Must restore the coroutine now to avoid corrupted pool */
202 static void coroutine_fn
mutex_fn(void *opaque
)
205 qemu_co_mutex_lock(m
);
208 qemu_coroutine_yield();
210 qemu_co_mutex_unlock(m
);
214 static void coroutine_fn
lockable_fn(void *opaque
)
216 QemuLockable
*x
= opaque
;
217 qemu_lockable_lock(x
);
220 qemu_coroutine_yield();
222 qemu_lockable_unlock(x
);
226 static void do_test_co_mutex(CoroutineEntry
*entry
, void *opaque
)
228 Coroutine
*c1
= qemu_coroutine_create(entry
, opaque
);
229 Coroutine
*c2
= qemu_coroutine_create(entry
, opaque
);
232 qemu_coroutine_enter(c1
);
234 qemu_coroutine_enter(c2
);
236 /* Unlock queues c2. It is then started automatically when c1 yields or
239 qemu_coroutine_enter(c1
);
240 g_assert_cmpint(done
, ==, 1);
243 qemu_coroutine_enter(c2
);
244 g_assert_cmpint(done
, ==, 2);
248 static void test_co_mutex(void)
252 qemu_co_mutex_init(&m
);
253 do_test_co_mutex(mutex_fn
, &m
);
256 static void test_co_mutex_lockable(void)
259 CoMutex
*null_pointer
= NULL
;
261 qemu_co_mutex_init(&m
);
262 do_test_co_mutex(lockable_fn
, QEMU_MAKE_LOCKABLE(&m
));
264 g_assert(QEMU_MAKE_LOCKABLE(null_pointer
) == NULL
);
268 * Check that creation, enter, and return work
271 static void coroutine_fn
set_and_exit(void *opaque
)
278 static void test_lifecycle(void)
280 Coroutine
*coroutine
;
283 /* Create, enter, and return from coroutine */
284 coroutine
= qemu_coroutine_create(set_and_exit
, &done
);
285 qemu_coroutine_enter(coroutine
);
286 g_assert(done
); /* expect done to be true (first time) */
288 /* Repeat to check that no state affects this test */
290 coroutine
= qemu_coroutine_create(set_and_exit
, &done
);
291 qemu_coroutine_enter(coroutine
);
292 g_assert(done
); /* expect done to be true (second time) */
296 #define RECORD_SIZE 10 /* Leave some room for expansion */
297 struct coroutine_position
{
301 static struct coroutine_position records
[RECORD_SIZE
];
302 static unsigned record_pos
;
304 static void record_push(int func
, int state
)
306 struct coroutine_position
*cp
= &records
[record_pos
++];
307 g_assert_cmpint(record_pos
, <, RECORD_SIZE
);
312 static void coroutine_fn
co_order_test(void *opaque
)
315 g_assert(qemu_in_coroutine());
316 qemu_coroutine_yield();
318 g_assert(qemu_in_coroutine());
321 static void do_order_test(void)
325 co
= qemu_coroutine_create(co_order_test
, NULL
);
327 qemu_coroutine_enter(co
);
329 g_assert(!qemu_in_coroutine());
330 qemu_coroutine_enter(co
);
332 g_assert(!qemu_in_coroutine());
335 static void test_order(void)
338 const struct coroutine_position expected_pos
[] = {
339 {1, 1,}, {2, 1}, {1, 2}, {2, 2}, {1, 3}
342 g_assert_cmpint(record_pos
, ==, 5);
343 for (i
= 0; i
< record_pos
; i
++) {
344 g_assert_cmpint(records
[i
].func
, ==, expected_pos
[i
].func
);
345 g_assert_cmpint(records
[i
].state
, ==, expected_pos
[i
].state
);
349 * Lifecycle benchmark
352 static void coroutine_fn
empty_coroutine(void *opaque
)
357 static void perf_lifecycle(void)
359 Coroutine
*coroutine
;
365 g_test_timer_start();
366 for (i
= 0; i
< max
; i
++) {
367 coroutine
= qemu_coroutine_create(empty_coroutine
, NULL
);
368 qemu_coroutine_enter(coroutine
);
370 duration
= g_test_timer_elapsed();
372 g_test_message("Lifecycle %u iterations: %f s", max
, duration
);
375 static void perf_nesting(void)
377 unsigned int i
, maxcycles
, maxnesting
;
384 g_test_timer_start();
385 for (i
= 0; i
< maxcycles
; i
++) {
391 root
= qemu_coroutine_create(nest
, &nd
);
392 qemu_coroutine_enter(root
);
394 duration
= g_test_timer_elapsed();
396 g_test_message("Nesting %u iterations of %u depth each: %f s",
397 maxcycles
, maxnesting
, duration
);
404 static void coroutine_fn
yield_loop(void *opaque
)
406 unsigned int *counter
= opaque
;
408 while ((*counter
) > 0) {
410 qemu_coroutine_yield();
414 static void perf_yield(void)
416 unsigned int i
, maxcycles
;
419 maxcycles
= 100000000;
421 Coroutine
*coroutine
= qemu_coroutine_create(yield_loop
, &i
);
423 g_test_timer_start();
425 qemu_coroutine_enter(coroutine
);
427 duration
= g_test_timer_elapsed();
429 g_test_message("Yield %u iterations: %f s", maxcycles
, duration
);
432 static __attribute__((noinline
)) void dummy(unsigned *i
)
437 static void perf_baseline(void)
439 unsigned int i
, maxcycles
;
442 maxcycles
= 100000000;
445 g_test_timer_start();
449 duration
= g_test_timer_elapsed();
451 g_test_message("Function call %u iterations: %f s", maxcycles
, duration
);
454 static __attribute__((noinline
)) void perf_cost_func(void *opaque
)
456 qemu_coroutine_yield();
459 static void perf_cost(void)
461 const unsigned long maxcycles
= 40000000;
467 g_test_timer_start();
468 while (i
++ < maxcycles
) {
469 co
= qemu_coroutine_create(perf_cost_func
, &i
);
470 qemu_coroutine_enter(co
);
471 qemu_coroutine_enter(co
);
473 duration
= g_test_timer_elapsed();
474 ops
= (long)(maxcycles
/ (duration
* 1000));
476 g_test_message("Run operation %lu iterations %f s, %luK operations/s, "
477 "%luns per coroutine",
480 (unsigned long)(1000000000.0 * duration
/ maxcycles
));
483 int main(int argc
, char **argv
)
485 g_test_init(&argc
, &argv
, NULL
);
487 /* This test assumes there is a freelist and marks freed coroutine memory
488 * with a sentinel value. If there is no freelist this would legitimately
491 if (CONFIG_COROUTINE_POOL
) {
492 g_test_add_func("/basic/no-dangling-access", test_no_dangling_access
);
495 g_test_add_func("/basic/lifecycle", test_lifecycle
);
496 g_test_add_func("/basic/yield", test_yield
);
497 g_test_add_func("/basic/nesting", test_nesting
);
498 g_test_add_func("/basic/self", test_self
);
499 g_test_add_func("/basic/entered", test_entered
);
500 g_test_add_func("/basic/in_coroutine", test_in_coroutine
);
501 g_test_add_func("/basic/order", test_order
);
502 g_test_add_func("/locking/co-mutex", test_co_mutex
);
503 g_test_add_func("/locking/co-mutex/lockable", test_co_mutex_lockable
);
505 g_test_add_func("/perf/lifecycle", perf_lifecycle
);
506 g_test_add_func("/perf/nesting", perf_nesting
);
507 g_test_add_func("/perf/yield", perf_yield
);
508 g_test_add_func("/perf/function-call", perf_baseline
);
509 g_test_add_func("/perf/cost", perf_cost
);