2 * Xen hypercall batching.
4 * Xen allows multiple hypercalls to be issued at once, using the
5 * multicall interface. This allows the cost of trapping into the
6 * hypervisor to be amortized over several calls.
8 * This file implements a simple interface for multicalls. There's a
9 * per-cpu buffer of outstanding multicalls. When you want to queue a
10 * multicall for issuing, you can allocate a multicall slot for the
11 * call and its arguments, along with storage for space which is
12 * pointed to by the arguments (for passing pointers to structures,
13 * etc). When the multicall is actually issued, all the space for the
14 * commands and allocated memory is freed for reuse.
16 * Multicalls are flushed whenever any of the buffers get full, or
17 * when explicitly requested. There's no way to get per-multicall
18 * return results back. It will BUG if any of the multicalls fail.
20 * Jeremy Fitzhardinge <jeremy@xensource.com>, XenSource Inc, 2007
22 #include <linux/percpu.h>
23 #include <linux/hardirq.h>
24 #include <linux/debugfs.h>
26 #include <asm/xen/hypercall.h>
28 #include "multicalls.h"
35 #define MC_ARGS (MC_BATCH * 16)
39 unsigned mcidx
, argidx
, cbidx
;
40 struct multicall_entry entries
[MC_BATCH
];
42 struct multicall_entry debug
[MC_BATCH
];
43 void *caller
[MC_BATCH
];
45 unsigned char args
[MC_ARGS
];
49 } callbacks
[MC_BATCH
];
52 static DEFINE_PER_CPU(struct mc_buffer
, mc_buffer
);
53 DEFINE_PER_CPU(unsigned long, xen_mc_irq_flags
);
55 void xen_mc_flush(void)
57 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
58 struct multicall_entry
*mc
;
63 BUG_ON(preemptible());
65 /* Disable interrupts in case someone comes in and queues
66 something in the middle */
67 local_irq_save(flags
);
69 trace_xen_mc_flush(b
->mcidx
, b
->argidx
, b
->cbidx
);
74 BUG_ON(b
->argidx
!= 0);
78 /* Singleton multicall - bypass multicall machinery
79 and just do the call directly. */
82 mc
->result
= privcmd_call(mc
->op
,
83 mc
->args
[0], mc
->args
[1], mc
->args
[2],
84 mc
->args
[3], mc
->args
[4]);
90 memcpy(b
->debug
, b
->entries
,
91 b
->mcidx
* sizeof(struct multicall_entry
));
94 if (HYPERVISOR_multicall(b
->entries
, b
->mcidx
) != 0)
96 for (i
= 0; i
< b
->mcidx
; i
++)
97 if (b
->entries
[i
].result
< 0)
102 printk(KERN_ERR
"%d multicall(s) failed: cpu %d\n",
103 ret
, smp_processor_id());
105 for (i
= 0; i
< b
->mcidx
; i
++) {
106 printk(KERN_DEBUG
" call %2d/%d: op=%lu arg=[%lx] result=%ld\t%pF\n",
110 b
->entries
[i
].result
,
120 for (i
= 0; i
< b
->cbidx
; i
++) {
121 struct callback
*cb
= &b
->callbacks
[i
];
127 local_irq_restore(flags
);
132 struct multicall_space
__xen_mc_entry(size_t args
)
134 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
135 struct multicall_space ret
;
136 unsigned argidx
= roundup(b
->argidx
, sizeof(u64
));
138 trace_xen_mc_entry_alloc(args
);
140 BUG_ON(preemptible());
141 BUG_ON(b
->argidx
>= MC_ARGS
);
143 if (unlikely(b
->mcidx
== MC_BATCH
||
144 (argidx
+ args
) >= MC_ARGS
)) {
145 trace_xen_mc_flush_reason((b
->mcidx
== MC_BATCH
) ?
146 XEN_MC_FL_BATCH
: XEN_MC_FL_ARGS
);
148 argidx
= roundup(b
->argidx
, sizeof(u64
));
151 ret
.mc
= &b
->entries
[b
->mcidx
];
153 b
->caller
[b
->mcidx
] = __builtin_return_address(0);
156 ret
.args
= &b
->args
[argidx
];
157 b
->argidx
= argidx
+ args
;
159 BUG_ON(b
->argidx
>= MC_ARGS
);
163 struct multicall_space
xen_mc_extend_args(unsigned long op
, size_t size
)
165 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
166 struct multicall_space ret
= { NULL
, NULL
};
168 BUG_ON(preemptible());
169 BUG_ON(b
->argidx
>= MC_ARGS
);
171 if (unlikely(b
->mcidx
== 0 ||
172 b
->entries
[b
->mcidx
- 1].op
!= op
)) {
173 trace_xen_mc_extend_args(op
, size
, XEN_MC_XE_BAD_OP
);
177 if (unlikely((b
->argidx
+ size
) >= MC_ARGS
)) {
178 trace_xen_mc_extend_args(op
, size
, XEN_MC_XE_NO_SPACE
);
182 ret
.mc
= &b
->entries
[b
->mcidx
- 1];
183 ret
.args
= &b
->args
[b
->argidx
];
186 BUG_ON(b
->argidx
>= MC_ARGS
);
188 trace_xen_mc_extend_args(op
, size
, XEN_MC_XE_OK
);
193 void xen_mc_callback(void (*fn
)(void *), void *data
)
195 struct mc_buffer
*b
= &__get_cpu_var(mc_buffer
);
198 if (b
->cbidx
== MC_BATCH
) {
199 trace_xen_mc_flush_reason(XEN_MC_FL_CALLBACK
);
203 trace_xen_mc_callback(fn
, data
);
205 cb
= &b
->callbacks
[b
->cbidx
++];