4 * handles SCLP event types
5 * - Signal Quiesce - system power down
6 * - ASCII Console Data - VT220 read and write
8 * Copyright IBM, Corp. 2012
11 * Heinz Graalfs <graalfs@de.ibm.com>
13 * This work is licensed under the terms of the GNU GPL, version 2 or (at your
14 * option) any later version. See the COPYING file in the top-level directory.
18 #include "qemu/osdep.h"
19 #include "qapi/error.h"
20 #include "qemu/module.h"
22 #include "hw/s390x/sclp.h"
23 #include "migration/vmstate.h"
24 #include "hw/s390x/event-facility.h"
26 typedef struct SCLPEventsBus
{
30 /* we need to save 32 bit chunks for compatibility */
31 #ifdef HOST_WORDS_BIGENDIAN
32 #define RECV_MASK_LOWER 1
33 #define RECV_MASK_UPPER 0
34 #else /* little endian host */
35 #define RECV_MASK_LOWER 0
36 #define RECV_MASK_UPPER 1
39 struct SCLPEventFacility
{
40 SysBusDevice parent_obj
;
42 SCLPEvent quiesce
, cpu_hotplug
;
43 /* guest's receive mask */
45 uint32_t receive_mask_pieces
[2];
46 sccb_mask_t receive_mask
;
49 * when false, we keep the same broken, backwards compatible behaviour as
50 * before, allowing only masks of size exactly 4; when true, we implement
51 * the architecture correctly, allowing all valid mask sizes. Needed for
52 * migration toward older versions.
54 bool allow_all_mask_sizes
;
55 /* length of the receive mask */
59 /* return true if any child has event pending set */
60 static bool event_pending(SCLPEventFacility
*ef
)
64 SCLPEventClass
*event_class
;
66 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
67 DeviceState
*qdev
= kid
->child
;
68 event
= DO_UPCAST(SCLPEvent
, qdev
, qdev
);
69 event_class
= SCLP_EVENT_GET_CLASS(event
);
70 if (event
->event_pending
&&
71 event_class
->get_send_mask() & ef
->receive_mask
) {
78 static sccb_mask_t
get_host_send_mask(SCLPEventFacility
*ef
)
82 SCLPEventClass
*child
;
86 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
87 DeviceState
*qdev
= kid
->child
;
88 child
= SCLP_EVENT_GET_CLASS((SCLPEvent
*) qdev
);
89 mask
|= child
->get_send_mask();
94 static sccb_mask_t
get_host_receive_mask(SCLPEventFacility
*ef
)
98 SCLPEventClass
*child
;
102 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
103 DeviceState
*qdev
= kid
->child
;
104 child
= SCLP_EVENT_GET_CLASS((SCLPEvent
*) qdev
);
105 mask
|= child
->get_receive_mask();
110 static uint16_t write_event_length_check(SCCB
*sccb
)
114 EventBufferHeader
*event
;
115 WriteEventData
*wed
= (WriteEventData
*) sccb
;
117 event
= (EventBufferHeader
*) &wed
->ebh
;
118 for (slen
= sccb_data_len(sccb
); slen
> 0; slen
-= elen
) {
119 elen
= be16_to_cpu(event
->length
);
120 if (elen
< sizeof(*event
) || elen
> slen
) {
121 return SCLP_RC_EVENT_BUFFER_SYNTAX_ERROR
;
123 event
= (void *) event
+ elen
;
126 return SCLP_RC_INCONSISTENT_LENGTHS
;
128 return SCLP_RC_NORMAL_COMPLETION
;
131 static uint16_t handle_write_event_buf(SCLPEventFacility
*ef
,
132 EventBufferHeader
*event_buf
, SCCB
*sccb
)
139 rc
= SCLP_RC_INVALID_FUNCTION
;
141 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
142 DeviceState
*qdev
= kid
->child
;
143 event
= (SCLPEvent
*) qdev
;
144 ec
= SCLP_EVENT_GET_CLASS(event
);
146 if (ec
->write_event_data
&&
147 ec
->can_handle_event(event_buf
->type
)) {
148 rc
= ec
->write_event_data(event
, event_buf
);
155 static uint16_t handle_sccb_write_events(SCLPEventFacility
*ef
, SCCB
*sccb
)
160 EventBufferHeader
*event_buf
;
161 WriteEventData
*wed
= (WriteEventData
*) sccb
;
163 event_buf
= &wed
->ebh
;
164 rc
= SCLP_RC_NORMAL_COMPLETION
;
166 /* loop over all contained event buffers */
167 for (slen
= sccb_data_len(sccb
); slen
> 0; slen
-= elen
) {
168 elen
= be16_to_cpu(event_buf
->length
);
170 /* in case of a previous error mark all trailing buffers
172 if (rc
!= SCLP_RC_NORMAL_COMPLETION
) {
173 event_buf
->flags
&= ~(SCLP_EVENT_BUFFER_ACCEPTED
);
175 rc
= handle_write_event_buf(ef
, event_buf
, sccb
);
177 event_buf
= (void *) event_buf
+ elen
;
182 static void write_event_data(SCLPEventFacility
*ef
, SCCB
*sccb
)
184 if (sccb
->h
.function_code
!= SCLP_FC_NORMAL_WRITE
) {
185 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INVALID_FUNCTION
);
188 if (be16_to_cpu(sccb
->h
.length
) < 8) {
189 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INSUFFICIENT_SCCB_LENGTH
);
192 /* first do a sanity check of the write events */
193 sccb
->h
.response_code
= cpu_to_be16(write_event_length_check(sccb
));
195 /* if no early error, then execute */
196 if (sccb
->h
.response_code
== be16_to_cpu(SCLP_RC_NORMAL_COMPLETION
)) {
197 sccb
->h
.response_code
=
198 cpu_to_be16(handle_sccb_write_events(ef
, sccb
));
202 static uint16_t handle_sccb_read_events(SCLPEventFacility
*ef
, SCCB
*sccb
,
211 EventBufferHeader
*event_buf
;
212 ReadEventData
*red
= (ReadEventData
*) sccb
;
214 event_buf
= &red
->ebh
;
215 event_buf
->length
= 0;
216 slen
= sccb_data_len(sccb
);
218 rc
= SCLP_RC_NO_EVENT_BUFFERS_STORED
;
220 QTAILQ_FOREACH(kid
, &ef
->sbus
.qbus
.children
, sibling
) {
221 DeviceState
*qdev
= kid
->child
;
222 event
= (SCLPEvent
*) qdev
;
223 ec
= SCLP_EVENT_GET_CLASS(event
);
225 if (mask
& ec
->get_send_mask()) {
226 if (ec
->read_event_data(event
, event_buf
, &slen
)) {
227 elen
= be16_to_cpu(event_buf
->length
);
228 event_buf
= (EventBufferHeader
*) ((char *)event_buf
+ elen
);
229 rc
= SCLP_RC_NORMAL_COMPLETION
;
234 if (sccb
->h
.control_mask
[2] & SCLP_VARIABLE_LENGTH_RESPONSE
) {
235 /* architecture suggests to reset variable-length-response bit */
236 sccb
->h
.control_mask
[2] &= ~SCLP_VARIABLE_LENGTH_RESPONSE
;
237 /* with a new length value */
238 sccb
->h
.length
= cpu_to_be16(SCCB_SIZE
- slen
);
243 /* copy up to src_len bytes and fill the rest of dst with zeroes */
244 static void copy_mask(uint8_t *dst
, uint8_t *src
, uint16_t dst_len
,
249 for (i
= 0; i
< dst_len
; i
++) {
250 dst
[i
] = i
< src_len
? src
[i
] : 0;
254 static void read_event_data(SCLPEventFacility
*ef
, SCCB
*sccb
)
256 sccb_mask_t sclp_active_selection_mask
;
257 sccb_mask_t sclp_cp_receive_mask
;
259 ReadEventData
*red
= (ReadEventData
*) sccb
;
261 if (be16_to_cpu(sccb
->h
.length
) != SCCB_SIZE
) {
262 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INSUFFICIENT_SCCB_LENGTH
);
266 switch (sccb
->h
.function_code
) {
267 case SCLP_UNCONDITIONAL_READ
:
268 sccb
->h
.response_code
= cpu_to_be16(
269 handle_sccb_read_events(ef
, sccb
, ef
->receive_mask
));
271 case SCLP_SELECTIVE_READ
:
272 /* get active selection mask */
273 sclp_cp_receive_mask
= ef
->receive_mask
;
275 copy_mask((uint8_t *)&sclp_active_selection_mask
, (uint8_t *)&red
->mask
,
276 sizeof(sclp_active_selection_mask
), ef
->mask_length
);
277 sclp_active_selection_mask
= be64_to_cpu(sclp_active_selection_mask
);
278 if (!sclp_cp_receive_mask
||
279 (sclp_active_selection_mask
& ~sclp_cp_receive_mask
)) {
280 sccb
->h
.response_code
=
281 cpu_to_be16(SCLP_RC_INVALID_SELECTION_MASK
);
283 sccb
->h
.response_code
= cpu_to_be16(
284 handle_sccb_read_events(ef
, sccb
, sclp_active_selection_mask
));
288 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INVALID_FUNCTION
);
292 static void write_event_mask(SCLPEventFacility
*ef
, SCCB
*sccb
)
294 WriteEventMask
*we_mask
= (WriteEventMask
*) sccb
;
295 uint16_t mask_length
= be16_to_cpu(we_mask
->mask_length
);
296 sccb_mask_t tmp_mask
;
298 if (!mask_length
|| (mask_length
> SCLP_EVENT_MASK_LEN_MAX
) ||
299 ((mask_length
!= 4) && !ef
->allow_all_mask_sizes
)) {
300 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_INVALID_MASK_LENGTH
);
305 * Note: We currently only support masks up to 8 byte length;
306 * the remainder is filled up with zeroes. Older Linux
307 * kernels use a 4 byte mask length, newer ones can use both
308 * 8 or 4 depending on what is available on the host.
311 /* keep track of the guest's capability masks */
312 copy_mask((uint8_t *)&tmp_mask
, WEM_CP_RECEIVE_MASK(we_mask
, mask_length
),
313 sizeof(tmp_mask
), mask_length
);
314 ef
->receive_mask
= be64_to_cpu(tmp_mask
);
316 /* return the SCLP's capability masks to the guest */
317 tmp_mask
= cpu_to_be64(get_host_receive_mask(ef
));
318 copy_mask(WEM_RECEIVE_MASK(we_mask
, mask_length
), (uint8_t *)&tmp_mask
,
319 mask_length
, sizeof(tmp_mask
));
320 tmp_mask
= cpu_to_be64(get_host_send_mask(ef
));
321 copy_mask(WEM_SEND_MASK(we_mask
, mask_length
), (uint8_t *)&tmp_mask
,
322 mask_length
, sizeof(tmp_mask
));
324 sccb
->h
.response_code
= cpu_to_be16(SCLP_RC_NORMAL_COMPLETION
);
325 ef
->mask_length
= mask_length
;
328 /* qemu object creation and initialization functions */
330 #define TYPE_SCLP_EVENTS_BUS "s390-sclp-events-bus"
332 static const TypeInfo sclp_events_bus_info
= {
333 .name
= TYPE_SCLP_EVENTS_BUS
,
337 static void command_handler(SCLPEventFacility
*ef
, SCCB
*sccb
, uint64_t code
)
339 switch (code
& SCLP_CMD_CODE_MASK
) {
340 case SCLP_CMD_READ_EVENT_DATA
:
341 read_event_data(ef
, sccb
);
343 case SCLP_CMD_WRITE_EVENT_DATA
:
344 write_event_data(ef
, sccb
);
346 case SCLP_CMD_WRITE_EVENT_MASK
:
347 write_event_mask(ef
, sccb
);
352 static bool vmstate_event_facility_mask64_needed(void *opaque
)
354 SCLPEventFacility
*ef
= opaque
;
356 return (ef
->receive_mask
& 0xFFFFFFFF) != 0;
359 static bool vmstate_event_facility_mask_length_needed(void *opaque
)
361 SCLPEventFacility
*ef
= opaque
;
363 return ef
->allow_all_mask_sizes
;
366 static const VMStateDescription vmstate_event_facility_mask64
= {
367 .name
= "vmstate-event-facility/mask64",
369 .minimum_version_id
= 0,
370 .needed
= vmstate_event_facility_mask64_needed
,
371 .fields
= (VMStateField
[]) {
372 VMSTATE_UINT32(receive_mask_pieces
[RECV_MASK_LOWER
], SCLPEventFacility
),
373 VMSTATE_END_OF_LIST()
377 static const VMStateDescription vmstate_event_facility_mask_length
= {
378 .name
= "vmstate-event-facility/mask_length",
380 .minimum_version_id
= 0,
381 .needed
= vmstate_event_facility_mask_length_needed
,
382 .fields
= (VMStateField
[]) {
383 VMSTATE_UINT16(mask_length
, SCLPEventFacility
),
384 VMSTATE_END_OF_LIST()
388 static const VMStateDescription vmstate_event_facility
= {
389 .name
= "vmstate-event-facility",
391 .minimum_version_id
= 0,
392 .fields
= (VMStateField
[]) {
393 VMSTATE_UINT32(receive_mask_pieces
[RECV_MASK_UPPER
], SCLPEventFacility
),
394 VMSTATE_END_OF_LIST()
396 .subsections
= (const VMStateDescription
* []) {
397 &vmstate_event_facility_mask64
,
398 &vmstate_event_facility_mask_length
,
403 static void sclp_event_set_allow_all_mask_sizes(Object
*obj
, bool value
,
406 SCLPEventFacility
*ef
= (SCLPEventFacility
*)obj
;
408 ef
->allow_all_mask_sizes
= value
;
411 static bool sclp_event_get_allow_all_mask_sizes(Object
*obj
, Error
**errp
)
413 SCLPEventFacility
*ef
= (SCLPEventFacility
*)obj
;
415 return ef
->allow_all_mask_sizes
;
418 static void init_event_facility(Object
*obj
)
420 SCLPEventFacility
*event_facility
= EVENT_FACILITY(obj
);
421 DeviceState
*sdev
= DEVICE(obj
);
423 event_facility
->mask_length
= 4;
424 event_facility
->allow_all_mask_sizes
= true;
425 object_property_add_bool(obj
, "allow_all_mask_sizes",
426 sclp_event_get_allow_all_mask_sizes
,
427 sclp_event_set_allow_all_mask_sizes
);
429 /* Spawn a new bus for SCLP events */
430 qbus_create_inplace(&event_facility
->sbus
, sizeof(event_facility
->sbus
),
431 TYPE_SCLP_EVENTS_BUS
, sdev
, NULL
);
433 object_initialize_child(obj
, TYPE_SCLP_QUIESCE
,
434 &event_facility
->quiesce
,
437 object_initialize_child(obj
, TYPE_SCLP_CPU_HOTPLUG
,
438 &event_facility
->cpu_hotplug
,
439 TYPE_SCLP_CPU_HOTPLUG
);
442 static void realize_event_facility(DeviceState
*dev
, Error
**errp
)
444 SCLPEventFacility
*event_facility
= EVENT_FACILITY(dev
);
446 if (!qdev_realize(DEVICE(&event_facility
->quiesce
),
447 BUS(&event_facility
->sbus
), errp
)) {
450 if (!qdev_realize(DEVICE(&event_facility
->cpu_hotplug
),
451 BUS(&event_facility
->sbus
), errp
)) {
452 qdev_unrealize(DEVICE(&event_facility
->quiesce
));
457 static void reset_event_facility(DeviceState
*dev
)
459 SCLPEventFacility
*sdev
= EVENT_FACILITY(dev
);
461 sdev
->receive_mask
= 0;
464 static void init_event_facility_class(ObjectClass
*klass
, void *data
)
466 SysBusDeviceClass
*sbdc
= SYS_BUS_DEVICE_CLASS(klass
);
467 DeviceClass
*dc
= DEVICE_CLASS(sbdc
);
468 SCLPEventFacilityClass
*k
= EVENT_FACILITY_CLASS(dc
);
470 dc
->realize
= realize_event_facility
;
471 dc
->reset
= reset_event_facility
;
472 dc
->vmsd
= &vmstate_event_facility
;
473 set_bit(DEVICE_CATEGORY_MISC
, dc
->categories
);
474 k
->command_handler
= command_handler
;
475 k
->event_pending
= event_pending
;
478 static const TypeInfo sclp_event_facility_info
= {
479 .name
= TYPE_SCLP_EVENT_FACILITY
,
480 .parent
= TYPE_SYS_BUS_DEVICE
,
481 .instance_init
= init_event_facility
,
482 .instance_size
= sizeof(SCLPEventFacility
),
483 .class_init
= init_event_facility_class
,
484 .class_size
= sizeof(SCLPEventFacilityClass
),
487 static void event_realize(DeviceState
*qdev
, Error
**errp
)
489 SCLPEvent
*event
= SCLP_EVENT(qdev
);
490 SCLPEventClass
*child
= SCLP_EVENT_GET_CLASS(event
);
493 int rc
= child
->init(event
);
495 error_setg(errp
, "SCLP event initialization failed.");
501 static void event_class_init(ObjectClass
*klass
, void *data
)
503 DeviceClass
*dc
= DEVICE_CLASS(klass
);
505 dc
->bus_type
= TYPE_SCLP_EVENTS_BUS
;
506 dc
->realize
= event_realize
;
509 static const TypeInfo sclp_event_type_info
= {
510 .name
= TYPE_SCLP_EVENT
,
511 .parent
= TYPE_DEVICE
,
512 .instance_size
= sizeof(SCLPEvent
),
513 .class_init
= event_class_init
,
514 .class_size
= sizeof(SCLPEventClass
),
518 static void register_types(void)
520 type_register_static(&sclp_events_bus_info
);
521 type_register_static(&sclp_event_facility_info
);
522 type_register_static(&sclp_event_type_info
);
525 type_init(register_types
)
527 BusState
*sclp_get_event_facility_bus(void)
532 busobj
= object_resolve_path_type("", TYPE_SCLP_EVENTS_BUS
, NULL
);
533 sbus
= OBJECT_CHECK(SCLPEventsBus
, busobj
, TYPE_SCLP_EVENTS_BUS
);