const: constify remaining dev_pm_ops
[linux-2.6.git] / drivers / s390 / char / vmlogrdr.c
blob7dfa5412d5a8bc9d5c0b6ddcacffc6fe6d958981
1 /*
2 * drivers/s390/char/vmlogrdr.c
3 * character device driver for reading z/VM system service records
6 * Copyright IBM Corp. 2004, 2009
7 * character device driver for reading z/VM system service records,
8 * Version 1.0
9 * Author(s): Xenia Tkatschow <xenia@us.ibm.com>
10 * Stefan Weinhuber <wein@de.ibm.com>
14 #define KMSG_COMPONENT "vmlogrdr"
15 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
17 #include <linux/module.h>
18 #include <linux/init.h>
19 #include <linux/errno.h>
20 #include <linux/types.h>
21 #include <linux/interrupt.h>
22 #include <linux/spinlock.h>
23 #include <asm/atomic.h>
24 #include <asm/uaccess.h>
25 #include <asm/cpcmd.h>
26 #include <asm/debug.h>
27 #include <asm/ebcdic.h>
28 #include <net/iucv/iucv.h>
29 #include <linux/kmod.h>
30 #include <linux/cdev.h>
31 #include <linux/device.h>
32 #include <linux/smp_lock.h>
33 #include <linux/string.h>
35 MODULE_AUTHOR
36 ("(C) 2004 IBM Corporation by Xenia Tkatschow (xenia@us.ibm.com)\n"
37 " Stefan Weinhuber (wein@de.ibm.com)");
38 MODULE_DESCRIPTION ("Character device driver for reading z/VM "
39 "system service records.");
40 MODULE_LICENSE("GPL");
44 * The size of the buffer for iucv data transfer is one page,
45 * but in addition to the data we read from iucv we also
46 * place an integer and some characters into that buffer,
47 * so the maximum size for record data is a little less then
48 * one page.
50 #define NET_BUFFER_SIZE (PAGE_SIZE - sizeof(int) - sizeof(FENCE))
53 * The elements that are concurrently accessed by bottom halves are
54 * connection_established, iucv_path_severed, local_interrupt_buffer
55 * and receive_ready. The first three can be protected by
56 * priv_lock. receive_ready is atomic, so it can be incremented and
57 * decremented without holding a lock.
58 * The variable dev_in_use needs to be protected by the lock, since
59 * it's a flag used by open to make sure that the device is opened only
60 * by one user at the same time.
62 struct vmlogrdr_priv_t {
63 char system_service[8];
64 char internal_name[8];
65 char recording_name[8];
66 struct iucv_path *path;
67 int connection_established;
68 int iucv_path_severed;
69 struct iucv_message local_interrupt_buffer;
70 atomic_t receive_ready;
71 int minor_num;
72 char * buffer;
73 char * current_position;
74 int remaining;
75 ulong residual_length;
76 int buffer_free;
77 int dev_in_use; /* 1: already opened, 0: not opened*/
78 spinlock_t priv_lock;
79 struct device *device;
80 struct device *class_device;
81 int autorecording;
82 int autopurge;
87 * File operation structure for vmlogrdr devices
89 static int vmlogrdr_open(struct inode *, struct file *);
90 static int vmlogrdr_release(struct inode *, struct file *);
91 static ssize_t vmlogrdr_read (struct file *filp, char __user *data,
92 size_t count, loff_t * ppos);
94 static const struct file_operations vmlogrdr_fops = {
95 .owner = THIS_MODULE,
96 .open = vmlogrdr_open,
97 .release = vmlogrdr_release,
98 .read = vmlogrdr_read,
102 static void vmlogrdr_iucv_path_complete(struct iucv_path *, u8 ipuser[16]);
103 static void vmlogrdr_iucv_path_severed(struct iucv_path *, u8 ipuser[16]);
104 static void vmlogrdr_iucv_message_pending(struct iucv_path *,
105 struct iucv_message *);
108 static struct iucv_handler vmlogrdr_iucv_handler = {
109 .path_complete = vmlogrdr_iucv_path_complete,
110 .path_severed = vmlogrdr_iucv_path_severed,
111 .message_pending = vmlogrdr_iucv_message_pending,
115 static DECLARE_WAIT_QUEUE_HEAD(conn_wait_queue);
116 static DECLARE_WAIT_QUEUE_HEAD(read_wait_queue);
119 * pointer to system service private structure
120 * minor number 0 --> logrec
121 * minor number 1 --> account
122 * minor number 2 --> symptom
125 static struct vmlogrdr_priv_t sys_ser[] = {
126 { .system_service = "*LOGREC ",
127 .internal_name = "logrec",
128 .recording_name = "EREP",
129 .minor_num = 0,
130 .buffer_free = 1,
131 .priv_lock = __SPIN_LOCK_UNLOCKED(sys_ser[0].priv_lock),
132 .autorecording = 1,
133 .autopurge = 1,
135 { .system_service = "*ACCOUNT",
136 .internal_name = "account",
137 .recording_name = "ACCOUNT",
138 .minor_num = 1,
139 .buffer_free = 1,
140 .priv_lock = __SPIN_LOCK_UNLOCKED(sys_ser[1].priv_lock),
141 .autorecording = 1,
142 .autopurge = 1,
144 { .system_service = "*SYMPTOM",
145 .internal_name = "symptom",
146 .recording_name = "SYMPTOM",
147 .minor_num = 2,
148 .buffer_free = 1,
149 .priv_lock = __SPIN_LOCK_UNLOCKED(sys_ser[2].priv_lock),
150 .autorecording = 1,
151 .autopurge = 1,
155 #define MAXMINOR (sizeof(sys_ser)/sizeof(struct vmlogrdr_priv_t))
157 static char FENCE[] = {"EOR"};
158 static int vmlogrdr_major = 0;
159 static struct cdev *vmlogrdr_cdev = NULL;
160 static int recording_class_AB;
163 static void vmlogrdr_iucv_path_complete(struct iucv_path *path, u8 ipuser[16])
165 struct vmlogrdr_priv_t * logptr = path->private;
167 spin_lock(&logptr->priv_lock);
168 logptr->connection_established = 1;
169 spin_unlock(&logptr->priv_lock);
170 wake_up(&conn_wait_queue);
174 static void vmlogrdr_iucv_path_severed(struct iucv_path *path, u8 ipuser[16])
176 struct vmlogrdr_priv_t * logptr = path->private;
177 u8 reason = (u8) ipuser[8];
179 pr_err("vmlogrdr: connection severed with reason %i\n", reason);
181 iucv_path_sever(path, NULL);
182 kfree(path);
183 logptr->path = NULL;
185 spin_lock(&logptr->priv_lock);
186 logptr->connection_established = 0;
187 logptr->iucv_path_severed = 1;
188 spin_unlock(&logptr->priv_lock);
190 wake_up(&conn_wait_queue);
191 /* just in case we're sleeping waiting for a record */
192 wake_up_interruptible(&read_wait_queue);
196 static void vmlogrdr_iucv_message_pending(struct iucv_path *path,
197 struct iucv_message *msg)
199 struct vmlogrdr_priv_t * logptr = path->private;
202 * This function is the bottom half so it should be quick.
203 * Copy the external interrupt data into our local eib and increment
204 * the usage count
206 spin_lock(&logptr->priv_lock);
207 memcpy(&logptr->local_interrupt_buffer, msg, sizeof(*msg));
208 atomic_inc(&logptr->receive_ready);
209 spin_unlock(&logptr->priv_lock);
210 wake_up_interruptible(&read_wait_queue);
214 static int vmlogrdr_get_recording_class_AB(void)
216 char cp_command[]="QUERY COMMAND RECORDING ";
217 char cp_response[80];
218 char *tail;
219 int len,i;
221 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL);
222 len = strnlen(cp_response,sizeof(cp_response));
223 // now the parsing
224 tail=strnchr(cp_response,len,'=');
225 if (!tail)
226 return 0;
227 tail++;
228 if (!strncmp("ANY",tail,3))
229 return 1;
230 if (!strncmp("NONE",tail,4))
231 return 0;
233 * expect comma separated list of classes here, if one of them
234 * is A or B return 1 otherwise 0
236 for (i=tail-cp_response; i<len; i++)
237 if ( cp_response[i]=='A' || cp_response[i]=='B' )
238 return 1;
239 return 0;
243 static int vmlogrdr_recording(struct vmlogrdr_priv_t * logptr,
244 int action, int purge)
247 char cp_command[80];
248 char cp_response[160];
249 char *onoff, *qid_string;
251 memset(cp_command, 0x00, sizeof(cp_command));
252 memset(cp_response, 0x00, sizeof(cp_response));
254 onoff = ((action == 1) ? "ON" : "OFF");
255 qid_string = ((recording_class_AB == 1) ? " QID * " : "");
258 * The recording commands needs to be called with option QID
259 * for guests that have previlege classes A or B.
260 * Purging has to be done as separate step, because recording
261 * can't be switched on as long as records are on the queue.
262 * Doing both at the same time doesn't work.
265 if (purge) {
266 snprintf(cp_command, sizeof(cp_command),
267 "RECORDING %s PURGE %s",
268 logptr->recording_name,
269 qid_string);
271 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL);
274 memset(cp_command, 0x00, sizeof(cp_command));
275 memset(cp_response, 0x00, sizeof(cp_response));
276 snprintf(cp_command, sizeof(cp_command), "RECORDING %s %s %s",
277 logptr->recording_name,
278 onoff,
279 qid_string);
281 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL);
282 /* The recording command will usually answer with 'Command complete'
283 * on success, but when the specific service was never connected
284 * before then there might be an additional informational message
285 * 'HCPCRC8072I Recording entry not found' before the
286 * 'Command complete'. So I use strstr rather then the strncmp.
288 if (strstr(cp_response,"Command complete"))
289 return 0;
290 else
291 return -EIO;
296 static int vmlogrdr_open (struct inode *inode, struct file *filp)
298 int dev_num = 0;
299 struct vmlogrdr_priv_t * logptr = NULL;
300 int connect_rc = 0;
301 int ret;
303 dev_num = iminor(inode);
304 if (dev_num > MAXMINOR)
305 return -ENODEV;
306 logptr = &sys_ser[dev_num];
309 * only allow for blocking reads to be open
311 if (filp->f_flags & O_NONBLOCK)
312 return -ENOSYS;
314 /* Besure this device hasn't already been opened */
315 spin_lock_bh(&logptr->priv_lock);
316 if (logptr->dev_in_use) {
317 spin_unlock_bh(&logptr->priv_lock);
318 return -EBUSY;
320 logptr->dev_in_use = 1;
321 logptr->connection_established = 0;
322 logptr->iucv_path_severed = 0;
323 atomic_set(&logptr->receive_ready, 0);
324 logptr->buffer_free = 1;
325 spin_unlock_bh(&logptr->priv_lock);
327 /* set the file options */
328 filp->private_data = logptr;
329 filp->f_op = &vmlogrdr_fops;
331 /* start recording for this service*/
332 if (logptr->autorecording) {
333 ret = vmlogrdr_recording(logptr,1,logptr->autopurge);
334 if (ret)
335 pr_warning("vmlogrdr: failed to start "
336 "recording automatically\n");
339 /* create connection to the system service */
340 logptr->path = iucv_path_alloc(10, 0, GFP_KERNEL);
341 if (!logptr->path)
342 goto out_dev;
343 connect_rc = iucv_path_connect(logptr->path, &vmlogrdr_iucv_handler,
344 logptr->system_service, NULL, NULL,
345 logptr);
346 if (connect_rc) {
347 pr_err("vmlogrdr: iucv connection to %s "
348 "failed with rc %i \n",
349 logptr->system_service, connect_rc);
350 goto out_path;
353 /* We've issued the connect and now we must wait for a
354 * ConnectionComplete or ConnectinSevered Interrupt
355 * before we can continue to process.
357 wait_event(conn_wait_queue, (logptr->connection_established)
358 || (logptr->iucv_path_severed));
359 if (logptr->iucv_path_severed)
360 goto out_record;
361 nonseekable_open(inode, filp);
362 return 0;
364 out_record:
365 if (logptr->autorecording)
366 vmlogrdr_recording(logptr,0,logptr->autopurge);
367 out_path:
368 kfree(logptr->path); /* kfree(NULL) is ok. */
369 logptr->path = NULL;
370 out_dev:
371 logptr->dev_in_use = 0;
372 return -EIO;
376 static int vmlogrdr_release (struct inode *inode, struct file *filp)
378 int ret;
380 struct vmlogrdr_priv_t * logptr = filp->private_data;
382 iucv_path_sever(logptr->path, NULL);
383 kfree(logptr->path);
384 logptr->path = NULL;
385 if (logptr->autorecording) {
386 ret = vmlogrdr_recording(logptr,0,logptr->autopurge);
387 if (ret)
388 pr_warning("vmlogrdr: failed to stop "
389 "recording automatically\n");
391 logptr->dev_in_use = 0;
393 return 0;
397 static int vmlogrdr_receive_data(struct vmlogrdr_priv_t *priv)
399 int rc, *temp;
400 /* we need to keep track of two data sizes here:
401 * The number of bytes we need to receive from iucv and
402 * the total number of bytes we actually write into the buffer.
404 int user_data_count, iucv_data_count;
405 char * buffer;
407 if (atomic_read(&priv->receive_ready)) {
408 spin_lock_bh(&priv->priv_lock);
409 if (priv->residual_length){
410 /* receive second half of a record */
411 iucv_data_count = priv->residual_length;
412 user_data_count = 0;
413 buffer = priv->buffer;
414 } else {
415 /* receive a new record:
416 * We need to return the total length of the record
417 * + size of FENCE in the first 4 bytes of the buffer.
419 iucv_data_count = priv->local_interrupt_buffer.length;
420 user_data_count = sizeof(int);
421 temp = (int*)priv->buffer;
422 *temp= iucv_data_count + sizeof(FENCE);
423 buffer = priv->buffer + sizeof(int);
426 * If the record is bigger than our buffer, we receive only
427 * a part of it. We can get the rest later.
429 if (iucv_data_count > NET_BUFFER_SIZE)
430 iucv_data_count = NET_BUFFER_SIZE;
431 rc = iucv_message_receive(priv->path,
432 &priv->local_interrupt_buffer,
433 0, buffer, iucv_data_count,
434 &priv->residual_length);
435 spin_unlock_bh(&priv->priv_lock);
436 /* An rc of 5 indicates that the record was bigger than
437 * the buffer, which is OK for us. A 9 indicates that the
438 * record was purged befor we could receive it.
440 if (rc == 5)
441 rc = 0;
442 if (rc == 9)
443 atomic_set(&priv->receive_ready, 0);
444 } else {
445 rc = 1;
447 if (!rc) {
448 priv->buffer_free = 0;
449 user_data_count += iucv_data_count;
450 priv->current_position = priv->buffer;
451 if (priv->residual_length == 0){
452 /* the whole record has been captured,
453 * now add the fence */
454 atomic_dec(&priv->receive_ready);
455 buffer = priv->buffer + user_data_count;
456 memcpy(buffer, FENCE, sizeof(FENCE));
457 user_data_count += sizeof(FENCE);
459 priv->remaining = user_data_count;
462 return rc;
466 static ssize_t vmlogrdr_read(struct file *filp, char __user *data,
467 size_t count, loff_t * ppos)
469 int rc;
470 struct vmlogrdr_priv_t * priv = filp->private_data;
472 while (priv->buffer_free) {
473 rc = vmlogrdr_receive_data(priv);
474 if (rc) {
475 rc = wait_event_interruptible(read_wait_queue,
476 atomic_read(&priv->receive_ready));
477 if (rc)
478 return rc;
481 /* copy only up to end of record */
482 if (count > priv->remaining)
483 count = priv->remaining;
485 if (copy_to_user(data, priv->current_position, count))
486 return -EFAULT;
488 *ppos += count;
489 priv->current_position += count;
490 priv->remaining -= count;
492 /* if all data has been transferred, set buffer free */
493 if (priv->remaining == 0)
494 priv->buffer_free = 1;
496 return count;
499 static ssize_t vmlogrdr_autopurge_store(struct device * dev,
500 struct device_attribute *attr,
501 const char * buf, size_t count)
503 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev);
504 ssize_t ret = count;
506 switch (buf[0]) {
507 case '0':
508 priv->autopurge=0;
509 break;
510 case '1':
511 priv->autopurge=1;
512 break;
513 default:
514 ret = -EINVAL;
516 return ret;
520 static ssize_t vmlogrdr_autopurge_show(struct device *dev,
521 struct device_attribute *attr,
522 char *buf)
524 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev);
525 return sprintf(buf, "%u\n", priv->autopurge);
529 static DEVICE_ATTR(autopurge, 0644, vmlogrdr_autopurge_show,
530 vmlogrdr_autopurge_store);
533 static ssize_t vmlogrdr_purge_store(struct device * dev,
534 struct device_attribute *attr,
535 const char * buf, size_t count)
538 char cp_command[80];
539 char cp_response[80];
540 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev);
542 if (buf[0] != '1')
543 return -EINVAL;
545 memset(cp_command, 0x00, sizeof(cp_command));
546 memset(cp_response, 0x00, sizeof(cp_response));
549 * The recording command needs to be called with option QID
550 * for guests that have previlege classes A or B.
551 * Other guests will not recognize the command and we have to
552 * issue the same command without the QID parameter.
555 if (recording_class_AB)
556 snprintf(cp_command, sizeof(cp_command),
557 "RECORDING %s PURGE QID * ",
558 priv->recording_name);
559 else
560 snprintf(cp_command, sizeof(cp_command),
561 "RECORDING %s PURGE ",
562 priv->recording_name);
564 cpcmd(cp_command, cp_response, sizeof(cp_response), NULL);
566 return count;
570 static DEVICE_ATTR(purge, 0200, NULL, vmlogrdr_purge_store);
573 static ssize_t vmlogrdr_autorecording_store(struct device *dev,
574 struct device_attribute *attr,
575 const char *buf, size_t count)
577 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev);
578 ssize_t ret = count;
580 switch (buf[0]) {
581 case '0':
582 priv->autorecording=0;
583 break;
584 case '1':
585 priv->autorecording=1;
586 break;
587 default:
588 ret = -EINVAL;
590 return ret;
594 static ssize_t vmlogrdr_autorecording_show(struct device *dev,
595 struct device_attribute *attr,
596 char *buf)
598 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev);
599 return sprintf(buf, "%u\n", priv->autorecording);
603 static DEVICE_ATTR(autorecording, 0644, vmlogrdr_autorecording_show,
604 vmlogrdr_autorecording_store);
607 static ssize_t vmlogrdr_recording_store(struct device * dev,
608 struct device_attribute *attr,
609 const char * buf, size_t count)
611 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev);
612 ssize_t ret;
614 switch (buf[0]) {
615 case '0':
616 ret = vmlogrdr_recording(priv,0,0);
617 break;
618 case '1':
619 ret = vmlogrdr_recording(priv,1,0);
620 break;
621 default:
622 ret = -EINVAL;
624 if (ret)
625 return ret;
626 else
627 return count;
632 static DEVICE_ATTR(recording, 0200, NULL, vmlogrdr_recording_store);
635 static ssize_t vmlogrdr_recording_status_show(struct device_driver *driver,
636 char *buf)
639 char cp_command[] = "QUERY RECORDING ";
640 int len;
642 cpcmd(cp_command, buf, 4096, NULL);
643 len = strlen(buf);
644 return len;
648 static DRIVER_ATTR(recording_status, 0444, vmlogrdr_recording_status_show,
649 NULL);
651 static struct attribute *vmlogrdr_attrs[] = {
652 &dev_attr_autopurge.attr,
653 &dev_attr_purge.attr,
654 &dev_attr_autorecording.attr,
655 &dev_attr_recording.attr,
656 NULL,
659 static int vmlogrdr_pm_prepare(struct device *dev)
661 int rc;
662 struct vmlogrdr_priv_t *priv = dev_get_drvdata(dev);
664 rc = 0;
665 if (priv) {
666 spin_lock_bh(&priv->priv_lock);
667 if (priv->dev_in_use)
668 rc = -EBUSY;
669 spin_unlock_bh(&priv->priv_lock);
671 if (rc)
672 pr_err("vmlogrdr: device %s is busy. Refuse to suspend.\n",
673 dev_name(dev));
674 return rc;
678 static const struct dev_pm_ops vmlogrdr_pm_ops = {
679 .prepare = vmlogrdr_pm_prepare,
682 static struct attribute_group vmlogrdr_attr_group = {
683 .attrs = vmlogrdr_attrs,
686 static struct class *vmlogrdr_class;
687 static struct device_driver vmlogrdr_driver = {
688 .name = "vmlogrdr",
689 .bus = &iucv_bus,
690 .pm = &vmlogrdr_pm_ops,
694 static int vmlogrdr_register_driver(void)
696 int ret;
698 /* Register with iucv driver */
699 ret = iucv_register(&vmlogrdr_iucv_handler, 1);
700 if (ret)
701 goto out;
703 ret = driver_register(&vmlogrdr_driver);
704 if (ret)
705 goto out_iucv;
707 ret = driver_create_file(&vmlogrdr_driver,
708 &driver_attr_recording_status);
709 if (ret)
710 goto out_driver;
712 vmlogrdr_class = class_create(THIS_MODULE, "vmlogrdr");
713 if (IS_ERR(vmlogrdr_class)) {
714 ret = PTR_ERR(vmlogrdr_class);
715 vmlogrdr_class = NULL;
716 goto out_attr;
718 return 0;
720 out_attr:
721 driver_remove_file(&vmlogrdr_driver, &driver_attr_recording_status);
722 out_driver:
723 driver_unregister(&vmlogrdr_driver);
724 out_iucv:
725 iucv_unregister(&vmlogrdr_iucv_handler, 1);
726 out:
727 return ret;
731 static void vmlogrdr_unregister_driver(void)
733 class_destroy(vmlogrdr_class);
734 vmlogrdr_class = NULL;
735 driver_remove_file(&vmlogrdr_driver, &driver_attr_recording_status);
736 driver_unregister(&vmlogrdr_driver);
737 iucv_unregister(&vmlogrdr_iucv_handler, 1);
741 static int vmlogrdr_register_device(struct vmlogrdr_priv_t *priv)
743 struct device *dev;
744 int ret;
746 dev = kzalloc(sizeof(struct device), GFP_KERNEL);
747 if (dev) {
748 dev_set_name(dev, priv->internal_name);
749 dev->bus = &iucv_bus;
750 dev->parent = iucv_root;
751 dev->driver = &vmlogrdr_driver;
752 dev_set_drvdata(dev, priv);
754 * The release function could be called after the
755 * module has been unloaded. It's _only_ task is to
756 * free the struct. Therefore, we specify kfree()
757 * directly here. (Probably a little bit obfuscating
758 * but legitime ...).
760 dev->release = (void (*)(struct device *))kfree;
761 } else
762 return -ENOMEM;
763 ret = device_register(dev);
764 if (ret) {
765 put_device(dev);
766 return ret;
769 ret = sysfs_create_group(&dev->kobj, &vmlogrdr_attr_group);
770 if (ret) {
771 device_unregister(dev);
772 return ret;
774 priv->class_device = device_create(vmlogrdr_class, dev,
775 MKDEV(vmlogrdr_major,
776 priv->minor_num),
777 priv, "%s", dev_name(dev));
778 if (IS_ERR(priv->class_device)) {
779 ret = PTR_ERR(priv->class_device);
780 priv->class_device=NULL;
781 sysfs_remove_group(&dev->kobj, &vmlogrdr_attr_group);
782 device_unregister(dev);
783 return ret;
785 priv->device = dev;
786 return 0;
790 static int vmlogrdr_unregister_device(struct vmlogrdr_priv_t *priv)
792 device_destroy(vmlogrdr_class, MKDEV(vmlogrdr_major, priv->minor_num));
793 if (priv->device != NULL) {
794 sysfs_remove_group(&priv->device->kobj, &vmlogrdr_attr_group);
795 device_unregister(priv->device);
796 priv->device=NULL;
798 return 0;
802 static int vmlogrdr_register_cdev(dev_t dev)
804 int rc = 0;
805 vmlogrdr_cdev = cdev_alloc();
806 if (!vmlogrdr_cdev) {
807 return -ENOMEM;
809 vmlogrdr_cdev->owner = THIS_MODULE;
810 vmlogrdr_cdev->ops = &vmlogrdr_fops;
811 vmlogrdr_cdev->dev = dev;
812 rc = cdev_add(vmlogrdr_cdev, vmlogrdr_cdev->dev, MAXMINOR);
813 if (!rc)
814 return 0;
816 // cleanup: cdev is not fully registered, no cdev_del here!
817 kobject_put(&vmlogrdr_cdev->kobj);
818 vmlogrdr_cdev=NULL;
819 return rc;
823 static void vmlogrdr_cleanup(void)
825 int i;
827 if (vmlogrdr_cdev) {
828 cdev_del(vmlogrdr_cdev);
829 vmlogrdr_cdev=NULL;
831 for (i=0; i < MAXMINOR; ++i ) {
832 vmlogrdr_unregister_device(&sys_ser[i]);
833 free_page((unsigned long)sys_ser[i].buffer);
835 vmlogrdr_unregister_driver();
836 if (vmlogrdr_major) {
837 unregister_chrdev_region(MKDEV(vmlogrdr_major, 0), MAXMINOR);
838 vmlogrdr_major=0;
843 static int __init vmlogrdr_init(void)
845 int rc;
846 int i;
847 dev_t dev;
849 if (! MACHINE_IS_VM) {
850 pr_err("not running under VM, driver not loaded.\n");
851 return -ENODEV;
854 recording_class_AB = vmlogrdr_get_recording_class_AB();
856 rc = alloc_chrdev_region(&dev, 0, MAXMINOR, "vmlogrdr");
857 if (rc)
858 return rc;
859 vmlogrdr_major = MAJOR(dev);
861 rc=vmlogrdr_register_driver();
862 if (rc)
863 goto cleanup;
865 for (i=0; i < MAXMINOR; ++i ) {
866 sys_ser[i].buffer = (char *) get_zeroed_page(GFP_KERNEL);
867 if (!sys_ser[i].buffer) {
868 rc = -ENOMEM;
869 break;
871 sys_ser[i].current_position = sys_ser[i].buffer;
872 rc=vmlogrdr_register_device(&sys_ser[i]);
873 if (rc)
874 break;
876 if (rc)
877 goto cleanup;
879 rc = vmlogrdr_register_cdev(dev);
880 if (rc)
881 goto cleanup;
882 return 0;
884 cleanup:
885 vmlogrdr_cleanup();
886 return rc;
890 static void __exit vmlogrdr_exit(void)
892 vmlogrdr_cleanup();
893 return;
897 module_init(vmlogrdr_init);
898 module_exit(vmlogrdr_exit);