4 * @remark Copyright 2002 OProfile authors
5 * @remark Read the file COPYING
7 * @author John Levon <levon@movementarian.org>
9 * This is the global event buffer that the user-space
10 * daemon reads from. The event buffer is an untyped array
11 * of unsigned longs. Entries are prefixed by the
12 * escape value ESCAPE_CODE followed by an identifying code.
15 #include <linux/vmalloc.h>
16 #include <linux/oprofile.h>
17 #include <linux/sched.h>
18 #include <linux/capability.h>
19 #include <linux/dcookies.h>
21 #include <asm/uaccess.h>
24 #include "event_buffer.h"
25 #include "oprofile_stats.h"
27 DEFINE_MUTEX(buffer_mutex
);
29 static unsigned long buffer_opened
;
30 static DECLARE_WAIT_QUEUE_HEAD(buffer_wait
);
31 static unsigned long * event_buffer
;
32 static unsigned long buffer_size
;
33 static unsigned long buffer_watershed
;
34 static size_t buffer_pos
;
35 /* atomic_t because wait_event checks it outside of buffer_mutex */
36 static atomic_t buffer_ready
= ATOMIC_INIT(0);
38 /* Add an entry to the event buffer. When we
39 * get near to the end we wake up the process
40 * sleeping on the read() of the file.
42 void add_event_entry(unsigned long value
)
44 if (buffer_pos
== buffer_size
) {
45 atomic_inc(&oprofile_stats
.event_lost_overflow
);
49 event_buffer
[buffer_pos
] = value
;
50 if (++buffer_pos
== buffer_size
- buffer_watershed
) {
51 atomic_set(&buffer_ready
, 1);
52 wake_up(&buffer_wait
);
57 /* Wake up the waiting process if any. This happens
58 * on "echo 0 >/dev/oprofile/enable" so the daemon
59 * processes the data remaining in the event buffer.
61 void wake_up_buffer_waiter(void)
63 mutex_lock(&buffer_mutex
);
64 atomic_set(&buffer_ready
, 1);
65 wake_up(&buffer_wait
);
66 mutex_unlock(&buffer_mutex
);
70 int alloc_event_buffer(void)
75 spin_lock_irqsave(&oprofilefs_lock
, flags
);
76 buffer_size
= fs_buffer_size
;
77 buffer_watershed
= fs_buffer_watershed
;
78 spin_unlock_irqrestore(&oprofilefs_lock
, flags
);
80 if (buffer_watershed
>= buffer_size
)
83 event_buffer
= vmalloc(sizeof(unsigned long) * buffer_size
);
93 void free_event_buffer(void)
99 static int event_buffer_open(struct inode
* inode
, struct file
* file
)
103 if (!capable(CAP_SYS_ADMIN
))
106 if (test_and_set_bit(0, &buffer_opened
))
109 /* Register as a user of dcookies
110 * to ensure they persist for the lifetime of
111 * the open event file
114 file
->private_data
= dcookie_register();
115 if (!file
->private_data
)
118 if ((err
= oprofile_setup()))
121 /* NB: the actual start happens from userspace
122 * echo 1 >/dev/oprofile/enable
128 dcookie_unregister(file
->private_data
);
130 clear_bit(0, &buffer_opened
);
135 static int event_buffer_release(struct inode
* inode
, struct file
* file
)
139 dcookie_unregister(file
->private_data
);
141 atomic_set(&buffer_ready
, 0);
142 clear_bit(0, &buffer_opened
);
147 static ssize_t
event_buffer_read(struct file
* file
, char __user
* buf
,
148 size_t count
, loff_t
* offset
)
150 int retval
= -EINVAL
;
151 size_t const max
= buffer_size
* sizeof(unsigned long);
153 /* handling partial reads is more trouble than it's worth */
154 if (count
!= max
|| *offset
)
157 wait_event_interruptible(buffer_wait
, atomic_read(&buffer_ready
));
159 if (signal_pending(current
))
162 /* can't currently happen */
163 if (!atomic_read(&buffer_ready
))
166 mutex_lock(&buffer_mutex
);
168 atomic_set(&buffer_ready
, 0);
172 count
= buffer_pos
* sizeof(unsigned long);
174 if (copy_to_user(buf
, event_buffer
, count
))
181 mutex_unlock(&buffer_mutex
);
185 const struct file_operations event_buffer_fops
= {
186 .open
= event_buffer_open
,
187 .release
= event_buffer_release
,
188 .read
= event_buffer_read
,