4 * @remark Copyright 2002 OProfile authors
5 * @remark Read the file COPYING
7 * @author John Levon <levon@movementarian.org>
9 * This is the global event buffer that the user-space
10 * daemon reads from. The event buffer is an untyped array
11 * of unsigned longs. Entries are prefixed by the
12 * escape value ESCAPE_CODE followed by an identifying code.
15 #include <linux/vmalloc.h>
16 #include <linux/oprofile.h>
17 #include <linux/sched.h>
18 #include <linux/capability.h>
19 #include <linux/dcookies.h>
21 #include <asm/uaccess.h>
24 #include "event_buffer.h"
25 #include "oprofile_stats.h"
27 DEFINE_MUTEX(buffer_mutex
);
29 static unsigned long buffer_opened
;
30 static DECLARE_WAIT_QUEUE_HEAD(buffer_wait
);
31 static unsigned long * event_buffer
;
32 static unsigned long buffer_size
;
33 static unsigned long buffer_watershed
;
34 static size_t buffer_pos
;
35 /* atomic_t because wait_event checks it outside of buffer_mutex */
36 static atomic_t buffer_ready
= ATOMIC_INIT(0);
38 /* Add an entry to the event buffer. When we
39 * get near to the end we wake up the process
40 * sleeping on the read() of the file.
42 void add_event_entry(unsigned long value
)
44 if (buffer_pos
== buffer_size
) {
45 atomic_inc(&oprofile_stats
.event_lost_overflow
);
49 event_buffer
[buffer_pos
] = value
;
50 if (++buffer_pos
== buffer_size
- buffer_watershed
) {
51 atomic_set(&buffer_ready
, 1);
52 wake_up(&buffer_wait
);
57 /* Wake up the waiting process if any. This happens
58 * on "echo 0 >/dev/oprofile/enable" so the daemon
59 * processes the data remaining in the event buffer.
61 void wake_up_buffer_waiter(void)
63 mutex_lock(&buffer_mutex
);
64 atomic_set(&buffer_ready
, 1);
65 wake_up(&buffer_wait
);
66 mutex_unlock(&buffer_mutex
);
70 int alloc_event_buffer(void)
75 spin_lock_irqsave(&oprofilefs_lock
, flags
);
76 buffer_size
= fs_buffer_size
;
77 buffer_watershed
= fs_buffer_watershed
;
78 spin_unlock_irqrestore(&oprofilefs_lock
, flags
);
80 if (buffer_watershed
>= buffer_size
)
83 event_buffer
= vmalloc(sizeof(unsigned long) * buffer_size
);
93 void free_event_buffer(void)
101 static int event_buffer_open(struct inode
* inode
, struct file
* file
)
105 if (!capable(CAP_SYS_ADMIN
))
108 if (test_and_set_bit(0, &buffer_opened
))
111 /* Register as a user of dcookies
112 * to ensure they persist for the lifetime of
113 * the open event file
116 file
->private_data
= dcookie_register();
117 if (!file
->private_data
)
120 if ((err
= oprofile_setup()))
123 /* NB: the actual start happens from userspace
124 * echo 1 >/dev/oprofile/enable
130 dcookie_unregister(file
->private_data
);
132 clear_bit(0, &buffer_opened
);
137 static int event_buffer_release(struct inode
* inode
, struct file
* file
)
141 dcookie_unregister(file
->private_data
);
143 atomic_set(&buffer_ready
, 0);
144 clear_bit(0, &buffer_opened
);
149 static ssize_t
event_buffer_read(struct file
* file
, char __user
* buf
,
150 size_t count
, loff_t
* offset
)
152 int retval
= -EINVAL
;
153 size_t const max
= buffer_size
* sizeof(unsigned long);
155 /* handling partial reads is more trouble than it's worth */
156 if (count
!= max
|| *offset
)
159 wait_event_interruptible(buffer_wait
, atomic_read(&buffer_ready
));
161 if (signal_pending(current
))
164 /* can't currently happen */
165 if (!atomic_read(&buffer_ready
))
168 mutex_lock(&buffer_mutex
);
170 atomic_set(&buffer_ready
, 0);
174 count
= buffer_pos
* sizeof(unsigned long);
176 if (copy_to_user(buf
, event_buffer
, count
))
183 mutex_unlock(&buffer_mutex
);
187 const struct file_operations event_buffer_fops
= {
188 .open
= event_buffer_open
,
189 .release
= event_buffer_release
,
190 .read
= event_buffer_read
,