2 * proc/fs/generic.c --- generic routines for the proc-fs
4 * This file contains generic proc-fs routines for handling
5 * directories and files.
7 * Copyright (C) 1991, 1992 Linus Torvalds.
8 * Copyright (C) 1997 Theodore Ts'o
11 #include <linux/errno.h>
12 #include <linux/time.h>
13 #include <linux/proc_fs.h>
14 #include <linux/stat.h>
15 #include <linux/module.h>
16 #include <linux/mount.h>
17 #include <linux/smp_lock.h>
18 #include <linux/init.h>
19 #include <linux/idr.h>
20 #include <linux/namei.h>
21 #include <linux/bitops.h>
22 #include <linux/spinlock.h>
23 #include <linux/completion.h>
24 #include <asm/uaccess.h>
28 DEFINE_SPINLOCK(proc_subdir_lock
);
30 static int proc_match(int len
, const char *name
, struct proc_dir_entry
*de
)
32 if (de
->namelen
!= len
)
34 return !memcmp(name
, de
->name
, len
);
37 /* buffer size is one page but our output routines use some slack for overruns */
38 #define PROC_BLOCK_SIZE (PAGE_SIZE - 1024)
41 proc_file_read(struct file
*file
, char __user
*buf
, size_t nbytes
,
44 struct inode
* inode
= file
->f_path
.dentry
->d_inode
;
50 struct proc_dir_entry
* dp
;
51 unsigned long long pos
;
54 * Gaah, please just use "seq_file" instead. The legacy /proc
55 * interfaces cut loff_t down to off_t for reads, and ignore
56 * the offset entirely for writes..
59 if (pos
> MAX_NON_LFS
)
61 if (nbytes
> MAX_NON_LFS
- pos
)
62 nbytes
= MAX_NON_LFS
- pos
;
65 if (!(page
= (char*) __get_free_page(GFP_TEMPORARY
)))
68 while ((nbytes
> 0) && !eof
) {
69 count
= min_t(size_t, PROC_BLOCK_SIZE
, nbytes
);
74 * How to be a proc read function
75 * ------------------------------
77 * int f(char *buffer, char **start, off_t offset,
78 * int count, int *peof, void *dat)
80 * Assume that the buffer is "count" bytes in size.
82 * If you know you have supplied all the data you
85 * You have three ways to return data:
86 * 0) Leave *start = NULL. (This is the default.)
87 * Put the data of the requested offset at that
88 * offset within the buffer. Return the number (n)
89 * of bytes there are from the beginning of the
90 * buffer up to the last byte of data. If the
91 * number of supplied bytes (= n - offset) is
92 * greater than zero and you didn't signal eof
93 * and the reader is prepared to take more data
94 * you will be called again with the requested
95 * offset advanced by the number of bytes
96 * absorbed. This interface is useful for files
97 * no larger than the buffer.
98 * 1) Set *start = an unsigned long value less than
99 * the buffer address but greater than zero.
100 * Put the data of the requested offset at the
101 * beginning of the buffer. Return the number of
102 * bytes of data placed there. If this number is
103 * greater than zero and you didn't signal eof
104 * and the reader is prepared to take more data
105 * you will be called again with the requested
106 * offset advanced by *start. This interface is
107 * useful when you have a large file consisting
108 * of a series of blocks which you want to count
109 * and return as wholes.
110 * (Hack by Paul.Russell@rustcorp.com.au)
111 * 2) Set *start = an address within the buffer.
112 * Put the data of the requested offset at *start.
113 * Return the number of bytes of data placed there.
114 * If this number is greater than zero and you
115 * didn't signal eof and the reader is prepared to
116 * take more data you will be called again with the
117 * requested offset advanced by the number of bytes
120 n
= dp
->read_proc(page
, &start
, *ppos
,
121 count
, &eof
, dp
->data
);
125 if (n
== 0) /* end of file */
127 if (n
< 0) { /* error */
136 "proc_file_read: Apparent buffer overflow!\n");
144 start
= page
+ *ppos
;
145 } else if (start
< page
) {
148 "proc_file_read: Apparent buffer overflow!\n");
153 * Don't reduce n because doing so might
154 * cut off part of a data block.
157 "proc_file_read: Read count exceeded\n");
159 } else /* start >= page */ {
160 unsigned long startoff
= (unsigned long)(start
- page
);
161 if (n
> (PAGE_SIZE
- startoff
)) {
163 "proc_file_read: Apparent buffer overflow!\n");
164 n
= PAGE_SIZE
- startoff
;
170 n
-= copy_to_user(buf
, start
< page
? page
: start
, n
);
177 *ppos
+= start
< page
? (unsigned long)start
: n
;
182 free_page((unsigned long) page
);
187 proc_file_write(struct file
*file
, const char __user
*buffer
,
188 size_t count
, loff_t
*ppos
)
190 struct inode
*inode
= file
->f_path
.dentry
->d_inode
;
191 struct proc_dir_entry
* dp
;
198 /* FIXME: does this routine need ppos? probably... */
199 return dp
->write_proc(file
, buffer
, count
, dp
->data
);
204 proc_file_lseek(struct file
*file
, loff_t offset
, int orig
)
206 loff_t retval
= -EINVAL
;
209 offset
+= file
->f_pos
;
212 if (offset
< 0 || offset
> MAX_NON_LFS
)
214 file
->f_pos
= retval
= offset
;
219 static const struct file_operations proc_file_operations
= {
220 .llseek
= proc_file_lseek
,
221 .read
= proc_file_read
,
222 .write
= proc_file_write
,
225 static int proc_notify_change(struct dentry
*dentry
, struct iattr
*iattr
)
227 struct inode
*inode
= dentry
->d_inode
;
228 struct proc_dir_entry
*de
= PDE(inode
);
231 error
= inode_change_ok(inode
, iattr
);
235 error
= inode_setattr(inode
, iattr
);
239 de
->uid
= inode
->i_uid
;
240 de
->gid
= inode
->i_gid
;
241 de
->mode
= inode
->i_mode
;
246 static int proc_getattr(struct vfsmount
*mnt
, struct dentry
*dentry
,
249 struct inode
*inode
= dentry
->d_inode
;
250 struct proc_dir_entry
*de
= PROC_I(inode
)->pde
;
252 inode
->i_nlink
= de
->nlink
;
254 generic_fillattr(inode
, stat
);
258 static const struct inode_operations proc_file_inode_operations
= {
259 .setattr
= proc_notify_change
,
263 * This function parses a name such as "tty/driver/serial", and
264 * returns the struct proc_dir_entry for "/proc/tty/driver", and
265 * returns "serial" in residual.
267 static int xlate_proc_name(const char *name
,
268 struct proc_dir_entry
**ret
, const char **residual
)
270 const char *cp
= name
, *next
;
271 struct proc_dir_entry
*de
;
279 spin_lock(&proc_subdir_lock
);
281 next
= strchr(cp
, '/');
286 for (de
= de
->subdir
; de
; de
= de
->next
) {
287 if (proc_match(len
, cp
, de
))
299 spin_unlock(&proc_subdir_lock
);
303 static DEFINE_IDR(proc_inum_idr
);
304 static DEFINE_SPINLOCK(proc_inum_lock
); /* protects the above */
306 #define PROC_DYNAMIC_FIRST 0xF0000000UL
309 * Return an inode number between PROC_DYNAMIC_FIRST and
310 * 0xffffffff, or zero on failure.
312 static unsigned int get_inode_number(void)
318 if (idr_pre_get(&proc_inum_idr
, GFP_KERNEL
) == 0)
321 spin_lock(&proc_inum_lock
);
322 error
= idr_get_new(&proc_inum_idr
, NULL
, &i
);
323 spin_unlock(&proc_inum_lock
);
324 if (error
== -EAGAIN
)
329 inum
= (i
& MAX_ID_MASK
) + PROC_DYNAMIC_FIRST
;
331 /* inum will never be more than 0xf0ffffff, so no check
338 static void release_inode_number(unsigned int inum
)
340 int id
= (inum
- PROC_DYNAMIC_FIRST
) | ~MAX_ID_MASK
;
342 spin_lock(&proc_inum_lock
);
343 idr_remove(&proc_inum_idr
, id
);
344 spin_unlock(&proc_inum_lock
);
347 static void *proc_follow_link(struct dentry
*dentry
, struct nameidata
*nd
)
349 nd_set_link(nd
, PDE(dentry
->d_inode
)->data
);
353 static const struct inode_operations proc_link_inode_operations
= {
354 .readlink
= generic_readlink
,
355 .follow_link
= proc_follow_link
,
359 * As some entries in /proc are volatile, we want to
360 * get rid of unused dentries. This could be made
361 * smarter: we could keep a "volatile" flag in the
362 * inode to indicate which ones to keep.
364 static int proc_delete_dentry(struct dentry
* dentry
)
369 static struct dentry_operations proc_dentry_operations
=
371 .d_delete
= proc_delete_dentry
,
375 * Don't create negative dentries here, return -ENOENT by hand
378 struct dentry
*proc_lookup_de(struct proc_dir_entry
*de
, struct inode
*dir
,
379 struct dentry
*dentry
)
381 struct inode
*inode
= NULL
;
385 spin_lock(&proc_subdir_lock
);
386 for (de
= de
->subdir
; de
; de
= de
->next
) {
387 if (de
->namelen
!= dentry
->d_name
.len
)
389 if (!memcmp(dentry
->d_name
.name
, de
->name
, de
->namelen
)) {
394 spin_unlock(&proc_subdir_lock
);
396 inode
= proc_get_inode(dir
->i_sb
, ino
, de
);
400 spin_unlock(&proc_subdir_lock
);
405 dentry
->d_op
= &proc_dentry_operations
;
406 d_add(dentry
, inode
);
411 return ERR_PTR(error
);
414 struct dentry
*proc_lookup(struct inode
*dir
, struct dentry
*dentry
,
415 struct nameidata
*nd
)
417 return proc_lookup_de(PDE(dir
), dir
, dentry
);
421 * This returns non-zero if at EOF, so that the /proc
422 * root directory can use this and check if it should
423 * continue with the <pid> entries..
425 * Note that the VFS-layer doesn't care about the return
426 * value of the readdir() call, as long as it's non-negative
429 int proc_readdir_de(struct proc_dir_entry
*de
, struct file
*filp
, void *dirent
,
434 struct inode
*inode
= filp
->f_path
.dentry
->d_inode
;
443 if (filldir(dirent
, ".", 1, i
, ino
, DT_DIR
) < 0)
449 if (filldir(dirent
, "..", 2, i
,
450 parent_ino(filp
->f_path
.dentry
),
457 spin_lock(&proc_subdir_lock
);
463 spin_unlock(&proc_subdir_lock
);
473 struct proc_dir_entry
*next
;
475 /* filldir passes info to user space */
477 spin_unlock(&proc_subdir_lock
);
478 if (filldir(dirent
, de
->name
, de
->namelen
, filp
->f_pos
,
479 de
->low_ino
, de
->mode
>> 12) < 0) {
483 spin_lock(&proc_subdir_lock
);
489 spin_unlock(&proc_subdir_lock
);
492 out
: unlock_kernel();
496 int proc_readdir(struct file
*filp
, void *dirent
, filldir_t filldir
)
498 struct inode
*inode
= filp
->f_path
.dentry
->d_inode
;
500 return proc_readdir_de(PDE(inode
), filp
, dirent
, filldir
);
504 * These are the generic /proc directory operations. They
505 * use the in-memory "struct proc_dir_entry" tree to parse
506 * the /proc directory.
508 static const struct file_operations proc_dir_operations
= {
509 .read
= generic_read_dir
,
510 .readdir
= proc_readdir
,
514 * proc directories can do almost nothing..
516 static const struct inode_operations proc_dir_inode_operations
= {
517 .lookup
= proc_lookup
,
518 .getattr
= proc_getattr
,
519 .setattr
= proc_notify_change
,
522 static int proc_register(struct proc_dir_entry
* dir
, struct proc_dir_entry
* dp
)
525 struct proc_dir_entry
*tmp
;
527 i
= get_inode_number();
532 if (S_ISDIR(dp
->mode
)) {
533 if (dp
->proc_iops
== NULL
) {
534 dp
->proc_fops
= &proc_dir_operations
;
535 dp
->proc_iops
= &proc_dir_inode_operations
;
538 } else if (S_ISLNK(dp
->mode
)) {
539 if (dp
->proc_iops
== NULL
)
540 dp
->proc_iops
= &proc_link_inode_operations
;
541 } else if (S_ISREG(dp
->mode
)) {
542 if (dp
->proc_fops
== NULL
)
543 dp
->proc_fops
= &proc_file_operations
;
544 if (dp
->proc_iops
== NULL
)
545 dp
->proc_iops
= &proc_file_inode_operations
;
548 spin_lock(&proc_subdir_lock
);
550 for (tmp
= dir
->subdir
; tmp
; tmp
= tmp
->next
)
551 if (strcmp(tmp
->name
, dp
->name
) == 0) {
552 printk(KERN_WARNING
"proc_dir_entry '%s' already "
553 "registered\n", dp
->name
);
558 dp
->next
= dir
->subdir
;
561 spin_unlock(&proc_subdir_lock
);
566 static struct proc_dir_entry
*__proc_create(struct proc_dir_entry
**parent
,
571 struct proc_dir_entry
*ent
= NULL
;
572 const char *fn
= name
;
575 /* make sure name is valid */
576 if (!name
|| !strlen(name
)) goto out
;
578 if (xlate_proc_name(name
, parent
, &fn
) != 0)
581 /* At this point there must not be any '/' characters beyond *fn */
587 ent
= kmalloc(sizeof(struct proc_dir_entry
) + len
+ 1, GFP_KERNEL
);
590 memset(ent
, 0, sizeof(struct proc_dir_entry
));
591 memcpy(((char *) ent
) + sizeof(struct proc_dir_entry
), fn
, len
+ 1);
592 ent
->name
= ((char *) ent
) + sizeof(*ent
);
596 atomic_set(&ent
->count
, 1);
598 spin_lock_init(&ent
->pde_unload_lock
);
599 ent
->pde_unload_completion
= NULL
;
604 struct proc_dir_entry
*proc_symlink(const char *name
,
605 struct proc_dir_entry
*parent
, const char *dest
)
607 struct proc_dir_entry
*ent
;
609 ent
= __proc_create(&parent
, name
,
610 (S_IFLNK
| S_IRUGO
| S_IWUGO
| S_IXUGO
),1);
613 ent
->data
= kmalloc((ent
->size
=strlen(dest
))+1, GFP_KERNEL
);
615 strcpy((char*)ent
->data
,dest
);
616 if (proc_register(parent
, ent
) < 0) {
629 struct proc_dir_entry
*proc_mkdir_mode(const char *name
, mode_t mode
,
630 struct proc_dir_entry
*parent
)
632 struct proc_dir_entry
*ent
;
634 ent
= __proc_create(&parent
, name
, S_IFDIR
| mode
, 2);
636 if (proc_register(parent
, ent
) < 0) {
644 struct proc_dir_entry
*proc_net_mkdir(struct net
*net
, const char *name
,
645 struct proc_dir_entry
*parent
)
647 struct proc_dir_entry
*ent
;
649 ent
= __proc_create(&parent
, name
, S_IFDIR
| S_IRUGO
| S_IXUGO
, 2);
652 if (proc_register(parent
, ent
) < 0) {
659 EXPORT_SYMBOL_GPL(proc_net_mkdir
);
661 struct proc_dir_entry
*proc_mkdir(const char *name
,
662 struct proc_dir_entry
*parent
)
664 return proc_mkdir_mode(name
, S_IRUGO
| S_IXUGO
, parent
);
667 struct proc_dir_entry
*create_proc_entry(const char *name
, mode_t mode
,
668 struct proc_dir_entry
*parent
)
670 struct proc_dir_entry
*ent
;
674 if ((mode
& S_IALLUGO
) == 0)
675 mode
|= S_IRUGO
| S_IXUGO
;
678 if ((mode
& S_IFMT
) == 0)
680 if ((mode
& S_IALLUGO
) == 0)
685 ent
= __proc_create(&parent
, name
, mode
, nlink
);
687 if (proc_register(parent
, ent
) < 0) {
695 struct proc_dir_entry
*proc_create_data(const char *name
, mode_t mode
,
696 struct proc_dir_entry
*parent
,
697 const struct file_operations
*proc_fops
,
700 struct proc_dir_entry
*pde
;
704 if ((mode
& S_IALLUGO
) == 0)
705 mode
|= S_IRUGO
| S_IXUGO
;
708 if ((mode
& S_IFMT
) == 0)
710 if ((mode
& S_IALLUGO
) == 0)
715 pde
= __proc_create(&parent
, name
, mode
, nlink
);
718 pde
->proc_fops
= proc_fops
;
720 if (proc_register(parent
, pde
) < 0)
729 void free_proc_entry(struct proc_dir_entry
*de
)
731 unsigned int ino
= de
->low_ino
;
733 if (ino
< PROC_DYNAMIC_FIRST
)
736 release_inode_number(ino
);
738 if (S_ISLNK(de
->mode
))
744 * Remove a /proc entry and free it if it's not currently in use.
746 void remove_proc_entry(const char *name
, struct proc_dir_entry
*parent
)
748 struct proc_dir_entry
**p
;
749 struct proc_dir_entry
*de
= NULL
;
750 const char *fn
= name
;
753 if (xlate_proc_name(name
, &parent
, &fn
) != 0)
757 spin_lock(&proc_subdir_lock
);
758 for (p
= &parent
->subdir
; *p
; p
=&(*p
)->next
) {
759 if (proc_match(len
, fn
, *p
)) {
766 spin_unlock(&proc_subdir_lock
);
770 spin_lock(&de
->pde_unload_lock
);
772 * Stop accepting new callers into module. If you're
773 * dynamically allocating ->proc_fops, save a pointer somewhere.
775 de
->proc_fops
= NULL
;
776 /* Wait until all existing callers into module are done. */
777 if (de
->pde_users
> 0) {
778 DECLARE_COMPLETION_ONSTACK(c
);
780 if (!de
->pde_unload_completion
)
781 de
->pde_unload_completion
= &c
;
783 spin_unlock(&de
->pde_unload_lock
);
785 wait_for_completion(de
->pde_unload_completion
);
787 goto continue_removing
;
789 spin_unlock(&de
->pde_unload_lock
);
792 if (S_ISDIR(de
->mode
))
796 printk(KERN_WARNING
"%s: removing non-empty directory "
797 "'%s/%s', leaking at least '%s'\n", __func__
,
798 de
->parent
->name
, de
->name
, de
->subdir
->name
);
801 if (atomic_dec_and_test(&de
->count
))