2 * fs/proc/vmcore.c Interface for accessing the crash
3 * dump from the system's previous life.
4 * Heavily borrowed from fs/proc/kcore.c
5 * Created by: Hariprasad Nellitheertha (hari@in.ibm.com)
6 * Copyright (C) IBM Corporation, 2004. All rights reserved
11 #include <linux/proc_fs.h>
12 #include <linux/user.h>
13 #include <linux/elf.h>
14 #include <linux/elfcore.h>
15 #include <linux/slab.h>
16 #include <linux/highmem.h>
17 #include <linux/bootmem.h>
18 #include <linux/init.h>
19 #include <linux/crash_dump.h>
20 #include <linux/list.h>
21 #include <asm/uaccess.h>
24 /* List representing chunks of contiguous memory areas and their offsets in
27 static LIST_HEAD(vmcore_list
);
29 /* Stores the pointer to the buffer containing kernel elf core headers. */
30 static char *elfcorebuf
;
31 static size_t elfcorebuf_sz
;
33 /* Total size of vmcore file. */
34 static u64 vmcore_size
;
36 static struct proc_dir_entry
*proc_vmcore
= NULL
;
39 * Returns > 0 for RAM pages, 0 for non-RAM pages, < 0 on error
40 * The called function has to take care of module refcounting.
42 static int (*oldmem_pfn_is_ram
)(unsigned long pfn
);
44 int register_oldmem_pfn_is_ram(int (*fn
)(unsigned long pfn
))
46 if (oldmem_pfn_is_ram
)
48 oldmem_pfn_is_ram
= fn
;
51 EXPORT_SYMBOL_GPL(register_oldmem_pfn_is_ram
);
53 void unregister_oldmem_pfn_is_ram(void)
55 oldmem_pfn_is_ram
= NULL
;
58 EXPORT_SYMBOL_GPL(unregister_oldmem_pfn_is_ram
);
60 static int pfn_is_ram(unsigned long pfn
)
62 int (*fn
)(unsigned long pfn
);
63 /* pfn is ram unless fn() checks pagetype */
67 * Ask hypervisor if the pfn is really ram.
68 * A ballooned page contains no data and reading from such a page
69 * will cause high load in the hypervisor.
71 fn
= oldmem_pfn_is_ram
;
78 /* Reads a page from the oldmem device from given offset. */
79 static ssize_t
read_from_oldmem(char *buf
, size_t count
,
80 u64
*ppos
, int userbuf
)
82 unsigned long pfn
, offset
;
84 ssize_t read
= 0, tmp
;
89 offset
= (unsigned long)(*ppos
% PAGE_SIZE
);
90 pfn
= (unsigned long)(*ppos
/ PAGE_SIZE
);
93 if (count
> (PAGE_SIZE
- offset
))
94 nr_bytes
= PAGE_SIZE
- offset
;
98 /* If pfn is not ram, return zeros for sparse dump files */
99 if (pfn_is_ram(pfn
) == 0)
100 memset(buf
, 0, nr_bytes
);
102 tmp
= copy_oldmem_page(pfn
, buf
, nr_bytes
,
118 /* Maps vmcore file offset to respective physical address in memroy. */
119 static u64
map_offset_to_paddr(loff_t offset
, struct list_head
*vc_list
,
120 struct vmcore
**m_ptr
)
125 list_for_each_entry(m
, vc_list
, list
) {
128 end
= m
->offset
+ m
->size
- 1;
129 if (offset
>= start
&& offset
<= end
) {
130 paddr
= m
->paddr
+ offset
- start
;
139 /* Read from the ELF header and then the crash dump. On error, negative value is
140 * returned otherwise number of bytes read are returned.
142 static ssize_t
read_vmcore(struct file
*file
, char __user
*buffer
,
143 size_t buflen
, loff_t
*fpos
)
145 ssize_t acc
= 0, tmp
;
148 struct vmcore
*curr_m
= NULL
;
150 if (buflen
== 0 || *fpos
>= vmcore_size
)
153 /* trim buflen to not go beyond EOF */
154 if (buflen
> vmcore_size
- *fpos
)
155 buflen
= vmcore_size
- *fpos
;
157 /* Read ELF core header */
158 if (*fpos
< elfcorebuf_sz
) {
159 tsz
= elfcorebuf_sz
- *fpos
;
162 if (copy_to_user(buffer
, elfcorebuf
+ *fpos
, tsz
))
169 /* leave now if filled buffer already */
174 start
= map_offset_to_paddr(*fpos
, &vmcore_list
, &curr_m
);
177 if ((tsz
= (PAGE_SIZE
- (start
& ~PAGE_MASK
))) > buflen
)
180 /* Calculate left bytes in current memory segment. */
181 nr_bytes
= (curr_m
->size
- (start
- curr_m
->paddr
));
186 tmp
= read_from_oldmem(buffer
, tsz
, &start
, 1);
193 if (start
>= (curr_m
->paddr
+ curr_m
->size
)) {
194 if (curr_m
->list
.next
== &vmcore_list
)
196 curr_m
= list_entry(curr_m
->list
.next
,
197 struct vmcore
, list
);
198 start
= curr_m
->paddr
;
200 if ((tsz
= (PAGE_SIZE
- (start
& ~PAGE_MASK
))) > buflen
)
202 /* Calculate left bytes in current memory segment. */
203 nr_bytes
= (curr_m
->size
- (start
- curr_m
->paddr
));
210 static const struct file_operations proc_vmcore_operations
= {
212 .llseek
= default_llseek
,
215 static struct vmcore
* __init
get_new_element(void)
217 return kzalloc(sizeof(struct vmcore
), GFP_KERNEL
);
220 static u64 __init
get_vmcore_size_elf64(char *elfptr
)
224 Elf64_Ehdr
*ehdr_ptr
;
225 Elf64_Phdr
*phdr_ptr
;
227 ehdr_ptr
= (Elf64_Ehdr
*)elfptr
;
228 phdr_ptr
= (Elf64_Phdr
*)(elfptr
+ sizeof(Elf64_Ehdr
));
229 size
= sizeof(Elf64_Ehdr
) + ((ehdr_ptr
->e_phnum
) * sizeof(Elf64_Phdr
));
230 for (i
= 0; i
< ehdr_ptr
->e_phnum
; i
++) {
231 size
+= phdr_ptr
->p_memsz
;
237 static u64 __init
get_vmcore_size_elf32(char *elfptr
)
241 Elf32_Ehdr
*ehdr_ptr
;
242 Elf32_Phdr
*phdr_ptr
;
244 ehdr_ptr
= (Elf32_Ehdr
*)elfptr
;
245 phdr_ptr
= (Elf32_Phdr
*)(elfptr
+ sizeof(Elf32_Ehdr
));
246 size
= sizeof(Elf32_Ehdr
) + ((ehdr_ptr
->e_phnum
) * sizeof(Elf32_Phdr
));
247 for (i
= 0; i
< ehdr_ptr
->e_phnum
; i
++) {
248 size
+= phdr_ptr
->p_memsz
;
254 /* Merges all the PT_NOTE headers into one. */
255 static int __init
merge_note_headers_elf64(char *elfptr
, size_t *elfsz
,
256 struct list_head
*vc_list
)
258 int i
, nr_ptnote
=0, rc
=0;
260 Elf64_Ehdr
*ehdr_ptr
;
261 Elf64_Phdr phdr
, *phdr_ptr
;
262 Elf64_Nhdr
*nhdr_ptr
;
263 u64 phdr_sz
= 0, note_off
;
265 ehdr_ptr
= (Elf64_Ehdr
*)elfptr
;
266 phdr_ptr
= (Elf64_Phdr
*)(elfptr
+ sizeof(Elf64_Ehdr
));
267 for (i
= 0; i
< ehdr_ptr
->e_phnum
; i
++, phdr_ptr
++) {
271 u64 offset
, max_sz
, sz
, real_sz
= 0;
272 if (phdr_ptr
->p_type
!= PT_NOTE
)
275 max_sz
= phdr_ptr
->p_memsz
;
276 offset
= phdr_ptr
->p_offset
;
277 notes_section
= kmalloc(max_sz
, GFP_KERNEL
);
280 rc
= read_from_oldmem(notes_section
, max_sz
, &offset
, 0);
282 kfree(notes_section
);
285 nhdr_ptr
= notes_section
;
286 for (j
= 0; j
< max_sz
; j
+= sz
) {
287 if (nhdr_ptr
->n_namesz
== 0)
289 sz
= sizeof(Elf64_Nhdr
) +
290 ((nhdr_ptr
->n_namesz
+ 3) & ~3) +
291 ((nhdr_ptr
->n_descsz
+ 3) & ~3);
293 nhdr_ptr
= (Elf64_Nhdr
*)((char*)nhdr_ptr
+ sz
);
296 /* Add this contiguous chunk of notes section to vmcore list.*/
297 new = get_new_element();
299 kfree(notes_section
);
302 new->paddr
= phdr_ptr
->p_offset
;
304 list_add_tail(&new->list
, vc_list
);
306 kfree(notes_section
);
309 /* Prepare merged PT_NOTE program header. */
310 phdr
.p_type
= PT_NOTE
;
312 note_off
= sizeof(Elf64_Ehdr
) +
313 (ehdr_ptr
->e_phnum
- nr_ptnote
+1) * sizeof(Elf64_Phdr
);
314 phdr
.p_offset
= note_off
;
315 phdr
.p_vaddr
= phdr
.p_paddr
= 0;
316 phdr
.p_filesz
= phdr
.p_memsz
= phdr_sz
;
319 /* Add merged PT_NOTE program header*/
320 tmp
= elfptr
+ sizeof(Elf64_Ehdr
);
321 memcpy(tmp
, &phdr
, sizeof(phdr
));
324 /* Remove unwanted PT_NOTE program headers. */
325 i
= (nr_ptnote
- 1) * sizeof(Elf64_Phdr
);
327 memmove(tmp
, tmp
+i
, ((*elfsz
)-sizeof(Elf64_Ehdr
)-sizeof(Elf64_Phdr
)));
329 /* Modify e_phnum to reflect merged headers. */
330 ehdr_ptr
->e_phnum
= ehdr_ptr
->e_phnum
- nr_ptnote
+ 1;
335 /* Merges all the PT_NOTE headers into one. */
336 static int __init
merge_note_headers_elf32(char *elfptr
, size_t *elfsz
,
337 struct list_head
*vc_list
)
339 int i
, nr_ptnote
=0, rc
=0;
341 Elf32_Ehdr
*ehdr_ptr
;
342 Elf32_Phdr phdr
, *phdr_ptr
;
343 Elf32_Nhdr
*nhdr_ptr
;
344 u64 phdr_sz
= 0, note_off
;
346 ehdr_ptr
= (Elf32_Ehdr
*)elfptr
;
347 phdr_ptr
= (Elf32_Phdr
*)(elfptr
+ sizeof(Elf32_Ehdr
));
348 for (i
= 0; i
< ehdr_ptr
->e_phnum
; i
++, phdr_ptr
++) {
352 u64 offset
, max_sz
, sz
, real_sz
= 0;
353 if (phdr_ptr
->p_type
!= PT_NOTE
)
356 max_sz
= phdr_ptr
->p_memsz
;
357 offset
= phdr_ptr
->p_offset
;
358 notes_section
= kmalloc(max_sz
, GFP_KERNEL
);
361 rc
= read_from_oldmem(notes_section
, max_sz
, &offset
, 0);
363 kfree(notes_section
);
366 nhdr_ptr
= notes_section
;
367 for (j
= 0; j
< max_sz
; j
+= sz
) {
368 if (nhdr_ptr
->n_namesz
== 0)
370 sz
= sizeof(Elf32_Nhdr
) +
371 ((nhdr_ptr
->n_namesz
+ 3) & ~3) +
372 ((nhdr_ptr
->n_descsz
+ 3) & ~3);
374 nhdr_ptr
= (Elf32_Nhdr
*)((char*)nhdr_ptr
+ sz
);
377 /* Add this contiguous chunk of notes section to vmcore list.*/
378 new = get_new_element();
380 kfree(notes_section
);
383 new->paddr
= phdr_ptr
->p_offset
;
385 list_add_tail(&new->list
, vc_list
);
387 kfree(notes_section
);
390 /* Prepare merged PT_NOTE program header. */
391 phdr
.p_type
= PT_NOTE
;
393 note_off
= sizeof(Elf32_Ehdr
) +
394 (ehdr_ptr
->e_phnum
- nr_ptnote
+1) * sizeof(Elf32_Phdr
);
395 phdr
.p_offset
= note_off
;
396 phdr
.p_vaddr
= phdr
.p_paddr
= 0;
397 phdr
.p_filesz
= phdr
.p_memsz
= phdr_sz
;
400 /* Add merged PT_NOTE program header*/
401 tmp
= elfptr
+ sizeof(Elf32_Ehdr
);
402 memcpy(tmp
, &phdr
, sizeof(phdr
));
405 /* Remove unwanted PT_NOTE program headers. */
406 i
= (nr_ptnote
- 1) * sizeof(Elf32_Phdr
);
408 memmove(tmp
, tmp
+i
, ((*elfsz
)-sizeof(Elf32_Ehdr
)-sizeof(Elf32_Phdr
)));
410 /* Modify e_phnum to reflect merged headers. */
411 ehdr_ptr
->e_phnum
= ehdr_ptr
->e_phnum
- nr_ptnote
+ 1;
416 /* Add memory chunks represented by program headers to vmcore list. Also update
417 * the new offset fields of exported program headers. */
418 static int __init
process_ptload_program_headers_elf64(char *elfptr
,
420 struct list_head
*vc_list
)
423 Elf64_Ehdr
*ehdr_ptr
;
424 Elf64_Phdr
*phdr_ptr
;
428 ehdr_ptr
= (Elf64_Ehdr
*)elfptr
;
429 phdr_ptr
= (Elf64_Phdr
*)(elfptr
+ sizeof(Elf64_Ehdr
)); /* PT_NOTE hdr */
431 /* First program header is PT_NOTE header. */
432 vmcore_off
= sizeof(Elf64_Ehdr
) +
433 (ehdr_ptr
->e_phnum
) * sizeof(Elf64_Phdr
) +
434 phdr_ptr
->p_memsz
; /* Note sections */
436 for (i
= 0; i
< ehdr_ptr
->e_phnum
; i
++, phdr_ptr
++) {
437 if (phdr_ptr
->p_type
!= PT_LOAD
)
440 /* Add this contiguous chunk of memory to vmcore list.*/
441 new = get_new_element();
444 new->paddr
= phdr_ptr
->p_offset
;
445 new->size
= phdr_ptr
->p_memsz
;
446 list_add_tail(&new->list
, vc_list
);
448 /* Update the program header offset. */
449 phdr_ptr
->p_offset
= vmcore_off
;
450 vmcore_off
= vmcore_off
+ phdr_ptr
->p_memsz
;
455 static int __init
process_ptload_program_headers_elf32(char *elfptr
,
457 struct list_head
*vc_list
)
460 Elf32_Ehdr
*ehdr_ptr
;
461 Elf32_Phdr
*phdr_ptr
;
465 ehdr_ptr
= (Elf32_Ehdr
*)elfptr
;
466 phdr_ptr
= (Elf32_Phdr
*)(elfptr
+ sizeof(Elf32_Ehdr
)); /* PT_NOTE hdr */
468 /* First program header is PT_NOTE header. */
469 vmcore_off
= sizeof(Elf32_Ehdr
) +
470 (ehdr_ptr
->e_phnum
) * sizeof(Elf32_Phdr
) +
471 phdr_ptr
->p_memsz
; /* Note sections */
473 for (i
= 0; i
< ehdr_ptr
->e_phnum
; i
++, phdr_ptr
++) {
474 if (phdr_ptr
->p_type
!= PT_LOAD
)
477 /* Add this contiguous chunk of memory to vmcore list.*/
478 new = get_new_element();
481 new->paddr
= phdr_ptr
->p_offset
;
482 new->size
= phdr_ptr
->p_memsz
;
483 list_add_tail(&new->list
, vc_list
);
485 /* Update the program header offset */
486 phdr_ptr
->p_offset
= vmcore_off
;
487 vmcore_off
= vmcore_off
+ phdr_ptr
->p_memsz
;
492 /* Sets offset fields of vmcore elements. */
493 static void __init
set_vmcore_list_offsets_elf64(char *elfptr
,
494 struct list_head
*vc_list
)
497 Elf64_Ehdr
*ehdr_ptr
;
500 ehdr_ptr
= (Elf64_Ehdr
*)elfptr
;
502 /* Skip Elf header and program headers. */
503 vmcore_off
= sizeof(Elf64_Ehdr
) +
504 (ehdr_ptr
->e_phnum
) * sizeof(Elf64_Phdr
);
506 list_for_each_entry(m
, vc_list
, list
) {
507 m
->offset
= vmcore_off
;
508 vmcore_off
+= m
->size
;
512 /* Sets offset fields of vmcore elements. */
513 static void __init
set_vmcore_list_offsets_elf32(char *elfptr
,
514 struct list_head
*vc_list
)
517 Elf32_Ehdr
*ehdr_ptr
;
520 ehdr_ptr
= (Elf32_Ehdr
*)elfptr
;
522 /* Skip Elf header and program headers. */
523 vmcore_off
= sizeof(Elf32_Ehdr
) +
524 (ehdr_ptr
->e_phnum
) * sizeof(Elf32_Phdr
);
526 list_for_each_entry(m
, vc_list
, list
) {
527 m
->offset
= vmcore_off
;
528 vmcore_off
+= m
->size
;
532 static int __init
parse_crash_elf64_headers(void)
538 addr
= elfcorehdr_addr
;
540 /* Read Elf header */
541 rc
= read_from_oldmem((char*)&ehdr
, sizeof(Elf64_Ehdr
), &addr
, 0);
545 /* Do some basic Verification. */
546 if (memcmp(ehdr
.e_ident
, ELFMAG
, SELFMAG
) != 0 ||
547 (ehdr
.e_type
!= ET_CORE
) ||
548 !vmcore_elf64_check_arch(&ehdr
) ||
549 ehdr
.e_ident
[EI_CLASS
] != ELFCLASS64
||
550 ehdr
.e_ident
[EI_VERSION
] != EV_CURRENT
||
551 ehdr
.e_version
!= EV_CURRENT
||
552 ehdr
.e_ehsize
!= sizeof(Elf64_Ehdr
) ||
553 ehdr
.e_phentsize
!= sizeof(Elf64_Phdr
) ||
555 printk(KERN_WARNING
"Warning: Core image elf header is not"
560 /* Read in all elf headers. */
561 elfcorebuf_sz
= sizeof(Elf64_Ehdr
) + ehdr
.e_phnum
* sizeof(Elf64_Phdr
);
562 elfcorebuf
= kmalloc(elfcorebuf_sz
, GFP_KERNEL
);
565 addr
= elfcorehdr_addr
;
566 rc
= read_from_oldmem(elfcorebuf
, elfcorebuf_sz
, &addr
, 0);
572 /* Merge all PT_NOTE headers into one. */
573 rc
= merge_note_headers_elf64(elfcorebuf
, &elfcorebuf_sz
, &vmcore_list
);
578 rc
= process_ptload_program_headers_elf64(elfcorebuf
, elfcorebuf_sz
,
584 set_vmcore_list_offsets_elf64(elfcorebuf
, &vmcore_list
);
588 static int __init
parse_crash_elf32_headers(void)
594 addr
= elfcorehdr_addr
;
596 /* Read Elf header */
597 rc
= read_from_oldmem((char*)&ehdr
, sizeof(Elf32_Ehdr
), &addr
, 0);
601 /* Do some basic Verification. */
602 if (memcmp(ehdr
.e_ident
, ELFMAG
, SELFMAG
) != 0 ||
603 (ehdr
.e_type
!= ET_CORE
) ||
604 !elf_check_arch(&ehdr
) ||
605 ehdr
.e_ident
[EI_CLASS
] != ELFCLASS32
||
606 ehdr
.e_ident
[EI_VERSION
] != EV_CURRENT
||
607 ehdr
.e_version
!= EV_CURRENT
||
608 ehdr
.e_ehsize
!= sizeof(Elf32_Ehdr
) ||
609 ehdr
.e_phentsize
!= sizeof(Elf32_Phdr
) ||
611 printk(KERN_WARNING
"Warning: Core image elf header is not"
616 /* Read in all elf headers. */
617 elfcorebuf_sz
= sizeof(Elf32_Ehdr
) + ehdr
.e_phnum
* sizeof(Elf32_Phdr
);
618 elfcorebuf
= kmalloc(elfcorebuf_sz
, GFP_KERNEL
);
621 addr
= elfcorehdr_addr
;
622 rc
= read_from_oldmem(elfcorebuf
, elfcorebuf_sz
, &addr
, 0);
628 /* Merge all PT_NOTE headers into one. */
629 rc
= merge_note_headers_elf32(elfcorebuf
, &elfcorebuf_sz
, &vmcore_list
);
634 rc
= process_ptload_program_headers_elf32(elfcorebuf
, elfcorebuf_sz
,
640 set_vmcore_list_offsets_elf32(elfcorebuf
, &vmcore_list
);
644 static int __init
parse_crash_elf_headers(void)
646 unsigned char e_ident
[EI_NIDENT
];
650 addr
= elfcorehdr_addr
;
651 rc
= read_from_oldmem(e_ident
, EI_NIDENT
, &addr
, 0);
654 if (memcmp(e_ident
, ELFMAG
, SELFMAG
) != 0) {
655 printk(KERN_WARNING
"Warning: Core image elf header"
660 if (e_ident
[EI_CLASS
] == ELFCLASS64
) {
661 rc
= parse_crash_elf64_headers();
665 /* Determine vmcore size. */
666 vmcore_size
= get_vmcore_size_elf64(elfcorebuf
);
667 } else if (e_ident
[EI_CLASS
] == ELFCLASS32
) {
668 rc
= parse_crash_elf32_headers();
672 /* Determine vmcore size. */
673 vmcore_size
= get_vmcore_size_elf32(elfcorebuf
);
675 printk(KERN_WARNING
"Warning: Core image elf header is not"
682 /* Init function for vmcore module. */
683 static int __init
vmcore_init(void)
687 /* If elfcorehdr= has been passed in cmdline, then capture the dump.*/
688 if (!(is_vmcore_usable()))
690 rc
= parse_crash_elf_headers();
692 printk(KERN_WARNING
"Kdump: vmcore not initialized\n");
696 proc_vmcore
= proc_create("vmcore", S_IRUSR
, NULL
, &proc_vmcore_operations
);
698 proc_vmcore
->size
= vmcore_size
;
701 module_init(vmcore_init
)