2 * sparse memory mappings.
4 #include <linux/config.h>
6 #include <linux/mmzone.h>
7 #include <linux/bootmem.h>
8 #include <linux/module.h>
12 * Permanent SPARSEMEM data:
14 * 1) mem_section - memory sections, mem_map's for valid memory
16 struct mem_section mem_section
[NR_MEM_SECTIONS
];
17 EXPORT_SYMBOL(mem_section
);
19 /* Record a memory area against a node. */
20 void memory_present(int nid
, unsigned long start
, unsigned long end
)
24 start
&= PAGE_SECTION_MASK
;
25 for (pfn
= start
; pfn
< end
; pfn
+= PAGES_PER_SECTION
) {
26 unsigned long section
= pfn_to_section_nr(pfn
);
27 if (!mem_section
[section
].section_mem_map
)
28 mem_section
[section
].section_mem_map
= SECTION_MARKED_PRESENT
;
33 * Only used by the i386 NUMA architecures, but relatively
36 unsigned long __init
node_memmap_size_bytes(int nid
, unsigned long start_pfn
,
37 unsigned long end_pfn
)
40 unsigned long nr_pages
= 0;
42 for (pfn
= start_pfn
; pfn
< end_pfn
; pfn
+= PAGES_PER_SECTION
) {
43 if (nid
!= early_pfn_to_nid(pfn
))
47 nr_pages
+= PAGES_PER_SECTION
;
50 return nr_pages
* sizeof(struct page
);
54 * Subtle, we encode the real pfn into the mem_map such that
55 * the identity pfn - section_mem_map will return the actual
56 * physical page frame number.
58 static unsigned long sparse_encode_mem_map(struct page
*mem_map
, unsigned long pnum
)
60 return (unsigned long)(mem_map
- (section_nr_to_pfn(pnum
)));
64 * We need this if we ever free the mem_maps. While not implemented yet,
65 * this function is included for parity with its sibling.
67 static __attribute((unused
))
68 struct page
*sparse_decode_mem_map(unsigned long coded_mem_map
, unsigned long pnum
)
70 return ((struct page
*)coded_mem_map
) + section_nr_to_pfn(pnum
);
73 static int sparse_init_one_section(struct mem_section
*ms
,
74 unsigned long pnum
, struct page
*mem_map
)
76 if (!valid_section(ms
))
79 ms
->section_mem_map
|= sparse_encode_mem_map(mem_map
, pnum
);
84 static struct page
*sparse_early_mem_map_alloc(unsigned long pnum
)
87 int nid
= early_pfn_to_nid(section_nr_to_pfn(pnum
));
89 map
= alloc_remap(nid
, sizeof(struct page
) * PAGES_PER_SECTION
);
93 map
= alloc_bootmem_node(NODE_DATA(nid
),
94 sizeof(struct page
) * PAGES_PER_SECTION
);
98 printk(KERN_WARNING
"%s: allocation failed\n", __FUNCTION__
);
99 mem_section
[pnum
].section_mem_map
= 0;
104 * Allocate the accumulated non-linear sections, allocate a mem_map
105 * for each and record the physical to section mapping.
107 void sparse_init(void)
112 for (pnum
= 0; pnum
< NR_MEM_SECTIONS
; pnum
++) {
113 if (!valid_section_nr(pnum
))
116 map
= sparse_early_mem_map_alloc(pnum
);
118 sparse_init_one_section(&mem_section
[pnum
], pnum
, map
);
123 * returns the number of sections whose mem_maps were properly
124 * set. If this is <=0, then that means that the passed-in
125 * map was not consumed and must be freed.
127 int sparse_add_one_section(unsigned long start_pfn
, int nr_pages
, struct page
*map
)
129 struct mem_section
*ms
= __pfn_to_section(start_pfn
);
131 if (ms
->section_mem_map
& SECTION_MARKED_PRESENT
)
134 ms
->section_mem_map
|= SECTION_MARKED_PRESENT
;
136 return sparse_init_one_section(ms
, pfn_to_section_nr(start_pfn
), map
);