2 * sparse memory mappings.
4 #include <linux/config.h>
6 #include <linux/mmzone.h>
7 #include <linux/bootmem.h>
8 #include <linux/module.h>
12 * Permanent SPARSEMEM data:
14 * 1) mem_section - memory sections, mem_map's for valid memory
16 #ifdef CONFIG_ARCH_SPARSEMEM_EXTREME
17 struct mem_section
*mem_section
[NR_SECTION_ROOTS
]
18 ____cacheline_maxaligned_in_smp
;
20 static void sparse_index_init(unsigned long section
, int nid
)
22 unsigned long root
= SECTION_TO_ROOT(section
);
24 if (mem_section
[root
])
26 mem_section
[root
] = alloc_bootmem_node(NODE_DATA(nid
), PAGE_SIZE
);
27 if (mem_section
[root
])
28 memset(mem_section
[root
], 0, PAGE_SIZE
);
30 panic("memory_present: NO MEMORY\n");
33 struct mem_section mem_section
[NR_MEM_SECTIONS
]
34 ____cacheline_maxaligned_in_smp
;
36 EXPORT_SYMBOL(mem_section
);
38 /* Record a memory area against a node. */
39 void memory_present(int nid
, unsigned long start
, unsigned long end
)
43 start
&= PAGE_SECTION_MASK
;
44 for (pfn
= start
; pfn
< end
; pfn
+= PAGES_PER_SECTION
) {
45 unsigned long section
= pfn_to_section_nr(pfn
);
46 struct mem_section
*ms
;
48 sparse_index_init(section
, nid
);
50 ms
= __nr_to_section(section
);
51 if (!ms
->section_mem_map
)
52 ms
->section_mem_map
= SECTION_MARKED_PRESENT
;
57 * Only used by the i386 NUMA architecures, but relatively
60 unsigned long __init
node_memmap_size_bytes(int nid
, unsigned long start_pfn
,
61 unsigned long end_pfn
)
64 unsigned long nr_pages
= 0;
66 for (pfn
= start_pfn
; pfn
< end_pfn
; pfn
+= PAGES_PER_SECTION
) {
67 if (nid
!= early_pfn_to_nid(pfn
))
71 nr_pages
+= PAGES_PER_SECTION
;
74 return nr_pages
* sizeof(struct page
);
78 * Subtle, we encode the real pfn into the mem_map such that
79 * the identity pfn - section_mem_map will return the actual
80 * physical page frame number.
82 static unsigned long sparse_encode_mem_map(struct page
*mem_map
, unsigned long pnum
)
84 return (unsigned long)(mem_map
- (section_nr_to_pfn(pnum
)));
88 * We need this if we ever free the mem_maps. While not implemented yet,
89 * this function is included for parity with its sibling.
91 static __attribute((unused
))
92 struct page
*sparse_decode_mem_map(unsigned long coded_mem_map
, unsigned long pnum
)
94 return ((struct page
*)coded_mem_map
) + section_nr_to_pfn(pnum
);
97 static int sparse_init_one_section(struct mem_section
*ms
,
98 unsigned long pnum
, struct page
*mem_map
)
100 if (!valid_section(ms
))
103 ms
->section_mem_map
|= sparse_encode_mem_map(mem_map
, pnum
);
108 static struct page
*sparse_early_mem_map_alloc(unsigned long pnum
)
111 int nid
= early_pfn_to_nid(section_nr_to_pfn(pnum
));
112 struct mem_section
*ms
= __nr_to_section(pnum
);
114 map
= alloc_remap(nid
, sizeof(struct page
) * PAGES_PER_SECTION
);
118 map
= alloc_bootmem_node(NODE_DATA(nid
),
119 sizeof(struct page
) * PAGES_PER_SECTION
);
123 printk(KERN_WARNING
"%s: allocation failed\n", __FUNCTION__
);
124 ms
->section_mem_map
= 0;
129 * Allocate the accumulated non-linear sections, allocate a mem_map
130 * for each and record the physical to section mapping.
132 void sparse_init(void)
137 for (pnum
= 0; pnum
< NR_MEM_SECTIONS
; pnum
++) {
138 if (!valid_section_nr(pnum
))
141 map
= sparse_early_mem_map_alloc(pnum
);
144 sparse_init_one_section(__nr_to_section(pnum
), pnum
, map
);
149 * returns the number of sections whose mem_maps were properly
150 * set. If this is <=0, then that means that the passed-in
151 * map was not consumed and must be freed.
153 int sparse_add_one_section(unsigned long start_pfn
, int nr_pages
, struct page
*map
)
155 struct mem_section
*ms
= __pfn_to_section(start_pfn
);
157 if (ms
->section_mem_map
& SECTION_MARKED_PRESENT
)
160 ms
->section_mem_map
|= SECTION_MARKED_PRESENT
;
162 return sparse_init_one_section(ms
, pfn_to_section_nr(start_pfn
), map
);