2 * Copyright 2011 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com>
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License, version 2, as
6 * published by the Free Software Foundation.
9 #include <linux/kvm_host.h>
10 #include <linux/preempt.h>
11 #include <linux/sched.h>
12 #include <linux/spinlock.h>
13 #include <linux/bootmem.h>
14 #include <linux/init.h>
16 #include <asm/cputable.h>
17 #include <asm/kvm_ppc.h>
18 #include <asm/kvm_book3s.h>
21 * This maintains a list of RMAs (real mode areas) for KVM guests to use.
22 * Each RMA has to be physically contiguous and of a size that the
23 * hardware supports. PPC970 and POWER7 support 64MB, 128MB and 256MB,
24 * and other larger sizes. Since we are unlikely to be allocate that
25 * much physically contiguous memory after the system is up and running,
26 * we preallocate a set of RMAs in early boot for KVM to use.
28 static unsigned long kvm_rma_size
= 64 << 20; /* 64MB */
29 static unsigned long kvm_rma_count
;
31 static int __init
early_parse_rma_size(char *p
)
36 kvm_rma_size
= memparse(p
, &p
);
40 early_param("kvm_rma_size", early_parse_rma_size
);
42 static int __init
early_parse_rma_count(char *p
)
47 kvm_rma_count
= simple_strtoul(p
, NULL
, 0);
51 early_param("kvm_rma_count", early_parse_rma_count
);
53 static struct kvmppc_rma_info
*rma_info
;
54 static LIST_HEAD(free_rmas
);
55 static DEFINE_SPINLOCK(rma_lock
);
57 /* Work out RMLS (real mode limit selector) field value for a given RMA size.
59 static inline int lpcr_rmls(unsigned long rma_size
)
62 case 32ul << 20: /* 32 MB */
64 case 64ul << 20: /* 64 MB */
66 case 128ul << 20: /* 128 MB */
68 case 256ul << 20: /* 256 MB */
70 case 1ul << 30: /* 1 GB */
72 case 16ul << 30: /* 16 GB */
74 case 256ul << 30: /* 256 GB */
82 * Called at boot time while the bootmem allocator is active,
83 * to allocate contiguous physical memory for the real memory
86 void kvm_rma_init(void)
89 unsigned long j
, npages
;
93 /* Only do this on POWER7 in HV mode */
94 if (!cpu_has_feature(CPU_FTR_HVMODE_206
))
97 if (!kvm_rma_size
|| !kvm_rma_count
)
100 /* Check that the requested size is one supported in hardware */
101 if (lpcr_rmls(kvm_rma_size
) < 0) {
102 pr_err("RMA size of 0x%lx not supported\n", kvm_rma_size
);
106 npages
= kvm_rma_size
>> PAGE_SHIFT
;
107 rma_info
= alloc_bootmem(kvm_rma_count
* sizeof(struct kvmppc_rma_info
));
108 for (i
= 0; i
< kvm_rma_count
; ++i
) {
109 rma
= alloc_bootmem_align(kvm_rma_size
, kvm_rma_size
);
110 pr_info("Allocated KVM RMA at %p (%ld MB)\n", rma
,
112 rma_info
[i
].base_virt
= rma
;
113 rma_info
[i
].base_pfn
= __pa(rma
) >> PAGE_SHIFT
;
114 rma_info
[i
].npages
= npages
;
115 list_add_tail(&rma_info
[i
].list
, &free_rmas
);
116 atomic_set(&rma_info
[i
].use_count
, 0);
118 pg
= pfn_to_page(rma_info
[i
].base_pfn
);
119 for (j
= 0; j
< npages
; ++j
) {
120 atomic_inc(&pg
->_count
);
126 struct kvmppc_rma_info
*kvm_alloc_rma(void)
128 struct kvmppc_rma_info
*ri
;
131 spin_lock(&rma_lock
);
132 if (!list_empty(&free_rmas
)) {
133 ri
= list_first_entry(&free_rmas
, struct kvmppc_rma_info
, list
);
135 atomic_inc(&ri
->use_count
);
137 spin_unlock(&rma_lock
);
140 EXPORT_SYMBOL_GPL(kvm_alloc_rma
);
142 void kvm_release_rma(struct kvmppc_rma_info
*ri
)
144 if (atomic_dec_and_test(&ri
->use_count
)) {
145 spin_lock(&rma_lock
);
146 list_add_tail(&ri
->list
, &free_rmas
);
147 spin_unlock(&rma_lock
);
151 EXPORT_SYMBOL_GPL(kvm_release_rma
);