1 #ifndef __ARCH_SPARC64_PERCPU__
2 #define __ARCH_SPARC64_PERCPU__
4 #include <linux/compiler.h>
9 # define PERCPU_MODULE_RESERVE 8192
11 # define PERCPU_MODULE_RESERVE 0
14 #define PERCPU_ENOUGH_ROOM \
15 (ALIGN(__per_cpu_end - __per_cpu_start, SMP_CACHE_BYTES) + \
16 PERCPU_MODULE_RESERVE)
18 extern void setup_per_cpu_areas(void);
20 extern unsigned long __per_cpu_base
;
21 extern unsigned long __per_cpu_shift
;
22 #define __per_cpu_offset(__cpu) \
23 (__per_cpu_base + ((unsigned long)(__cpu) << __per_cpu_shift))
24 #define per_cpu_offset(x) (__per_cpu_offset(x))
26 /* Separate out the type, so (int[3], foo) works. */
27 #define DEFINE_PER_CPU(type, name) \
28 __attribute__((__section__(".data.percpu"))) __typeof__(type) per_cpu__##name
30 register unsigned long __local_per_cpu_offset
asm("g5");
32 /* var is in discarded region: offset to particular copy we want */
33 #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu)))
34 #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __local_per_cpu_offset))
35 #define __raw_get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __local_per_cpu_offset))
37 /* A macro to avoid #include hell... */
38 #define percpu_modcopy(pcpudst, src, size) \
41 for_each_possible_cpu(__i) \
42 memcpy((pcpudst)+__per_cpu_offset(__i), \
47 #define DEFINE_PER_CPU(type, name) \
48 __typeof__(type) per_cpu__##name
50 #define per_cpu(var, cpu) (*((void)cpu, &per_cpu__##var))
51 #define __get_cpu_var(var) per_cpu__##var
52 #define __raw_get_cpu_var(var) per_cpu__##var
56 #define DECLARE_PER_CPU(type, name) extern __typeof__(type) per_cpu__##name
58 #define EXPORT_PER_CPU_SYMBOL(var) EXPORT_SYMBOL(per_cpu__##var)
59 #define EXPORT_PER_CPU_SYMBOL_GPL(var) EXPORT_SYMBOL_GPL(per_cpu__##var)
61 #endif /* __ARCH_SPARC64_PERCPU__ */