1 #ifndef _ASM_POWERPC_PERCPU_H_
2 #define _ASM_POWERPC_PERCPU_H_
4 #include <linux/compiler.h>
7 * Same as asm-generic/percpu.h, except that we store the per cpu offset
8 * in the paca. Based on the x86-64 implementation.
15 #define __per_cpu_offset(cpu) (paca[cpu].data_offset)
16 #define __my_cpu_offset() get_paca()->data_offset
17 #define per_cpu_offset(x) (__per_cpu_offset(x))
19 /* var is in discarded region: offset to particular copy we want */
20 #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu)))
21 #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset()))
22 #define __raw_get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, local_paca->data_offset))
24 /* A macro to avoid #include hell... */
25 #define percpu_modcopy(pcpudst, src, size) \
28 for_each_possible_cpu(__i) \
29 memcpy((pcpudst)+__per_cpu_offset(__i), \
33 extern void setup_per_cpu_areas(void);
37 #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var))
38 #define __get_cpu_var(var) per_cpu__##var
39 #define __raw_get_cpu_var(var) per_cpu__##var
43 #define DECLARE_PER_CPU(type, name) extern __typeof__(type) per_cpu__##name
46 #include <asm-generic/percpu.h>
49 #endif /* _ASM_POWERPC_PERCPU_H_ */