@@ -40,13 +40,13 @@ static __always_inline unsigned long __percpu_##op(void *ptr, \
40
40
switch (size) { \
41
41
case 4: \
42
42
__asm__ __volatile__( \
43
- "am"#asm_op".w" " %[ret], %[val], %[ptr] \n" \
43
+ "am"#asm_op".w" " %[ret], %[val], %[ptr] \n" \
44
44
: [ret] "=&r" (ret), [ptr] "+ZB"(*(u32 *)ptr) \
45
45
: [val] "r" (val)); \
46
46
break; \
47
47
case 8: \
48
48
__asm__ __volatile__( \
49
- "am"#asm_op".d" " %[ret], %[val], %[ptr] \n" \
49
+ "am"#asm_op".d" " %[ret], %[val], %[ptr] \n" \
50
50
: [ret] "=&r" (ret), [ptr] "+ZB"(*(u64 *)ptr) \
51
51
: [val] "r" (val)); \
52
52
break; \
@@ -63,7 +63,7 @@ PERCPU_OP(and, and, &)
63
63
PERCPU_OP (or , or , |)
64
64
#undef PERCPU_OP
65
65
66
- static __always_inline unsigned long __percpu_read (void * ptr , int size )
66
+ static __always_inline unsigned long __percpu_read (void __percpu * ptr , int size )
67
67
{
68
68
unsigned long ret ;
69
69
@@ -100,7 +100,7 @@ static __always_inline unsigned long __percpu_read(void *ptr, int size)
100
100
return ret ;
101
101
}
102
102
103
- static __always_inline void __percpu_write (void * ptr , unsigned long val , int size )
103
+ static __always_inline void __percpu_write (void __percpu * ptr , unsigned long val , int size )
104
104
{
105
105
switch (size ) {
106
106
case 1 :
@@ -132,8 +132,7 @@ static __always_inline void __percpu_write(void *ptr, unsigned long val, int siz
132
132
}
133
133
}
134
134
135
- static __always_inline unsigned long __percpu_xchg (void * ptr , unsigned long val ,
136
- int size )
135
+ static __always_inline unsigned long __percpu_xchg (void * ptr , unsigned long val , int size )
137
136
{
138
137
switch (size ) {
139
138
case 1 :
0 commit comments