@@ -33,7 +33,7 @@ static union {
3333 char pad [L1_CACHE_BYTES ];
3434} atomic64_lock [NR_LOCKS ] __cacheline_aligned_in_smp ;
3535
36- static inline spinlock_t * lock_addr (const atomic64_t * v )
36+ static inline raw_spinlock_t * lock_addr (const atomic64_t * v )
3737{
3838 unsigned long addr = (unsigned long ) v ;
3939
@@ -45,7 +45,7 @@ static inline spinlock_t *lock_addr(const atomic64_t *v)
4545long long atomic64_read (const atomic64_t * v )
4646{
4747 unsigned long flags ;
48- spinlock_t * lock = lock_addr (v );
48+ raw_spinlock_t * lock = lock_addr (v );
4949 long long val ;
5050
5151 raw_spin_lock_irqsave (lock , flags );
@@ -58,7 +58,7 @@ EXPORT_SYMBOL(atomic64_read);
5858void atomic64_set (atomic64_t * v , long long i )
5959{
6060 unsigned long flags ;
61- spinlock_t * lock = lock_addr (v );
61+ raw_spinlock_t * lock = lock_addr (v );
6262
6363 raw_spin_lock_irqsave (lock , flags );
6464 v -> counter = i ;
@@ -69,7 +69,7 @@ EXPORT_SYMBOL(atomic64_set);
6969void atomic64_add (long long a , atomic64_t * v )
7070{
7171 unsigned long flags ;
72- spinlock_t * lock = lock_addr (v );
72+ raw_spinlock_t * lock = lock_addr (v );
7373
7474 raw_spin_lock_irqsave (lock , flags );
7575 v -> counter += a ;
@@ -80,7 +80,7 @@ EXPORT_SYMBOL(atomic64_add);
8080long long atomic64_add_return (long long a , atomic64_t * v )
8181{
8282 unsigned long flags ;
83- spinlock_t * lock = lock_addr (v );
83+ raw_spinlock_t * lock = lock_addr (v );
8484 long long val ;
8585
8686 raw_spin_lock_irqsave (lock , flags );
@@ -93,7 +93,7 @@ EXPORT_SYMBOL(atomic64_add_return);
9393void atomic64_sub (long long a , atomic64_t * v )
9494{
9595 unsigned long flags ;
96- spinlock_t * lock = lock_addr (v );
96+ raw_spinlock_t * lock = lock_addr (v );
9797
9898 raw_spin_lock_irqsave (lock , flags );
9999 v -> counter -= a ;
@@ -104,7 +104,7 @@ EXPORT_SYMBOL(atomic64_sub);
104104long long atomic64_sub_return (long long a , atomic64_t * v )
105105{
106106 unsigned long flags ;
107- spinlock_t * lock = lock_addr (v );
107+ raw_spinlock_t * lock = lock_addr (v );
108108 long long val ;
109109
110110 raw_spin_lock_irqsave (lock , flags );
@@ -117,7 +117,7 @@ EXPORT_SYMBOL(atomic64_sub_return);
117117long long atomic64_dec_if_positive (atomic64_t * v )
118118{
119119 unsigned long flags ;
120- spinlock_t * lock = lock_addr (v );
120+ raw_spinlock_t * lock = lock_addr (v );
121121 long long val ;
122122
123123 raw_spin_lock_irqsave (lock , flags );
@@ -132,7 +132,7 @@ EXPORT_SYMBOL(atomic64_dec_if_positive);
132132long long atomic64_cmpxchg (atomic64_t * v , long long o , long long n )
133133{
134134 unsigned long flags ;
135- spinlock_t * lock = lock_addr (v );
135+ raw_spinlock_t * lock = lock_addr (v );
136136 long long val ;
137137
138138 raw_spin_lock_irqsave (lock , flags );
@@ -147,7 +147,7 @@ EXPORT_SYMBOL(atomic64_cmpxchg);
147147long long atomic64_xchg (atomic64_t * v , long long new )
148148{
149149 unsigned long flags ;
150- spinlock_t * lock = lock_addr (v );
150+ raw_spinlock_t * lock = lock_addr (v );
151151 long long val ;
152152
153153 raw_spin_lock_irqsave (lock , flags );
@@ -161,7 +161,7 @@ EXPORT_SYMBOL(atomic64_xchg);
161161int atomic64_add_unless (atomic64_t * v , long long a , long long u )
162162{
163163 unsigned long flags ;
164- spinlock_t * lock = lock_addr (v );
164+ raw_spinlock_t * lock = lock_addr (v );
165165 int ret = 0 ;
166166
167167 raw_spin_lock_irqsave (lock , flags );
0 commit comments