projects
/
linux-2.6-microblaze.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
tools headers UAPI: Synch KVM's svm.h header with the kernel
[linux-2.6-microblaze.git]
/
lib
/
atomic64.c
diff --git
a/lib/atomic64.c
b/lib/atomic64.c
index
e98c85a
..
3df6539
100644
(file)
--- a/
lib/atomic64.c
+++ b/
lib/atomic64.c
@@
-42,7
+42,7
@@
static inline raw_spinlock_t *lock_addr(const atomic64_t *v)
return &atomic64_lock[addr & (NR_LOCKS - 1)].lock;
}
return &atomic64_lock[addr & (NR_LOCKS - 1)].lock;
}
-s64 atomic64_read(const atomic64_t *v)
+s64
generic_
atomic64_read(const atomic64_t *v)
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
@@
-53,9
+53,9
@@
s64 atomic64_read(const atomic64_t *v)
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
-EXPORT_SYMBOL(atomic64_read);
+EXPORT_SYMBOL(
generic_
atomic64_read);
-void atomic64_set(atomic64_t *v, s64 i)
+void
generic_
atomic64_set(atomic64_t *v, s64 i)
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
@@
-64,10
+64,10
@@
void atomic64_set(atomic64_t *v, s64 i)
v->counter = i;
raw_spin_unlock_irqrestore(lock, flags);
}
v->counter = i;
raw_spin_unlock_irqrestore(lock, flags);
}
-EXPORT_SYMBOL(atomic64_set);
+EXPORT_SYMBOL(
generic_
atomic64_set);
#define ATOMIC64_OP(op, c_op) \
#define ATOMIC64_OP(op, c_op) \
-void
atomic64_##op(s64 a, atomic64_t *v)
\
+void
generic_atomic64_##op(s64 a, atomic64_t *v)
\
{ \
unsigned long flags; \
raw_spinlock_t *lock = lock_addr(v); \
{ \
unsigned long flags; \
raw_spinlock_t *lock = lock_addr(v); \
@@
-76,10
+76,10
@@
void atomic64_##op(s64 a, atomic64_t *v) \
v->counter c_op a; \
raw_spin_unlock_irqrestore(lock, flags); \
} \
v->counter c_op a; \
raw_spin_unlock_irqrestore(lock, flags); \
} \
-EXPORT_SYMBOL(atomic64_##op);
+EXPORT_SYMBOL(
generic_
atomic64_##op);
#define ATOMIC64_OP_RETURN(op, c_op) \
#define ATOMIC64_OP_RETURN(op, c_op) \
-s64
atomic64_##op##_return(s64 a, atomic64_t *v)
\
+s64
generic_atomic64_##op##_return(s64 a, atomic64_t *v)
\
{ \
unsigned long flags; \
raw_spinlock_t *lock = lock_addr(v); \
{ \
unsigned long flags; \
raw_spinlock_t *lock = lock_addr(v); \
@@
-90,10
+90,10
@@
s64 atomic64_##op##_return(s64 a, atomic64_t *v) \
raw_spin_unlock_irqrestore(lock, flags); \
return val; \
} \
raw_spin_unlock_irqrestore(lock, flags); \
return val; \
} \
-EXPORT_SYMBOL(atomic64_##op##_return);
+EXPORT_SYMBOL(
generic_
atomic64_##op##_return);
#define ATOMIC64_FETCH_OP(op, c_op) \
#define ATOMIC64_FETCH_OP(op, c_op) \
-s64
atomic64_fetch_##op(s64 a, atomic64_t *v)
\
+s64
generic_atomic64_fetch_##op(s64 a, atomic64_t *v)
\
{ \
unsigned long flags; \
raw_spinlock_t *lock = lock_addr(v); \
{ \
unsigned long flags; \
raw_spinlock_t *lock = lock_addr(v); \
@@
-105,7
+105,7
@@
s64 atomic64_fetch_##op(s64 a, atomic64_t *v) \
raw_spin_unlock_irqrestore(lock, flags); \
return val; \
} \
raw_spin_unlock_irqrestore(lock, flags); \
return val; \
} \
-EXPORT_SYMBOL(atomic64_fetch_##op);
+EXPORT_SYMBOL(
generic_
atomic64_fetch_##op);
#define ATOMIC64_OPS(op, c_op) \
ATOMIC64_OP(op, c_op) \
#define ATOMIC64_OPS(op, c_op) \
ATOMIC64_OP(op, c_op) \
@@
-130,7
+130,7
@@
ATOMIC64_OPS(xor, ^=)
#undef ATOMIC64_OP_RETURN
#undef ATOMIC64_OP
#undef ATOMIC64_OP_RETURN
#undef ATOMIC64_OP
-s64 atomic64_dec_if_positive(atomic64_t *v)
+s64
generic_
atomic64_dec_if_positive(atomic64_t *v)
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
@@
-143,9
+143,9
@@
s64 atomic64_dec_if_positive(atomic64_t *v)
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
-EXPORT_SYMBOL(atomic64_dec_if_positive);
+EXPORT_SYMBOL(
generic_
atomic64_dec_if_positive);
-s64 atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
+s64
generic_
atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
@@
-158,9
+158,9
@@
s64 atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
-EXPORT_SYMBOL(atomic64_cmpxchg);
+EXPORT_SYMBOL(
generic_
atomic64_cmpxchg);
-s64 atomic64_xchg(atomic64_t *v, s64 new)
+s64
generic_
atomic64_xchg(atomic64_t *v, s64 new)
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
@@
-172,9
+172,9
@@
s64 atomic64_xchg(atomic64_t *v, s64 new)
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
raw_spin_unlock_irqrestore(lock, flags);
return val;
}
-EXPORT_SYMBOL(atomic64_xchg);
+EXPORT_SYMBOL(
generic_
atomic64_xchg);
-s64 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
+s64
generic_
atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
{
unsigned long flags;
raw_spinlock_t *lock = lock_addr(v);
@@
-188,4
+188,4
@@
s64 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
return val;
}
return val;
}
-EXPORT_SYMBOL(atomic64_fetch_add_unless);
+EXPORT_SYMBOL(
generic_
atomic64_fetch_add_unless);