git://git.onelab.eu
/
linux-2.6.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
vserver 1.9.5.x5
[linux-2.6.git]
/
include
/
asm-parisc
/
bitops.h
diff --git
a/include/asm-parisc/bitops.h
b/include/asm-parisc/bitops.h
index
914f70c
..
5badc61
100644
(file)
--- a/
include/asm-parisc/bitops.h
+++ b/
include/asm-parisc/bitops.h
@@
-38,9
+38,9
@@
static __inline__ void set_bit(int nr, volatile unsigned long * address)
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
-
atomic_spin_lock_irqsave(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_lock_irqsave(addr
, flags);
*addr |= mask;
*addr |= mask;
-
atomic_spin_unlock_irqrestore(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_unlock_irqrestore(addr
, flags);
}
static __inline__ void __set_bit(int nr, volatile unsigned long * address)
}
static __inline__ void __set_bit(int nr, volatile unsigned long * address)
@@
-61,9
+61,9
@@
static __inline__ void clear_bit(int nr, volatile unsigned long * address)
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
-
atomic_spin_lock_irqsave(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_lock_irqsave(addr
, flags);
*addr &= ~mask;
*addr &= ~mask;
-
atomic_spin_unlock_irqrestore(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_unlock_irqrestore(addr
, flags);
}
static __inline__ void __clear_bit(unsigned long nr, volatile unsigned long * address)
}
static __inline__ void __clear_bit(unsigned long nr, volatile unsigned long * address)
@@
-84,9
+84,9
@@
static __inline__ void change_bit(int nr, volatile unsigned long * address)
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
-
atomic_spin_lock_irqsave(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_lock_irqsave(addr
, flags);
*addr ^= mask;
*addr ^= mask;
-
atomic_spin_unlock_irqrestore(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_unlock_irqrestore(addr
, flags);
}
static __inline__ void __change_bit(int nr, volatile unsigned long * address)
}
static __inline__ void __change_bit(int nr, volatile unsigned long * address)
@@
-108,10
+108,10
@@
static __inline__ int test_and_set_bit(int nr, volatile unsigned long * address)
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
-
atomic_spin_lock_irqsave(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_lock_irqsave(addr
, flags);
oldbit = (*addr & mask) ? 1 : 0;
*addr |= mask;
oldbit = (*addr & mask) ? 1 : 0;
*addr |= mask;
-
atomic_spin_unlock_irqrestore(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_unlock_irqrestore(addr
, flags);
return oldbit;
}
return oldbit;
}
@@
-139,10
+139,10
@@
static __inline__ int test_and_clear_bit(int nr, volatile unsigned long * addres
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
-
atomic_spin_lock_irqsave(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_lock_irqsave(addr
, flags);
oldbit = (*addr & mask) ? 1 : 0;
*addr &= ~mask;
oldbit = (*addr & mask) ? 1 : 0;
*addr &= ~mask;
-
atomic_spin_unlock_irqrestore(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_unlock_irqrestore(addr
, flags);
return oldbit;
}
return oldbit;
}
@@
-170,10
+170,10
@@
static __inline__ int test_and_change_bit(int nr, volatile unsigned long * addre
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
addr += (nr >> SHIFT_PER_LONG);
mask = 1L << CHOP_SHIFTCOUNT(nr);
-
atomic_spin_lock_irqsave(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_lock_irqsave(addr
, flags);
oldbit = (*addr & mask) ? 1 : 0;
*addr ^= mask;
oldbit = (*addr & mask) ? 1 : 0;
*addr ^= mask;
-
atomic_spin_unlock_irqrestore(ATOMIC_HASH(addr)
, flags);
+
_atomic_spin_unlock_irqrestore(addr
, flags);
return oldbit;
}
return oldbit;
}