Home | History | Annotate | Download | only in asm
      1 /*
      2  * bitops.h: Bit string operations on the ppc
      3  */
      4 
      5 #ifndef _PPC_BITOPS_H
      6 #define _PPC_BITOPS_H
      7 
      8 #include <asm/byteorder.h>
      9 #include <asm-generic/bitops/__ffs.h>
     10 
     11 /*
     12  * Arguably these bit operations don't imply any memory barrier or
     13  * SMP ordering, but in fact a lot of drivers expect them to imply
     14  * both, since they do on x86 cpus.
     15  */
     16 #ifdef CONFIG_SMP
     17 #define SMP_WMB		"eieio\n"
     18 #define SMP_MB		"\nsync"
     19 #else
     20 #define SMP_WMB
     21 #define SMP_MB
     22 #endif /* CONFIG_SMP */
     23 
     24 #define __INLINE_BITOPS	1
     25 
     26 #if __INLINE_BITOPS
     27 /*
     28  * These used to be if'd out here because using : "cc" as a constraint
     29  * resulted in errors from egcs.  Things may be OK with gcc-2.95.
     30  */
     31 static __inline__ void set_bit(int nr, volatile void * addr)
     32 {
     33 	unsigned long old;
     34 	unsigned long mask = 1 << (nr & 0x1f);
     35 	unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
     36 
     37 	__asm__ __volatile__(SMP_WMB "\
     38 1:	lwarx	%0,0,%3\n\
     39 	or	%0,%0,%2\n\
     40 	stwcx.	%0,0,%3\n\
     41 	bne	1b"
     42 	SMP_MB
     43 	: "=&r" (old), "=m" (*p)
     44 	: "r" (mask), "r" (p), "m" (*p)
     45 	: "cc" );
     46 }
     47 
     48 static __inline__ void clear_bit(int nr, volatile void *addr)
     49 {
     50 	unsigned long old;
     51 	unsigned long mask = 1 << (nr & 0x1f);
     52 	unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
     53 
     54 	__asm__ __volatile__(SMP_WMB "\
     55 1:	lwarx	%0,0,%3\n\
     56 	andc	%0,%0,%2\n\
     57 	stwcx.	%0,0,%3\n\
     58 	bne	1b"
     59 	SMP_MB
     60 	: "=&r" (old), "=m" (*p)
     61 	: "r" (mask), "r" (p), "m" (*p)
     62 	: "cc");
     63 }
     64 
     65 static __inline__ void change_bit(int nr, volatile void *addr)
     66 {
     67 	unsigned long old;
     68 	unsigned long mask = 1 << (nr & 0x1f);
     69 	unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
     70 
     71 	__asm__ __volatile__(SMP_WMB "\
     72 1:	lwarx	%0,0,%3\n\
     73 	xor	%0,%0,%2\n\
     74 	stwcx.	%0,0,%3\n\
     75 	bne	1b"
     76 	SMP_MB
     77 	: "=&r" (old), "=m" (*p)
     78 	: "r" (mask), "r" (p), "m" (*p)
     79 	: "cc");
     80 }
     81 
     82 static __inline__ int test_and_set_bit(int nr, volatile void *addr)
     83 {
     84 	unsigned int old, t;
     85 	unsigned int mask = 1 << (nr & 0x1f);
     86 	volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
     87 
     88 	__asm__ __volatile__(SMP_WMB "\
     89 1:	lwarx	%0,0,%4\n\
     90 	or	%1,%0,%3\n\
     91 	stwcx.	%1,0,%4\n\
     92 	bne	1b"
     93 	SMP_MB
     94 	: "=&r" (old), "=&r" (t), "=m" (*p)
     95 	: "r" (mask), "r" (p), "m" (*p)
     96 	: "cc");
     97 
     98 	return (old & mask) != 0;
     99 }
    100 
    101 static __inline__ int test_and_clear_bit(int nr, volatile void *addr)
    102 {
    103 	unsigned int old, t;
    104 	unsigned int mask = 1 << (nr & 0x1f);
    105 	volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
    106 
    107 	__asm__ __volatile__(SMP_WMB "\
    108 1:	lwarx	%0,0,%4\n\
    109 	andc	%1,%0,%3\n\
    110 	stwcx.	%1,0,%4\n\
    111 	bne	1b"
    112 	SMP_MB
    113 	: "=&r" (old), "=&r" (t), "=m" (*p)
    114 	: "r" (mask), "r" (p), "m" (*p)
    115 	: "cc");
    116 
    117 	return (old & mask) != 0;
    118 }
    119 
    120 static __inline__ int test_and_change_bit(int nr, volatile void *addr)
    121 {
    122 	unsigned int old, t;
    123 	unsigned int mask = 1 << (nr & 0x1f);
    124 	volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
    125 
    126 	__asm__ __volatile__(SMP_WMB "\
    127 1:	lwarx	%0,0,%4\n\
    128 	xor	%1,%0,%3\n\
    129 	stwcx.	%1,0,%4\n\
    130 	bne	1b"
    131 	SMP_MB
    132 	: "=&r" (old), "=&r" (t), "=m" (*p)
    133 	: "r" (mask), "r" (p), "m" (*p)
    134 	: "cc");
    135 
    136 	return (old & mask) != 0;
    137 }
    138 #endif /* __INLINE_BITOPS */
    139 
    140 static __inline__ int test_bit(int nr, __const__ volatile void *addr)
    141 {
    142 	__const__ unsigned int *p = (__const__ unsigned int *) addr;
    143 
    144 	return ((p[nr >> 5] >> (nr & 0x1f)) & 1) != 0;
    145 }
    146 
    147 /* Return the bit position of the most significant 1 bit in a word */
    148 /* - the result is undefined when x == 0 */
    149 static __inline__ int __ilog2(unsigned int x)
    150 {
    151 	int lz;
    152 
    153 	asm ("cntlzw %0,%1" : "=r" (lz) : "r" (x));
    154 	return 31 - lz;
    155 }
    156 
    157 static __inline__ int ffz(unsigned int x)
    158 {
    159 	if ((x = ~x) == 0)
    160 		return 32;
    161 	return __ilog2(x & -x);
    162 }
    163 
    164 /*
    165  * fls: find last (most-significant) bit set.
    166  * Note fls(0) = 0, fls(1) = 1, fls(0x80000000) = 32.
    167  *
    168  * On powerpc, __ilog2(0) returns -1, but this is not safe in general
    169  */
    170 static __inline__ int fls(unsigned int x)
    171 {
    172 	return __ilog2(x) + 1;
    173 }
    174 #define PLATFORM_FLS
    175 
    176 /**
    177  * fls64 - find last set bit in a 64-bit word
    178  * @x: the word to search
    179  *
    180  * This is defined in a similar way as the libc and compiler builtin
    181  * ffsll, but returns the position of the most significant set bit.
    182  *
    183  * fls64(value) returns 0 if value is 0 or the position of the last
    184  * set bit if value is nonzero. The last (most significant) bit is
    185  * at position 64.
    186  */
    187 #if BITS_PER_LONG == 32
    188 static inline int fls64(__u64 x)
    189 {
    190 	__u32 h = x >> 32;
    191 	if (h)
    192 		return fls(h) + 32;
    193 	return fls(x);
    194 }
    195 #elif BITS_PER_LONG == 64
    196 static inline int fls64(__u64 x)
    197 {
    198 	if (x == 0)
    199 		return 0;
    200 	return __ilog2(x) + 1;
    201 }
    202 #else
    203 #error BITS_PER_LONG not 32 or 64
    204 #endif
    205 
    206 #ifdef __KERNEL__
    207 
    208 /*
    209  * ffs: find first bit set. This is defined the same way as
    210  * the libc and compiler builtin ffs routines, therefore
    211  * differs in spirit from the above ffz (man ffs).
    212  */
    213 static __inline__ int ffs(int x)
    214 {
    215 	return __ilog2(x & -x) + 1;
    216 }
    217 #define PLATFORM_FFS
    218 
    219 /*
    220  * hweightN: returns the hamming weight (i.e. the number
    221  * of bits set) of a N-bit word
    222  */
    223 
    224 #define hweight32(x) generic_hweight32(x)
    225 #define hweight16(x) generic_hweight16(x)
    226 #define hweight8(x) generic_hweight8(x)
    227 
    228 #endif /* __KERNEL__ */
    229 
    230 /*
    231  * This implementation of find_{first,next}_zero_bit was stolen from
    232  * Linus' asm-alpha/bitops.h.
    233  */
    234 #define find_first_zero_bit(addr, size) \
    235 	find_next_zero_bit((addr), (size), 0)
    236 
    237 static __inline__ unsigned long find_next_zero_bit(void * addr,
    238 	unsigned long size, unsigned long offset)
    239 {
    240 	unsigned int * p = ((unsigned int *) addr) + (offset >> 5);
    241 	unsigned int result = offset & ~31UL;
    242 	unsigned int tmp;
    243 
    244 	if (offset >= size)
    245 		return size;
    246 	size -= result;
    247 	offset &= 31UL;
    248 	if (offset) {
    249 		tmp = *p++;
    250 		tmp |= ~0UL >> (32-offset);
    251 		if (size < 32)
    252 			goto found_first;
    253 		if (tmp != ~0U)
    254 			goto found_middle;
    255 		size -= 32;
    256 		result += 32;
    257 	}
    258 	while (size >= 32) {
    259 		if ((tmp = *p++) != ~0U)
    260 			goto found_middle;
    261 		result += 32;
    262 		size -= 32;
    263 	}
    264 	if (!size)
    265 		return result;
    266 	tmp = *p;
    267 found_first:
    268 	tmp |= ~0UL << size;
    269 found_middle:
    270 	return result + ffz(tmp);
    271 }
    272 
    273 
    274 #define _EXT2_HAVE_ASM_BITOPS_
    275 
    276 #ifdef __KERNEL__
    277 /*
    278  * test_and_{set,clear}_bit guarantee atomicity without
    279  * disabling interrupts.
    280  */
    281 #define ext2_set_bit(nr, addr)		test_and_set_bit((nr) ^ 0x18, addr)
    282 #define ext2_clear_bit(nr, addr)	test_and_clear_bit((nr) ^ 0x18, addr)
    283 
    284 #else
    285 static __inline__ int ext2_set_bit(int nr, void * addr)
    286 {
    287 	int		mask;
    288 	unsigned char	*ADDR = (unsigned char *) addr;
    289 	int oldbit;
    290 
    291 	ADDR += nr >> 3;
    292 	mask = 1 << (nr & 0x07);
    293 	oldbit = (*ADDR & mask) ? 1 : 0;
    294 	*ADDR |= mask;
    295 	return oldbit;
    296 }
    297 
    298 static __inline__ int ext2_clear_bit(int nr, void * addr)
    299 {
    300 	int		mask;
    301 	unsigned char	*ADDR = (unsigned char *) addr;
    302 	int oldbit;
    303 
    304 	ADDR += nr >> 3;
    305 	mask = 1 << (nr & 0x07);
    306 	oldbit = (*ADDR & mask) ? 1 : 0;
    307 	*ADDR = *ADDR & ~mask;
    308 	return oldbit;
    309 }
    310 #endif	/* __KERNEL__ */
    311 
    312 static __inline__ int ext2_test_bit(int nr, __const__ void * addr)
    313 {
    314 	__const__ unsigned char	*ADDR = (__const__ unsigned char *) addr;
    315 
    316 	return (ADDR[nr >> 3] >> (nr & 7)) & 1;
    317 }
    318 
    319 /*
    320  * This implementation of ext2_find_{first,next}_zero_bit was stolen from
    321  * Linus' asm-alpha/bitops.h and modified for a big-endian machine.
    322  */
    323 
    324 #define ext2_find_first_zero_bit(addr, size) \
    325 	ext2_find_next_zero_bit((addr), (size), 0)
    326 
    327 static __inline__ unsigned long ext2_find_next_zero_bit(void *addr,
    328 	unsigned long size, unsigned long offset)
    329 {
    330 	unsigned int *p = ((unsigned int *) addr) + (offset >> 5);
    331 	unsigned int result = offset & ~31UL;
    332 	unsigned int tmp;
    333 
    334 	if (offset >= size)
    335 		return size;
    336 	size -= result;
    337 	offset &= 31UL;
    338 	if (offset) {
    339 		tmp = cpu_to_le32p(p++);
    340 		tmp |= ~0UL >> (32-offset);
    341 		if (size < 32)
    342 			goto found_first;
    343 		if (tmp != ~0U)
    344 			goto found_middle;
    345 		size -= 32;
    346 		result += 32;
    347 	}
    348 	while (size >= 32) {
    349 		if ((tmp = cpu_to_le32p(p++)) != ~0U)
    350 			goto found_middle;
    351 		result += 32;
    352 		size -= 32;
    353 	}
    354 	if (!size)
    355 		return result;
    356 	tmp = cpu_to_le32p(p);
    357 found_first:
    358 	tmp |= ~0U << size;
    359 found_middle:
    360 	return result + ffz(tmp);
    361 }
    362 
    363 /* Bitmap functions for the minix filesystem.  */
    364 #define minix_test_and_set_bit(nr,addr) ext2_set_bit(nr,addr)
    365 #define minix_set_bit(nr,addr) ((void)ext2_set_bit(nr,addr))
    366 #define minix_test_and_clear_bit(nr,addr) ext2_clear_bit(nr,addr)
    367 #define minix_test_bit(nr,addr) ext2_test_bit(nr,addr)
    368 #define minix_find_first_zero_bit(addr,size) ext2_find_first_zero_bit(addr,size)
    369 
    370 #endif /* _PPC_BITOPS_H */
    371