2008-02-08 15:19:31 +03:00
/* MN10300 bit operations
*
* Copyright ( C ) 2007 Red Hat , Inc . All Rights Reserved .
* Written by David Howells ( dhowells @ redhat . com )
*
* This program is free software ; you can redistribute it and / or
* modify it under the terms of the GNU General Public Licence
* as published by the Free Software Foundation ; either version
* 2 of the Licence , or ( at your option ) any later version .
*
* These have to be done with inline assembly : that way the bit - setting
* is guaranteed to be atomic . All bit operations return 0 if the bit
* was cleared before the operation and ! = 0 if it was not .
*
* bit 0 is the LSB of addr ; bit 32 is the LSB of ( addr + 1 ) .
*/
# ifndef __ASM_BITOPS_H
# define __ASM_BITOPS_H
# include <asm/cpu-regs.h>
# define smp_mb__before_clear_bit() barrier()
# define smp_mb__after_clear_bit() barrier()
/*
* set bit
*/
# define __set_bit(nr, addr) \
( { \
volatile unsigned char * _a = ( unsigned char * ) ( addr ) ; \
const unsigned shift = ( nr ) & 7 ; \
_a + = ( nr ) > > 3 ; \
\
asm volatile ( " bset %2,(%1) # set_bit reg " \
: " =m " ( * _a ) \
: " a " ( _a ) , " d " ( 1 < < shift ) , " m " ( * _a ) \
: " memory " , " cc " ) ; \
} )
# define set_bit(nr, addr) __set_bit((nr), (addr))
/*
* clear bit
*/
# define ___clear_bit(nr, addr) \
( { \
volatile unsigned char * _a = ( unsigned char * ) ( addr ) ; \
const unsigned shift = ( nr ) & 7 ; \
_a + = ( nr ) > > 3 ; \
\
asm volatile ( " bclr %2,(%1) # clear_bit reg " \
: " =m " ( * _a ) \
: " a " ( _a ) , " d " ( 1 < < shift ) , " m " ( * _a ) \
: " memory " , " cc " ) ; \
} )
# define clear_bit(nr, addr) ___clear_bit((nr), (addr))
static inline void __clear_bit ( int nr , volatile void * addr )
{
unsigned int * a = ( unsigned int * ) addr ;
int mask ;
a + = nr > > 5 ;
mask = 1 < < ( nr & 0x1f ) ;
* a & = ~ mask ;
}
/*
* test bit
*/
static inline int test_bit ( int nr , const volatile void * addr )
{
return 1UL & ( ( ( const unsigned int * ) addr ) [ nr > > 5 ] > > ( nr & 31 ) ) ;
}
/*
* change bit
*/
static inline void __change_bit ( int nr , volatile void * addr )
{
int mask ;
unsigned int * a = ( unsigned int * ) addr ;
a + = nr > > 5 ;
mask = 1 < < ( nr & 0x1f ) ;
* a ^ = mask ;
}
extern void change_bit ( int nr , volatile void * addr ) ;
/*
* test and set bit
*/
# define __test_and_set_bit(nr,addr) \
( { \
volatile unsigned char * _a = ( unsigned char * ) ( addr ) ; \
const unsigned shift = ( nr ) & 7 ; \
unsigned epsw ; \
_a + = ( nr ) > > 3 ; \
\
asm volatile ( " bset %3,(%2) # test_set_bit reg \n " \
" mov epsw,%1 " \
: " =m " ( * _a ) , " =d " ( epsw ) \
: " a " ( _a ) , " d " ( 1 < < shift ) , " m " ( * _a ) \
: " memory " , " cc " ) ; \
\
! ( epsw & EPSW_FLAG_Z ) ; \
} )
# define test_and_set_bit(nr, addr) __test_and_set_bit((nr), (addr))
/*
* test and clear bit
*/
# define __test_and_clear_bit(nr, addr) \
( { \
volatile unsigned char * _a = ( unsigned char * ) ( addr ) ; \
const unsigned shift = ( nr ) & 7 ; \
unsigned epsw ; \
_a + = ( nr ) > > 3 ; \
\
asm volatile ( " bclr %3,(%2) # test_clear_bit reg \n " \
" mov epsw,%1 " \
: " =m " ( * _a ) , " =d " ( epsw ) \
: " a " ( _a ) , " d " ( 1 < < shift ) , " m " ( * _a ) \
: " memory " , " cc " ) ; \
\
! ( epsw & EPSW_FLAG_Z ) ; \
} )
# define test_and_clear_bit(nr, addr) __test_and_clear_bit((nr), (addr))
/*
* test and change bit
*/
static inline int __test_and_change_bit ( int nr , volatile void * addr )
{
int mask , retval ;
unsigned int * a = ( unsigned int * ) addr ;
a + = nr > > 5 ;
mask = 1 < < ( nr & 0x1f ) ;
retval = ( mask & * a ) ! = 0 ;
* a ^ = mask ;
return retval ;
}
extern int test_and_change_bit ( int nr , volatile void * addr ) ;
# include <asm-generic/bitops/lock.h>
# ifdef __KERNEL__
/**
* __ffs - find first bit set
* @ x : the word to search
*
* - return 31. .0 to indicate bit 31. .0 most least significant bit set
* - if no bits are set in x , the result is undefined
*/
static inline __attribute__ ( ( const ) )
unsigned long __ffs ( unsigned long x )
{
int bit ;
asm ( " bsch %2,%0 " : " =r " ( bit ) : " 0 " ( 0 ) , " r " ( x & - x ) ) ;
return bit ;
}
/*
* special slimline version of fls ( ) for calculating ilog2_u32 ( )
* - note : no protection against n = = 0
*/
static inline __attribute__ ( ( const ) )
int __ilog2_u32 ( u32 n )
{
int bit ;
asm ( " bsch %2,%0 " : " =r " ( bit ) : " 0 " ( 0 ) , " r " ( n ) ) ;
return bit ;
}
/**
* fls - find last bit set
* @ x : the word to search
*
* This is defined the same way as ffs :
* - return 32. .1 to indicate bit 31. .0 most significant bit set
* - return 0 to indicate no bits set
*/
static inline __attribute__ ( ( const ) )
int fls ( int x )
{
return ( x ! = 0 ) ? __ilog2_u32 ( x ) + 1 : 0 ;
}
2009-01-03 08:49:03 +03:00
/**
* __fls - find last ( most - significant ) set bit in a long word
* @ word : the word to search
*
* Undefined if no set bit exists , so code should check against 0 first .
*/
static inline unsigned long __fls ( unsigned long word )
{
return __ilog2_u32 ( word ) ;
}
2008-02-08 15:19:31 +03:00
/**
* ffs - find first bit set
* @ x : the word to search
*
* - return 32. .1 to indicate bit 31. .0 most least significant bit set
* - return 0 to indicate no bits set
*/
static inline __attribute__ ( ( const ) )
int ffs ( int x )
{
/* Note: (x & -x) gives us a mask that is the least significant
* ( rightmost ) 1 - bit of the value in x .
*/
return fls ( x & - x ) ;
}
# include <asm-generic/bitops/ffz.h>
# include <asm-generic/bitops/fls64.h>
# include <asm-generic/bitops/find.h>
# include <asm-generic/bitops/sched.h>
# include <asm-generic/bitops/hweight.h>
# define ext2_set_bit_atomic(lock, nr, addr) \
test_and_set_bit ( ( nr ) ^ 0x18 , ( addr ) )
# define ext2_clear_bit_atomic(lock, nr, addr) \
test_and_clear_bit ( ( nr ) ^ 0x18 , ( addr ) )
# include <asm-generic/bitops/ext2-non-atomic.h>
# include <asm-generic/bitops/minix-le.h>
# endif /* __KERNEL__ */
# endif /* __ASM_BITOPS_H */