#ifndef __ASM_SH_BITOPS_GRB_H
#define __ASM_SH_BITOPS_GRB_H
static inline void set_bit(int nr, volatile void * addr)
{
int mask;
volatile unsigned int *a = addr;
unsigned long tmp;
a += nr >> 5;
mask = 1 << (nr & 0x1f);
__asm__ __volatile__ (
" .align 2 \n\t"
" mova 1f, r0 \n\t"
" mov r15, r1 \n\t"
" mov #-6, r15 \n\t"
" mov.l @%1, %0 \n\t"
" or %2, %0 \n\t"
" mov.l %0, @%1 \n\t"
"1: mov r1, r15 \n\t"
: "=&r" (tmp),
"+r" (a)
: "r" (mask)
: "memory" , "r0", "r1");
}
static inline void clear_bit(int nr, volatile void * addr)
{
int mask;
volatile unsigned int *a = addr;
unsigned long tmp;
a += nr >> 5;
mask = ~(1 << (nr & 0x1f));
__asm__ __volatile__ (
" .align 2 \n\t"
" mova 1f, r0 \n\t"
" mov r15, r1 \n\t"
" mov #-6, r15 \n\t"
" mov.l @%1, %0 \n\t"
" and %2, %0 \n\t"
" mov.l %0, @%1 \n\t"
"1: mov r1, r15 \n\t"
: "=&r" (tmp),
"+r" (a)
: "r" (mask)
: "memory" , "r0", "r1");
}
static inline void change_bit(int nr, volatile void * addr)
{
int mask;
volatile unsigned int *a = addr;
unsigned long tmp;
a += nr >> 5;
mask = 1 << (nr & 0x1f);
__asm__ __volatile__ (
" .align 2 \n\t"
" mova 1f, r0 \n\t"
" mov r15, r1 \n\t"
" mov #-6, r15 \n\t"
" mov.l @%1, %0 \n\t"
" xor %2, %0 \n\t"
" mov.l %0, @%1 \n\t"
"1: mov r1, r15 \n\t"
: "=&r" (tmp),
"+r" (a)
: "r" (mask)
: "memory" , "r0", "r1");
}
static inline int test_and_set_bit(int nr, volatile void * addr)
{
int mask, retval;
volatile unsigned int *a = addr;
unsigned long tmp;
a += nr >> 5;
mask = 1 << (nr & 0x1f);
__asm__ __volatile__ (
" .align 2 \n\t"
" mova 1f, r0 \n\t"
" mov r15, r1 \n\t"
" mov #-14, r15 \n\t"
" mov.l @%2, %0 \n\t"
" mov %0, %1 \n\t"
" tst %1, %3 \n\t"
" mov #-1, %1 \n\t"
" negc %1, %1 \n\t"
" or %3, %0 \n\t"
" mov.l %0, @%2 \n\t"
"1: mov r1, r15 \n\t"
: "=&r" (tmp),
"=&r" (retval),
"+r" (a)
: "r" (mask)
: "memory" , "r0", "r1" ,"t");
return retval;
}
static inline int test_and_clear_bit(int nr, volatile void * addr)
{
int mask, retval,not_mask;
volatile unsigned int *a = addr;
unsigned long tmp;
a += nr >> 5;
mask = 1 << (nr & 0x1f);
not_mask = ~mask;
__asm__ __volatile__ (
" .align 2 \n\t"
" mova 1f, r0 \n\t"
" mov r15, r1 \n\t"
" mov #-14, r15 \n\t"
" mov.l @%2, %0 \n\t"
" mov %0, %1 \n\t"
" tst %1, %3 \n\t"
" mov #-1, %1 \n\t"
" negc %1, %1 \n\t"
" and %4, %0 \n\t"
" mov.l %0, @%2 \n\t"
"1: mov r1, r15 \n\t"
: "=&r" (tmp),
"=&r" (retval),
"+r" (a)
: "r" (mask),
"r" (not_mask)
: "memory" , "r0", "r1", "t");
return retval;
}
static inline int test_and_change_bit(int nr, volatile void * addr)
{
int mask, retval;
volatile unsigned int *a = addr;
unsigned long tmp;
a += nr >> 5;
mask = 1 << (nr & 0x1f);
__asm__ __volatile__ (
" .align 2 \n\t"
" mova 1f, r0 \n\t"
" mov r15, r1 \n\t"
" mov #-14, r15 \n\t"
" mov.l @%2, %0 \n\t"
" mov %0, %1 \n\t"
" tst %1, %3 \n\t"
" mov #-1, %1 \n\t"
" negc %1, %1 \n\t"
" xor %3, %0 \n\t"
" mov.l %0, @%2 \n\t"
"1: mov r1, r15 \n\t"
: "=&r" (tmp),
"=&r" (retval),
"+r" (a)
: "r" (mask)
: "memory" , "r0", "r1", "t");
return retval;
}
#include <asm-generic/bitops/non-atomic.h>
#endif