1da177e4c3
Initial git repository build. I'm not bothering with the full history, even though we have it. We can create a separate "historical" git archive of that later if we want to, and in the meantime it's about 3.2GB when imported into git - space that would just make the early git days unnecessarily complicated, when we don't have a lot of good infrastructure for it. Let it rip!
111 lines
2.2 KiB
ArmAsm
111 lines
2.2 KiB
ArmAsm
/* bitops.S: Low level assembler bit operations.
|
|
*
|
|
* Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu)
|
|
*/
|
|
|
|
#include <linux/config.h>
|
|
#include <asm/ptrace.h>
|
|
#include <asm/psr.h>
|
|
|
|
.text
|
|
.align 4
|
|
|
|
.globl __bitops_begin
|
|
__bitops_begin:
|
|
|
|
/* Take bits in %g2 and set them in word at %g1,
|
|
* return whether bits were set in original value
|
|
* in %g2. %g4 holds value to restore into %o7
|
|
* in delay slot of jmpl return, %g3 + %g5 + %g7 can be
|
|
* used as temporaries and thus is considered clobbered
|
|
* by all callers.
|
|
*/
|
|
.globl ___set_bit
|
|
___set_bit:
|
|
rd %psr, %g3
|
|
nop; nop; nop;
|
|
or %g3, PSR_PIL, %g5
|
|
wr %g5, 0x0, %psr
|
|
nop; nop; nop
|
|
#ifdef CONFIG_SMP
|
|
set bitops_spinlock, %g5
|
|
2: ldstub [%g5], %g7 ! Spin on the byte lock for SMP.
|
|
orcc %g7, 0x0, %g0 ! Did we get it?
|
|
bne 2b ! Nope...
|
|
#endif
|
|
ld [%g1], %g7
|
|
or %g7, %g2, %g5
|
|
and %g7, %g2, %g2
|
|
#ifdef CONFIG_SMP
|
|
st %g5, [%g1]
|
|
set bitops_spinlock, %g5
|
|
stb %g0, [%g5]
|
|
#else
|
|
st %g5, [%g1]
|
|
#endif
|
|
wr %g3, 0x0, %psr
|
|
nop; nop; nop
|
|
jmpl %o7, %g0
|
|
mov %g4, %o7
|
|
|
|
/* Same as above, but clears the bits from %g2 instead. */
|
|
.globl ___clear_bit
|
|
___clear_bit:
|
|
rd %psr, %g3
|
|
nop; nop; nop
|
|
or %g3, PSR_PIL, %g5
|
|
wr %g5, 0x0, %psr
|
|
nop; nop; nop
|
|
#ifdef CONFIG_SMP
|
|
set bitops_spinlock, %g5
|
|
2: ldstub [%g5], %g7 ! Spin on the byte lock for SMP.
|
|
orcc %g7, 0x0, %g0 ! Did we get it?
|
|
bne 2b ! Nope...
|
|
#endif
|
|
ld [%g1], %g7
|
|
andn %g7, %g2, %g5
|
|
and %g7, %g2, %g2
|
|
#ifdef CONFIG_SMP
|
|
st %g5, [%g1]
|
|
set bitops_spinlock, %g5
|
|
stb %g0, [%g5]
|
|
#else
|
|
st %g5, [%g1]
|
|
#endif
|
|
wr %g3, 0x0, %psr
|
|
nop; nop; nop
|
|
jmpl %o7, %g0
|
|
mov %g4, %o7
|
|
|
|
/* Same thing again, but this time toggles the bits from %g2. */
|
|
.globl ___change_bit
|
|
___change_bit:
|
|
rd %psr, %g3
|
|
nop; nop; nop
|
|
or %g3, PSR_PIL, %g5
|
|
wr %g5, 0x0, %psr
|
|
nop; nop; nop
|
|
#ifdef CONFIG_SMP
|
|
set bitops_spinlock, %g5
|
|
2: ldstub [%g5], %g7 ! Spin on the byte lock for SMP.
|
|
orcc %g7, 0x0, %g0 ! Did we get it?
|
|
bne 2b ! Nope...
|
|
#endif
|
|
ld [%g1], %g7
|
|
xor %g7, %g2, %g5
|
|
and %g7, %g2, %g2
|
|
#ifdef CONFIG_SMP
|
|
st %g5, [%g1]
|
|
set bitops_spinlock, %g5
|
|
stb %g0, [%g5]
|
|
#else
|
|
st %g5, [%g1]
|
|
#endif
|
|
wr %g3, 0x0, %psr
|
|
nop; nop; nop
|
|
jmpl %o7, %g0
|
|
mov %g4, %o7
|
|
|
|
.globl __bitops_end
|
|
__bitops_end:
|