Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
25b160e3
Commit
25b160e3
authored
Feb 03, 2004
by
Yoshinori Sato
Committed by
Linus Torvalds
Feb 03, 2004
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[PATCH] H8/300 support update (3/3): bitops
o Cleanup reduced and faster code
parent
8432cdd0
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
126 additions
and
147 deletions
+126
-147
include/asm-h8300/bitops.h
include/asm-h8300/bitops.h
+126
-147
No files found.
include/asm-h8300/bitops.h
View file @
25b160e3
...
@@ -35,169 +35,148 @@ static __inline__ unsigned long ffz(unsigned long word)
...
@@ -35,169 +35,148 @@ static __inline__ unsigned long ffz(unsigned long word)
return
result
;
return
result
;
}
}
static
__inline__
void
set_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
#define H8300_GEN_BITOP_CONST(OP,BIT) \
{
case BIT: \
volatile
unsigned
char
*
b_addr
;
__asm__(OP " #" #BIT ",@%0"::"r"(b_addr):"memory"); \
b_addr
=
(
volatile
unsigned
char
*
)
addr
+
((
nr
>>
3
)
^
3
);
break;
__asm__
(
"mov.l %1,er0
\n\t
"
"bset r0l,%0"
#define H8300_GEN_BITOP(FNAME,OP) \
:
"+m"
(
*
b_addr
)
static __inline__ void FNAME(int nr, volatile unsigned long* addr) \
:
"g"
(
nr
&
7
),
"m"
(
*
b_addr
)
{ \
:
"er0"
);
volatile unsigned char *b_addr; \
b_addr = (volatile unsigned char *)addr + ((nr >> 3) ^ 3); \
if (__builtin_constant_p(nr)) { \
switch(nr & 7) { \
H8300_GEN_BITOP_CONST(OP,0) \
H8300_GEN_BITOP_CONST(OP,1) \
H8300_GEN_BITOP_CONST(OP,2) \
H8300_GEN_BITOP_CONST(OP,3) \
H8300_GEN_BITOP_CONST(OP,4) \
H8300_GEN_BITOP_CONST(OP,5) \
H8300_GEN_BITOP_CONST(OP,6) \
H8300_GEN_BITOP_CONST(OP,7) \
} \
} else { \
__asm__(OP " %w0,@%1"::"r"(nr),"r"(b_addr):"memory"); \
} \
}
}
/* Bigendian is complexed... */
#define __set_bit(nr, addr) set_bit((nr), (addr))
/*
/*
* clear_bit() doesn't provide any barrier for the compiler.
* clear_bit() doesn't provide any barrier for the compiler.
*/
*/
#define smp_mb__before_clear_bit() barrier()
#define smp_mb__before_clear_bit() barrier()
#define smp_mb__after_clear_bit() barrier()
#define smp_mb__after_clear_bit() barrier()
static
__inline__
void
clear_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
H8300_GEN_BITOP
(
set_bit
,
"bset"
)
{
H8300_GEN_BITOP
(
clear_bit
,
"bclr"
)
volatile
unsigned
char
*
b_addr
;
H8300_GEN_BITOP
(
change_bit
,
"bnot"
)
b_addr
=
(
volatile
unsigned
char
*
)
addr
+
((
nr
>>
3
)
^
3
);
#define __set_bit(nr,addr) set_bit((nr),(addr))
__asm__
(
"mov.l %1,er0
\n\t
"
#define __clear_bit(nr,addr) clear_bit((nr),(addr))
"bclr r0l,%0"
#define __change_bit(nr,addr) change_bit((nr),(addr))
:
"+m"
(
*
b_addr
)
:
"g"
(
nr
&
7
),
"m"
(
*
b_addr
)
:
"er0"
);
}
#define __clear_bit(nr, addr) clear_bit((nr), (addr))
static
__inline__
void
change_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
{
volatile
unsigned
char
*
b_addr
;
b_addr
=
(
volatile
unsigned
char
*
)
addr
+
((
nr
>>
3
)
^
3
);
__asm__
(
"mov.l %1,er0
\n\t
"
"bnot r0l,%0"
:
"+m"
(
*
b_addr
)
:
"g"
(
nr
&
7
),
"m"
(
*
b_addr
)
:
"er0"
);
}
#define __change_bit(nr, addr) change_bit((nr), (addr))
#undef H8300_GEN_BITOP
#undef H8300_GEN_BITOP_CONST
static
__inline__
int
test_bit
(
int
nr
,
const
unsigned
long
*
addr
)
static
__inline__
int
test_bit
(
int
nr
,
const
unsigned
long
*
addr
)
{
{
return
(
*
((
volatile
unsigned
char
*
)
addr
+
((
nr
>>
3
)
^
3
))
&
(
1UL
<<
(
nr
&
7
)))
!=
0
;
return
(
*
((
volatile
unsigned
char
*
)
addr
+
((
nr
>>
3
)
^
3
))
&
(
1UL
<<
(
nr
&
7
)))
!=
0
;
}
}
#define __test_bit(nr, addr) test_bit(nr, addr)
#define __test_bit(nr, addr) test_bit(nr, addr)
static
__inline__
int
test_and_set_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
#define H8300_GEN_TEST_BITOP_CONST_INT(OP,BIT) \
{
case BIT: \
int
retval
=
0
;
__asm__("stc ccr,%w1\n\t" \
volatile
unsigned
char
*
a
;
"orc #0x80,ccr\n\t" \
"bld #" #BIT ",@%3\n\t" \
a
=
(
volatile
unsigned
char
*
)
addr
+=
((
nr
>>
3
)
^
3
);
\
OP " #" #BIT ",@%3\n\t" \
__asm__
(
"mov.l %4,er3
\n\t
"
"rotxl.l %0\n\t" \
"stc ccr,r3h
\n\t
"
"ldc %w1,ccr" \
"orc #0x80,ccr
\n\t
"
: "=r"(retval),"=&r"(ccrsave) \
"btst r3l,%1
\n\t
"
: "0" (retval),"r" (b_addr) \
"bset r3l,%1
\n\t
"
: "memory"); \
"beq 1f
\n\t
"
break;
"inc.l #1,%0
\n\t
"
"1:"
#define H8300_GEN_TEST_BITOP_CONST(OP,BIT) \
"ldc r3h,ccr"
case BIT: \
:
"=r"
(
retval
),
"+m"
(
*
a
)
__asm__("bld #" #BIT ",@%2\n\t" \
:
"0"
(
retval
),
"m"
(
*
a
),
"g"
(
nr
&
7
)
:
"er3"
,
"memory"
);
OP " #" #BIT ",@%2\n\t" \
return
retval
;
"rotxl.l %0\n\t" \
: "=r"(retval) \
: "0" (retval),"r" (b_addr) \
: "memory"); \
break;
#define H8300_GEN_TEST_BITOP(FNNAME,OP) \
static __inline__ int FNNAME(int nr, volatile void * addr) \
{ \
int retval = 0; \
char ccrsave; \
volatile unsigned char *b_addr; \
b_addr = (volatile unsigned char *)addr + ((nr >> 3) ^ 3); \
if (__builtin_constant_p(nr)) { \
switch(nr & 7) { \
H8300_GEN_TEST_BITOP_CONST_INT(OP,0) \
H8300_GEN_TEST_BITOP_CONST_INT(OP,1) \
H8300_GEN_TEST_BITOP_CONST_INT(OP,2) \
H8300_GEN_TEST_BITOP_CONST_INT(OP,3) \
H8300_GEN_TEST_BITOP_CONST_INT(OP,4) \
H8300_GEN_TEST_BITOP_CONST_INT(OP,5) \
H8300_GEN_TEST_BITOP_CONST_INT(OP,6) \
H8300_GEN_TEST_BITOP_CONST_INT(OP,7) \
} \
} else { \
__asm__("stc ccr,%w1\n\t" \
"orc #0x80,ccr\n\t" \
"btst %w4,@%3\n\t" \
OP " %w4,@%3\n\t" \
"beq 1f\n\t" \
"inc.l #1,%0\n" \
"1:\n\t" \
"ldc %w1,ccr" \
: "=r"(retval),"=&r"(ccrsave) \
: "0" (retval),"r" (b_addr),"r"(nr) \
: "memory"); \
} \
return retval; \
} \
\
static __inline__ int __ ## FNNAME(int nr, volatile void * addr) \
{ \
int retval = 0; \
volatile unsigned char *b_addr; \
b_addr = (volatile unsigned char *)addr + ((nr >> 3) ^ 3); \
if (__builtin_constant_p(nr)) { \
switch(nr & 7) { \
H8300_GEN_TEST_BITOP_CONST(OP,0) \
H8300_GEN_TEST_BITOP_CONST(OP,1) \
H8300_GEN_TEST_BITOP_CONST(OP,2) \
H8300_GEN_TEST_BITOP_CONST(OP,3) \
H8300_GEN_TEST_BITOP_CONST(OP,4) \
H8300_GEN_TEST_BITOP_CONST(OP,5) \
H8300_GEN_TEST_BITOP_CONST(OP,6) \
H8300_GEN_TEST_BITOP_CONST(OP,7) \
} \
} else { \
__asm__("btst %w3,@%2\n\t" \
OP " %w3,@%2\n\t" \
"beq 1f\n\t" \
"inc.l #1,%0\n" \
"1:" \
: "=r"(retval) \
: "0" (retval),"r" (b_addr),"r"(nr) \
: "memory"); \
} \
return retval; \
}
}
static
__inline__
int
__test_and_set_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
H8300_GEN_TEST_BITOP
(
test_and_set_bit
,
"bset"
)
{
H8300_GEN_TEST_BITOP
(
test_and_clear_bit
,
"bclr"
)
int
retval
=
0
;
H8300_GEN_TEST_BITOP
(
test_and_change_bit
,
"bnot"
)
volatile
unsigned
char
*
a
;
#undef H8300_GEN_TEST_BITOP_CONST
#undef H8300_GEN_TEST_BITOP_CONST_INT
a
=
(
volatile
unsigned
char
*
)
addr
+=
((
nr
>>
3
)
^
3
);
\
#undef H8300_GEN_TEST_BITOP
__asm__
(
"mov.l %4,er3
\n\t
"
"btst r3l,%1
\n\t
"
"bset r3l,%1
\n\t
"
"beq 1f
\n\t
"
"inc.l #1,%0
\n\t
"
"1:"
:
"=r"
(
retval
),
"+m"
(
*
a
)
:
"0"
(
retval
),
"m"
(
*
a
),
"g"
(
nr
&
7
)
:
"er3"
,
"memory"
);
return
retval
;
}
static
__inline__
int
test_and_clear_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
{
int
retval
=
0
;
volatile
unsigned
char
*
a
;
a
=
(
volatile
unsigned
char
*
)
addr
+=
((
nr
>>
3
)
^
3
);
\
__asm__
(
"mov.l %4,er3
\n\t
"
"stc ccr,r3h
\n\t
"
"orc #0x80,ccr
\n\t
"
"btst r3l,%1
\n\t
"
"bclr r3l,%1
\n\t
"
"beq 1f
\n\t
"
"inc.l #1,%0
\n\t
"
"1:"
"ldc r3h,ccr"
:
"=r"
(
retval
),
"+m"
(
*
a
)
:
"0"
(
retval
),
"m"
(
*
a
),
"g"
(
nr
&
7
)
:
"er3"
,
"memory"
);
return
retval
;
}
static
__inline__
int
__test_and_clear_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
{
int
retval
=
0
;
volatile
unsigned
char
*
a
;
a
=
(
volatile
unsigned
char
*
)
addr
+=
((
nr
>>
3
)
^
3
);
\
__asm__
(
"mov.l %4,er3
\n\t
"
"btst r3l,%1
\n\t
"
"bclr r3l,%1
\n\t
"
"beq 1f
\n\t
"
"inc.l #1,%0
\n\t
"
"1:"
:
"=r"
(
retval
),
"+m"
(
*
a
)
:
"0"
(
retval
),
"m"
(
*
a
),
"g"
(
nr
&
7
)
:
"er3"
,
"memory"
);
return
retval
;
}
static
__inline__
int
test_and_change_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
{
int
retval
=
0
;
volatile
unsigned
char
*
a
;
a
=
(
volatile
unsigned
char
*
)
addr
+=
((
nr
>>
3
)
^
3
);
\
__asm__
(
"mov.l %4,er3
\n\t
"
"stc ccr,r3h
\n\t
"
"orc #0x80,ccr
\n\t
"
"btst r3l,%1
\n\t
"
"bnot r3l,%1
\n\t
"
"beq 1f
\n\t
"
"inc.l #1,%0
\n\t
"
"1:"
"ldc r3h,ccr"
:
"=r"
(
retval
),
"+m"
(
*
a
)
:
"0"
(
retval
),
"m"
(
*
a
),
"g"
(
nr
&
7
)
:
"er3"
,
"memory"
);
return
retval
;
}
static
__inline__
int
__test_and_change_bit
(
int
nr
,
volatile
unsigned
long
*
addr
)
{
int
retval
=
0
;
volatile
unsigned
char
*
a
;
a
=
(
volatile
unsigned
char
*
)
addr
+=
((
nr
>>
3
)
^
3
);
\
__asm__
(
"mov.l %4,er3
\n\t
"
"btst r3l,%1
\n\t
"
"bnot r3l,%1
\n\t
"
"beq 1f
\n\t
"
"inc.l #1,%0
\n\t
"
"1:"
:
"=r"
(
retval
),
"+m"
(
*
a
)
:
"0"
(
retval
),
"m"
(
*
a
),
"g"
(
nr
&
7
)
:
"er3"
,
"memory"
);
return
retval
;
}
#define find_first_zero_bit(addr, size) \
#define find_first_zero_bit(addr, size) \
find_next_zero_bit((addr), (size), 0)
find_next_zero_bit((addr), (size), 0)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment