Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
5472e862
Commit
5472e862
authored
Nov 25, 2011
by
Cong Wang
Committed by
Cong Wang
Mar 20, 2012
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
arm: remove the second argument of k[un]map_atomic()
Signed-off-by:
Cong Wang
<
amwang@redhat.com
>
parent
1ec9c5dd
Changes
9
Show whitespace changes
Inline
Side-by-side
Showing
9 changed files
with
50 additions
and
50 deletions
+50
-50
arch/arm/mm/copypage-fa.c
arch/arm/mm/copypage-fa.c
+6
-6
arch/arm/mm/copypage-feroceon.c
arch/arm/mm/copypage-feroceon.c
+6
-6
arch/arm/mm/copypage-v3.c
arch/arm/mm/copypage-v3.c
+6
-6
arch/arm/mm/copypage-v4mc.c
arch/arm/mm/copypage-v4mc.c
+4
-4
arch/arm/mm/copypage-v4wb.c
arch/arm/mm/copypage-v4wb.c
+6
-6
arch/arm/mm/copypage-v4wt.c
arch/arm/mm/copypage-v4wt.c
+6
-6
arch/arm/mm/copypage-v6.c
arch/arm/mm/copypage-v6.c
+6
-6
arch/arm/mm/copypage-xsc3.c
arch/arm/mm/copypage-xsc3.c
+6
-6
arch/arm/mm/copypage-xscale.c
arch/arm/mm/copypage-xscale.c
+4
-4
No files found.
arch/arm/mm/copypage-fa.c
View file @
5472e862
...
...
@@ -44,11 +44,11 @@ void fa_copy_user_highpage(struct page *to, struct page *from,
{
void
*
kto
,
*
kfrom
;
kto
=
kmap_atomic
(
to
,
KM_USER0
);
kfrom
=
kmap_atomic
(
from
,
KM_USER1
);
kto
=
kmap_atomic
(
to
);
kfrom
=
kmap_atomic
(
from
);
fa_copy_user_page
(
kto
,
kfrom
);
kunmap_atomic
(
kfrom
,
KM_USER1
);
kunmap_atomic
(
kto
,
KM_USER0
);
kunmap_atomic
(
kfrom
);
kunmap_atomic
(
kto
);
}
/*
...
...
@@ -58,7 +58,7 @@ void fa_copy_user_highpage(struct page *to, struct page *from,
*/
void
fa_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"\
mov r1, %2 @ 1
\n
\
mov r2, #0 @ 1
\n
\
...
...
@@ -77,7 +77,7 @@ void fa_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
32
)
:
"r1"
,
"r2"
,
"r3"
,
"ip"
,
"lr"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
fa_user_fns
__initdata
=
{
...
...
arch/arm/mm/copypage-feroceon.c
View file @
5472e862
...
...
@@ -72,17 +72,17 @@ void feroceon_copy_user_highpage(struct page *to, struct page *from,
{
void
*
kto
,
*
kfrom
;
kto
=
kmap_atomic
(
to
,
KM_USER0
);
kfrom
=
kmap_atomic
(
from
,
KM_USER1
);
kto
=
kmap_atomic
(
to
);
kfrom
=
kmap_atomic
(
from
);
flush_cache_page
(
vma
,
vaddr
,
page_to_pfn
(
from
));
feroceon_copy_user_page
(
kto
,
kfrom
);
kunmap_atomic
(
kfrom
,
KM_USER1
);
kunmap_atomic
(
kto
,
KM_USER0
);
kunmap_atomic
(
kfrom
);
kunmap_atomic
(
kto
);
}
void
feroceon_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"\
mov r1, %2
\n
\
mov r2, #0
\n
\
...
...
@@ -102,7 +102,7 @@ void feroceon_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
32
)
:
"r1"
,
"r2"
,
"r3"
,
"r4"
,
"r5"
,
"r6"
,
"r7"
,
"ip"
,
"lr"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
feroceon_user_fns
__initdata
=
{
...
...
arch/arm/mm/copypage-v3.c
View file @
5472e862
...
...
@@ -42,11 +42,11 @@ void v3_copy_user_highpage(struct page *to, struct page *from,
{
void
*
kto
,
*
kfrom
;
kto
=
kmap_atomic
(
to
,
KM_USER0
);
kfrom
=
kmap_atomic
(
from
,
KM_USER1
);
kto
=
kmap_atomic
(
to
);
kfrom
=
kmap_atomic
(
from
);
v3_copy_user_page
(
kto
,
kfrom
);
kunmap_atomic
(
kfrom
,
KM_USER1
);
kunmap_atomic
(
kto
,
KM_USER0
);
kunmap_atomic
(
kfrom
);
kunmap_atomic
(
kto
);
}
/*
...
...
@@ -56,7 +56,7 @@ void v3_copy_user_highpage(struct page *to, struct page *from,
*/
void
v3_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"
\n
\
mov r1, %2 @ 1
\n
\
mov r2, #0 @ 1
\n
\
...
...
@@ -72,7 +72,7 @@ void v3_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
64
)
:
"r1"
,
"r2"
,
"r3"
,
"ip"
,
"lr"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
v3_user_fns
__initdata
=
{
...
...
arch/arm/mm/copypage-v4mc.c
View file @
5472e862
...
...
@@ -71,7 +71,7 @@ mc_copy_user_page(void *from, void *to)
void
v4_mc_copy_user_highpage
(
struct
page
*
to
,
struct
page
*
from
,
unsigned
long
vaddr
,
struct
vm_area_struct
*
vma
)
{
void
*
kto
=
kmap_atomic
(
to
,
KM_USER1
);
void
*
kto
=
kmap_atomic
(
to
);
if
(
!
test_and_set_bit
(
PG_dcache_clean
,
&
from
->
flags
))
__flush_dcache_page
(
page_mapping
(
from
),
from
);
...
...
@@ -85,7 +85,7 @@ void v4_mc_copy_user_highpage(struct page *to, struct page *from,
raw_spin_unlock
(
&
minicache_lock
);
kunmap_atomic
(
kto
,
KM_USER1
);
kunmap_atomic
(
kto
);
}
/*
...
...
@@ -93,7 +93,7 @@ void v4_mc_copy_user_highpage(struct page *to, struct page *from,
*/
void
v4_mc_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"\
mov r1, %2 @ 1
\n
\
mov r2, #0 @ 1
\n
\
...
...
@@ -111,7 +111,7 @@ void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
64
)
:
"r1"
,
"r2"
,
"r3"
,
"ip"
,
"lr"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
v4_mc_user_fns
__initdata
=
{
...
...
arch/arm/mm/copypage-v4wb.c
View file @
5472e862
...
...
@@ -52,12 +52,12 @@ void v4wb_copy_user_highpage(struct page *to, struct page *from,
{
void
*
kto
,
*
kfrom
;
kto
=
kmap_atomic
(
to
,
KM_USER0
);
kfrom
=
kmap_atomic
(
from
,
KM_USER1
);
kto
=
kmap_atomic
(
to
);
kfrom
=
kmap_atomic
(
from
);
flush_cache_page
(
vma
,
vaddr
,
page_to_pfn
(
from
));
v4wb_copy_user_page
(
kto
,
kfrom
);
kunmap_atomic
(
kfrom
,
KM_USER1
);
kunmap_atomic
(
kto
,
KM_USER0
);
kunmap_atomic
(
kfrom
);
kunmap_atomic
(
kto
);
}
/*
...
...
@@ -67,7 +67,7 @@ void v4wb_copy_user_highpage(struct page *to, struct page *from,
*/
void
v4wb_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"\
mov r1, %2 @ 1
\n
\
mov r2, #0 @ 1
\n
\
...
...
@@ -86,7 +86,7 @@ void v4wb_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
64
)
:
"r1"
,
"r2"
,
"r3"
,
"ip"
,
"lr"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
v4wb_user_fns
__initdata
=
{
...
...
arch/arm/mm/copypage-v4wt.c
View file @
5472e862
...
...
@@ -48,11 +48,11 @@ void v4wt_copy_user_highpage(struct page *to, struct page *from,
{
void
*
kto
,
*
kfrom
;
kto
=
kmap_atomic
(
to
,
KM_USER0
);
kfrom
=
kmap_atomic
(
from
,
KM_USER1
);
kto
=
kmap_atomic
(
to
);
kfrom
=
kmap_atomic
(
from
);
v4wt_copy_user_page
(
kto
,
kfrom
);
kunmap_atomic
(
kfrom
,
KM_USER1
);
kunmap_atomic
(
kto
,
KM_USER0
);
kunmap_atomic
(
kfrom
);
kunmap_atomic
(
kto
);
}
/*
...
...
@@ -62,7 +62,7 @@ void v4wt_copy_user_highpage(struct page *to, struct page *from,
*/
void
v4wt_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"\
mov r1, %2 @ 1
\n
\
mov r2, #0 @ 1
\n
\
...
...
@@ -79,7 +79,7 @@ void v4wt_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
64
)
:
"r1"
,
"r2"
,
"r3"
,
"ip"
,
"lr"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
v4wt_user_fns
__initdata
=
{
...
...
arch/arm/mm/copypage-v6.c
View file @
5472e862
...
...
@@ -38,11 +38,11 @@ static void v6_copy_user_highpage_nonaliasing(struct page *to,
{
void
*
kto
,
*
kfrom
;
kfrom
=
kmap_atomic
(
from
,
KM_USER0
);
kto
=
kmap_atomic
(
to
,
KM_USER1
);
kfrom
=
kmap_atomic
(
from
);
kto
=
kmap_atomic
(
to
);
copy_page
(
kto
,
kfrom
);
kunmap_atomic
(
kto
,
KM_USER1
);
kunmap_atomic
(
kfrom
,
KM_USER0
);
kunmap_atomic
(
kto
);
kunmap_atomic
(
kfrom
);
}
/*
...
...
@@ -51,9 +51,9 @@ static void v6_copy_user_highpage_nonaliasing(struct page *to,
*/
static
void
v6_clear_user_highpage_nonaliasing
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
kaddr
=
kmap_atomic
(
page
);
clear_page
(
kaddr
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
/*
...
...
arch/arm/mm/copypage-xsc3.c
View file @
5472e862
...
...
@@ -75,12 +75,12 @@ void xsc3_mc_copy_user_highpage(struct page *to, struct page *from,
{
void
*
kto
,
*
kfrom
;
kto
=
kmap_atomic
(
to
,
KM_USER0
);
kfrom
=
kmap_atomic
(
from
,
KM_USER1
);
kto
=
kmap_atomic
(
to
);
kfrom
=
kmap_atomic
(
from
);
flush_cache_page
(
vma
,
vaddr
,
page_to_pfn
(
from
));
xsc3_mc_copy_user_page
(
kto
,
kfrom
);
kunmap_atomic
(
kfrom
,
KM_USER1
);
kunmap_atomic
(
kto
,
KM_USER0
);
kunmap_atomic
(
kfrom
);
kunmap_atomic
(
kto
);
}
/*
...
...
@@ -90,7 +90,7 @@ void xsc3_mc_copy_user_highpage(struct page *to, struct page *from,
*/
void
xsc3_mc_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"\
mov r1, %2
\n
\
mov r2, #0
\n
\
...
...
@@ -105,7 +105,7 @@ void xsc3_mc_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
32
)
:
"r1"
,
"r2"
,
"r3"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
xsc3_mc_user_fns
__initdata
=
{
...
...
arch/arm/mm/copypage-xscale.c
View file @
5472e862
...
...
@@ -93,7 +93,7 @@ mc_copy_user_page(void *from, void *to)
void
xscale_mc_copy_user_highpage
(
struct
page
*
to
,
struct
page
*
from
,
unsigned
long
vaddr
,
struct
vm_area_struct
*
vma
)
{
void
*
kto
=
kmap_atomic
(
to
,
KM_USER1
);
void
*
kto
=
kmap_atomic
(
to
);
if
(
!
test_and_set_bit
(
PG_dcache_clean
,
&
from
->
flags
))
__flush_dcache_page
(
page_mapping
(
from
),
from
);
...
...
@@ -107,7 +107,7 @@ void xscale_mc_copy_user_highpage(struct page *to, struct page *from,
raw_spin_unlock
(
&
minicache_lock
);
kunmap_atomic
(
kto
,
KM_USER1
);
kunmap_atomic
(
kto
);
}
/*
...
...
@@ -116,7 +116,7 @@ void xscale_mc_copy_user_highpage(struct page *to, struct page *from,
void
xscale_mc_clear_user_highpage
(
struct
page
*
page
,
unsigned
long
vaddr
)
{
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
,
KM_USER0
);
void
*
ptr
,
*
kaddr
=
kmap_atomic
(
page
);
asm
volatile
(
"mov r1, %2
\n
\
mov r2, #0
\n
\
...
...
@@ -133,7 +133,7 @@ xscale_mc_clear_user_highpage(struct page *page, unsigned long vaddr)
:
"=r"
(
ptr
)
:
"0"
(
kaddr
),
"I"
(
PAGE_SIZE
/
32
)
:
"r1"
,
"r2"
,
"r3"
,
"ip"
);
kunmap_atomic
(
kaddr
,
KM_USER0
);
kunmap_atomic
(
kaddr
);
}
struct
cpu_user_fns
xscale_mc_user_fns
__initdata
=
{
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment