Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
32738fbf
Commit
32738fbf
authored
Feb 05, 2003
by
Andrew Morton
Committed by
Linus Torvalds
Feb 05, 2003
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[PATCH] mm/mremap.c whitespace cleanup
- Not everyone uses 160-column xterms. - Coding style consistency
parent
df79ea40
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
20 additions
and
15 deletions
+20
-15
mm/mremap.c
mm/mremap.c
+20
-15
No files found.
mm/mremap.c
View file @
32738fbf
...
@@ -24,9 +24,9 @@
...
@@ -24,9 +24,9 @@
static
pte_t
*
get_one_pte_map_nested
(
struct
mm_struct
*
mm
,
unsigned
long
addr
)
static
pte_t
*
get_one_pte_map_nested
(
struct
mm_struct
*
mm
,
unsigned
long
addr
)
{
{
pgd_t
*
pgd
;
pgd_t
*
pgd
;
pmd_t
*
pmd
;
pmd_t
*
pmd
;
pte_t
*
pte
=
NULL
;
pte_t
*
pte
=
NULL
;
pgd
=
pgd_offset
(
mm
,
addr
);
pgd
=
pgd_offset
(
mm
,
addr
);
if
(
pgd_none
(
*
pgd
))
if
(
pgd_none
(
*
pgd
))
...
@@ -73,8 +73,8 @@ static inline int page_table_present(struct mm_struct *mm, unsigned long addr)
...
@@ -73,8 +73,8 @@ static inline int page_table_present(struct mm_struct *mm, unsigned long addr)
static
inline
pte_t
*
alloc_one_pte_map
(
struct
mm_struct
*
mm
,
unsigned
long
addr
)
static
inline
pte_t
*
alloc_one_pte_map
(
struct
mm_struct
*
mm
,
unsigned
long
addr
)
{
{
pmd_t
*
pmd
;
pmd_t
*
pmd
;
pte_t
*
pte
=
NULL
;
pte_t
*
pte
=
NULL
;
pmd
=
pmd_alloc
(
mm
,
pgd_offset
(
mm
,
addr
),
addr
);
pmd
=
pmd_alloc
(
mm
,
pgd_offset
(
mm
,
addr
),
addr
);
if
(
pmd
)
if
(
pmd
)
...
@@ -88,7 +88,7 @@ copy_one_pte(struct mm_struct *mm, pte_t *src, pte_t *dst,
...
@@ -88,7 +88,7 @@ copy_one_pte(struct mm_struct *mm, pte_t *src, pte_t *dst,
{
{
int
error
=
0
;
int
error
=
0
;
pte_t
pte
;
pte_t
pte
;
struct
page
*
page
=
NULL
;
struct
page
*
page
=
NULL
;
if
(
pte_present
(
*
src
))
if
(
pte_present
(
*
src
))
page
=
pte_page
(
*
src
);
page
=
pte_page
(
*
src
);
...
@@ -183,12 +183,12 @@ static int move_page_tables(struct vm_area_struct *vma,
...
@@ -183,12 +183,12 @@ static int move_page_tables(struct vm_area_struct *vma,
return
-
1
;
return
-
1
;
}
}
static
unsigned
long
move_vma
(
struct
vm_area_struct
*
vma
,
static
unsigned
long
move_vma
(
struct
vm_area_struct
*
vma
,
unsigned
long
addr
,
unsigned
long
old_len
,
unsigned
long
new_len
,
unsigned
long
addr
,
unsigned
long
old_len
,
unsigned
long
new_len
,
unsigned
long
new_addr
)
unsigned
long
new_addr
)
{
{
struct
mm_struct
*
mm
=
vma
->
vm_mm
;
struct
mm_struct
*
mm
=
vma
->
vm_mm
;
struct
vm_area_struct
*
new_vma
,
*
next
,
*
prev
;
struct
vm_area_struct
*
new_vma
,
*
next
,
*
prev
;
int
allocated_vma
;
int
allocated_vma
;
int
split
=
0
;
int
split
=
0
;
...
@@ -196,14 +196,16 @@ static unsigned long move_vma(struct vm_area_struct * vma,
...
@@ -196,14 +196,16 @@ static unsigned long move_vma(struct vm_area_struct * vma,
next
=
find_vma_prev
(
mm
,
new_addr
,
&
prev
);
next
=
find_vma_prev
(
mm
,
new_addr
,
&
prev
);
if
(
next
)
{
if
(
next
)
{
if
(
prev
&&
prev
->
vm_end
==
new_addr
&&
if
(
prev
&&
prev
->
vm_end
==
new_addr
&&
can_vma_merge
(
prev
,
vma
->
vm_flags
)
&&
!
vma
->
vm_file
&&
!
(
vma
->
vm_flags
&
VM_SHARED
))
{
can_vma_merge
(
prev
,
vma
->
vm_flags
)
&&
!
vma
->
vm_file
&&
!
(
vma
->
vm_flags
&
VM_SHARED
))
{
spin_lock
(
&
mm
->
page_table_lock
);
spin_lock
(
&
mm
->
page_table_lock
);
prev
->
vm_end
=
new_addr
+
new_len
;
prev
->
vm_end
=
new_addr
+
new_len
;
spin_unlock
(
&
mm
->
page_table_lock
);
spin_unlock
(
&
mm
->
page_table_lock
);
new_vma
=
prev
;
new_vma
=
prev
;
if
(
next
!=
prev
->
vm_next
)
if
(
next
!=
prev
->
vm_next
)
BUG
();
BUG
();
if
(
prev
->
vm_end
==
next
->
vm_start
&&
can_vma_merge
(
next
,
prev
->
vm_flags
))
{
if
(
prev
->
vm_end
==
next
->
vm_start
&&
can_vma_merge
(
next
,
prev
->
vm_flags
))
{
spin_lock
(
&
mm
->
page_table_lock
);
spin_lock
(
&
mm
->
page_table_lock
);
prev
->
vm_end
=
next
->
vm_end
;
prev
->
vm_end
=
next
->
vm_end
;
__vma_unlink
(
mm
,
next
,
prev
);
__vma_unlink
(
mm
,
next
,
prev
);
...
@@ -214,7 +216,8 @@ static unsigned long move_vma(struct vm_area_struct * vma,
...
@@ -214,7 +216,8 @@ static unsigned long move_vma(struct vm_area_struct * vma,
kmem_cache_free
(
vm_area_cachep
,
next
);
kmem_cache_free
(
vm_area_cachep
,
next
);
}
}
}
else
if
(
next
->
vm_start
==
new_addr
+
new_len
&&
}
else
if
(
next
->
vm_start
==
new_addr
+
new_len
&&
can_vma_merge
(
next
,
vma
->
vm_flags
)
&&
!
vma
->
vm_file
&&
!
(
vma
->
vm_flags
&
VM_SHARED
))
{
can_vma_merge
(
next
,
vma
->
vm_flags
)
&&
!
vma
->
vm_file
&&
!
(
vma
->
vm_flags
&
VM_SHARED
))
{
spin_lock
(
&
mm
->
page_table_lock
);
spin_lock
(
&
mm
->
page_table_lock
);
next
->
vm_start
=
new_addr
;
next
->
vm_start
=
new_addr
;
spin_unlock
(
&
mm
->
page_table_lock
);
spin_unlock
(
&
mm
->
page_table_lock
);
...
@@ -223,7 +226,8 @@ static unsigned long move_vma(struct vm_area_struct * vma,
...
@@ -223,7 +226,8 @@ static unsigned long move_vma(struct vm_area_struct * vma,
}
else
{
}
else
{
prev
=
find_vma
(
mm
,
new_addr
-
1
);
prev
=
find_vma
(
mm
,
new_addr
-
1
);
if
(
prev
&&
prev
->
vm_end
==
new_addr
&&
if
(
prev
&&
prev
->
vm_end
==
new_addr
&&
can_vma_merge
(
prev
,
vma
->
vm_flags
)
&&
!
vma
->
vm_file
&&
!
(
vma
->
vm_flags
&
VM_SHARED
))
{
can_vma_merge
(
prev
,
vma
->
vm_flags
)
&&
!
vma
->
vm_file
&&
!
(
vma
->
vm_flags
&
VM_SHARED
))
{
spin_lock
(
&
mm
->
page_table_lock
);
spin_lock
(
&
mm
->
page_table_lock
);
prev
->
vm_end
=
new_addr
+
new_len
;
prev
->
vm_end
=
new_addr
+
new_len
;
spin_unlock
(
&
mm
->
page_table_lock
);
spin_unlock
(
&
mm
->
page_table_lock
);
...
@@ -249,7 +253,7 @@ static unsigned long move_vma(struct vm_area_struct * vma,
...
@@ -249,7 +253,7 @@ static unsigned long move_vma(struct vm_area_struct * vma,
INIT_LIST_HEAD
(
&
new_vma
->
shared
);
INIT_LIST_HEAD
(
&
new_vma
->
shared
);
new_vma
->
vm_start
=
new_addr
;
new_vma
->
vm_start
=
new_addr
;
new_vma
->
vm_end
=
new_addr
+
new_len
;
new_vma
->
vm_end
=
new_addr
+
new_len
;
new_vma
->
vm_pgoff
+=
(
addr
-
vma
->
vm_start
)
>>
PAGE_SHIFT
;
new_vma
->
vm_pgoff
+=
(
addr
-
vma
->
vm_start
)
>>
PAGE_SHIFT
;
if
(
new_vma
->
vm_file
)
if
(
new_vma
->
vm_file
)
get_file
(
new_vma
->
vm_file
);
get_file
(
new_vma
->
vm_file
);
if
(
new_vma
->
vm_ops
&&
new_vma
->
vm_ops
->
open
)
if
(
new_vma
->
vm_ops
&&
new_vma
->
vm_ops
->
open
)
...
@@ -428,7 +432,8 @@ unsigned long do_mremap(unsigned long addr,
...
@@ -428,7 +432,8 @@ unsigned long do_mremap(unsigned long addr,
if
(
vma
->
vm_flags
&
VM_SHARED
)
if
(
vma
->
vm_flags
&
VM_SHARED
)
map_flags
|=
MAP_SHARED
;
map_flags
|=
MAP_SHARED
;
new_addr
=
get_unmapped_area
(
vma
->
vm_file
,
0
,
new_len
,
vma
->
vm_pgoff
,
map_flags
);
new_addr
=
get_unmapped_area
(
vma
->
vm_file
,
0
,
new_len
,
vma
->
vm_pgoff
,
map_flags
);
ret
=
new_addr
;
ret
=
new_addr
;
if
(
new_addr
&
~
PAGE_MASK
)
if
(
new_addr
&
~
PAGE_MASK
)
goto
out
;
goto
out
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment