Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
67a670bd
Commit
67a670bd
authored
May 18, 2002
by
Linus Torvalds
Browse files
Options
Browse Files
Download
Plain Diff
Merge
http://linux-isdn.bkbits.net/linux-2.5.make-next
into home.transmeta.com:/home/torvalds/v2.5/linux
parents
8876c643
66ebd50b
Changes
21
Show whitespace changes
Inline
Side-by-side
Showing
21 changed files
with
961 additions
and
904 deletions
+961
-904
drivers/md/md.c
drivers/md/md.c
+45
-47
drivers/md/raid5.c
drivers/md/raid5.c
+442
-370
drivers/md/xor.c
drivers/md/xor.c
+8
-10
fs/intermezzo/ext_attr.c
fs/intermezzo/ext_attr.c
+2
-3
fs/intermezzo/kml.c
fs/intermezzo/kml.c
+18
-22
fs/intermezzo/psdev.c
fs/intermezzo/psdev.c
+109
-141
fs/nfsd/export.c
fs/nfsd/export.c
+10
-8
fs/nfsd/vfs.c
fs/nfsd/vfs.c
+21
-33
fs/read_write.c
fs/read_write.c
+81
-98
fs/reiserfs/inode.c
fs/reiserfs/inode.c
+78
-69
fs/reiserfs/journal.c
fs/reiserfs/journal.c
+2
-2
fs/reiserfs/namei.c
fs/reiserfs/namei.c
+91
-67
include/asm-i386/xor.h
include/asm-i386/xor.h
+3
-1
include/linux/fs.h
include/linux/fs.h
+3
-0
include/linux/raid/md_k.h
include/linux/raid/md_k.h
+3
-2
include/linux/raid/raid5.h
include/linux/raid/raid5.h
+28
-20
include/linux/raid/xor.h
include/linux/raid/xor.h
+1
-1
include/linux/reiserfs_fs.h
include/linux/reiserfs_fs.h
+3
-3
kernel/exit.c
kernel/exit.c
+9
-3
kernel/ksyms.c
kernel/ksyms.c
+2
-0
net/sunrpc/svcsock.c
net/sunrpc/svcsock.c
+2
-4
No files found.
drivers/md/md.c
View file @
67a670bd
...
...
@@ -436,14 +436,15 @@ static int alloc_array_sb(mddev_t * mddev)
static
int
alloc_disk_sb
(
mdk_rdev_t
*
rdev
)
{
if
(
rdev
->
sb
)
if
(
rdev
->
sb
_page
)
MD_BUG
();
rdev
->
sb
=
(
mdp_super_t
*
)
__get_free
_page
(
GFP_KERNEL
);
if
(
!
rdev
->
sb
)
{
rdev
->
sb
_page
=
alloc
_page
(
GFP_KERNEL
);
if
(
!
rdev
->
sb
_page
)
{
printk
(
OUT_OF_MEM
);
return
-
EINVAL
;
}
rdev
->
sb
=
(
mdp_super_t
*
)
page_address
(
rdev
->
sb_page
);
clear_page
(
rdev
->
sb
);
return
0
;
...
...
@@ -451,9 +452,10 @@ static int alloc_disk_sb(mdk_rdev_t * rdev)
static
void
free_disk_sb
(
mdk_rdev_t
*
rdev
)
{
if
(
rdev
->
sb
)
{
free_page
((
unsigned
long
)
rdev
->
sb
);
if
(
rdev
->
sb
_page
)
{
page_cache_release
(
rdev
->
sb_page
);
rdev
->
sb
=
NULL
;
rdev
->
sb_page
=
NULL
;
rdev
->
sb_offset
=
0
;
rdev
->
size
=
0
;
}
else
{
...
...
@@ -462,13 +464,42 @@ static void free_disk_sb(mdk_rdev_t * rdev)
}
}
static
void
bi_complete
(
struct
bio
*
bio
)
{
complete
((
struct
completion
*
)
bio
->
bi_private
);
}
static
int
sync_page_io
(
struct
block_device
*
bdev
,
sector_t
sector
,
int
size
,
struct
page
*
page
,
int
rw
)
{
struct
bio
bio
;
struct
bio_vec
vec
;
struct
completion
event
;
bio_init
(
&
bio
);
bio
.
bi_io_vec
=
&
vec
;
vec
.
bv_page
=
page
;
vec
.
bv_len
=
size
;
vec
.
bv_offset
=
0
;
bio
.
bi_vcnt
=
1
;
bio
.
bi_idx
=
0
;
bio
.
bi_size
=
size
;
bio
.
bi_bdev
=
bdev
;
bio
.
bi_sector
=
sector
;
init_completion
(
&
event
);
bio
.
bi_private
=
&
event
;
bio
.
bi_end_io
=
bi_complete
;
submit_bio
(
rw
,
&
bio
);
run_task_queue
(
&
tq_disk
);
wait_for_completion
(
&
event
);
return
test_bit
(
BIO_UPTODATE
,
&
bio
.
bi_flags
);
}
static
int
read_disk_sb
(
mdk_rdev_t
*
rdev
)
{
struct
address_space
*
mapping
=
rdev
->
bdev
->
bd_inode
->
i_mapping
;
struct
page
*
page
;
char
*
p
;
unsigned
long
sb_offset
;
int
n
=
PAGE_CACHE_SIZE
/
BLOCK_SIZE
;
if
(
!
rdev
->
sb
)
{
MD_BUG
();
...
...
@@ -483,24 +514,14 @@ static int read_disk_sb(mdk_rdev_t * rdev)
*/
sb_offset
=
calc_dev_sboffset
(
rdev
->
dev
,
rdev
->
mddev
,
1
);
rdev
->
sb_offset
=
sb_offset
;
page
=
read_cache_page
(
mapping
,
sb_offset
/
n
,
(
filler_t
*
)
mapping
->
a_ops
->
readpage
,
NULL
);
if
(
IS_ERR
(
page
))
goto
out
;
wait_on_page_locked
(
page
);
if
(
!
PageUptodate
(
page
))
goto
fail
;
if
(
PageError
(
page
))
if
(
!
sync_page_io
(
rdev
->
bdev
,
sb_offset
<<
1
,
MD_SB_BYTES
,
rdev
->
sb_page
,
READ
))
goto
fail
;
p
=
(
char
*
)
page_address
(
page
)
+
BLOCK_SIZE
*
(
sb_offset
%
n
);
memcpy
((
char
*
)
rdev
->
sb
,
p
,
MD_SB_BYTES
);
page_cache_release
(
page
);
printk
(
KERN_INFO
" [events: %08lx]
\n
"
,
(
unsigned
long
)
rdev
->
sb
->
events_lo
);
return
0
;
fail:
page_cache_release
(
page
);
out:
printk
(
NO_SB
,
partition_name
(
rdev
->
dev
));
return
-
EINVAL
;
}
...
...
@@ -893,11 +914,6 @@ static mdk_rdev_t * find_rdev_all(kdev_t dev)
static
int
write_disk_sb
(
mdk_rdev_t
*
rdev
)
{
struct
block_device
*
bdev
=
rdev
->
bdev
;
struct
address_space
*
mapping
=
bdev
->
bd_inode
->
i_mapping
;
struct
page
*
page
;
unsigned
offs
;
int
error
;
kdev_t
dev
=
rdev
->
dev
;
unsigned
long
sb_offset
,
size
;
...
...
@@ -933,29 +949,11 @@ static int write_disk_sb(mdk_rdev_t * rdev)
}
printk
(
KERN_INFO
"(write) %s's sb offset: %ld
\n
"
,
partition_name
(
dev
),
sb_offset
);
fsync_bdev
(
bdev
);
page
=
grab_cache_page
(
mapping
,
sb_offset
/
(
PAGE_CACHE_SIZE
/
BLOCK_SIZE
));
offs
=
sb_offset
%
(
PAGE_CACHE_SIZE
/
BLOCK_SIZE
);
if
(
!
page
)
if
(
!
sync_page_io
(
rdev
->
bdev
,
sb_offset
<<
1
,
MD_SB_BYTES
,
rdev
->
sb_page
,
WRITE
))
goto
fail
;
error
=
mapping
->
a_ops
->
prepare_write
(
NULL
,
page
,
offs
,
offs
+
MD_SB_BYTES
);
if
(
error
)
goto
unlock
;
memcpy
((
char
*
)
page_address
(
page
)
+
offs
,
rdev
->
sb
,
MD_SB_BYTES
);
error
=
mapping
->
a_ops
->
commit_write
(
NULL
,
page
,
offs
,
offs
+
MD_SB_BYTES
);
if
(
error
)
goto
unlock
;
unlock_page
(
page
);
wait_on_page_locked
(
page
);
page_cache_release
(
page
);
fsync_bdev
(
bdev
);
skip:
return
0
;
unlock:
unlock_page
(
page
);
page_cache_release
(
page
);
fail:
printk
(
"md: write_disk_sb failed for device %s
\n
"
,
partition_name
(
dev
));
return
1
;
...
...
drivers/md/raid5.c
View file @
67a670bd
...
...
@@ -24,19 +24,19 @@
#include <asm/bitops.h>
#include <asm/atomic.h>
static
mdk_personality_t
raid5_personality
;
/*
* Stripe cache
*/
#define NR_STRIPES 256
#define STRIPE_SIZE PAGE_SIZE
#define STRIPE_SECTORS (STRIPE_SIZE>>9)
#define IO_THRESHOLD 1
#define HASH_PAGES 1
#define HASH_PAGES_ORDER 0
#define NR_HASH (HASH_PAGES * PAGE_SIZE / sizeof(struct stripe_head *))
#define HASH_MASK (NR_HASH - 1)
#define stripe_hash(conf, sect) ((conf)->stripe_hashtbl[((sect) /
((conf)->buffer_size >> 9)
) & HASH_MASK])
#define stripe_hash(conf, sect) ((conf)->stripe_hashtbl[((sect) /
STRIPE_SECTORS
) & HASH_MASK])
/*
* The following can be used to debug the driver
...
...
@@ -142,47 +142,36 @@ static struct stripe_head *get_free_stripe(raid5_conf_t *conf)
static
void
shrink_buffers
(
struct
stripe_head
*
sh
,
int
num
)
{
struct
buffer_head
*
bh
;
struct
page
*
p
;
int
i
;
for
(
i
=
0
;
i
<
num
;
i
++
)
{
bh
=
sh
->
bh_cache
[
i
];
if
(
!
bh
)
return
;
sh
->
bh_cache
[
i
]
=
NULL
;
free_page
((
unsigned
long
)
bh
->
b_data
);
kfree
(
bh
);
p
=
sh
->
dev
[
i
].
page
;
if
(
!
p
)
continue
;
sh
->
dev
[
i
].
page
=
NULL
;
page_cache_release
(
p
);
}
}
static
int
grow_buffers
(
struct
stripe_head
*
sh
,
int
num
,
int
b_size
,
int
priority
)
static
int
grow_buffers
(
struct
stripe_head
*
sh
,
int
num
)
{
struct
buffer_head
*
bh
;
int
i
;
for
(
i
=
0
;
i
<
num
;
i
++
)
{
struct
page
*
page
;
bh
=
kmalloc
(
sizeof
(
struct
buffer_head
),
priority
);
if
(
!
bh
)
return
1
;
memset
(
bh
,
0
,
sizeof
(
struct
buffer_head
));
if
((
page
=
alloc_page
(
priority
)))
bh
->
b_data
=
page_address
(
page
);
else
{
kfree
(
bh
);
if
(
!
(
page
=
alloc_page
(
GFP_KERNEL
)))
{
return
1
;
}
atomic_set
(
&
bh
->
b_count
,
0
);
bh
->
b_page
=
page
;
sh
->
bh_cache
[
i
]
=
bh
;
sh
->
dev
[
i
].
page
=
page
;
}
return
0
;
}
static
struct
buffer_head
*
raid5_build_block
(
struct
stripe_head
*
sh
,
int
i
);
static
void
raid5_build_block
(
struct
stripe_head
*
sh
,
int
i
);
static
inline
void
init_stripe
(
struct
stripe_head
*
sh
,
unsigned
long
sector
)
static
inline
void
init_stripe
(
struct
stripe_head
*
sh
,
unsigned
long
sector
,
int
pd_idx
)
{
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
int
disks
=
conf
->
raid_disks
,
i
;
...
...
@@ -198,40 +187,26 @@ static inline void init_stripe(struct stripe_head *sh, unsigned long sector)
remove_hash
(
sh
);
sh
->
sector
=
sector
;
sh
->
size
=
conf
->
buffer_size
;
sh
->
pd_idx
=
pd_idx
;
sh
->
state
=
0
;
for
(
i
=
disks
;
i
--
;
)
{
if
(
sh
->
bh_read
[
i
]
||
sh
->
bh_write
[
i
]
||
sh
->
bh_written
[
i
]
||
buffer_locked
(
sh
->
bh_cache
[
i
]))
{
struct
r5dev
*
dev
=
&
sh
->
dev
[
i
];
if
(
dev
->
toread
||
dev
->
towrite
||
dev
->
written
||
test_bit
(
R5_LOCKED
,
&
dev
->
flags
))
{
printk
(
"sector=%lx i=%d %p %p %p %d
\n
"
,
sh
->
sector
,
i
,
sh
->
bh_read
[
i
]
,
sh
->
bh_write
[
i
],
sh
->
bh_written
[
i
]
,
buffer_locked
(
sh
->
bh_cache
[
i
]
));
sh
->
sector
,
i
,
dev
->
toread
,
dev
->
towrite
,
dev
->
written
,
test_bit
(
R5_LOCKED
,
&
dev
->
flags
));
BUG
();
}
clear_buffer_uptodate
(
sh
->
bh_cache
[
i
])
;
dev
->
flags
=
0
;
raid5_build_block
(
sh
,
i
);
}
insert_hash
(
conf
,
sh
);
}
/* the buffer size has changed, so unhash all stripes
* as active stripes complete, they will go onto inactive list
*/
static
void
shrink_stripe_cache
(
raid5_conf_t
*
conf
)
{
int
i
;
CHECK_DEVLOCK
();
if
(
atomic_read
(
&
conf
->
active_stripes
))
BUG
();
for
(
i
=
0
;
i
<
NR_HASH
;
i
++
)
{
struct
stripe_head
*
sh
;
while
((
sh
=
conf
->
stripe_hashtbl
[
i
]))
remove_hash
(
sh
);
}
}
static
struct
stripe_head
*
__find_stripe
(
raid5_conf_t
*
conf
,
unsigned
long
sector
)
{
struct
stripe_head
*
sh
;
...
...
@@ -245,53 +220,16 @@ static struct stripe_head *__find_stripe(raid5_conf_t *conf, unsigned long secto
return
NULL
;
}
static
struct
stripe_head
*
get_active_stripe
(
raid5_conf_t
*
conf
,
unsigned
long
sector
,
int
size
,
int
noblock
)
static
struct
stripe_head
*
get_active_stripe
(
raid5_conf_t
*
conf
,
unsigned
long
sector
,
int
pd_idx
,
int
noblock
)
{
struct
stripe_head
*
sh
;
PRINTK
(
"get_stripe, sector %lu
\n
"
,
sector
);
md_
spin_lock_irq
(
&
conf
->
device_lock
);
spin_lock_irq
(
&
conf
->
device_lock
);
do
{
if
(
conf
->
buffer_size
==
0
||
(
size
&&
size
!=
conf
->
buffer_size
))
{
/* either the size is being changed (buffer_size==0) or
* we need to change it.
* If size==0, we can proceed as soon as buffer_size gets set.
* If size>0, we can proceed when active_stripes reaches 0, or
* when someone else sets the buffer_size to size.
* If someone sets the buffer size to something else, we will need to
* assert that we want to change it again
*/
int
oldsize
=
conf
->
buffer_size
;
PRINTK
(
"get_stripe %ld/%d buffer_size is %d, %d active
\n
"
,
sector
,
size
,
conf
->
buffer_size
,
atomic_read
(
&
conf
->
active_stripes
));
if
(
size
==
0
)
wait_event_lock_irq
(
conf
->
wait_for_stripe
,
conf
->
buffer_size
,
conf
->
device_lock
);
else
{
while
(
conf
->
buffer_size
!=
size
&&
atomic_read
(
&
conf
->
active_stripes
))
{
conf
->
buffer_size
=
0
;
wait_event_lock_irq
(
conf
->
wait_for_stripe
,
atomic_read
(
&
conf
->
active_stripes
)
==
0
||
conf
->
buffer_size
,
conf
->
device_lock
);
PRINTK
(
"waited and now %ld/%d buffer_size is %d - %d active
\n
"
,
sector
,
size
,
conf
->
buffer_size
,
atomic_read
(
&
conf
->
active_stripes
));
}
if
(
conf
->
buffer_size
!=
size
)
{
printk
(
"raid5: switching cache buffer size, %d --> %d
\n
"
,
oldsize
,
size
);
shrink_stripe_cache
(
conf
);
if
(
size
==
0
)
BUG
();
conf
->
buffer_size
=
size
;
PRINTK
(
"size now %d
\n
"
,
conf
->
buffer_size
);
}
}
}
if
(
size
==
0
)
sector
-=
sector
&
((
conf
->
buffer_size
>>
9
)
-
1
);
sh
=
__find_stripe
(
conf
,
sector
);
if
(
!
sh
)
{
if
(
!
conf
->
inactive_blocked
)
...
...
@@ -307,7 +245,7 @@ static struct stripe_head *get_active_stripe(raid5_conf_t *conf, unsigned long s
conf
->
device_lock
);
conf
->
inactive_blocked
=
0
;
}
else
init_stripe
(
sh
,
sector
);
init_stripe
(
sh
,
sector
,
pd_idx
);
}
else
{
if
(
atomic_read
(
&
sh
->
count
))
{
if
(
!
list_empty
(
&
sh
->
lru
))
...
...
@@ -325,25 +263,35 @@ static struct stripe_head *get_active_stripe(raid5_conf_t *conf, unsigned long s
if
(
sh
)
atomic_inc
(
&
sh
->
count
);
md_
spin_unlock_irq
(
&
conf
->
device_lock
);
spin_unlock_irq
(
&
conf
->
device_lock
);
return
sh
;
}
static
int
grow_stripes
(
raid5_conf_t
*
conf
,
int
num
,
int
priority
)
static
int
grow_stripes
(
raid5_conf_t
*
conf
,
int
num
)
{
struct
stripe_head
*
sh
;
kmem_cache_t
*
sc
;
int
devs
=
conf
->
raid_disks
;
sprintf
(
conf
->
cache_name
,
"md/raid5-%d"
,
conf
->
mddev
->
__minor
);
sc
=
kmem_cache_create
(
conf
->
cache_name
,
sizeof
(
struct
stripe_head
)
+
(
devs
-
1
)
*
sizeof
(
struct
r5dev
),
0
,
0
,
NULL
,
NULL
);
if
(
!
sc
)
return
1
;
conf
->
slab_cache
=
sc
;
while
(
num
--
)
{
sh
=
km
alloc
(
sizeof
(
struct
stripe_head
),
priority
);
sh
=
km
em_cache_alloc
(
sc
,
GFP_KERNEL
);
if
(
!
sh
)
return
1
;
memset
(
sh
,
0
,
sizeof
(
*
sh
));
memset
(
sh
,
0
,
sizeof
(
*
sh
)
+
(
devs
-
1
)
*
sizeof
(
struct
r5dev
)
);
sh
->
raid_conf
=
conf
;
sh
->
lock
=
SPIN_LOCK_UNLOCKED
;
if
(
grow_buffers
(
sh
,
conf
->
raid_disks
,
PAGE_SIZE
,
priority
))
{
if
(
grow_buffers
(
sh
,
conf
->
raid_disks
))
{
shrink_buffers
(
sh
,
conf
->
raid_disks
);
k
free
(
sh
);
k
mem_cache_free
(
sc
,
sh
);
return
1
;
}
/* we just created an active stripe so... */
...
...
@@ -355,11 +303,11 @@ static int grow_stripes(raid5_conf_t *conf, int num, int priority)
return
0
;
}
static
void
shrink_stripes
(
raid5_conf_t
*
conf
,
int
num
)
static
void
shrink_stripes
(
raid5_conf_t
*
conf
)
{
struct
stripe_head
*
sh
;
while
(
num
--
)
{
while
(
1
)
{
spin_lock_irq
(
&
conf
->
device_lock
);
sh
=
get_free_stripe
(
conf
);
spin_unlock_irq
(
&
conf
->
device_lock
);
...
...
@@ -368,21 +316,22 @@ static void shrink_stripes(raid5_conf_t *conf, int num)
if
(
atomic_read
(
&
sh
->
count
))
BUG
();
shrink_buffers
(
sh
,
conf
->
raid_disks
);
k
free
(
sh
);
k
mem_cache_free
(
conf
->
slab_cache
,
sh
);
atomic_dec
(
&
conf
->
active_stripes
);
}
kmem_cache_destroy
(
conf
->
slab_cache
);
conf
->
slab_cache
=
NULL
;
}
static
void
raid5_end_read_request
(
struct
buffer_head
*
bh
,
int
uptodate
)
static
void
raid5_end_read_request
(
struct
bio
*
bi
)
{
struct
stripe_head
*
sh
=
b
h
->
b
_private
;
struct
stripe_head
*
sh
=
b
i
->
bi
_private
;
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
int
disks
=
conf
->
raid_disks
,
i
;
unsigned
long
flags
;
int
uptodate
=
test_bit
(
BIO_UPTODATE
,
&
bi
->
bi_flags
)
;
for
(
i
=
0
;
i
<
disks
;
i
++
)
if
(
b
h
==
sh
->
bh_cache
[
i
]
)
if
(
b
i
==
&
sh
->
dev
[
i
].
req
)
break
;
PRINTK
(
"end_read_request %lu/%d, count: %d, uptodate %d.
\n
"
,
sh
->
sector
,
i
,
atomic_read
(
&
sh
->
count
),
uptodate
);
...
...
@@ -392,7 +341,9 @@ static void raid5_end_read_request (struct buffer_head * bh, int uptodate)
}
if
(
uptodate
)
{
struct
buffer_head
*
buffer
;
#if 0
struct bio *bio;
unsigned long flags;
spin_lock_irqsave(&conf->device_lock, flags);
/* we can return a buffer if we bypassed the cache or
* if the top buffer is not in highmem. If there are
...
...
@@ -409,38 +360,43 @@ static void raid5_end_read_request (struct buffer_head * bh, int uptodate)
} else
buffer = NULL;
spin_unlock_irqrestore(&conf->device_lock, flags);
if
(
sh
->
bh_page
[
i
]
==
NULL
)
if (sh->bh_page[i]==
bh->b_page
)
set_buffer_uptodate(bh);
if (buffer) {
if (buffer->b_page != bh->b_page)
memcpy(buffer->b_data, bh->b_data, bh->b_size);
buffer->b_end_io(buffer, 1);
}
#else
set_bit
(
R5_UPTODATE
,
&
sh
->
dev
[
i
].
flags
);
#endif
}
else
{
md_error
(
conf
->
mddev
,
b
h
->
b
_bdev
);
clear_b
uffer_uptodate
(
bh
);
md_error
(
conf
->
mddev
,
b
i
->
bi
_bdev
);
clear_b
it
(
R5_UPTODATE
,
&
sh
->
dev
[
i
].
flags
);
}
#if 0
/* must restore b_page before unlocking buffer... */
if
(
sh
->
bh_page
[
i
])
{
if (sh->bh_page[i]
!= bh->b_page
) {
bh->b_page = sh->bh_page[i];
bh->b_data = page_address(bh->b_page);
sh
->
bh_page
[
i
]
=
NULL
;
clear_buffer_uptodate(bh);
}
clear_buffer_locked
(
bh
);
#endif
clear_bit
(
R5_LOCKED
,
&
sh
->
dev
[
i
].
flags
);
set_bit
(
STRIPE_HANDLE
,
&
sh
->
state
);
release_stripe
(
sh
);
}
static
void
raid5_end_write_request
(
struct
b
uffer_head
*
bh
,
int
uptodate
)
static
void
raid5_end_write_request
(
struct
b
io
*
bi
)
{
struct
stripe_head
*
sh
=
b
h
->
b
_private
;
struct
stripe_head
*
sh
=
b
i
->
bi
_private
;
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
int
disks
=
conf
->
raid_disks
,
i
;
unsigned
long
flags
;
int
uptodate
=
test_bit
(
BIO_UPTODATE
,
&
bi
->
bi_flags
);
for
(
i
=
0
;
i
<
disks
;
i
++
)
if
(
b
h
==
sh
->
bh_cache
[
i
]
)
if
(
b
i
==
&
sh
->
dev
[
i
].
req
)
break
;
PRINTK
(
"end_write_request %lu/%d, count %d, uptodate: %d.
\n
"
,
sh
->
sector
,
i
,
atomic_read
(
&
sh
->
count
),
uptodate
);
...
...
@@ -449,41 +405,48 @@ static void raid5_end_write_request (struct buffer_head *bh, int uptodate)
return
;
}
md_
spin_lock_irqsave
(
&
conf
->
device_lock
,
flags
);
spin_lock_irqsave
(
&
conf
->
device_lock
,
flags
);
if
(
!
uptodate
)
md_error
(
conf
->
mddev
,
bh
->
b_bdev
);
clear_buffer_locked
(
bh
);
md_error
(
conf
->
mddev
,
bi
->
bi_bdev
);
clear_bit
(
R5_LOCKED
,
&
sh
->
dev
[
i
].
flags
);
set_bit
(
STRIPE_HANDLE
,
&
sh
->
state
);
__release_stripe
(
conf
,
sh
);
md_
spin_unlock_irqrestore
(
&
conf
->
device_lock
,
flags
);
spin_unlock_irqrestore
(
&
conf
->
device_lock
,
flags
);
}
static
unsigned
long
compute_blocknr
(
struct
stripe_head
*
sh
,
int
i
);
static
struct
buffer_head
*
raid5_build_block
(
struct
stripe_head
*
sh
,
int
i
)
static
void
raid5_build_block
(
struct
stripe_head
*
sh
,
int
i
)
{
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
struct
buffer_head
*
bh
=
sh
->
bh_cache
[
i
];
unsigned
long
block
=
sh
->
sector
/
(
sh
->
size
>>
9
);
init_buffer
(
bh
,
raid5_end_read_request
,
sh
);
bh
->
b_dev
=
conf
->
disks
[
i
].
dev
;
/* FIXME - later we will need bdev here */
bh
->
b_blocknr
=
block
;
bh
->
b_state
=
(
1
<<
BH_Req
)
|
(
1
<<
BH_Mapped
);
bh
->
b_size
=
sh
->
size
;
return
bh
;
struct
r5dev
*
dev
=
&
sh
->
dev
[
i
];
bio_init
(
&
dev
->
req
);
dev
->
req
.
bi_io_vec
=
&
dev
->
vec
;
dev
->
req
.
bi_vcnt
++
;
dev
->
vec
.
bv_page
=
dev
->
page
;
dev
->
vec
.
bv_len
=
STRIPE_SIZE
;
dev
->
vec
.
bv_offset
=
0
;
dev
->
req
.
bi_bdev
=
conf
->
disks
[
i
].
bdev
;
dev
->
req
.
bi_sector
=
sh
->
sector
;
dev
->
req
.
bi_private
=
sh
;
dev
->
flags
=
0
;
if
(
i
!=
sh
->
pd_idx
)
dev
->
sector
=
compute_blocknr
(
sh
,
i
);
}
static
int
raid5_
error
(
mddev_t
*
mddev
,
kdev_t
dev
)
static
int
error
(
mddev_t
*
mddev
,
kdev_t
dev
)
{
raid5_conf_t
*
conf
=
(
raid5_conf_t
*
)
mddev
->
private
;
mdp_super_t
*
sb
=
mddev
->
sb
;
struct
disk_info
*
disk
;
int
i
;
PRINTK
(
"raid5
_
error called
\n
"
);
PRINTK
(
"raid5
:
error called
\n
"
);
for
(
i
=
0
,
disk
=
conf
->
disks
;
i
<
conf
->
raid_disks
;
i
++
,
disk
++
)
{
if
(
kdev_same
(
disk
->
dev
,
dev
))
{
...
...
@@ -544,14 +507,14 @@ static int raid5_error (mddev_t *mddev, kdev_t dev)
* Input: a 'big' sector number,
* Output: index of the data and parity disk, and the sector # in them.
*/
static
unsigned
long
raid5_compute_sector
(
unsigned
long
r_sector
,
unsigned
int
raid_disks
,
static
unsigned
long
raid5_compute_sector
(
sector_t
r_sector
,
unsigned
int
raid_disks
,
unsigned
int
data_disks
,
unsigned
int
*
dd_idx
,
unsigned
int
*
pd_idx
,
raid5_conf_t
*
conf
)
{
unsigned
long
stripe
;
sector_t
stripe
;
unsigned
long
chunk_number
;
unsigned
int
chunk_offset
;
unsigned
long
new_sector
;
sector_t
new_sector
;
int
sectors_per_chunk
=
conf
->
chunk_size
>>
9
;
/* First compute the information on this sector */
...
...
@@ -607,17 +570,17 @@ static unsigned long raid5_compute_sector(unsigned long r_sector, unsigned int r
return
new_sector
;
}
#if 0
static
unsigned long
compute_blocknr(struct stripe_head *sh, int i)
static
sector_t
compute_blocknr
(
struct
stripe_head
*
sh
,
int
i
)
{
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
int
raid_disks
=
conf
->
raid_disks
,
data_disks
=
raid_disks
-
1
;
unsigned long
new_sector = sh->sector, check;
sector_t
new_sector
=
sh
->
sector
,
check
;
int
sectors_per_chunk
=
conf
->
chunk_size
>>
9
;
unsigned long
stripe = new_sector / sectors_per_chunk;
sector_t
stripe
=
new_sector
/
sectors_per_chunk
;
int
chunk_offset
=
new_sector
%
sectors_per_chunk
;
int
chunk_number
,
dummy1
,
dummy2
,
dd_idx
=
i
;
unsigned long r_sector, blockn
r;
sector_t
r_secto
r
;
switch
(
conf
->
algorithm
)
{
case
ALGORITHM_LEFT_ASYMMETRIC
:
...
...
@@ -637,20 +600,70 @@ static unsigned long compute_blocknr(struct stripe_head *sh, int i)
chunk_number
=
stripe
*
data_disks
+
i
;
r_sector
=
chunk_number
*
sectors_per_chunk
+
chunk_offset
;
blocknr = r_sector / (sh->size >> 9);
check
=
raid5_compute_sector
(
r_sector
,
raid_disks
,
data_disks
,
&
dummy1
,
&
dummy2
,
conf
);
if
(
check
!=
sh
->
sector
||
dummy1
!=
dd_idx
||
dummy2
!=
sh
->
pd_idx
)
{
printk
(
"compute_blocknr: map not correct
\n
"
);
return
0
;
}
return blocknr;
return
r_sector
;
}
/*
* Copy data between a page in the stripe cache, and one or more bion
* The page could align with the middle of the bio, or there could be
* several bion, each with several bio_vecs, which cover part of the page
* Multiple bion are linked together on bi_next. There may be extras
* at the end of this list. We ignore them.
*/
static
void
copy_data
(
int
frombio
,
struct
bio
*
bio
,
struct
page
*
page
,
sector_t
sector
)
{
char
*
pa
=
page_address
(
page
);
struct
bio_vec
*
bvl
;
int
i
;
for
(;
bio
&&
bio
->
bi_sector
<
sector
+
STRIPE_SECTORS
;
bio
=
bio
->
bi_next
)
{
int
page_offset
;
if
(
bio
->
bi_sector
>=
sector
)
page_offset
=
(
signed
)(
bio
->
bi_sector
-
sector
)
*
512
;
else
page_offset
=
(
signed
)(
sector
-
bio
->
bi_sector
)
*
-
512
;
bio_for_each_segment
(
bvl
,
bio
,
i
)
{
char
*
ba
=
__bio_kmap
(
bio
,
i
);
int
len
=
bio_iovec_idx
(
bio
,
i
)
->
bv_len
;
int
clen
;
int
b_offset
=
0
;
if
(
page_offset
<
0
)
{
b_offset
=
-
page_offset
;
page_offset
+=
b_offset
;
len
-=
b_offset
;
}
if
(
len
>
0
&&
page_offset
+
len
>
STRIPE_SIZE
)
clen
=
STRIPE_SIZE
-
page_offset
;
else
clen
=
len
;
if
(
len
>
0
)
{
if
(
frombio
)
memcpy
(
pa
+
page_offset
,
ba
+
b_offset
,
clen
);
else
memcpy
(
ba
+
b_offset
,
pa
+
page_offset
,
clen
);
}
__bio_kunmap
(
bio
,
i
);
page_offset
+=
len
;
}
}
}
#endif
#define check_xor() do { \
if (count == MAX_XOR_BLOCKS) { \
xor_block(count,
bh_
ptr); \
xor_block(count,
STRIPE_SIZE,
ptr); \
count = 1; \
} \
} while(0)
...
...
@@ -660,88 +673,84 @@ static void compute_block(struct stripe_head *sh, int dd_idx)
{
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
int
i
,
count
,
disks
=
conf
->
raid_disks
;
struct
buffer_head
*
bh_ptr
[
MAX_XOR_BLOCKS
],
*
bh
;
void
*
ptr
[
MAX_XOR_BLOCKS
],
*
p
;
PRINTK
(
"compute_block, stripe %lu, idx %d
\n
"
,
sh
->
sector
,
dd_idx
);
memset
(
sh
->
bh_cache
[
dd_idx
]
->
b_data
,
0
,
sh
->
size
);
bh_ptr
[
0
]
=
sh
->
bh_cache
[
dd_idx
];
ptr
[
0
]
=
page_address
(
sh
->
dev
[
dd_idx
].
page
);
memset
(
ptr
[
0
],
0
,
STRIPE_SIZE
);
count
=
1
;
for
(
i
=
disks
;
i
--
;
)
{
if
(
i
==
dd_idx
)
continue
;
bh
=
sh
->
bh_cache
[
i
]
;
if
(
buffer_uptodate
(
bh
))
bh_ptr
[
count
++
]
=
bh
;
p
=
page_address
(
sh
->
dev
[
i
].
page
)
;
if
(
test_bit
(
R5_UPTODATE
,
&
sh
->
dev
[
i
].
flags
))
ptr
[
count
++
]
=
p
;
else
printk
(
"compute_block() %d, stripe %lu, %d not present
\n
"
,
dd_idx
,
sh
->
sector
,
i
);
check_xor
();
}
if
(
count
!=
1
)
xor_block
(
count
,
bh_
ptr
);
set_b
uffer_uptodate
(
sh
->
bh_cache
[
dd_idx
]
);
xor_block
(
count
,
STRIPE_SIZE
,
ptr
);
set_b
it
(
R5_UPTODATE
,
&
sh
->
dev
[
i
].
flags
);
}
static
void
compute_parity
(
struct
stripe_head
*
sh
,
int
method
)
{
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
int
i
,
pd_idx
=
sh
->
pd_idx
,
disks
=
conf
->
raid_disks
,
count
;
struct
buffer_head
*
bh_
ptr
[
MAX_XOR_BLOCKS
];
struct
b
uffer_head
*
chosen
[
MD_SB_DISKS
];
void
*
ptr
[
MAX_XOR_BLOCKS
];
struct
b
io
*
chosen
[
MD_SB_DISKS
];
PRINTK
(
"compute_parity, stripe %lu, method %d
\n
"
,
sh
->
sector
,
method
);
memset
(
chosen
,
0
,
sizeof
(
chosen
));
count
=
1
;
bh_ptr
[
0
]
=
sh
->
bh_cache
[
pd_idx
]
;
ptr
[
0
]
=
page_address
(
sh
->
dev
[
pd_idx
].
page
)
;
switch
(
method
)
{
case
READ_MODIFY_WRITE
:
if
(
!
buffer_uptodate
(
sh
->
bh_cache
[
pd_idx
]
))
if
(
!
test_bit
(
R5_UPTODATE
,
&
sh
->
dev
[
pd_idx
].
flags
))
BUG
();
for
(
i
=
disks
;
i
--
;)
{
if
(
i
==
pd_idx
)
continue
;
if
(
sh
->
bh_write
[
i
]
&&
buffer_uptodate
(
sh
->
bh_cache
[
i
]
))
{
bh_ptr
[
count
++
]
=
sh
->
bh_cache
[
i
]
;
chosen
[
i
]
=
sh
->
bh_write
[
i
]
;
sh
->
bh_write
[
i
]
=
sh
->
bh_write
[
i
]
->
b_reqnext
;
chosen
[
i
]
->
b_reqnext
=
sh
->
bh_written
[
i
]
;
sh
->
bh_written
[
i
]
=
chosen
[
i
];
if
(
sh
->
dev
[
i
].
towrite
&&
test_bit
(
R5_UPTODATE
,
&
sh
->
dev
[
i
].
flags
))
{
ptr
[
count
++
]
=
page_address
(
sh
->
dev
[
i
].
page
)
;
chosen
[
i
]
=
sh
->
dev
[
i
].
towrite
;
sh
->
dev
[
i
].
towrite
=
NULL
;
if
(
sh
->
dev
[
i
].
written
)
BUG
()
;
sh
->
dev
[
i
].
written
=
chosen
[
i
];
check_xor
();
}
}
break
;
case
RECONSTRUCT_WRITE
:
memset
(
sh
->
bh_cache
[
pd_idx
]
->
b_data
,
0
,
sh
->
size
);
memset
(
ptr
[
0
],
0
,
STRIPE_SIZE
);
for
(
i
=
disks
;
i
--
;)
if
(
i
!=
pd_idx
&&
sh
->
bh_write
[
i
]
)
{
chosen
[
i
]
=
sh
->
bh_write
[
i
]
;
sh
->
bh_write
[
i
]
=
sh
->
bh_write
[
i
]
->
b_reqnext
;
chosen
[
i
]
->
b_reqnext
=
sh
->
bh_written
[
i
]
;
sh
->
bh_written
[
i
]
=
chosen
[
i
];
if
(
i
!=
pd_idx
&&
sh
->
dev
[
i
].
towrite
)
{
chosen
[
i
]
=
sh
->
dev
[
i
].
towrite
;
sh
->
dev
[
i
].
towrite
=
NULL
;
if
(
sh
->
dev
[
i
].
written
)
BUG
()
;
sh
->
dev
[
i
].
written
=
chosen
[
i
];
}
break
;
case
CHECK_PARITY
:
break
;
}
if
(
count
>
1
)
{
xor_block
(
count
,
bh_
ptr
);
xor_block
(
count
,
STRIPE_SIZE
,
ptr
);
count
=
1
;
}
for
(
i
=
disks
;
i
--
;)
if
(
chosen
[
i
])
{
struct
buffer_head
*
bh
=
sh
->
bh_cache
[
i
];
char
*
bdata
;
bdata
=
bh_kmap
(
chosen
[
i
]);
memcpy
(
bh
->
b_data
,
bdata
,
sh
->
size
);
bh_kunmap
(
chosen
[
i
]);
set_buffer_locked
(
bh
);
set_buffer_uptodate
(
bh
);
sector_t
sector
=
sh
->
dev
[
i
].
sector
;
copy_data
(
1
,
chosen
[
i
],
sh
->
dev
[
i
].
page
,
sector
);
set_bit
(
R5_LOCKED
,
&
sh
->
dev
[
i
].
flags
);
set_bit
(
R5_UPTODATE
,
&
sh
->
dev
[
i
].
flags
);
}
switch
(
method
)
{
...
...
@@ -749,55 +758,74 @@ static void compute_parity(struct stripe_head *sh, int method)
case
CHECK_PARITY
:
for
(
i
=
disks
;
i
--
;)
if
(
i
!=
pd_idx
)
{
bh_ptr
[
count
++
]
=
sh
->
bh_cache
[
i
]
;
ptr
[
count
++
]
=
page_address
(
sh
->
dev
[
i
].
page
)
;
check_xor
();
}
break
;
case
READ_MODIFY_WRITE
:
for
(
i
=
disks
;
i
--
;)
if
(
chosen
[
i
])
{
bh_ptr
[
count
++
]
=
sh
->
bh_cache
[
i
]
;
ptr
[
count
++
]
=
page_address
(
sh
->
dev
[
i
].
page
)
;
check_xor
();
}
}
if
(
count
!=
1
)
xor_block
(
count
,
bh_
ptr
);
xor_block
(
count
,
STRIPE_SIZE
,
ptr
);
if
(
method
!=
CHECK_PARITY
)
{
set_b
uffer_uptodate
(
sh
->
bh_cache
[
pd_idx
]
);
set_b
uffer_locked
(
sh
->
bh_cache
[
pd_idx
]
);
set_b
it
(
R5_UPTODATE
,
&
sh
->
dev
[
pd_idx
].
flags
);
set_b
it
(
R5_LOCKED
,
&
sh
->
dev
[
pd_idx
].
flags
);
}
else
clear_b
uffer_uptodate
(
sh
->
bh_cache
[
pd_idx
]
);
clear_b
it
(
R5_UPTODATE
,
&
sh
->
dev
[
pd_idx
].
flags
);
}
static
void
add_stripe_bh
(
struct
stripe_head
*
sh
,
struct
buffer_head
*
bh
,
int
dd_idx
,
int
rw
)
/*
* Each stripe/dev can have one or more bion attached.
* toread/towrite point to the first in a chain.
* The bi_next chain must be in order.
*/
static
void
add_stripe_bio
(
struct
stripe_head
*
sh
,
struct
bio
*
bi
,
int
dd_idx
,
int
forwrite
)
{
struct
b
uffer_head
**
bh
p
;
struct
b
io
**
bi
p
;
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
PRINTK
(
"adding bh b#%lu to stripe s#%lu
\n
"
,
b
h
->
b_blockn
r
,
sh
->
sector
);
PRINTK
(
"adding bh b#%lu to stripe s#%lu
\n
"
,
b
i
->
bi_secto
r
,
sh
->
sector
);
spin_lock
(
&
sh
->
lock
);
spin_lock_irq
(
&
conf
->
device_lock
);
bh
->
b_reqnext
=
NULL
;
if
(
rw
==
READ
)
bhp
=
&
sh
->
bh_read
[
dd_idx
];
if
(
forwrite
)
bip
=
&
sh
->
dev
[
dd_idx
].
towrite
;
else
bhp
=
&
sh
->
bh_write
[
dd_idx
];
while
(
*
bhp
)
{
printk
(
KERN_NOTICE
"raid5: multiple %d requests for sector %ld
\n
"
,
rw
,
sh
->
sector
);
bhp
=
&
(
*
bhp
)
->
b_reqnext
;
}
*
bhp
=
bh
;
bip
=
&
sh
->
dev
[
dd_idx
].
toread
;
while
(
*
bip
&&
(
*
bip
)
->
bi_sector
<
bi
->
bi_sector
)
bip
=
&
(
*
bip
)
->
bi_next
;
/* FIXME do I need to worry about overlapping bion */
if
(
*
bip
&&
bi
->
bi_next
&&
(
*
bip
)
!=
bi
->
bi_next
)
BUG
();
if
(
*
bip
)
bi
->
bi_next
=
*
bip
;
*
bip
=
bi
;
bi
->
bi_phys_segments
++
;
spin_unlock_irq
(
&
conf
->
device_lock
);
spin_unlock
(
&
sh
->
lock
);
PRINTK
(
"added bh b#%lu to stripe s#%lu, disk %d.
\n
"
,
bh
->
b_blocknr
,
sh
->
sector
,
dd_idx
);
}
if
(
forwrite
)
{
/* check if page is coverred */
sector_t
sector
=
sh
->
dev
[
dd_idx
].
sector
;
for
(
bi
=
sh
->
dev
[
dd_idx
].
towrite
;
sector
<
sh
->
dev
[
dd_idx
].
sector
+
STRIPE_SECTORS
&&
bi
&&
bi
->
bi_sector
<=
sector
;
bi
=
bi
->
bi_next
)
{
if
(
bi
->
bi_sector
+
(
bi
->
bi_size
>>
9
)
>=
sector
)
sector
=
bi
->
bi_sector
+
(
bi
->
bi_size
>>
9
);
}
if
(
sector
>=
sh
->
dev
[
dd_idx
].
sector
+
STRIPE_SECTORS
)
set_bit
(
R5_OVERWRITE
,
&
sh
->
dev
[
dd_idx
].
flags
);
}
PRINTK
(
"added bi b#%lu to stripe s#%lu, disk %d.
\n
"
,
bi
->
bi_sector
,
sh
->
sector
,
dd_idx
);
}
/*
...
...
@@ -822,13 +850,14 @@ static void handle_stripe(struct stripe_head *sh)
{
raid5_conf_t
*
conf
=
sh
->
raid_conf
;
int
disks
=
conf
->
raid_disks
;
struct
buffer_head
*
return_ok
=
NULL
,
*
return_fail
=
NULL
;
struct
bio
*
return_bi
=
NULL
;
struct
bio
*
bi
;
int
action
[
MD_SB_DISKS
];
int
i
;
int
syncing
;
int
locked
=
0
,
uptodate
=
0
,
to_read
=
0
,
to_write
=
0
,
failed
=
0
,
written
=
0
;
int
failed_num
=
0
;
struct
buffer_head
*
bh
;
struct
r5dev
*
dev
;
PRINTK
(
"handling stripe %ld, cnt=%d, pd_idx=%d
\n
"
,
sh
->
sector
,
atomic_read
(
&
sh
->
count
),
sh
->
pd_idx
);
memset
(
action
,
0
,
sizeof
(
action
));
...
...
@@ -841,36 +870,38 @@ static void handle_stripe(struct stripe_head *sh)
/* Now to look around and see what can be done */
for
(
i
=
disks
;
i
--
;
)
{
bh
=
sh
->
bh_cache
[
i
];
PRINTK
(
"check %d: state 0x%lx read %p write %p written %p
\n
"
,
i
,
bh
->
b_state
,
sh
->
bh_read
[
i
],
sh
->
bh_write
[
i
],
sh
->
bh_written
[
i
]);
dev
=
&
sh
->
dev
[
i
];
PRINTK
(
"check %d: state 0x%lx read %p write %p written %p
\n
"
,
i
,
dev
->
flags
,
dev
->
toread
,
dev
->
towrite
,
dev
->
written
);
/* maybe we can reply to a read */
if
(
buffer_uptodate
(
bh
)
&&
sh
->
bh_read
[
i
]
)
{
struct
b
uffer_head
*
rbh
,
*
rbh
2
;
if
(
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
)
&&
dev
->
toread
)
{
struct
b
io
*
rbi
,
*
rbi
2
;
PRINTK
(
"Return read for disc %d
\n
"
,
i
);
spin_lock_irq
(
&
conf
->
device_lock
);
rb
h
=
sh
->
bh_read
[
i
]
;
sh
->
bh_read
[
i
]
=
NULL
;
rb
i
=
dev
->
toread
;
dev
->
toread
=
NULL
;
spin_unlock_irq
(
&
conf
->
device_lock
);
while
(
rbh
)
{
char
*
bdata
;
bdata
=
bh_kmap
(
rbh
);
memcpy
(
bdata
,
bh
->
b_data
,
bh
->
b_size
);
bh_kunmap
(
rbh
);
rbh2
=
rbh
->
b_reqnext
;
rbh
->
b_reqnext
=
return_ok
;
return_ok
=
rbh
;
rbh
=
rbh2
;
while
(
rbi
&&
rbi
->
bi_sector
<
dev
->
sector
+
STRIPE_SECTORS
)
{
copy_data
(
0
,
rbi
,
dev
->
page
,
dev
->
sector
);
rbi2
=
rbi
->
bi_next
;
spin_lock_irq
(
&
conf
->
device_lock
);
if
(
--
rbi
->
bi_phys_segments
==
0
)
{
rbi
->
bi_next
=
return_bi
;
return_bi
=
rbi
;
}
spin_unlock_irq
(
&
conf
->
device_lock
);
rbi
=
rbi2
;
}
}
/* now count some things */
if
(
buffer_locked
(
bh
))
locked
++
;
if
(
buffer_uptodate
(
bh
))
uptodate
++
;
if
(
test_bit
(
R5_LOCKED
,
&
dev
->
flags
))
locked
++
;
if
(
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
))
uptodate
++
;
if
(
sh
->
bh_read
[
i
]
)
to_read
++
;
if
(
sh
->
bh_write
[
i
]
)
to_write
++
;
if
(
sh
->
bh_written
[
i
]
)
written
++
;
if
(
dev
->
toread
)
to_read
++
;
if
(
dev
->
towrite
)
to_write
++
;
if
(
dev
->
written
)
written
++
;
if
(
!
conf
->
disks
[
i
].
operational
)
{
failed
++
;
failed_num
=
i
;
...
...
@@ -882,29 +913,42 @@ static void handle_stripe(struct stripe_head *sh)
* need to be failed
*/
if
(
failed
>
1
&&
to_read
+
to_write
)
{
spin_lock_irq
(
&
conf
->
device_lock
);
for
(
i
=
disks
;
i
--
;
)
{
/* fail all writes first */
if
(
sh
->
bh_write
[
i
])
to_write
--
;
while
((
bh
=
sh
->
bh_write
[
i
]))
{
sh
->
bh_write
[
i
]
=
bh
->
b_reqnext
;
bh
->
b_reqnext
=
return_fail
;
return_fail
=
bh
;
bi
=
sh
->
dev
[
i
].
towrite
;
sh
->
dev
[
i
].
towrite
=
NULL
;
if
(
bi
)
to_write
--
;
while
(
bi
&&
bi
->
bi_sector
<
sh
->
dev
[
i
].
sector
+
STRIPE_SECTORS
){
struct
bio
*
nextbi
=
bi
->
bi_next
;
clear_bit
(
BIO_UPTODATE
,
&
bi
->
bi_flags
);
if
(
--
bi
->
bi_phys_segments
==
0
)
{
bi
->
bi_next
=
return_bi
;
return_bi
=
bi
;
}
bi
=
nextbi
;
}
/* fail any reads if this device is non-operational */
if
(
!
conf
->
disks
[
i
].
operational
)
{
spin_lock_irq
(
&
conf
->
device_lock
);
if
(
sh
->
bh_read
[
i
])
to_read
--
;
while
((
bh
=
sh
->
bh_read
[
i
]))
{
sh
->
bh_read
[
i
]
=
bh
->
b_reqnext
;
bh
->
b_reqnext
=
return_fail
;
return_fail
=
bh
;
bi
=
sh
->
dev
[
i
].
toread
;
sh
->
dev
[
i
].
toread
=
NULL
;
if
(
bi
)
to_read
--
;
while
(
bi
&&
bi
->
bi_sector
<
sh
->
dev
[
i
].
sector
+
STRIPE_SECTORS
){
struct
bio
*
nextbi
=
bi
->
bi_next
;
clear_bit
(
BIO_UPTODATE
,
&
bi
->
bi_flags
);
if
(
--
bi
->
bi_phys_segments
==
0
)
{
bi
->
bi_next
=
return_bi
;
return_bi
=
bi
;
}
spin_unlock_irq
(
&
conf
->
device_lock
)
;
bi
=
nextbi
;
}
}
}
spin_unlock_irq
(
&
conf
->
device_lock
);
}
if
(
failed
>
1
&&
syncing
)
{
md_done_sync
(
conf
->
mddev
,
(
sh
->
size
>>
9
)
-
sh
->
sync_redone
,
0
);
md_done_sync
(
conf
->
mddev
,
STRIPE_SECTORS
,
0
);
clear_bit
(
STRIPE_SYNCING
,
&
sh
->
state
);
syncing
=
0
;
}
...
...
@@ -912,27 +956,30 @@ static void handle_stripe(struct stripe_head *sh)
/* might be able to return some write requests if the parity block
* is safe, or on a failed drive
*/
bh
=
sh
->
bh_cache
[
sh
->
pd_idx
];
dev
=
&
sh
->
dev
[
sh
->
pd_idx
];
if
(
written
&&
(
(
conf
->
disks
[
sh
->
pd_idx
].
operational
&&
!
buffer_locked
(
bh
)
&&
buffer_uptodate
(
bh
))
(
(
conf
->
disks
[
sh
->
pd_idx
].
operational
&&
!
test_bit
(
R5_LOCKED
,
&
dev
->
flags
)
&&
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
))
||
(
failed
==
1
&&
failed_num
==
sh
->
pd_idx
))
)
{
/* any written block on a uptodate or failed drive can be returned */
/* any written block on a
n
uptodate or failed drive can be returned */
for
(
i
=
disks
;
i
--
;
)
if
(
sh
->
bh_written
[
i
]
)
{
bh
=
sh
->
bh_cache
[
i
];
if
(
sh
->
dev
[
i
].
written
)
{
dev
=
&
sh
->
dev
[
i
];
if
(
!
conf
->
disks
[
sh
->
pd_idx
].
operational
||
(
!
buffer_locked
(
bh
)
&&
buffer_uptodate
(
bh
))
)
{
(
!
test_bit
(
R5_LOCKED
,
&
dev
->
flags
)
&&
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
))
)
{
/* maybe we can return some write requests */
struct
buffer_head
*
wbh
,
*
wbh
2
;
struct
bio
*
wbi
,
*
wbi
2
;
PRINTK
(
"Return write for disc %d
\n
"
,
i
);
wbh
=
sh
->
bh_written
[
i
];
sh
->
bh_written
[
i
]
=
NULL
;
while
(
wbh
)
{
wbh2
=
wbh
->
b_reqnext
;
wbh
->
b_reqnext
=
return_ok
;
return_ok
=
wbh
;
wbh
=
wbh2
;
wbi
=
dev
->
written
;
dev
->
written
=
NULL
;
while
(
wbi
&&
wbi
->
bi_sector
<
dev
->
sector
+
STRIPE_SECTORS
)
{
wbi2
=
wbi
->
bi_next
;
if
(
--
wbi
->
bi_phys_segments
==
0
)
{
wbi
->
bi_next
=
return_bi
;
return_bi
=
wbi
;
}
wbi
=
wbi2
;
}
}
}
...
...
@@ -943,9 +990,9 @@ static void handle_stripe(struct stripe_head *sh)
*/
if
(
to_read
||
(
syncing
&&
(
uptodate
+
failed
<
disks
)))
{
for
(
i
=
disks
;
i
--
;)
{
bh
=
sh
->
bh_cache
[
i
];
if
(
!
buffer_locked
(
bh
)
&&
!
buffer_uptodate
(
bh
)
&&
(
sh
->
bh_read
[
i
]
||
syncing
||
(
failed
&&
sh
->
bh_read
[
failed_num
]
)))
{
dev
=
&
sh
->
dev
[
i
];
if
(
!
test_bit
(
R5_LOCKED
,
&
dev
->
flags
)
&&
!
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
)
&&
(
dev
->
toread
||
syncing
||
(
failed
&&
sh
->
dev
[
failed_num
].
toread
)))
{
/* we would like to get this block, possibly
* by computing it, but we might not be able to
*/
...
...
@@ -954,21 +1001,21 @@ static void handle_stripe(struct stripe_head *sh)
compute_block
(
sh
,
i
);
uptodate
++
;
}
else
if
(
conf
->
disks
[
i
].
operational
)
{
set_b
uffer_locked
(
bh
);
set_b
it
(
R5_LOCKED
,
&
dev
->
flags
);
action
[
i
]
=
READ
+
1
;
#if 0
/* if I am just reading this block and we don't have
a failed drive, or any pending writes then sidestep the cache */
if
(
sh
->
bh_page
[
i
])
BUG
();
if (sh->bh_read[i] && !sh->bh_read[i]->b_reqnext &&
! syncing && !failed && !to_write) {
sh
->
bh_page
[
i
]
=
sh
->
bh_cache
[
i
]
->
b_page
;
sh->bh_cache[i]->b_page = sh->bh_read[i]->b_page;
sh->bh_cache[i]->b_data = sh->bh_read[i]->b_data;
}
#endif
locked
++
;
PRINTK
(
"Reading block %d (sync=%d)
\n
"
,
i
,
syncing
);
if
(
syncing
)
md_sync_acct
(
conf
->
disks
[
i
].
dev
,
bh
->
b_size
>>
9
);
md_sync_acct
(
conf
->
disks
[
i
].
dev
,
STRIPE_SECTORS
);
}
}
}
...
...
@@ -980,10 +1027,14 @@ static void handle_stripe(struct stripe_head *sh)
int
rmw
=
0
,
rcw
=
0
;
for
(
i
=
disks
;
i
--
;)
{
/* would I have to read this buffer for read_modify_write */
bh
=
sh
->
bh_cache
[
i
];
if
((
sh
->
bh_write
[
i
]
||
i
==
sh
->
pd_idx
)
&&
(
!
buffer_locked
(
bh
)
||
sh
->
bh_page
[
i
])
&&
!
buffer_uptodate
(
bh
))
{
dev
=
&
sh
->
dev
[
i
];
if
((
dev
->
towrite
||
i
==
sh
->
pd_idx
)
&&
(
!
test_bit
(
R5_LOCKED
,
&
dev
->
flags
)
#if 0
|| sh->bh_page[i]!=bh->b_page
#endif
)
&&
!
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
))
{
if
(
conf
->
disks
[
i
].
operational
/* && !(conf->resync_parity && i == sh->pd_idx) */
)
...
...
@@ -991,9 +1042,13 @@ static void handle_stripe(struct stripe_head *sh)
else
rmw
+=
2
*
disks
;
/* cannot read it */
}
/* Would I have to read this buffer for reconstruct_write */
if
(
!
sh
->
bh_write
[
i
]
&&
i
!=
sh
->
pd_idx
&&
(
!
buffer_locked
(
bh
)
||
sh
->
bh_page
[
i
])
&&
!
buffer_uptodate
(
bh
))
{
if
(
!
test_bit
(
R5_OVERWRITE
,
&
dev
->
flags
)
&&
i
!=
sh
->
pd_idx
&&
(
!
test_bit
(
R5_LOCKED
,
&
dev
->
flags
)
#if 0
|| sh->bh_page[i] != bh->b_page
#endif
)
&&
!
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
))
{
if
(
conf
->
disks
[
i
].
operational
)
rcw
++
;
else
rcw
+=
2
*
disks
;
}
...
...
@@ -1003,14 +1058,14 @@ static void handle_stripe(struct stripe_head *sh)
if
(
rmw
<
rcw
&&
rmw
>
0
)
/* prefer read-modify-write, but need to get some data */
for
(
i
=
disks
;
i
--
;)
{
bh
=
sh
->
bh_cache
[
i
];
if
((
sh
->
bh_write
[
i
]
||
i
==
sh
->
pd_idx
)
&&
!
buffer_locked
(
bh
)
&&
!
buffer_uptodate
(
bh
)
&&
dev
=
&
sh
->
dev
[
i
];
if
((
dev
->
towrite
||
i
==
sh
->
pd_idx
)
&&
!
test_bit
(
R5_LOCKED
,
&
dev
->
flags
)
&&
!
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
)
&&
conf
->
disks
[
i
].
operational
)
{
if
(
test_bit
(
STRIPE_PREREAD_ACTIVE
,
&
sh
->
state
))
{
PRINTK
(
"Read_old block %d for r-m-w
\n
"
,
i
);
set_b
uffer_locked
(
bh
);
set_b
it
(
R5_LOCKED
,
&
dev
->
flags
);
action
[
i
]
=
READ
+
1
;
locked
++
;
}
else
{
...
...
@@ -1022,14 +1077,14 @@ static void handle_stripe(struct stripe_head *sh)
if
(
rcw
<=
rmw
&&
rcw
>
0
)
/* want reconstruct write, but need to get some data */
for
(
i
=
disks
;
i
--
;)
{
bh
=
sh
->
bh_cache
[
i
];
if
(
!
sh
->
bh_write
[
i
]
&&
i
!=
sh
->
pd_idx
&&
!
buffer_locked
(
bh
)
&&
!
buffer_uptodate
(
bh
)
&&
dev
=
&
sh
->
dev
[
i
];
if
(
!
test_bit
(
R5_OVERWRITE
,
&
dev
->
flags
)
&&
i
!=
sh
->
pd_idx
&&
!
test_bit
(
R5_LOCKED
,
&
dev
->
flags
)
&&
!
test_bit
(
R5_UPTODATE
,
&
dev
->
flags
)
&&
conf
->
disks
[
i
].
operational
)
{
if
(
test_bit
(
STRIPE_PREREAD_ACTIVE
,
&
sh
->
state
))
{
PRINTK
(
"Read_old block %d for Reconstruct
\n
"
,
i
);
set_b
uffer_locked
(
bh
);
set_b
it
(
R5_LOCKED
,
&
dev
->
flags
);
action
[
i
]
=
READ
+
1
;
locked
++
;
}
else
{
...
...
@@ -1044,7 +1099,7 @@ static void handle_stripe(struct stripe_head *sh)
compute_parity
(
sh
,
rcw
==
0
?
RECONSTRUCT_WRITE
:
READ_MODIFY_WRITE
);
/* now every locked buffer is ready to be written */
for
(
i
=
disks
;
i
--
;)
if
(
buffer_locked
(
sh
->
bh_cache
[
i
]
))
{
if
(
test_bit
(
R5_LOCKED
,
&
sh
->
dev
[
i
].
flags
))
{
PRINTK
(
"Writing block %d
\n
"
,
i
);
locked
++
;
action
[
i
]
=
WRITE
+
1
;
...
...
@@ -1068,13 +1123,14 @@ static void handle_stripe(struct stripe_head *sh)
!
test_bit
(
STRIPE_INSYNC
,
&
sh
->
state
)
&&
failed
<=
1
)
{
set_bit
(
STRIPE_HANDLE
,
&
sh
->
state
);
if
(
failed
==
0
)
{
char
*
pagea
;
if
(
uptodate
!=
disks
)
BUG
();
compute_parity
(
sh
,
CHECK_PARITY
);
uptodate
--
;
bh
=
sh
->
bh_cache
[
sh
->
pd_idx
]
;
if
((
*
(
u32
*
)
bh
->
b_dat
a
)
==
0
&&
!
memcmp
(
bh
->
b_data
,
bh
->
b_data
+
4
,
bh
->
b_size
-
4
))
{
pagea
=
page_address
(
sh
->
dev
[
sh
->
pd_idx
].
page
)
;
if
((
*
(
u32
*
)
page
a
)
==
0
&&
!
memcmp
(
pagea
,
pagea
+
4
,
STRIPE_SIZE
-
4
))
{
/* parity is correct (on disc, not in buffer any more) */
set_bit
(
STRIPE_INSYNC
,
&
sh
->
state
);
}
...
...
@@ -1084,7 +1140,7 @@ static void handle_stripe(struct stripe_head *sh)
if
(
failed
==
0
)
failed_num
=
sh
->
pd_idx
;
/* should be able to compute the missing block and write it to spare */
if
(
!
buffer_uptodate
(
sh
->
bh_cache
[
failed_num
]
))
{
if
(
!
test_bit
(
R5_UPTODATE
,
&
sh
->
dev
[
failed_num
].
flags
))
{
if
(
uptodate
+
1
!=
disks
)
BUG
();
compute_block
(
sh
,
failed_num
);
...
...
@@ -1092,60 +1148,62 @@ static void handle_stripe(struct stripe_head *sh)
}
if
(
uptodate
!=
disks
)
BUG
();
bh
=
sh
->
bh_cache
[
failed_num
];
set_b
uffer_locked
(
bh
);
dev
=
&
sh
->
dev
[
failed_num
];
set_b
it
(
R5_LOCKED
,
&
dev
->
flags
);
action
[
failed_num
]
=
WRITE
+
1
;
locked
++
;
set_bit
(
STRIPE_INSYNC
,
&
sh
->
state
);
if
(
conf
->
disks
[
failed_num
].
operational
)
md_sync_acct
(
conf
->
disks
[
failed_num
].
dev
,
bh
->
b_size
>>
9
);
md_sync_acct
(
conf
->
disks
[
failed_num
].
dev
,
STRIPE_SECTORS
);
else
if
((
spare
=
conf
->
spare
))
md_sync_acct
(
spare
->
dev
,
bh
->
b_size
>>
9
);
md_sync_acct
(
spare
->
dev
,
STRIPE_SECTORS
);
}
}
if
(
syncing
&&
locked
==
0
&&
test_bit
(
STRIPE_INSYNC
,
&
sh
->
state
))
{
md_done_sync
(
conf
->
mddev
,
(
sh
->
size
>>
9
)
-
sh
->
sync_redone
,
1
);
md_done_sync
(
conf
->
mddev
,
STRIPE_SECTORS
,
1
);
clear_bit
(
STRIPE_SYNCING
,
&
sh
->
state
);
}
spin_unlock
(
&
sh
->
lock
);
while
((
bh
=
return_ok
))
{
return_ok
=
bh
->
b_reqnext
;
bh
->
b_reqnext
=
NULL
;
bh
->
b_end_io
(
bh
,
1
);
}
while
((
bh
=
return_fail
))
{
return_fail
=
bh
->
b_reqnext
;
bh
->
b_reqnext
=
NULL
;
bh
->
b_end_io
(
bh
,
0
);
while
((
bi
=
return_bi
))
{
return_bi
=
bi
->
bi_next
;
bi
->
bi_next
=
NULL
;
bi
->
bi_end_io
(
bi
);
}
for
(
i
=
disks
;
i
--
;)
if
(
action
[
i
])
{
struct
b
uffer_head
*
bh
=
sh
->
bh_cache
[
i
]
;
struct
b
io
*
bi
=
&
sh
->
dev
[
i
].
req
;
struct
disk_info
*
spare
=
conf
->
spare
;
int
skip
=
0
;
if
(
action
[
i
]
==
READ
+
1
)
b
h
->
b
_end_io
=
raid5_end_read_request
;
b
i
->
bi
_end_io
=
raid5_end_read_request
;
else
b
h
->
b
_end_io
=
raid5_end_write_request
;
b
i
->
bi
_end_io
=
raid5_end_write_request
;
if
(
conf
->
disks
[
i
].
operational
)
b
h
->
b_dev
=
conf
->
disks
[
i
].
dev
;
b
i
->
bi_bdev
=
conf
->
disks
[
i
].
b
dev
;
else
if
(
spare
&&
action
[
i
]
==
WRITE
+
1
)
b
h
->
b_dev
=
spare
->
dev
;
b
i
->
bi_bdev
=
spare
->
b
dev
;
else
skip
=
1
;
/* FIXME - later we will need bdev here */
if
(
!
skip
)
{
PRINTK
(
"for %ld schedule op %d on disc %d
\n
"
,
sh
->
sector
,
action
[
i
]
-
1
,
i
);
atomic_inc
(
&
sh
->
count
);
bh
->
b_rdev
=
bh
->
b_dev
;
bh
->
b_rsector
=
bh
->
b_blocknr
*
(
bh
->
b_size
>>
9
);
generic_make_request
(
action
[
i
]
-
1
,
bh
);
bi
->
bi_sector
=
sh
->
sector
;
if
(
action
[
i
]
==
READ
+
1
)
bi
->
bi_rw
=
0
;
else
bi
->
bi_rw
=
1
;
bi
->
bi_flags
=
0
;
bi
->
bi_vcnt
=
1
;
bi
->
bi_idx
=
0
;
bi
->
bi_io_vec
=
&
sh
->
dev
[
i
].
vec
;
bi
->
bi_size
=
STRIPE_SIZE
;
bi
->
bi_next
=
NULL
;
generic_make_request
(
bi
);
}
else
{
PRINTK
(
"skip op %d on disc %d for sector %ld
\n
"
,
action
[
i
]
-
1
,
i
,
sh
->
sector
);
clear_b
uffer_locked
(
bh
);
clear_b
it
(
R5_LOCKED
,
&
dev
->
flags
);
set_bit
(
STRIPE_HANDLE
,
&
sh
->
state
);
}
}
...
...
@@ -1192,13 +1250,14 @@ static inline void raid5_plug_device(raid5_conf_t *conf)
spin_unlock_irq
(
&
conf
->
device_lock
);
}
static
int
raid5_make_request
(
mddev_t
*
mddev
,
int
rw
,
struct
buffer_head
*
bh
)
static
int
make_request
(
mddev_t
*
mddev
,
int
rw
,
struct
bio
*
bi
)
{
raid5_conf_t
*
conf
=
(
raid5_conf_t
*
)
mddev
->
private
;
const
unsigned
int
raid_disks
=
conf
->
raid_disks
;
const
unsigned
int
data_disks
=
raid_disks
-
1
;
unsigned
int
dd_idx
,
pd_idx
;
unsigned
long
new_sector
;
sector_t
new_sector
;
sector_t
logical_sector
,
last_sector
;
int
read_ahead
=
0
;
struct
stripe_head
*
sh
;
...
...
@@ -1208,25 +1267,39 @@ static int raid5_make_request (mddev_t *mddev, int rw, struct buffer_head * bh)
read_ahead
=
1
;
}
new_sector
=
raid5_compute_sector
(
bh
->
b_rsector
,
logical_sector
=
bi
->
bi_sector
&
~
(
STRIPE_SECTORS
-
1
);
last_sector
=
bi
->
bi_sector
+
(
bi
->
bi_size
>>
9
);
bi
->
bi_next
=
NULL
;
set_bit
(
BIO_UPTODATE
,
&
bi
->
bi_flags
);
/* will be cleared if error detected */
bi
->
bi_phys_segments
=
1
;
/* over-loaded to count active stripes */
for
(;
logical_sector
<
last_sector
;
logical_sector
+=
STRIPE_SECTORS
)
{
new_sector
=
raid5_compute_sector
(
logical_sector
,
raid_disks
,
data_disks
,
&
dd_idx
,
&
pd_idx
,
conf
);
PRINTK
(
"raid5_make_request, sector %lu
\n
"
,
new_sector
);
sh
=
get_active_stripe
(
conf
,
new_sector
,
bh
->
b_size
,
read_ahead
);
PRINTK
(
"raid5: make_request, sector %ul logical %ul
\n
"
,
new_sector
,
logical_sector
);
sh
=
get_active_stripe
(
conf
,
new_sector
,
pd_idx
,
read_ahead
);
if
(
sh
)
{
sh
->
pd_idx
=
pd_idx
;
add_stripe_bh
(
sh
,
bh
,
dd_idx
,
rw
);
add_stripe_bio
(
sh
,
bi
,
dd_idx
,
rw
);
raid5_plug_device
(
conf
);
handle_stripe
(
sh
);
release_stripe
(
sh
);
}
else
bh
->
b_end_io
(
bh
,
buffer_uptodate
(
bh
));
}
}
spin_lock_irq
(
&
conf
->
device_lock
);
if
(
--
bi
->
bi_phys_segments
==
0
)
bi
->
bi_end_io
(
bi
);
spin_unlock_irq
(
&
conf
->
device_lock
);
return
0
;
}
static
int
raid5_sync_request
(
mddev_t
*
mddev
,
unsigned
long
sector_nr
)
/* FIXME go_faster isn't used */
static
int
sync_request
(
mddev_t
*
mddev
,
sector_t
sector_nr
,
int
go_faster
)
{
raid5_conf_t
*
conf
=
(
raid5_conf_t
*
)
mddev
->
private
;
struct
stripe_head
*
sh
;
...
...
@@ -1237,25 +1310,19 @@ static int raid5_sync_request (mddev_t *mddev, unsigned long sector_nr)
unsigned
long
first_sector
;
int
raid_disks
=
conf
->
raid_disks
;
int
data_disks
=
raid_disks
-
1
;
int
redone
=
0
;
int
bufsize
;
sh
=
get_active_stripe
(
conf
,
sector_nr
,
0
,
0
);
bufsize
=
sh
->
size
;
redone
=
sector_nr
-
sh
->
sector
;
first_sector
=
raid5_compute_sector
(
stripe
*
data_disks
*
sectors_per_chunk
+
chunk_offset
,
raid_disks
,
data_disks
,
&
dd_idx
,
&
pd_idx
,
conf
);
sh
->
pd_idx
=
pd_idx
;
sh
=
get_active_stripe
(
conf
,
sector_nr
,
pd_idx
,
0
)
;
spin_lock
(
&
sh
->
lock
);
set_bit
(
STRIPE_SYNCING
,
&
sh
->
state
);
clear_bit
(
STRIPE_INSYNC
,
&
sh
->
state
);
sh
->
sync_redone
=
redone
;
spin_unlock
(
&
sh
->
lock
);
handle_stripe
(
sh
);
release_stripe
(
sh
);
return
(
bufsize
>>
9
)
-
redone
;
return
STRIPE_SECTORS
;
}
/*
...
...
@@ -1280,7 +1347,7 @@ static void raid5d (void *data)
mddev
->
sb_dirty
=
0
;
md_update_sb
(
mddev
);
}
md_
spin_lock_irq
(
&
conf
->
device_lock
);
spin_lock_irq
(
&
conf
->
device_lock
);
while
(
1
)
{
struct
list_head
*
first
;
...
...
@@ -1300,17 +1367,17 @@ static void raid5d (void *data)
atomic_inc
(
&
sh
->
count
);
if
(
atomic_read
(
&
sh
->
count
)
!=
1
)
BUG
();
md_
spin_unlock_irq
(
&
conf
->
device_lock
);
spin_unlock_irq
(
&
conf
->
device_lock
);
handled
++
;
handle_stripe
(
sh
);
release_stripe
(
sh
);
md_
spin_lock_irq
(
&
conf
->
device_lock
);
spin_lock_irq
(
&
conf
->
device_lock
);
}
PRINTK
(
"%d stripes handled
\n
"
,
handled
);
md_
spin_unlock_irq
(
&
conf
->
device_lock
);
spin_unlock_irq
(
&
conf
->
device_lock
);
PRINTK
(
"--- raid5d inactive
\n
"
);
}
...
...
@@ -1340,7 +1407,7 @@ static void raid5syncd (void *data)
printk
(
"raid5: resync finished.
\n
"
);
}
static
int
r
aid5_r
un
(
mddev_t
*
mddev
)
static
int
run
(
mddev_t
*
mddev
)
{
raid5_conf_t
*
conf
;
int
i
,
j
,
raid_disk
,
memory
;
...
...
@@ -1348,7 +1415,7 @@ static int raid5_run (mddev_t *mddev)
mdp_disk_t
*
desc
;
mdk_rdev_t
*
rdev
;
struct
disk_info
*
disk
;
struct
md_
list_head
*
tmp
;
struct
list_head
*
tmp
;
int
start_recovery
=
0
;
MOD_INC_USE_COUNT
;
...
...
@@ -1365,25 +1432,24 @@ static int raid5_run (mddev_t *mddev)
memset
(
conf
,
0
,
sizeof
(
*
conf
));
conf
->
mddev
=
mddev
;
if
((
conf
->
stripe_hashtbl
=
(
struct
stripe_head
**
)
md
__get_free_pages
(
GFP_ATOMIC
,
HASH_PAGES_ORDER
))
==
NULL
)
if
((
conf
->
stripe_hashtbl
=
(
struct
stripe_head
**
)
__get_free_pages
(
GFP_ATOMIC
,
HASH_PAGES_ORDER
))
==
NULL
)
goto
abort
;
memset
(
conf
->
stripe_hashtbl
,
0
,
HASH_PAGES
*
PAGE_SIZE
);
conf
->
device_lock
=
MD_
SPIN_LOCK_UNLOCKED
;
md_
init_waitqueue_head
(
&
conf
->
wait_for_stripe
);
conf
->
device_lock
=
SPIN_LOCK_UNLOCKED
;
init_waitqueue_head
(
&
conf
->
wait_for_stripe
);
INIT_LIST_HEAD
(
&
conf
->
handle_list
);
INIT_LIST_HEAD
(
&
conf
->
delayed_list
);
INIT_LIST_HEAD
(
&
conf
->
inactive_list
);
atomic_set
(
&
conf
->
active_stripes
,
0
);
atomic_set
(
&
conf
->
preread_active_stripes
,
0
);
conf
->
buffer_size
=
PAGE_SIZE
;
/* good default for rebuild */
conf
->
plugged
=
0
;
conf
->
plug_tq
.
sync
=
0
;
conf
->
plug_tq
.
routine
=
&
raid5_unplug_device
;
conf
->
plug_tq
.
data
=
conf
;
PRINTK
(
"raid5
_
run(md%d) called.
\n
"
,
mdidx
(
mddev
));
PRINTK
(
"raid5
:
run(md%d) called.
\n
"
,
mdidx
(
mddev
));
ITERATE_RDEV
(
mddev
,
rdev
,
tmp
)
{
/*
...
...
@@ -1404,6 +1470,7 @@ static int raid5_run (mddev_t *mddev)
disk
->
number
=
desc
->
number
;
disk
->
raid_disk
=
raid_disk
;
disk
->
dev
=
rdev
->
dev
;
disk
->
bdev
=
rdev
->
bdev
;
disk
->
operational
=
0
;
disk
->
write_only
=
0
;
...
...
@@ -1430,6 +1497,7 @@ static int raid5_run (mddev_t *mddev)
disk
->
number
=
desc
->
number
;
disk
->
raid_disk
=
raid_disk
;
disk
->
dev
=
rdev
->
dev
;
disk
->
bdev
=
rdev
->
bdev
;
disk
->
operational
=
1
;
disk
->
used_slot
=
1
;
...
...
@@ -1442,6 +1510,7 @@ static int raid5_run (mddev_t *mddev)
disk
->
number
=
desc
->
number
;
disk
->
raid_disk
=
raid_disk
;
disk
->
dev
=
rdev
->
dev
;
disk
->
bdev
=
rdev
->
bdev
;
disk
->
operational
=
0
;
disk
->
write_only
=
0
;
...
...
@@ -1461,6 +1530,7 @@ static int raid5_run (mddev_t *mddev)
disk
->
number
=
desc
->
number
;
disk
->
raid_disk
=
raid_disk
;
disk
->
dev
=
NODEV
;
disk
->
bdev
=
NULL
;
disk
->
operational
=
0
;
disk
->
write_only
=
0
;
...
...
@@ -1518,9 +1588,9 @@ static int raid5_run (mddev_t *mddev)
memory
=
conf
->
max_nr_stripes
*
(
sizeof
(
struct
stripe_head
)
+
conf
->
raid_disks
*
((
sizeof
(
struct
buffer_head
)
+
PAGE_SIZE
)))
/
1024
;
if
(
grow_stripes
(
conf
,
conf
->
max_nr_stripes
,
GFP_KERNEL
))
{
if
(
grow_stripes
(
conf
,
conf
->
max_nr_stripes
))
{
printk
(
KERN_ERR
"raid5: couldn't allocate %dkB for buffers
\n
"
,
memory
);
shrink_stripes
(
conf
,
conf
->
max_nr_stripes
);
shrink_stripes
(
conf
);
goto
abort
;
}
else
printk
(
KERN_INFO
"raid5: allocated %dkB for md%d
\n
"
,
memory
,
mdidx
(
mddev
));
...
...
@@ -1580,7 +1650,7 @@ static int raid5_run (mddev_t *mddev)
return
-
EIO
;
}
static
int
raid5_
stop_resync
(
mddev_t
*
mddev
)
static
int
stop_resync
(
mddev_t
*
mddev
)
{
raid5_conf_t
*
conf
=
mddev_to_conf
(
mddev
);
mdk_thread_t
*
thread
=
conf
->
resync_thread
;
...
...
@@ -1597,7 +1667,7 @@ static int raid5_stop_resync (mddev_t *mddev)
return
0
;
}
static
int
r
aid5_r
estart_resync
(
mddev_t
*
mddev
)
static
int
restart_resync
(
mddev_t
*
mddev
)
{
raid5_conf_t
*
conf
=
mddev_to_conf
(
mddev
);
...
...
@@ -1616,14 +1686,14 @@ static int raid5_restart_resync (mddev_t *mddev)
}
static
int
raid5_
stop
(
mddev_t
*
mddev
)
static
int
stop
(
mddev_t
*
mddev
)
{
raid5_conf_t
*
conf
=
(
raid5_conf_t
*
)
mddev
->
private
;
if
(
conf
->
resync_thread
)
md_unregister_thread
(
conf
->
resync_thread
);
md_unregister_thread
(
conf
->
thread
);
shrink_stripes
(
conf
,
conf
->
max_nr_stripes
);
shrink_stripes
(
conf
);
free_pages
((
unsigned
long
)
conf
->
stripe_hashtbl
,
HASH_PAGES_ORDER
);
kfree
(
conf
);
mddev
->
private
=
NULL
;
...
...
@@ -1636,12 +1706,11 @@ static void print_sh (struct stripe_head *sh)
{
int
i
;
printk
(
"sh %lu,
size %d, pd_idx %d, state %ld.
\n
"
,
sh
->
sector
,
sh
->
size
,
sh
->
pd_idx
,
sh
->
state
);
printk
(
"sh %lu,
pd_idx %d, state %ld.
\n
"
,
sh
->
sector
,
sh
->
pd_idx
,
sh
->
state
);
printk
(
"sh %lu, count %d.
\n
"
,
sh
->
sector
,
atomic_read
(
&
sh
->
count
));
printk
(
"sh %lu, "
,
sh
->
sector
);
for
(
i
=
0
;
i
<
MD_SB_DISKS
;
i
++
)
{
if
(
sh
->
bh_cache
[
i
])
printk
(
"(cache%d: %p %ld) "
,
i
,
sh
->
bh_cache
[
i
],
sh
->
bh_cache
[
i
]
->
b_state
);
for
(
i
=
0
;
i
<
sh
->
raid_conf
->
raid_disks
;
i
++
)
{
printk
(
"(cache%d: %p %ld) "
,
i
,
sh
->
dev
[
i
].
page
,
sh
->
dev
[
i
].
flags
);
}
printk
(
"
\n
"
);
}
...
...
@@ -1651,7 +1720,7 @@ static void printall (raid5_conf_t *conf)
struct
stripe_head
*
sh
;
int
i
;
md_
spin_lock_irq
(
&
conf
->
device_lock
);
spin_lock_irq
(
&
conf
->
device_lock
);
for
(
i
=
0
;
i
<
NR_HASH
;
i
++
)
{
sh
=
conf
->
stripe_hashtbl
[
i
];
for
(;
sh
;
sh
=
sh
->
hash_next
)
{
...
...
@@ -1660,13 +1729,13 @@ static void printall (raid5_conf_t *conf)
print_sh
(
sh
);
}
}
md_
spin_unlock_irq
(
&
conf
->
device_lock
);
spin_unlock_irq
(
&
conf
->
device_lock
);
PRINTK
(
"--- raid5d inactive
\n
"
);
}
#endif
static
int
raid5_
status
(
char
*
page
,
mddev_t
*
mddev
)
static
int
status
(
char
*
page
,
mddev_t
*
mddev
)
{
raid5_conf_t
*
conf
=
(
raid5_conf_t
*
)
mddev
->
private
;
mdp_super_t
*
sb
=
mddev
->
sb
;
...
...
@@ -1711,7 +1780,7 @@ static void print_raid5_conf (raid5_conf_t *conf)
}
}
static
int
raid5_
diskop
(
mddev_t
*
mddev
,
mdp_disk_t
**
d
,
int
state
)
static
int
diskop
(
mddev_t
*
mddev
,
mdp_disk_t
**
d
,
int
state
)
{
int
err
=
0
;
int
i
,
failed_disk
=-
1
,
spare_disk
=-
1
,
removed_disk
=-
1
,
added_disk
=-
1
;
...
...
@@ -1722,7 +1791,7 @@ static int raid5_diskop(mddev_t *mddev, mdp_disk_t **d, int state)
mdk_rdev_t
*
spare_rdev
,
*
failed_rdev
;
print_raid5_conf
(
conf
);
md_
spin_lock_irq
(
&
conf
->
device_lock
);
spin_lock_irq
(
&
conf
->
device_lock
);
/*
* find the disk ...
*/
...
...
@@ -1948,6 +2017,7 @@ static int raid5_diskop(mddev_t *mddev, mdp_disk_t **d, int state)
goto
abort
;
}
rdisk
->
dev
=
NODEV
;
rdisk
->
bdev
=
NULL
;
rdisk
->
used_slot
=
0
;
break
;
...
...
@@ -1965,6 +2035,8 @@ static int raid5_diskop(mddev_t *mddev, mdp_disk_t **d, int state)
adisk
->
number
=
added_desc
->
number
;
adisk
->
raid_disk
=
added_desc
->
raid_disk
;
adisk
->
dev
=
mk_kdev
(
added_desc
->
major
,
added_desc
->
minor
);
/* it will be held open by rdev */
adisk
->
bdev
=
bdget
(
kdev_t_to_nr
(
adisk
->
dev
));
adisk
->
operational
=
0
;
adisk
->
write_only
=
0
;
...
...
@@ -1980,7 +2052,7 @@ static int raid5_diskop(mddev_t *mddev, mdp_disk_t **d, int state)
goto
abort
;
}
abort:
md_
spin_unlock_irq
(
&
conf
->
device_lock
);
spin_unlock_irq
(
&
conf
->
device_lock
);
print_raid5_conf
(
conf
);
return
err
;
}
...
...
@@ -1988,18 +2060,18 @@ static int raid5_diskop(mddev_t *mddev, mdp_disk_t **d, int state)
static
mdk_personality_t
raid5_personality
=
{
name:
"raid5"
,
make_request:
raid5_
make_request
,
run:
r
aid5_r
un
,
stop:
raid5_
stop
,
status:
raid5_
status
,
error_handler:
raid5_
error
,
diskop:
raid5_
diskop
,
stop_resync:
raid5_
stop_resync
,
restart_resync:
r
aid5_r
estart_resync
,
sync_request:
raid5_
sync_request
make_request:
make_request
,
run:
run
,
stop:
stop
,
status:
status
,
error_handler:
error
,
diskop:
diskop
,
stop_resync:
stop_resync
,
restart_resync:
restart_resync
,
sync_request:
sync_request
};
static
int
md
__init
raid5_init
(
void
)
static
int
__init
raid5_init
(
void
)
{
return
register_md_personality
(
RAID5
,
&
raid5_personality
);
}
...
...
drivers/md/xor.c
View file @
67a670bd
...
...
@@ -19,7 +19,6 @@
#define BH_TRACE 0
#include <linux/module.h>
#include <linux/raid/md.h>
#include <linux/raid/md_compatible.h>
#include <linux/raid/xor.h>
#include <asm/xor.h>
...
...
@@ -27,31 +26,30 @@
static
struct
xor_block_template
*
active_template
;
void
xor_block
(
unsigned
int
count
,
struct
buffer_head
**
bh_
ptr
)
xor_block
(
unsigned
int
count
,
unsigned
int
bytes
,
void
**
ptr
)
{
unsigned
long
*
p0
,
*
p1
,
*
p2
,
*
p3
,
*
p4
;
unsigned
long
bytes
=
bh_ptr
[
0
]
->
b_size
;
p0
=
(
unsigned
long
*
)
bh_ptr
[
0
]
->
b_data
;
p1
=
(
unsigned
long
*
)
bh_ptr
[
1
]
->
b_data
;
p0
=
(
unsigned
long
*
)
ptr
[
0
]
;
p1
=
(
unsigned
long
*
)
ptr
[
1
]
;
if
(
count
==
2
)
{
active_template
->
do_2
(
bytes
,
p0
,
p1
);
return
;
}
p2
=
(
unsigned
long
*
)
bh_ptr
[
2
]
->
b_data
;
p2
=
(
unsigned
long
*
)
ptr
[
2
]
;
if
(
count
==
3
)
{
active_template
->
do_3
(
bytes
,
p0
,
p1
,
p2
);
return
;
}
p3
=
(
unsigned
long
*
)
bh_ptr
[
3
]
->
b_data
;
p3
=
(
unsigned
long
*
)
ptr
[
3
]
;
if
(
count
==
4
)
{
active_template
->
do_4
(
bytes
,
p0
,
p1
,
p2
,
p3
);
return
;
}
p4
=
(
unsigned
long
*
)
bh_ptr
[
4
]
->
b_data
;
p4
=
(
unsigned
long
*
)
ptr
[
4
]
;
active_template
->
do_5
(
bytes
,
p0
,
p1
,
p2
,
p3
,
p4
);
}
...
...
@@ -103,7 +101,7 @@ calibrate_xor_block(void)
void
*
b1
,
*
b2
;
struct
xor_block_template
*
f
,
*
fastest
;
b1
=
(
void
*
)
md
__get_free_pages
(
GFP_KERNEL
,
2
);
b1
=
(
void
*
)
__get_free_pages
(
GFP_KERNEL
,
2
);
if
(
!
b1
)
{
printk
(
"raid5: Yikes! No memory available.
\n
"
);
return
-
ENOMEM
;
...
...
@@ -137,7 +135,7 @@ calibrate_xor_block(void)
return
0
;
}
MD_
EXPORT_SYMBOL
(
xor_block
);
EXPORT_SYMBOL
(
xor_block
);
MODULE_LICENSE
(
"GPL"
);
module_init
(
calibrate_xor_block
);
fs/intermezzo/ext_attr.c
View file @
67a670bd
...
...
@@ -105,9 +105,8 @@ int presto_set_ext_attr(struct inode *inode,
printk
(
"InterMezzo: out of memory!!!
\n
"
);
return
-
ENOMEM
;
}
error
=
copy_from_user
(
buf
,
buffer
,
buffer_len
);
if
(
error
)
return
error
;
if
(
copy_from_user
(
buf
,
buffer
,
buffer_len
))
return
-
EFAULT
;
}
else
buf
=
buffer
;
}
else
...
...
fs/intermezzo/kml.c
View file @
67a670bd
...
...
@@ -31,10 +31,9 @@ int begin_kml_reint (struct file *file, unsigned long arg)
ENTRY
;
/* allocate buffer & copy it to kernel space */
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
if
(
input
.
reclen
>
kml_fsdata
->
kml_maxsize
)
...
...
@@ -45,11 +44,10 @@ int begin_kml_reint (struct file *file, unsigned long arg)
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
volname
,
input
.
namelen
);
if
(
error
)
{
if
(
copy_from_user
(
path
,
input
.
volname
,
input
.
namelen
))
{
PRESTO_FREE
(
path
,
input
.
namelen
+
1
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
path
[
input
.
namelen
]
=
'\0'
;
fset
=
kml_getfset
(
path
);
...
...
@@ -57,10 +55,9 @@ int begin_kml_reint (struct file *file, unsigned long arg)
kml_fsdata
=
FSET_GET_KMLDATA
(
fset
);
/* read the buf from user memory here */
error
=
copy_from_user
(
kml_fsdata
->
kml_buf
,
input
.
recbuf
,
input
.
reclen
);
if
(
error
)
{
if
(
copy_from_user
(
kml_fsdata
->
kml_buf
,
input
.
recbuf
,
input
.
reclen
))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
kml_fsdata
->
kml_len
=
input
.
reclen
;
...
...
@@ -94,21 +91,19 @@ int do_kml_reint (struct file *file, unsigned long arg)
struct
presto_file_set
*
fset
;
ENTRY
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
path
,
char
*
,
input
.
namelen
+
1
);
if
(
!
path
)
{
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
volname
,
input
.
namelen
);
if
(
error
)
{
if
(
copy_from_user
(
path
,
input
.
volname
,
input
.
namelen
))
{
PRESTO_FREE
(
path
,
input
.
namelen
+
1
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
path
[
input
.
namelen
]
=
'\0'
;
fset
=
kml_getfset
(
path
);
...
...
@@ -138,7 +133,8 @@ int do_kml_reint (struct file *file, unsigned long arg)
strlen
(
close
->
path
)
+
1
,
input
.
pathlen
);
error
=
-
ENOMEM
;
}
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
));
if
(
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
)))
return
-
EFAULT
;
}
return
error
;
}
...
...
@@ -161,10 +157,9 @@ int end_kml_reint (struct file *file, unsigned long arg)
char
*
path
;
ENTRY
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
path
,
char
*
,
input
.
namelen
+
1
);
...
...
@@ -172,11 +167,11 @@ int end_kml_reint (struct file *file, unsigned long arg)
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
volname
,
input
.
namelen
);
if
(
copy_from_user
(
path
,
input
.
volname
,
input
.
namelen
))
{
if
(
error
)
{
PRESTO_FREE
(
path
,
input
.
namelen
+
1
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
path
[
input
.
namelen
]
=
'\0'
;
fset
=
kml_getfset
(
path
);
...
...
@@ -193,7 +188,8 @@ int end_kml_reint (struct file *file, unsigned long arg)
#if 0
input.newpos = kml_upc->newpos;
input.count = kml_upc->count;
copy_to_user((char *)arg, &input, sizeof (input));
if (copy_to_user((char *)arg, &input, sizeof (input)))
return -EFAULT;
#endif
return
error
;
}
fs/intermezzo/psdev.c
View file @
67a670bd
...
...
@@ -149,9 +149,8 @@ static ssize_t presto_psdev_write(struct file *file, const char *buf,
return
-
EINVAL
;
}
error
=
copy_from_user
(
&
hdr
,
buf
,
sizeof
(
hdr
));
if
(
error
)
return
error
;
if
(
copy_from_user
(
&
hdr
,
buf
,
sizeof
(
hdr
)))
return
-
EFAULT
;
CDEBUG
(
D_PSDEV
,
"(process,opc,uniq)=(%d,%d,%d)
\n
"
,
current
->
pid
,
hdr
.
opcode
,
hdr
.
unique
);
...
...
@@ -183,9 +182,8 @@ static ssize_t presto_psdev_write(struct file *file, const char *buf,
req
->
rq_bufsize
,
count
,
hdr
.
opcode
,
hdr
.
unique
);
count
=
req
->
rq_bufsize
;
/* don't have more space! */
}
error
=
copy_from_user
(
req
->
rq_data
,
buf
,
count
);
if
(
error
)
return
error
;
if
(
copy_from_user
(
req
->
rq_data
,
buf
,
count
))
return
-
EFAULT
;
/* adjust outsize: good upcalls can be aware of this */
req
->
rq_rep_size
=
count
;
...
...
@@ -280,14 +278,12 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
char
*
tmp
;
int
error
;
error
=
copy_from_user
(
&
readmount
,
(
void
*
)
arg
,
sizeof
(
readmount
));
if
(
error
)
{
if
(
copy_from_user
(
&
readmount
,
(
void
*
)
arg
,
sizeof
(
readmount
)))
printk
(
"psdev: can't copy %Zd bytes from %p to %p
\n
"
,
sizeof
(
readmount
),
(
struct
readmount
*
)
arg
,
&
readmount
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
len
=
readmount
.
io_len
;
...
...
@@ -307,15 +303,16 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
* I mean, let's let the compiler do a little work ...
* gcc suggested the extra ()
*/
error
=
copy_to_user
(
readmount
.
io_string
,
tmp
,
outlen
);
if
(
error
)
{
if
(
copy_to_user
(
readmount
.
io_string
,
tmp
,
outlen
))
{
CDEBUG
(
D_PSDEV
,
"Copy_to_user string 0x%p failed
\n
"
,
readmount
.
io_string
);
error
=
-
EFAULT
;
}
if
(
(
!
error
)
&&
(
error
=
copy_to_user
(
&
(
user_readmount
->
io_len
),
&
outlen
,
sizeof
(
int
)))
)
{
if
(
!
error
&&
copy_to_user
(
&
(
user_readmount
->
io_len
),
&
outlen
,
sizeof
(
int
))
)
{
CDEBUG
(
D_PSDEV
,
"Copy_to_user len @0x%p failed
\n
"
,
&
(
user_readmount
->
io_len
));
error
=
-
EFAULT
;
}
PRESTO_FREE
(
tmp
,
len
);
...
...
@@ -360,10 +357,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
path_len
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
path
,
char
*
,
input
.
path_len
+
1
);
...
...
@@ -371,11 +367,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
))
{
PRESTO_FREE
(
path
,
input
.
path_len
+
1
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
path
[
input
.
path_len
]
=
'\0'
;
CDEBUG
(
D_PSDEV
,
"clear_fsetroot: path %s
\n
"
,
path
);
...
...
@@ -401,10 +396,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
path_len
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
path
,
char
*
,
input
.
path_len
+
1
);
...
...
@@ -412,11 +406,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
))
{
PRESTO_FREE
(
path
,
input
.
path_len
+
1
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
path
[
input
.
path_len
]
=
'\0'
;
CDEBUG
(
D_PSDEV
,
"clear_all_fsetroot: path %s
\n
"
,
path
);
...
...
@@ -440,10 +433,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
path_len
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
path
,
char
*
,
input
.
path_len
+
1
);
...
...
@@ -451,11 +443,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
))
{
PRESTO_FREE
(
path
,
input
.
path_len
+
1
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
path
[
input
.
path_len
]
=
'\0'
;
CDEBUG
(
D_PSDEV
,
"get_kmlsize: len %d path %s
\n
"
,
...
...
@@ -474,7 +465,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
CDEBUG
(
D_PSDEV
,
"get_kmlsize: size = %Zd
\n
"
,
size
);
EXIT
;
return
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
));
if
(
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
)))
return
-
EFAULT
;
return
0
;
}
case
PRESTO_GET_RECNO
:
{
...
...
@@ -488,10 +481,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
path_len
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
path
,
char
*
,
input
.
path_len
+
1
);
...
...
@@ -499,11 +491,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
))
{
PRESTO_FREE
(
path
,
input
.
path_len
+
1
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
path
[
input
.
path_len
]
=
'\0'
;
CDEBUG
(
D_PSDEV
,
"get_recno: len %d path %s
\n
"
,
...
...
@@ -522,7 +513,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
CDEBUG
(
D_PSDEV
,
"get_recno: recno = %d
\n
"
,
(
int
)
recno
);
EXIT
;
return
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
));
if
(
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
)))
return
-
EFAULT
;
return
0
;
}
case
PRESTO_SET_FSETROOT
:
{
...
...
@@ -543,10 +536,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
flags
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
path
,
char
*
,
input
.
path_len
+
1
);
...
...
@@ -554,9 +546,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
path
,
input
.
path
,
input
.
path_len
))
{
EXIT
;
error
-
EFAULT
;
goto
exit_free_path
;
}
path
[
input
.
path_len
]
=
'\0'
;
...
...
@@ -567,9 +559,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
goto
exit_free_path
;
}
error
=
copy_from_user
(
fsetname
,
input
.
name
,
input
.
name_len
);
if
(
error
)
{
if
(
copy_from_user
(
fsetname
,
input
.
name
,
input
.
name_len
))
{
EXIT
;
error
=
-
EFAULT
;
goto
exit_free_fsetname
;
}
fsetname
[
input
.
name_len
]
=
'\0'
;
...
...
@@ -621,12 +613,11 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
struct
psdev_opt
*
user_opt
=
(
struct
psdev_opt
*
)
arg
;
int
error
;
error
=
copy_from_user
(
&
kopt
,
(
void
*
)
arg
,
sizeof
(
kopt
));
if
(
error
)
{
if
(
copy_from_user
(
&
kopt
,
(
void
*
)
arg
,
sizeof
(
kopt
)))
{
printk
(
"psdev: can't copyin %Zd bytes from %p to %p
\n
"
,
sizeof
(
kopt
),
(
struct
kopt
*
)
arg
,
&
kopt
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
minor
=
minor
(
dev
);
if
(
cmd
==
PRESTO_SETOPT
)
...
...
@@ -650,12 +641,11 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
return
error
;
}
error
=
copy_to_user
(
user_opt
,
&
kopt
,
sizeof
(
kopt
));
if
(
error
)
{
if
(
copy_to_user
(
user_opt
,
&
kopt
,
sizeof
(
kopt
)))
{
CDEBUG
(
D_PSDEV
,
"Copy_to_user opt 0x%p failed
\n
"
,
user_opt
);
EXIT
;
return
error
;
return
-
EFAULT
;
}
CDEBUG
(
D_PSDEV
,
"dosetopt minor %d, opt %d, val %d return %d
\n
"
,
minor
,
kopt
.
optname
,
kopt
.
optval
,
error
);
...
...
@@ -668,10 +658,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
struct
lento_input_attr
input
;
struct
iattr
iattr
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
iattr
.
ia_valid
=
input
.
valid
;
iattr
.
ia_mode
=
(
umode_t
)
input
.
mode
;
...
...
@@ -692,10 +681,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input_mode
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_create
(
input
.
name
,
input
.
mode
,
&
input
.
info
);
...
...
@@ -707,10 +695,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input_old_new
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_link
(
input
.
oldname
,
input
.
newname
,
&
input
.
info
);
...
...
@@ -722,10 +709,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_unlink
(
input
.
name
,
&
input
.
info
);
...
...
@@ -737,10 +723,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input_old_new
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_symlink
(
input
.
oldname
,
input
.
newname
,
&
input
.
info
);
...
...
@@ -752,10 +737,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input_mode
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_mkdir
(
input
.
name
,
input
.
mode
,
&
input
.
info
);
...
...
@@ -767,10 +751,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_rmdir
(
input
.
name
,
&
input
.
info
);
...
...
@@ -782,10 +765,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input_dev
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_mknod
(
input
.
name
,
input
.
mode
,
...
...
@@ -798,10 +780,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
int
error
;
struct
lento_input_old_new
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
error
=
lento_rename
(
input
.
oldname
,
input
.
newname
,
&
input
.
info
);
...
...
@@ -817,30 +798,27 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
char
*
name
;
char
*
buffer
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
/* Now setup the input parameters */
PRESTO_ALLOC
(
name
,
char
*
,
input
.
name_len
+
1
);
/* We need null terminated strings for attr names */
name
[
input
.
name_len
]
=
'\0'
;
error
=
copy_from_user
(
name
,
input
.
name
,
input
.
name_len
);
if
(
error
)
{
if
(
copy_from_user
(
name
,
input
.
name
,
input
.
name_len
))
{
EXIT
;
PRESTO_FREE
(
name
,
input
.
name_len
+
1
);
return
error
;
return
-
EFAULT
;
}
PRESTO_ALLOC
(
buffer
,
char
*
,
input
.
buffer_len
+
1
);
error
=
copy_from_user
(
buffer
,
input
.
buffer
,
input
.
buffer_len
);
if
(
error
)
{
if
(
copy_from_user
(
buffer
,
input
.
buffer
,
input
.
buffer_len
))
{
EXIT
;
PRESTO_FREE
(
name
,
input
.
name_len
+
1
);
PRESTO_FREE
(
buffer
,
input
.
buffer_len
+
1
);
return
error
;
return
-
EFAULT
;
}
/* Make null terminated for easy printing */
buffer
[
input
.
buffer_len
]
=
'\0'
;
...
...
@@ -869,21 +847,19 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
struct
lento_input_ext_attr
input
;
char
*
name
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
/* Now setup the input parameters */
PRESTO_ALLOC
(
name
,
char
*
,
input
.
name_len
+
1
);
/* We need null terminated strings for attr names */
name
[
input
.
name_len
]
=
'\0'
;
error
=
copy_from_user
(
name
,
input
.
name
,
input
.
name_len
);
if
(
error
)
{
if
(
copy_from_user
(
name
,
input
.
name
,
input
.
name_len
))
{
EXIT
;
PRESTO_FREE
(
name
,
input
.
name_len
+
1
);
return
error
;
return
-
EFAULT
;
}
CDEBUG
(
D_PSDEV
,
" delextattr params: name %s,"
...
...
@@ -907,10 +883,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
struct
lento_input_iopen
input
;
int
error
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
input
.
fd
=
lento_iopen
(
input
.
name
,
(
ino_t
)
input
.
ino
,
...
...
@@ -921,17 +896,18 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
return
input
.
fd
;
}
EXIT
;
return
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
));
if
(
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
)))
return
-
EFAULT
;
return
0
;
}
case
PRESTO_VFS_CLOSE
:
{
int
error
;
struct
lento_input_close
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
CDEBUG
(
D_PIOCTL
,
"lento_close file descriptor: %d
\n
"
,
input
.
fd
);
...
...
@@ -952,10 +928,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
struct
presto_version
remote_file_version
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
user_path
=
input
.
path
;
...
...
@@ -964,11 +939,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
))
{
EXIT
;
PRESTO_FREE
(
input
.
path
,
input
.
path_len
+
1
);
return
error
;
return
-
EFAULT
;
}
input
.
path
[
input
.
path_len
]
=
'\0'
;
...
...
@@ -996,10 +970,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
struct
lento_vfs_context
info
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
user_path
=
input
.
path
;
...
...
@@ -1008,11 +981,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
))
{
EXIT
;
PRESTO_FREE
(
input
.
path
,
input
.
path_len
+
1
);
return
error
;
return
-
EFAULT
;
}
input
.
path
[
input
.
path_len
]
=
'\0'
;
...
...
@@ -1035,10 +1007,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
__u32
path_len
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
user_path
=
input
.
path
;
...
...
@@ -1047,11 +1018,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
))
{
EXIT
;
PRESTO_FREE
(
input
.
path
,
input
.
path_len
+
1
);
return
error
;
return
-
EFAULT
;
}
input
.
path
[
input
.
path_len
]
=
'\0'
;
...
...
@@ -1072,10 +1042,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
__u32
recno
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
user_path
=
input
.
path
;
...
...
@@ -1084,11 +1053,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
))
{
EXIT
;
PRESTO_FREE
(
input
.
path
,
input
.
path_len
+
1
);
return
error
;
return
-
EFAULT
;
}
input
.
path
[
input
.
path_len
]
=
'\0'
;
...
...
@@ -1111,10 +1079,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
char
*
path
;
}
input
;
error
=
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
));
if
(
error
)
{
if
(
copy_from_user
(
&
input
,
(
char
*
)
arg
,
sizeof
(
input
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
user_path
=
input
.
path
;
...
...
@@ -1123,11 +1090,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
input
.
path
,
user_path
,
input
.
path_len
))
{
EXIT
;
PRESTO_FREE
(
input
.
path
,
input
.
path_len
+
1
);
return
error
;
return
-
EFAULT
;
}
input
.
path
[
input
.
path_len
]
=
'\0'
;
...
...
@@ -1190,7 +1156,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
}
/* return the correct cookie to wait for */
input
.
mark_what
=
res
;
return
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
));
if
(
copy_to_user
((
char
*
)
arg
,
&
input
,
sizeof
(
input
)))
return
-
EFAULT
;
return
0
;
}
#ifdef CONFIG_KREINT
...
...
@@ -1211,10 +1179,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
char
*
path
;
}
permit
;
error
=
copy_from_user
(
&
permit
,
(
char
*
)
arg
,
sizeof
(
permit
));
if
(
error
)
{
if
(
copy_from_user
(
&
permit
,
(
char
*
)
arg
,
sizeof
(
permit
)))
{
EXIT
;
return
error
;
return
-
EFAULT
;
}
user_path
=
permit
.
path
;
...
...
@@ -1223,11 +1190,10 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
EXIT
;
return
-
ENOMEM
;
}
error
=
copy_from_user
(
permit
.
path
,
user_path
,
permit
.
path_len
);
if
(
error
)
{
if
(
copy_from_user
(
permit
.
path
,
user_path
,
permit
.
path_len
))
{
EXIT
;
PRESTO_FREE
(
permit
.
path
,
permit
.
path_len
+
1
);
return
error
;
return
-
EFAULT
;
}
permit
.
path
[
permit
.
path_len
]
=
'\0'
;
...
...
@@ -1241,7 +1207,9 @@ static int presto_psdev_ioctl(struct inode *inode, struct file *file,
return
error
;
}
/* return the correct cookie to wait for */
return
copy_to_user
((
char
*
)
arg
,
&
permit
,
sizeof
(
permit
));
if
(
copy_to_user
((
char
*
)
arg
,
&
permit
,
sizeof
(
permit
)))
return
-
EFAULT
;
return
0
;
}
default:
...
...
fs/nfsd/export.c
View file @
67a670bd
...
...
@@ -315,17 +315,19 @@ exp_export(struct nfsctl_export *nxp)
* 2: We must be able to find an inode from a filehandle.
* This means that s_export_op must be set.
*/
if
(((
inode
->
i_sb
->
s_type
->
fs_flags
&
FS_REQUIRES_DEV
)
||
(
nxp
->
ex_flags
&
NFSEXP_FSID
))
&&
inode
->
i_sb
->
s_export_op
)
/* Ok, we can export it */
;
else
{
if
(
!
(
inode
->
i_sb
->
s_type
->
fs_flags
&
FS_REQUIRES_DEV
))
{
if
(
!
(
nxp
->
ex_flags
&
NFSEXP_FSID
))
{
dprintk
(
"exp_export: export of non-dev fs without fsid"
);
goto
finish
;
}
}
if
(
!
inode
->
i_sb
->
s_export_op
)
{
dprintk
(
"exp_export: export of invalid fs type.
\n
"
);
goto
finish
;
}
if
(
inode
->
i_sb
->
s_export_op
&&
!
inode
->
i_sb
->
s_export_op
->
find_exported_dentry
)
/* Ok, we can export it */
;
if
(
!
inode
->
i_sb
->
s_export_op
->
find_exported_dentry
)
inode
->
i_sb
->
s_export_op
->
find_exported_dentry
=
find_exported_dentry
;
...
...
fs/nfsd/vfs.c
View file @
67a670bd
...
...
@@ -414,11 +414,13 @@ nfsd_open(struct svc_rqst *rqstp, struct svc_fh *fhp, int type,
{
struct
dentry
*
dentry
;
struct
inode
*
inode
;
int
err
;
int
flags
=
O_RDONLY
|
O_LARGEFILE
,
mode
=
FMODE_READ
,
err
;
/* If we get here, then the client has already done an "open", and (hopefully)
* checked permission - so allow OWNER_OVERRIDE in case a chmod has now revoked
* permission */
/*
* If we get here, then the client has already done an "open",
* and (hopefully) checked permission - so allow OWNER_OVERRIDE
* in case a chmod has now revoked permission.
*/
err
=
fh_verify
(
rqstp
,
fhp
,
type
,
access
|
MAY_OWNER_OVERRIDE
);
if
(
err
)
goto
out
;
...
...
@@ -443,37 +445,24 @@ nfsd_open(struct svc_rqst *rqstp, struct svc_fh *fhp, int type,
if
(
err
)
goto
out_nfserr
;
if
((
access
&
MAY_WRITE
)
&&
(
err
=
get_write_access
(
inode
))
!=
0
)
if
(
access
&
MAY_WRITE
)
{
err
=
get_write_access
(
inode
);
if
(
err
)
goto
out_nfserr
;
memset
(
filp
,
0
,
sizeof
(
*
filp
));
filp
->
f_op
=
fops_get
(
inode
->
i_fop
);
atomic_set
(
&
filp
->
f_count
,
1
);
filp
->
f_dentry
=
dentry
;
filp
->
f_vfsmnt
=
fhp
->
fh_export
->
ex_mnt
;
if
(
access
&
MAY_WRITE
)
{
filp
->
f_flags
=
O_WRONLY
|
O_LARGEFILE
;
filp
->
f_mode
=
FMODE_WRITE
;
flags
=
O_WRONLY
|
O_LARGEFILE
;
mode
=
FMODE_WRITE
;
DQUOT_INIT
(
inode
);
}
else
{
filp
->
f_flags
=
O_RDONLY
|
O_LARGEFILE
;
filp
->
f_mode
=
FMODE_READ
;
}
err
=
0
;
if
(
filp
->
f_op
&&
filp
->
f_op
->
open
)
{
err
=
filp
->
f_op
->
open
(
inode
,
filp
);
if
(
err
)
{
fops_put
(
filp
->
f_op
);
if
(
access
&
MAY_WRITE
)
err
=
init_private_file
(
filp
,
dentry
,
mode
);
if
(
!
err
)
{
filp
->
f_flags
=
flags
;
filp
->
f_vfsmnt
=
fhp
->
fh_export
->
ex_mnt
;
}
else
if
(
access
&
MAY_WRITE
)
put_write_access
(
inode
);
/* I nearly added put_filp() call here, but this filp
* is really on callers stack frame. -DaveM
*/
atomic_dec
(
&
filp
->
f_count
);
}
}
out_nfserr:
if
(
err
)
err
=
nfserrno
(
err
);
...
...
@@ -490,9 +479,8 @@ nfsd_close(struct file *filp)
struct
dentry
*
dentry
=
filp
->
f_dentry
;
struct
inode
*
inode
=
dentry
->
d_inode
;
if
(
filp
->
f_op
&&
filp
->
f_op
->
release
)
if
(
filp
->
f_op
->
release
)
filp
->
f_op
->
release
(
inode
,
filp
);
fops_put
(
filp
->
f_op
);
if
(
filp
->
f_mode
&
FMODE_WRITE
)
put_write_access
(
inode
);
}
...
...
fs/read_write.c
View file @
67a670bd
...
...
@@ -162,57 +162,107 @@ asmlinkage long sys_llseek(unsigned int fd, unsigned long offset_high,
}
#endif
asmlinkage
ssize_t
sys_read
(
unsigned
int
fd
,
char
*
buf
,
size_t
count
)
ssize_t
vfs_read
(
struct
file
*
file
,
char
*
buf
,
size_t
count
,
loff_t
*
pos
)
{
struct
inode
*
inode
=
file
->
f_dentry
->
d_inode
;
ssize_t
ret
;
struct
file
*
file
;
ret
=
-
EBADF
;
file
=
fget
(
fd
);
if
(
file
)
{
if
(
file
->
f_mode
&
FMODE_READ
)
{
ret
=
locks_verify_area
(
FLOCK_VERIFY_READ
,
file
->
f_dentry
->
d_inode
,
file
,
file
->
f_pos
,
count
);
if
(
!
(
file
->
f_mode
&
FMODE_READ
))
return
-
EBADF
;
if
(
!
file
->
f_op
||
!
file
->
f_op
->
read
)
return
-
EINVAL
;
if
(
pos
<
0
)
return
-
EINVAL
;
ret
=
locks_verify_area
(
FLOCK_VERIFY_READ
,
inode
,
file
,
*
pos
,
count
);
if
(
!
ret
)
{
ssize_t
(
*
read
)(
struct
file
*
,
char
*
,
size_t
,
loff_t
*
);
ret
=
-
EINVAL
;
if
(
file
->
f_op
&&
(
read
=
file
->
f_op
->
read
)
!=
NULL
)
ret
=
read
(
file
,
buf
,
count
,
&
file
->
f_pos
);
}
}
ret
=
file
->
f_op
->
read
(
file
,
buf
,
count
,
pos
);
if
(
ret
>
0
)
dnotify_parent
(
file
->
f_dentry
,
DN_ACCESS
);
fput
(
file
);
}
return
ret
;
}
asmlinkage
ssize_t
sys_write
(
unsigned
int
fd
,
const
char
*
buf
,
size_t
count
)
ssize_t
vfs_write
(
struct
file
*
file
,
const
char
*
buf
,
size_t
count
,
loff_t
*
pos
)
{
struct
inode
*
inode
=
file
->
f_dentry
->
d_inode
;
ssize_t
ret
;
struct
file
*
file
;
ret
=
-
EBADF
;
if
(
!
(
file
->
f_mode
&
FMODE_WRITE
))
return
-
EBADF
;
if
(
!
file
->
f_op
||
!
file
->
f_op
->
write
)
return
-
EINVAL
;
if
(
pos
<
0
)
return
-
EINVAL
;
ret
=
locks_verify_area
(
FLOCK_VERIFY_WRITE
,
inode
,
file
,
*
pos
,
count
);
if
(
!
ret
)
{
ret
=
file
->
f_op
->
write
(
file
,
buf
,
count
,
pos
);
if
(
ret
>
0
)
dnotify_parent
(
file
->
f_dentry
,
DN_MODIFY
);
}
return
ret
;
}
asmlinkage
ssize_t
sys_read
(
unsigned
int
fd
,
char
*
buf
,
size_t
count
)
{
struct
file
*
file
;
ssize_t
ret
=
-
EBADF
;
file
=
fget
(
fd
);
if
(
file
)
{
if
(
file
->
f_mode
&
FMODE_WRITE
)
{
struct
inode
*
inode
=
file
->
f_dentry
->
d_inode
;
ret
=
locks_verify_area
(
FLOCK_VERIFY_WRITE
,
inode
,
file
,
file
->
f_pos
,
count
);
if
(
!
ret
)
{
ssize_t
(
*
write
)(
struct
file
*
,
const
char
*
,
size_t
,
loff_t
*
);
ret
=
-
EINVAL
;
if
(
file
->
f_op
&&
(
write
=
file
->
f_op
->
write
)
!=
NULL
)
ret
=
write
(
file
,
buf
,
count
,
&
file
->
f_pos
);
ret
=
vfs_read
(
file
,
buf
,
count
,
&
file
->
f_pos
);
fput
(
file
);
}
return
ret
;
}
asmlinkage
ssize_t
sys_write
(
unsigned
int
fd
,
const
char
*
buf
,
size_t
count
)
{
struct
file
*
file
;
ssize_t
ret
=
-
EBADF
;
file
=
fget
(
fd
);
if
(
file
)
{
ret
=
vfs_write
(
file
,
buf
,
count
,
&
file
->
f_pos
);
fput
(
file
);
}
if
(
ret
>
0
)
dnotify_parent
(
file
->
f_dentry
,
DN_MODIFY
);
return
ret
;
}
asmlinkage
ssize_t
sys_pread
(
unsigned
int
fd
,
char
*
buf
,
size_t
count
,
loff_t
pos
)
{
struct
file
*
file
;
ssize_t
ret
=
-
EBADF
;
file
=
fget
(
fd
);
if
(
file
)
{
ret
=
vfs_read
(
file
,
buf
,
count
,
&
pos
);
fput
(
file
);
}
return
ret
;
}
asmlinkage
ssize_t
sys_pwrite
(
unsigned
int
fd
,
const
char
*
buf
,
size_t
count
,
loff_t
pos
)
{
struct
file
*
file
;
ssize_t
ret
=
-
EBADF
;
file
=
fget
(
fd
);
if
(
file
)
{
ret
=
vfs_write
(
file
,
buf
,
count
,
&
pos
);
fput
(
file
);
}
return
ret
;
}
static
ssize_t
do_readv_writev
(
int
type
,
struct
file
*
file
,
const
struct
iovec
*
vector
,
...
...
@@ -355,70 +405,3 @@ asmlinkage ssize_t sys_writev(unsigned long fd, const struct iovec * vector,
bad_file:
return
ret
;
}
/* From the Single Unix Spec: pread & pwrite act like lseek to pos + op +
lseek back to original location. They fail just like lseek does on
non-seekable files. */
asmlinkage
ssize_t
sys_pread
(
unsigned
int
fd
,
char
*
buf
,
size_t
count
,
loff_t
pos
)
{
ssize_t
ret
;
struct
file
*
file
;
ssize_t
(
*
read
)(
struct
file
*
,
char
*
,
size_t
,
loff_t
*
);
ret
=
-
EBADF
;
file
=
fget
(
fd
);
if
(
!
file
)
goto
bad_file
;
if
(
!
(
file
->
f_mode
&
FMODE_READ
))
goto
out
;
ret
=
locks_verify_area
(
FLOCK_VERIFY_READ
,
file
->
f_dentry
->
d_inode
,
file
,
pos
,
count
);
if
(
ret
)
goto
out
;
ret
=
-
EINVAL
;
if
(
!
file
->
f_op
||
!
(
read
=
file
->
f_op
->
read
))
goto
out
;
if
(
pos
<
0
)
goto
out
;
ret
=
read
(
file
,
buf
,
count
,
&
pos
);
if
(
ret
>
0
)
dnotify_parent
(
file
->
f_dentry
,
DN_ACCESS
);
out:
fput
(
file
);
bad_file:
return
ret
;
}
asmlinkage
ssize_t
sys_pwrite
(
unsigned
int
fd
,
const
char
*
buf
,
size_t
count
,
loff_t
pos
)
{
ssize_t
ret
;
struct
file
*
file
;
ssize_t
(
*
write
)(
struct
file
*
,
const
char
*
,
size_t
,
loff_t
*
);
ret
=
-
EBADF
;
file
=
fget
(
fd
);
if
(
!
file
)
goto
bad_file
;
if
(
!
(
file
->
f_mode
&
FMODE_WRITE
))
goto
out
;
ret
=
locks_verify_area
(
FLOCK_VERIFY_WRITE
,
file
->
f_dentry
->
d_inode
,
file
,
pos
,
count
);
if
(
ret
)
goto
out
;
ret
=
-
EINVAL
;
if
(
!
file
->
f_op
||
!
(
write
=
file
->
f_op
->
write
))
goto
out
;
if
(
pos
<
0
)
goto
out
;
ret
=
write
(
file
,
buf
,
count
,
&
pos
);
if
(
ret
>
0
)
dnotify_parent
(
file
->
f_dentry
,
DN_MODIFY
);
out:
fput
(
file
);
bad_file:
return
ret
;
}
fs/reiserfs/inode.c
View file @
67a670bd
...
...
@@ -1484,13 +1484,19 @@ static int reiserfs_new_symlink (struct reiserfs_transaction_handle *th,
/* inserts the stat data into the tree, and then calls
reiserfs_new_directory (to insert ".", ".." item if new object is
directory) or reiserfs_new_symlink (to insert symlink body if new
object is symlink) or nothing (if new object is regular file) */
struct
inode
*
reiserfs_new_inode
(
struct
reiserfs_transaction_handle
*
th
,
object is symlink) or nothing (if new object is regular file)
NOTE! uid and gid must already be set in the inode. If we return
non-zero due to an error, we have to drop the quota previously allocated
for the fresh inode. This can only be done outside a transaction, so
if we return non-zero, we also end the transaction. */
int
reiserfs_new_inode
(
struct
reiserfs_transaction_handle
*
th
,
struct
inode
*
dir
,
int
mode
,
const
char
*
symname
,
int
i_size
,
/* 0 for regular, EMTRY_DIR_SIZE for dirs,
/* 0 for regular, EMTRY_DIR_SIZE for dirs,
strlen (symname) for symlinks)*/
struct
dentry
*
dentry
,
struct
inode
*
inode
,
int
*
err
)
loff_t
i_size
,
struct
dentry
*
dentry
,
struct
inode
*
inode
)
{
struct
super_block
*
sb
;
INITIALIZE_PATH
(
path_to_key
);
...
...
@@ -1498,72 +1504,40 @@ struct inode * reiserfs_new_inode (struct reiserfs_transaction_handle *th,
struct
item_head
ih
;
struct
stat_data
sd
;
int
retval
;
int
err
;
if
(
!
dir
||
!
dir
->
i_nlink
)
{
*
err
=
-
EPERM
;
iput
(
inode
)
;
return
NULL
;
err
=
-
EPERM
;
goto
out_bad_inode
;
}
sb
=
dir
->
i_sb
;
inode
->
i_flags
=
0
;
//inode->i_sb->s_flags;
/* item head of new item */
ih
.
ih_key
.
k_dir_id
=
INODE_PKEY
(
dir
)
->
k_objectid
;
ih
.
ih_key
.
k_objectid
=
cpu_to_le32
(
reiserfs_get_unused_objectid
(
th
));
if
(
!
ih
.
ih_key
.
k_objectid
)
{
iput
(
inode
)
;
*
err
=
-
ENOMEM
;
return
NULL
;
err
=
-
ENOMEM
;
goto
out_bad_inode
;
}
if
(
old_format_only
(
sb
))
/* not a perfect generation count, as object ids can be reused, but this
**
is as good as reiserfs can do right now.
/* not a perfect generation count, as object ids can be reused, but
** this
is as good as reiserfs can do right now.
** note that the private part of inode isn't filled in yet, we have
** to use the directory.
*/
inode
->
i_generation
=
le32_to_cpu
(
INODE_PKEY
(
dir
)
->
k_objectid
);
else
#if defined( USE_INODE_GENERATION_COUNTER )
inode
->
i_generation
=
le32_to_cpu
(
REISERFS_SB
(
sb
)
->
s_rs
->
s_inode_generation
);
inode
->
i_generation
=
le32_to_cpu
(
REISERFS_SB
(
sb
)
->
s_rs
->
s_inode_generation
);
#else
inode
->
i_generation
=
++
event
;
#endif
if
(
old_format_only
(
sb
))
make_le_item_head
(
&
ih
,
0
,
KEY_FORMAT_3_5
,
SD_OFFSET
,
TYPE_STAT_DATA
,
SD_V1_SIZE
,
MAX_US_INT
);
else
make_le_item_head
(
&
ih
,
0
,
KEY_FORMAT_3_6
,
SD_OFFSET
,
TYPE_STAT_DATA
,
SD_SIZE
,
MAX_US_INT
);
/* key to search for correct place for new stat data */
_make_cpu_key
(
&
key
,
KEY_FORMAT_3_6
,
le32_to_cpu
(
ih
.
ih_key
.
k_dir_id
),
le32_to_cpu
(
ih
.
ih_key
.
k_objectid
),
SD_OFFSET
,
TYPE_STAT_DATA
,
3
/*key length*/
);
/* find proper place for inserting of stat data */
retval
=
search_item
(
sb
,
&
key
,
&
path_to_key
);
if
(
retval
==
IO_ERROR
)
{
iput
(
inode
);
*
err
=
-
EIO
;
return
NULL
;
}
if
(
retval
==
ITEM_FOUND
)
{
pathrelse
(
&
path_to_key
);
iput
(
inode
);
*
err
=
-
EEXIST
;
return
NULL
;
}
/* fill stat data */
inode
->
i_mode
=
mode
;
inode
->
i_nlink
=
(
S_ISDIR
(
mode
)
?
2
:
1
);
inode
->
i_uid
=
current
->
fsuid
;
if
(
dir
->
i_mode
&
S_ISGID
)
{
inode
->
i_gid
=
dir
->
i_gid
;
if
(
S_ISDIR
(
mode
))
inode
->
i_mode
|=
S_ISGID
;
}
else
inode
->
i_gid
=
current
->
fsgid
;
/* uid and gid must already be set by the caller for quota init */
inode
->
i_mtime
=
inode
->
i_atime
=
inode
->
i_ctime
=
CURRENT_TIME
;
inode
->
i_size
=
i_size
;
...
...
@@ -1578,18 +1552,38 @@ struct inode * reiserfs_new_inode (struct reiserfs_transaction_handle *th,
REISERFS_I
(
inode
)
->
i_trans_id
=
0
;
REISERFS_I
(
inode
)
->
i_trans_index
=
0
;
if
(
old_format_only
(
sb
))
make_le_item_head
(
&
ih
,
0
,
KEY_FORMAT_3_5
,
SD_OFFSET
,
TYPE_STAT_DATA
,
SD_V1_SIZE
,
MAX_US_INT
);
else
make_le_item_head
(
&
ih
,
0
,
KEY_FORMAT_3_6
,
SD_OFFSET
,
TYPE_STAT_DATA
,
SD_SIZE
,
MAX_US_INT
);
/* key to search for correct place for new stat data */
_make_cpu_key
(
&
key
,
KEY_FORMAT_3_6
,
le32_to_cpu
(
ih
.
ih_key
.
k_dir_id
),
le32_to_cpu
(
ih
.
ih_key
.
k_objectid
),
SD_OFFSET
,
TYPE_STAT_DATA
,
3
/*key length*/
);
/* find proper place for inserting of stat data */
retval
=
search_item
(
sb
,
&
key
,
&
path_to_key
);
if
(
retval
==
IO_ERROR
)
{
err
=
-
EIO
;
goto
out_bad_inode
;
}
if
(
retval
==
ITEM_FOUND
)
{
pathrelse
(
&
path_to_key
);
err
=
-
EEXIST
;
goto
out_bad_inode
;
}
if
(
old_format_only
(
sb
))
{
if
(
inode
->
i_uid
&
~
0xffff
||
inode
->
i_gid
&
~
0xffff
)
{
pathrelse
(
&
path_to_key
);
/* i_uid or i_gid is too big to be stored in stat data v3.5 */
iput
(
inode
);
*
err
=
-
EINVAL
;
return
NULL
;
err
=
-
EINVAL
;
goto
out_bad_inode
;
}
inode2sd_v1
(
&
sd
,
inode
);
}
else
}
else
{
inode2sd
(
&
sd
,
inode
);
}
// these do not go to on-disk stat data
inode
->
i_ino
=
le32_to_cpu
(
ih
.
ih_key
.
k_objectid
);
inode
->
i_blksize
=
PAGE_SIZE
;
...
...
@@ -1610,10 +1604,9 @@ struct inode * reiserfs_new_inode (struct reiserfs_transaction_handle *th,
/* insert the stat data into the tree */
retval
=
reiserfs_insert_item
(
th
,
&
path_to_key
,
&
key
,
&
ih
,
(
char
*
)(
&
sd
));
if
(
retval
)
{
iput
(
inode
);
*
err
=
retval
;
err
=
retval
;
reiserfs_check_path
(
&
path_to_key
)
;
return
NULL
;
goto
out_bad_inode
;
}
if
(
S_ISDIR
(
mode
))
{
...
...
@@ -1628,19 +1621,35 @@ struct inode * reiserfs_new_inode (struct reiserfs_transaction_handle *th,
retval
=
reiserfs_new_symlink
(
th
,
&
ih
,
&
path_to_key
,
symname
,
i_size
);
}
if
(
retval
)
{
inode
->
i_nlink
=
0
;
iput
(
inode
);
*
err
=
retval
;
err
=
retval
;
reiserfs_check_path
(
&
path_to_key
)
;
return
NULL
;
journal_end
(
th
,
th
->
t_super
,
th
->
t_blocks_allocated
);
goto
out_inserted_sd
;
}
insert_inode_hash
(
inode
);
// we do not mark inode dirty: on disk content matches to the
// in-core one
reiserfs_update_sd
(
th
,
inode
);
reiserfs_check_path
(
&
path_to_key
)
;
return
inode
;
return
0
;
/* it looks like you can easily compress these two goto targets into
* one. Keeping it like this doesn't actually hurt anything, and they
* are place holders for what the quota code actually needs.
*/
out_bad_inode:
/* Invalidate the object, nothing was inserted yet */
INODE_PKEY
(
inode
)
->
k_objectid
=
0
;
/* dquot_drop must be done outside a transaction */
journal_end
(
th
,
th
->
t_super
,
th
->
t_blocks_allocated
)
;
make_bad_inode
(
inode
);
out_inserted_sd:
inode
->
i_nlink
=
0
;
th
->
t_trans_id
=
0
;
/* so the caller can't use this handle later */
iput
(
inode
);
return
err
;
}
/*
...
...
fs/reiserfs/journal.c
View file @
67a670bd
...
...
@@ -204,7 +204,7 @@ static int set_bit_in_list_bitmap(struct super_block *p_s_sb, int block,
if
(
!
jb
->
bitmaps
[
bmap_nr
])
{
jb
->
bitmaps
[
bmap_nr
]
=
get_bitmap_node
(
p_s_sb
)
;
}
set_bit
(
bit_nr
,
jb
->
bitmaps
[
bmap_nr
]
->
data
)
;
set_bit
(
bit_nr
,
(
unsigned
long
*
)
jb
->
bitmaps
[
bmap_nr
]
->
data
)
;
return
0
;
}
...
...
@@ -550,7 +550,7 @@ int reiserfs_in_journal(struct super_block *p_s_sb,
PROC_INFO_INC
(
p_s_sb
,
journal
.
in_journal_bitmap
);
jb
=
SB_JOURNAL
(
p_s_sb
)
->
j_list_bitmap
+
i
;
if
(
jb
->
journal_list
&&
jb
->
bitmaps
[
bmap_nr
]
&&
test_bit
(
bit_nr
,
jb
->
bitmaps
[
bmap_nr
]
->
data
))
{
test_bit
(
bit_nr
,
(
unsigned
long
*
)
jb
->
bitmaps
[
bmap_nr
]
->
data
))
{
tmp_bit
=
find_next_zero_bit
((
unsigned
long
*
)
(
jb
->
bitmaps
[
bmap_nr
]
->
data
),
p_s_sb
->
s_blocksize
<<
3
,
bit_nr
+
1
)
;
...
...
fs/reiserfs/namei.c
View file @
67a670bd
...
...
@@ -248,7 +248,7 @@ static int linear_search_in_dir_item (struct cpu_key * key, struct reiserfs_dir_
/* mark, that this generation number is used */
if
(
de
->
de_gen_number_bit_string
)
set_bit
(
GET_GENERATION_NUMBER
(
deh_offset
(
deh
)),
de
->
de_gen_number_bit_string
);
set_bit
(
GET_GENERATION_NUMBER
(
deh_offset
(
deh
)),
(
unsigned
long
*
)
de
->
de_gen_number_bit_string
);
// calculate pointer to name and namelen
de
->
de_entry_num
=
i
;
...
...
@@ -504,7 +504,7 @@ static int reiserfs_add_entry (struct reiserfs_transaction_handle *th, struct in
return
-
EEXIST
;
}
gen_number
=
find_first_zero_bit
(
bit_string
,
MAX_GENERATION_NUMBER
+
1
);
gen_number
=
find_first_zero_bit
(
(
unsigned
long
*
)
bit_string
,
MAX_GENERATION_NUMBER
+
1
);
if
(
gen_number
>
MAX_GENERATION_NUMBER
)
{
/* there is no free generation number */
reiserfs_warning
(
"reiserfs_add_entry: Congratulations! we have got hash function screwed up
\n
"
);
...
...
@@ -552,6 +552,40 @@ static int reiserfs_add_entry (struct reiserfs_transaction_handle *th, struct in
return
0
;
}
/* quota utility function, call if you've had to abort after calling
** new_inode_init, and have not called reiserfs_new_inode yet.
** This should only be called on inodes that do not hav stat data
** inserted into the tree yet.
*/
static
int
drop_new_inode
(
struct
inode
*
inode
)
{
make_bad_inode
(
inode
)
;
iput
(
inode
)
;
return
0
;
}
/* utility function that does setup for reiserfs_new_inode.
** DQUOT_ALLOC_INODE cannot be called inside a transaction, so we had
** to pull some bits of reiserfs_new_inode out into this func.
** Yes, the actual quota calls are missing, they are part of the quota
** patch.
*/
static
int
new_inode_init
(
struct
inode
*
inode
,
struct
inode
*
dir
,
int
mode
)
{
/* the quota init calls have to know who to charge the quota to, so
** we have to set uid and gid here
*/
inode
->
i_uid
=
current
->
fsuid
;
inode
->
i_mode
=
mode
;
if
(
dir
->
i_mode
&
S_ISGID
)
{
inode
->
i_gid
=
dir
->
i_gid
;
if
(
S_ISDIR
(
mode
))
inode
->
i_mode
|=
S_ISGID
;
}
else
{
inode
->
i_gid
=
current
->
fsgid
;
}
return
0
;
}
//
// a portion of this function, particularly the VFS interface portion,
...
...
@@ -564,7 +598,6 @@ static int reiserfs_create (struct inode * dir, struct dentry *dentry, int mode)
{
int
retval
;
struct
inode
*
inode
;
int
windex
;
int
jbegin_count
=
JOURNAL_PER_BALANCE_CNT
*
2
;
struct
reiserfs_transaction_handle
th
;
...
...
@@ -572,16 +605,16 @@ static int reiserfs_create (struct inode * dir, struct dentry *dentry, int mode)
if
(
!
inode
)
{
return
-
ENOMEM
;
}
retval
=
new_inode_init
(
inode
,
dir
,
mode
);
if
(
retval
)
return
retval
;
lock_kernel
();
journal_begin
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
th
.
t_caller
=
"create"
;
windex
=
push_journal_writer
(
"reiserfs_create"
)
;
inode
=
reiserfs_new_inode
(
&
th
,
dir
,
mode
,
0
,
0
/*i_size*/
,
dentry
,
inode
,
&
retval
);
if
(
!
inode
)
{
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
unlock_kernel
();
return
retval
;
retval
=
reiserfs_new_inode
(
&
th
,
dir
,
mode
,
0
,
0
/*i_size*/
,
dentry
,
inode
);
if
(
retval
)
{
goto
out_failed
;
}
inode
->
i_op
=
&
reiserfs_file_inode_operations
;
...
...
@@ -593,22 +626,19 @@ static int reiserfs_create (struct inode * dir, struct dentry *dentry, int mode)
if
(
retval
)
{
inode
->
i_nlink
--
;
reiserfs_update_sd
(
&
th
,
inode
);
pop_journal_writer
(
windex
)
;
// FIXME: should we put iput here and have stat data deleted
// in the same transactioin
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
iput
(
inode
);
unlock_kernel
();
return
retval
;
goto
out_failed
;
}
reiserfs_update_inode_transaction
(
inode
)
;
reiserfs_update_inode_transaction
(
dir
)
;
d_instantiate
(
dentry
,
inode
);
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
out_failed:
unlock_kernel
();
return
0
;
return
retval
;
}
...
...
@@ -623,7 +653,6 @@ static int reiserfs_mknod (struct inode * dir, struct dentry *dentry, int mode,
{
int
retval
;
struct
inode
*
inode
;
int
windex
;
struct
reiserfs_transaction_handle
th
;
int
jbegin_count
=
JOURNAL_PER_BALANCE_CNT
*
3
;
...
...
@@ -631,16 +660,16 @@ static int reiserfs_mknod (struct inode * dir, struct dentry *dentry, int mode,
if
(
!
inode
)
{
return
-
ENOMEM
;
}
retval
=
new_inode_init
(
inode
,
dir
,
mode
);
if
(
retval
)
return
retval
;
lock_kernel
();
journal_begin
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
windex
=
push_journal_writer
(
"reiserfs_mknod"
)
;
inode
=
reiserfs_new_inode
(
&
th
,
dir
,
mode
,
0
,
0
/*i_size*/
,
dentry
,
inode
,
&
retval
);
if
(
!
inode
)
{
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
unlock_kernel
();
return
retval
;
retval
=
reiserfs_new_inode
(
&
th
,
dir
,
mode
,
0
,
0
/*i_size*/
,
dentry
,
inode
);
if
(
retval
)
{
goto
out_failed
;
}
init_special_inode
(
inode
,
mode
,
rdev
)
;
...
...
@@ -656,18 +685,17 @@ static int reiserfs_mknod (struct inode * dir, struct dentry *dentry, int mode,
if
(
retval
)
{
inode
->
i_nlink
--
;
reiserfs_update_sd
(
&
th
,
inode
);
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
iput
(
inode
);
unlock_kernel
();
return
retval
;
goto
out_failed
;
}
d_instantiate
(
dentry
,
inode
);
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
out_failed:
unlock_kernel
();
return
0
;
return
retval
;
}
...
...
@@ -682,33 +710,33 @@ static int reiserfs_mkdir (struct inode * dir, struct dentry *dentry, int mode)
{
int
retval
;
struct
inode
*
inode
;
int
windex
;
struct
reiserfs_transaction_handle
th
;
int
jbegin_count
=
JOURNAL_PER_BALANCE_CNT
*
3
;
mode
=
S_IFDIR
|
mode
;
inode
=
new_inode
(
dir
->
i_sb
)
;
if
(
!
inode
)
{
return
-
ENOMEM
;
}
retval
=
new_inode_init
(
inode
,
dir
,
mode
);
if
(
retval
)
return
retval
;
lock_kernel
();
journal_begin
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
windex
=
push_journal_writer
(
"reiserfs_mkdir"
)
;
/* inc the link count now, so another writer doesn't overflow it while
** we sleep later on.
*/
INC_DIR_INODE_NLINK
(
dir
)
mode
=
S_IFDIR
|
mode
;
inode
=
reiserfs_new_inode
(
&
th
,
dir
,
mode
,
0
/*symlink*/
,
old_format_only
(
dir
->
i_sb
)
?
EMPTY_DIR_SIZE_V1
:
EMPTY_DIR_SIZE
,
dentry
,
inode
,
&
retval
);
if
(
!
inode
)
{
pop_journal_writer
(
windex
)
;
retval
=
reiserfs_new_inode
(
&
th
,
dir
,
mode
,
0
/*symlink*/
,
old_format_only
(
dir
->
i_sb
)
?
EMPTY_DIR_SIZE_V1
:
EMPTY_DIR_SIZE
,
dentry
,
inode
);
if
(
retval
)
{
dir
->
i_nlink
--
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
unlock_kernel
();
return
retval
;
goto
out_failed
;
}
reiserfs_update_inode_transaction
(
inode
)
;
reiserfs_update_inode_transaction
(
dir
)
;
...
...
@@ -723,21 +751,19 @@ static int reiserfs_mkdir (struct inode * dir, struct dentry *dentry, int mode)
inode
->
i_nlink
=
0
;
DEC_DIR_INODE_NLINK
(
dir
);
reiserfs_update_sd
(
&
th
,
inode
);
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
iput
(
inode
);
unlock_kernel
();
return
retval
;
goto
out_failed
;
}
// the above add_entry did not update dir's stat data
reiserfs_update_sd
(
&
th
,
dir
);
d_instantiate
(
dentry
,
inode
);
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
out_failed:
unlock_kernel
();
return
0
;
return
retval
;
}
static
inline
int
reiserfs_empty_dir
(
struct
inode
*
inode
)
{
...
...
@@ -942,43 +968,43 @@ static int reiserfs_symlink (struct inode * dir, struct dentry * dentry, const c
struct
inode
*
inode
;
char
*
name
;
int
item_len
;
int
windex
;
struct
reiserfs_transaction_handle
th
;
int
mode
=
S_IFLNK
|
S_IRWXUGO
;
int
jbegin_count
=
JOURNAL_PER_BALANCE_CNT
*
3
;
inode
=
new_inode
(
dir
->
i_sb
)
;
if
(
!
inode
)
{
return
-
ENOMEM
;
}
retval
=
new_inode_init
(
inode
,
dir
,
mode
);
if
(
retval
)
{
return
retval
;
}
lock_kernel
();
item_len
=
ROUND_UP
(
strlen
(
symname
));
if
(
item_len
>
MAX_DIRECT_ITEM_LEN
(
dir
->
i_sb
->
s_blocksize
))
{
iput
(
inode
)
;
return
-
ENAMETOOLONG
;
retval
=
-
ENAMETOOLONG
;
drop_new_inode
(
inode
);
goto
out_failed
;
}
lock_kernel
();
name
=
reiserfs_kmalloc
(
item_len
,
GFP_NOFS
,
dir
->
i_sb
);
if
(
!
name
)
{
iput
(
inode
)
;
unlock_kernel
()
;
return
-
ENOMEM
;
drop_new_inode
(
inode
)
;
retval
=
-
ENOMEM
;
goto
out_failed
;
}
memcpy
(
name
,
symname
,
strlen
(
symname
));
padd_item
(
name
,
item_len
,
strlen
(
symname
));
journal_begin
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
windex
=
push_journal_writer
(
"reiserfs_symlink"
)
;
inode
=
reiserfs_new_inode
(
&
th
,
dir
,
S_IFLNK
|
S_IRWXUGO
,
name
,
strlen
(
symname
),
dentry
,
inode
,
&
retval
);
retval
=
reiserfs_new_inode
(
&
th
,
dir
,
mode
,
name
,
strlen
(
symname
),
dentry
,
inode
);
reiserfs_kfree
(
name
,
item_len
,
dir
->
i_sb
);
if
(
inode
==
0
)
{
/* reiserfs_new_inode iputs for us */
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
unlock_kernel
();
return
retval
;
if
(
retval
)
{
/* reiserfs_new_inode iputs for us */
goto
out_failed
;
}
reiserfs_update_inode_transaction
(
inode
)
;
...
...
@@ -996,18 +1022,16 @@ static int reiserfs_symlink (struct inode * dir, struct dentry * dentry, const c
if
(
retval
)
{
inode
->
i_nlink
--
;
reiserfs_update_sd
(
&
th
,
inode
);
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
iput
(
inode
);
unlock_kernel
();
return
retval
;
goto
out_failed
;
}
d_instantiate
(
dentry
,
inode
);
pop_journal_writer
(
windex
)
;
journal_end
(
&
th
,
dir
->
i_sb
,
jbegin_count
)
;
out_failed:
unlock_kernel
();
return
0
;
return
retval
;
}
...
...
include/asm-i386/xor.h
View file @
67a670bd
...
...
@@ -839,6 +839,8 @@ static struct xor_block_template xor_block_pIII_sse = {
/* Also try the generic routines. */
#include <asm-generic/xor.h>
#define cpu_has_mmx (test_bit(X86_FEATURE_MMX, boot_cpu_data.x86_capability))
#undef XOR_TRY_TEMPLATES
#define XOR_TRY_TEMPLATES \
do { \
...
...
@@ -846,7 +848,7 @@ static struct xor_block_template xor_block_pIII_sse = {
xor_speed(&xor_block_32regs); \
if (cpu_has_xmm) \
xor_speed(&xor_block_pIII_sse); \
if (
md_cpu_has_mmx()
) { \
if (
cpu_has_mmx
) { \
xor_speed(&xor_block_pII_mmx); \
xor_speed(&xor_block_p5_mmx); \
} \
...
...
include/linux/fs.h
View file @
67a670bd
...
...
@@ -759,6 +759,9 @@ struct inode_operations {
struct
seq_file
;
extern
ssize_t
vfs_read
(
struct
file
*
,
char
*
,
size_t
,
loff_t
*
);
extern
ssize_t
vfs_write
(
struct
file
*
,
const
char
*
,
size_t
,
loff_t
*
);
/*
* NOTE: write_inode, delete_inode, clear_inode, put_inode can be called
* without the big kernel lock held in all filesystems.
...
...
include/linux/raid/md_k.h
View file @
67a670bd
...
...
@@ -169,6 +169,7 @@ struct mdk_rdev_s
struct
block_device
*
bdev
;
/* block device handle */
struct
page
*
sb_page
;
mdp_super_t
*
sb
;
unsigned
long
sb_offset
;
...
...
include/linux/raid/raid5.h
View file @
67a670bd
...
...
@@ -7,21 +7,21 @@
/*
*
* Each stripe contains one buffer per disc. Each buffer can be in
* one of a number of states
determined by bh_state
. Changes between
* one of a number of states
stored in "flags"
. Changes between
* these states happen *almost* exclusively under a per-stripe
* spinlock. Some very specific changes can happen in b_end_io, and
* spinlock. Some very specific changes can happen in b
i
_end_io, and
* these are not protected by the spin lock.
*
* The
bh_state
bits that are used to represent these states are:
*
BH_Uptodate, BH_Lock
* The
flag
bits that are used to represent these states are:
*
R5_UPTODATE and R5_LOCKED
*
* State Empty == !U
ptodate, !Lock
* State Empty == !U
PTODATE, !LOCK
* We have no data, and there is no active request
* State Want == !U
ptodate, Lock
* State Want == !U
PTODATE, LOCK
* A read request is being submitted for this block
* State Dirty == U
ptodate, Lock
* State Dirty == U
PTODATE, LOCK
* Some new data is in this buffer, and it is being written out
* State Clean == U
ptodate, !Lock
* State Clean == U
PTODATE, !LOCK
* We have valid data which is the same as on disc
*
* The possible state transitions are:
...
...
@@ -124,24 +124,29 @@
* plus raid5d if it is handling it, plus one for each active request
* on a cached buffer.
*/
struct
stripe_head
{
struct
stripe_head
*
hash_next
,
**
hash_pprev
;
/* hash pointers */
struct
list_head
lru
;
/* inactive_list or handle_list */
struct
raid5_private_data
*
raid_conf
;
struct
buffer_head
*
bh_cache
[
MD_SB_DISKS
];
/* buffered copy */
struct
buffer_head
*
bh_read
[
MD_SB_DISKS
];
/* read request buffers of the MD device */
struct
buffer_head
*
bh_write
[
MD_SB_DISKS
];
/* write request buffers of the MD device */
struct
buffer_head
*
bh_written
[
MD_SB_DISKS
];
/* write request buffers of the MD device that have been scheduled for write */
struct
page
*
bh_page
[
MD_SB_DISKS
];
/* saved bh_cache[n]->b_page when reading around the cache */
unsigned
long
sector
;
/* sector of this row */
int
size
;
/* buffers size */
sector_t
sector
;
/* sector of this row */
int
pd_idx
;
/* parity disk index */
unsigned
long
state
;
/* state flags */
atomic_t
count
;
/* nr of active thread/requests */
spinlock_t
lock
;
int
sync_redone
;
struct
r5dev
{
struct
bio
req
;
struct
bio_vec
vec
;
struct
page
*
page
;
struct
bio
*
toread
,
*
towrite
,
*
written
;
sector_t
sector
;
/* sector of this page */
unsigned
long
flags
;
}
dev
[
1
];
/* allocated with extra space depending of RAID geometry */
};
/* Flags */
#define R5_UPTODATE 0
/* page contains current data */
#define R5_LOCKED 1
/* IO has been submitted on "req" */
#define R5_OVERWRITE 2
/* towrite covers whole page */
/*
* Write method
...
...
@@ -187,6 +192,7 @@ struct stripe_head {
struct
disk_info
{
kdev_t
dev
;
struct
block_device
*
bdev
;
int
operational
;
int
number
;
int
raid_disk
;
...
...
@@ -201,7 +207,6 @@ struct raid5_private_data {
mdk_thread_t
*
thread
,
*
resync_thread
;
struct
disk_info
disks
[
MD_SB_DISKS
];
struct
disk_info
*
spare
;
int
buffer_size
;
int
chunk_size
,
level
,
algorithm
;
int
raid_disks
,
working_disks
,
failed_disks
;
int
resync_parity
;
...
...
@@ -210,16 +215,19 @@ struct raid5_private_data {
struct
list_head
handle_list
;
/* stripes needing handling */
struct
list_head
delayed_list
;
/* stripes that have plugged requests */
atomic_t
preread_active_stripes
;
/* stripes with scheduled io */
char
cache_name
[
20
];
kmem_cache_t
*
slab_cache
;
/* for allocating stripes */
/*
* Free stripes pool
*/
atomic_t
active_stripes
;
struct
list_head
inactive_list
;
md_
wait_queue_head_t
wait_for_stripe
;
wait_queue_head_t
wait_for_stripe
;
int
inactive_blocked
;
/* release of inactive stripes blocked,
* waiting for 25% to be free
*/
md_
spinlock_t
device_lock
;
spinlock_t
device_lock
;
int
plugged
;
struct
tq_struct
plug_tq
;
...
...
include/linux/raid/xor.h
View file @
67a670bd
...
...
@@ -5,7 +5,7 @@
#define MAX_XOR_BLOCKS 5
extern
void
xor_block
(
unsigned
int
count
,
struct
buffer_head
**
bh_
ptr
);
extern
void
xor_block
(
unsigned
int
count
,
unsigned
int
bytes
,
void
**
ptr
);
struct
xor_block_template
{
struct
xor_block_template
*
next
;
...
...
include/linux/reiserfs_fs.h
View file @
67a670bd
...
...
@@ -1841,10 +1841,10 @@ struct inode * reiserfs_iget (struct super_block * s,
const
struct
cpu_key
*
key
);
struct
inode
*
reiserfs_new_inode
(
struct
reiserfs_transaction_handle
*
th
,
int
reiserfs_new_inode
(
struct
reiserfs_transaction_handle
*
th
,
struct
inode
*
dir
,
int
mode
,
const
char
*
symname
,
int
item_len
,
struct
dentry
*
dentry
,
struct
inode
*
inode
,
int
*
err
);
const
char
*
symname
,
loff_t
i_size
,
struct
dentry
*
dentry
,
struct
inode
*
inode
);
int
reiserfs_sync_inode
(
struct
reiserfs_transaction_handle
*
th
,
struct
inode
*
inode
);
void
reiserfs_update_sd
(
struct
reiserfs_transaction_handle
*
th
,
struct
inode
*
inode
);
...
...
kernel/exit.c
View file @
67a670bd
...
...
@@ -231,7 +231,7 @@ void daemonize(void)
/*
* When we die, we re-parent all our children.
* Try to give them to another thread in our
process
* Try to give them to another thread in our
thread
* group, and if no such member exists, give it to
* the global child reaper process (ie "init")
*/
...
...
@@ -241,8 +241,14 @@ static inline void forget_original_parent(struct task_struct * father)
read_lock
(
&
tasklist_lock
);
/* Next in our thread group */
reaper
=
next_thread
(
father
);
/* Next in our thread group, if they're not already exiting */
reaper
=
father
;
do
{
reaper
=
next_thread
(
reaper
);
if
(
!
(
reaper
->
flags
&
PF_EXITING
))
break
;
}
while
(
reaper
!=
father
);
if
(
reaper
==
father
)
reaper
=
child_reaper
;
...
...
kernel/ksyms.c
View file @
67a670bd
...
...
@@ -243,6 +243,8 @@ EXPORT_SYMBOL(shrink_dcache_anon);
EXPORT_SYMBOL
(
find_inode_number
);
EXPORT_SYMBOL
(
is_subdir
);
EXPORT_SYMBOL
(
get_unused_fd
);
EXPORT_SYMBOL
(
vfs_read
);
EXPORT_SYMBOL
(
vfs_write
);
EXPORT_SYMBOL
(
vfs_create
);
EXPORT_SYMBOL
(
vfs_mkdir
);
EXPORT_SYMBOL
(
vfs_mknod
);
...
...
net/sunrpc/svcsock.c
View file @
67a670bd
...
...
@@ -949,12 +949,10 @@ svc_sock_update_bufs(struct svc_serv *serv)
if
(
sock
->
type
==
SOCK_DGRAM
)
{
/* udp sockets need large rcvbuf as all pending
* requests are still in that buffer.
* As outgoing requests do not wait for an
* ACK, only a moderate sndbuf is needed
*/
svc_sock_setbufsize
(
sock
,
5
*
serv
->
sv_bufsz
,
(
serv
->
sv_nrthreads
+
2
)
*
serv
->
sv_bufsz
);
(
serv
->
sv_nrthreads
+
3
)
*
serv
->
sv_bufsz
,
(
serv
->
sv_nrthreads
+
3
)
*
serv
->
sv_bufsz
);
}
else
if
(
svsk
->
sk_sk
->
state
!=
TCP_LISTEN
)
{
printk
(
KERN_ERR
"RPC update_bufs: permanent sock neither UDP or TCP_LISTEN
\n
"
);
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment