Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
8fe826f9
Commit
8fe826f9
authored
Feb 13, 2019
by
Kent Overstreet
Committed by
Kent Overstreet
Oct 22, 2023
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
bcachefs: Convert bucket invalidation to key marking path
Signed-off-by:
Kent Overstreet
<
kent.overstreet@linux.dev
>
parent
73c27c60
Changes
9
Show whitespace changes
Inline
Side-by-side
Showing
9 changed files
with
276 additions
and
49 deletions
+276
-49
fs/bcachefs/alloc_background.c
fs/bcachefs/alloc_background.c
+184
-27
fs/bcachefs/alloc_background.h
fs/bcachefs/alloc_background.h
+9
-0
fs/bcachefs/bcachefs_format.h
fs/bcachefs/bcachefs_format.h
+8
-8
fs/bcachefs/btree_types.h
fs/bcachefs/btree_types.h
+1
-0
fs/bcachefs/btree_update.h
fs/bcachefs/btree_update.h
+5
-4
fs/bcachefs/buckets.c
fs/bcachefs/buckets.c
+64
-7
fs/bcachefs/buckets.h
fs/bcachefs/buckets.h
+2
-1
fs/bcachefs/fifo.h
fs/bcachefs/fifo.h
+1
-1
fs/bcachefs/journal_io.c
fs/bcachefs/journal_io.c
+2
-1
No files found.
fs/bcachefs/alloc_background.c
View file @
8fe826f9
...
@@ -129,6 +129,34 @@ static inline void put_alloc_field(struct bkey_i_alloc *a, void **p,
...
@@ -129,6 +129,34 @@ static inline void put_alloc_field(struct bkey_i_alloc *a, void **p,
*
p
+=
bytes
;
*
p
+=
bytes
;
}
}
struct
bkey_alloc_unpacked
bch2_alloc_unpack
(
const
struct
bch_alloc
*
a
)
{
struct
bkey_alloc_unpacked
ret
=
{
.
gen
=
a
->
gen
};
const
void
*
d
=
a
->
data
;
unsigned
idx
=
0
;
#define x(_name, _bits) ret._name = get_alloc_field(a, &d, idx++);
BCH_ALLOC_FIELDS
()
#undef x
return
ret
;
}
static
void
bch2_alloc_pack
(
struct
bkey_i_alloc
*
dst
,
const
struct
bkey_alloc_unpacked
src
)
{
unsigned
idx
=
0
;
void
*
d
=
dst
->
v
.
data
;
dst
->
v
.
fields
=
0
;
dst
->
v
.
gen
=
src
.
gen
;
#define x(_name, _bits) put_alloc_field(dst, &d, idx++, src._name);
BCH_ALLOC_FIELDS
()
#undef x
set_bkey_val_bytes
(
&
dst
->
k
,
(
void
*
)
d
-
(
void
*
)
&
dst
->
v
);
}
static
unsigned
bch_alloc_val_u64s
(
const
struct
bch_alloc
*
a
)
static
unsigned
bch_alloc_val_u64s
(
const
struct
bch_alloc
*
a
)
{
{
unsigned
i
,
bytes
=
offsetof
(
struct
bch_alloc
,
data
);
unsigned
i
,
bytes
=
offsetof
(
struct
bch_alloc
,
data
);
...
@@ -174,16 +202,24 @@ void bch2_alloc_to_text(struct printbuf *out, struct bch_fs *c,
...
@@ -174,16 +202,24 @@ void bch2_alloc_to_text(struct printbuf *out, struct bch_fs *c,
static
void
__alloc_read_key
(
struct
bucket
*
g
,
const
struct
bch_alloc
*
a
)
static
void
__alloc_read_key
(
struct
bucket
*
g
,
const
struct
bch_alloc
*
a
)
{
{
const
void
*
d
=
a
->
data
;
const
void
*
d
=
a
->
data
;
unsigned
idx
=
0
;
unsigned
idx
=
0
,
data_type
,
dirty_sectors
,
cached_sectors
;
struct
bucket_mark
m
;
g
->
_mark
.
gen
=
a
->
gen
;
g
->
gen_valid
=
1
;
g
->
io_time
[
READ
]
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
io_time
[
READ
]
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
io_time
[
WRITE
]
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
io_time
[
WRITE
]
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
_mark
.
data_type
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
data_type
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
_mark
.
dirty_sectors
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
dirty_sectors
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
_mark
.
cached_sectors
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
cached_sectors
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
oldest_gen
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
g
->
oldest_gen
=
get_alloc_field
(
a
,
&
d
,
idx
++
);
bucket_cmpxchg
(
g
,
m
,
({
m
.
gen
=
a
->
gen
;
m
.
data_type
=
data_type
;
m
.
dirty_sectors
=
dirty_sectors
;
m
.
cached_sectors
=
cached_sectors
;
}));
g
->
gen_valid
=
1
;
}
}
static
void
__alloc_write_key
(
struct
bkey_i_alloc
*
a
,
struct
bucket
*
g
,
static
void
__alloc_write_key
(
struct
bkey_i_alloc
*
a
,
struct
bucket
*
g
,
...
@@ -318,6 +354,7 @@ static int __bch2_alloc_write_key(struct bch_fs *c, struct bch_dev *ca,
...
@@ -318,6 +354,7 @@ static int __bch2_alloc_write_key(struct bch_fs *c, struct bch_dev *ca,
BTREE_INSERT_NOFAIL
|
BTREE_INSERT_NOFAIL
|
BTREE_INSERT_USE_RESERVE
|
BTREE_INSERT_USE_RESERVE
|
BTREE_INSERT_USE_ALLOC_RESERVE
|
BTREE_INSERT_USE_ALLOC_RESERVE
|
BTREE_INSERT_NOMARK
|
flags
,
flags
,
BTREE_INSERT_ENTRY
(
iter
,
&
a
->
k_i
));
BTREE_INSERT_ENTRY
(
iter
,
&
a
->
k_i
));
if
(
ret
)
if
(
ret
)
...
@@ -361,7 +398,8 @@ int bch2_alloc_replay_key(struct bch_fs *c, struct bkey_i *k)
...
@@ -361,7 +398,8 @@ int bch2_alloc_replay_key(struct bch_fs *c, struct bkey_i *k)
?
0
?
0
:
bch2_btree_insert_at
(
c
,
NULL
,
NULL
,
:
bch2_btree_insert_at
(
c
,
NULL
,
NULL
,
BTREE_INSERT_NOFAIL
|
BTREE_INSERT_NOFAIL
|
BTREE_INSERT_JOURNAL_REPLAY
,
BTREE_INSERT_JOURNAL_REPLAY
|
BTREE_INSERT_NOMARK
,
BTREE_INSERT_ENTRY
(
&
iter
,
k
));
BTREE_INSERT_ENTRY
(
&
iter
,
k
));
err:
err:
bch2_btree_iter_unlock
(
&
iter
);
bch2_btree_iter_unlock
(
&
iter
);
...
@@ -827,6 +865,142 @@ static inline long next_alloc_bucket(struct bch_dev *ca)
...
@@ -827,6 +865,142 @@ static inline long next_alloc_bucket(struct bch_dev *ca)
return
-
1
;
return
-
1
;
}
}
/*
* returns sequence number of most recent journal entry that updated this
* bucket:
*/
static
u64
bucket_journal_seq
(
struct
bch_fs
*
c
,
struct
bucket_mark
m
)
{
if
(
m
.
journal_seq_valid
)
{
u64
journal_seq
=
atomic64_read
(
&
c
->
journal
.
seq
);
u64
bucket_seq
=
journal_seq
;
bucket_seq
&=
~
((
u64
)
U16_MAX
);
bucket_seq
|=
m
.
journal_seq
;
if
(
bucket_seq
>
journal_seq
)
bucket_seq
-=
1
<<
16
;
return
bucket_seq
;
}
else
{
return
0
;
}
}
static
int
bch2_invalidate_one_bucket2
(
struct
bch_fs
*
c
,
struct
bch_dev
*
ca
,
struct
btree_iter
*
iter
,
u64
*
journal_seq
,
unsigned
flags
)
{
#if 0
__BKEY_PADDED(k, BKEY_ALLOC_VAL_U64s_MAX) alloc_key;
#else
/* hack: */
__BKEY_PADDED
(
k
,
8
)
alloc_key
;
#endif
struct
bkey_i_alloc
*
a
;
struct
bkey_alloc_unpacked
u
;
struct
bucket_mark
m
;
struct
bkey_s_c
k
;
bool
invalidating_cached_data
;
size_t
b
;
int
ret
;
BUG_ON
(
!
ca
->
alloc_heap
.
used
||
!
ca
->
alloc_heap
.
data
[
0
].
nr
);
b
=
ca
->
alloc_heap
.
data
[
0
].
bucket
;
/* first, put on free_inc and mark as owned by allocator: */
percpu_down_read
(
&
c
->
mark_lock
);
spin_lock
(
&
c
->
freelist_lock
);
verify_not_on_freelist
(
c
,
ca
,
b
);
BUG_ON
(
!
fifo_push
(
&
ca
->
free_inc
,
b
));
bch2_mark_alloc_bucket
(
c
,
ca
,
b
,
true
,
gc_pos_alloc
(
c
,
NULL
),
0
);
m
=
bucket
(
ca
,
b
)
->
mark
;
spin_unlock
(
&
c
->
freelist_lock
);
percpu_up_read
(
&
c
->
mark_lock
);
bch2_btree_iter_cond_resched
(
iter
);
BUG_ON
(
BKEY_ALLOC_VAL_U64s_MAX
>
8
);
bch2_btree_iter_set_pos
(
iter
,
POS
(
ca
->
dev_idx
,
b
));
retry:
k
=
bch2_btree_iter_peek_slot
(
iter
);
ret
=
btree_iter_err
(
k
);
if
(
ret
)
return
ret
;
if
(
k
.
k
&&
k
.
k
->
type
==
KEY_TYPE_alloc
)
u
=
bch2_alloc_unpack
(
bkey_s_c_to_alloc
(
k
).
v
);
else
memset
(
&
u
,
0
,
sizeof
(
u
));
invalidating_cached_data
=
u
.
cached_sectors
!=
0
;
//BUG_ON(u.dirty_sectors);
u
.
data_type
=
0
;
u
.
dirty_sectors
=
0
;
u
.
cached_sectors
=
0
;
u
.
read_time
=
c
->
bucket_clock
[
READ
].
hand
;
u
.
write_time
=
c
->
bucket_clock
[
WRITE
].
hand
;
u
.
gen
++
;
a
=
bkey_alloc_init
(
&
alloc_key
.
k
);
a
->
k
.
p
=
iter
->
pos
;
bch2_alloc_pack
(
a
,
u
);
ret
=
bch2_btree_insert_at
(
c
,
NULL
,
invalidating_cached_data
?
journal_seq
:
NULL
,
BTREE_INSERT_ATOMIC
|
BTREE_INSERT_NOCHECK_RW
|
BTREE_INSERT_NOFAIL
|
BTREE_INSERT_USE_RESERVE
|
BTREE_INSERT_USE_ALLOC_RESERVE
|
flags
,
BTREE_INSERT_ENTRY
(
iter
,
&
a
->
k_i
));
if
(
ret
==
-
EINTR
)
goto
retry
;
if
(
!
ret
)
{
/* remove from alloc_heap: */
struct
alloc_heap_entry
e
,
*
top
=
ca
->
alloc_heap
.
data
;
top
->
bucket
++
;
top
->
nr
--
;
if
(
!
top
->
nr
)
heap_pop
(
&
ca
->
alloc_heap
,
e
,
bucket_alloc_cmp
,
NULL
);
/*
* Make sure we flush the last journal entry that updated this
* bucket (i.e. deleting the last reference) before writing to
* this bucket again:
*/
*
journal_seq
=
max
(
*
journal_seq
,
bucket_journal_seq
(
c
,
m
));
}
else
{
size_t
b2
;
/* remove from free_inc: */
percpu_down_read
(
&
c
->
mark_lock
);
spin_lock
(
&
c
->
freelist_lock
);
bch2_mark_alloc_bucket
(
c
,
ca
,
b
,
false
,
gc_pos_alloc
(
c
,
NULL
),
0
);
BUG_ON
(
!
fifo_pop_back
(
&
ca
->
free_inc
,
b2
));
BUG_ON
(
b
!=
b2
);
spin_unlock
(
&
c
->
freelist_lock
);
percpu_up_read
(
&
c
->
mark_lock
);
}
return
ret
;
}
static
bool
bch2_invalidate_one_bucket
(
struct
bch_fs
*
c
,
struct
bch_dev
*
ca
,
static
bool
bch2_invalidate_one_bucket
(
struct
bch_fs
*
c
,
struct
bch_dev
*
ca
,
size_t
bucket
,
u64
*
flush_seq
)
size_t
bucket
,
u64
*
flush_seq
)
{
{
...
@@ -847,18 +1021,7 @@ static bool bch2_invalidate_one_bucket(struct bch_fs *c, struct bch_dev *ca,
...
@@ -847,18 +1021,7 @@ static bool bch2_invalidate_one_bucket(struct bch_fs *c, struct bch_dev *ca,
percpu_up_read
(
&
c
->
mark_lock
);
percpu_up_read
(
&
c
->
mark_lock
);
if
(
m
.
journal_seq_valid
)
{
*
flush_seq
=
max
(
*
flush_seq
,
bucket_journal_seq
(
c
,
m
));
u64
journal_seq
=
atomic64_read
(
&
c
->
journal
.
seq
);
u64
bucket_seq
=
journal_seq
;
bucket_seq
&=
~
((
u64
)
U16_MAX
);
bucket_seq
|=
m
.
journal_seq
;
if
(
bucket_seq
>
journal_seq
)
bucket_seq
-=
1
<<
16
;
*
flush_seq
=
max
(
*
flush_seq
,
bucket_seq
);
}
return
m
.
cached_sectors
!=
0
;
return
m
.
cached_sectors
!=
0
;
}
}
...
@@ -871,7 +1034,6 @@ static int bch2_invalidate_buckets(struct bch_fs *c, struct bch_dev *ca)
...
@@ -871,7 +1034,6 @@ static int bch2_invalidate_buckets(struct bch_fs *c, struct bch_dev *ca)
struct
btree_iter
iter
;
struct
btree_iter
iter
;
u64
journal_seq
=
0
;
u64
journal_seq
=
0
;
int
ret
=
0
;
int
ret
=
0
;
long
b
;
bch2_btree_iter_init
(
&
iter
,
c
,
BTREE_ID_ALLOC
,
POS
(
ca
->
dev_idx
,
0
),
bch2_btree_iter_init
(
&
iter
,
c
,
BTREE_ID_ALLOC
,
POS
(
ca
->
dev_idx
,
0
),
BTREE_ITER_SLOTS
|
BTREE_ITER_INTENT
);
BTREE_ITER_SLOTS
|
BTREE_ITER_INTENT
);
...
@@ -879,16 +1041,11 @@ static int bch2_invalidate_buckets(struct bch_fs *c, struct bch_dev *ca)
...
@@ -879,16 +1041,11 @@ static int bch2_invalidate_buckets(struct bch_fs *c, struct bch_dev *ca)
/* Only use nowait if we've already invalidated at least one bucket: */
/* Only use nowait if we've already invalidated at least one bucket: */
while
(
!
ret
&&
while
(
!
ret
&&
!
fifo_full
(
&
ca
->
free_inc
)
&&
!
fifo_full
(
&
ca
->
free_inc
)
&&
(
b
=
next_alloc_bucket
(
ca
))
>=
0
)
{
ca
->
alloc_heap
.
used
)
bool
must_flush
=
ret
=
bch2_invalidate_one_bucket2
(
c
,
ca
,
&
iter
,
&
journal_seq
,
bch2_invalidate_one_bucket
(
c
,
ca
,
b
,
&
journal_seq
);
ret
=
__bch2_alloc_write_key
(
c
,
ca
,
b
,
&
iter
,
must_flush
?
&
journal_seq
:
NULL
,
BTREE_INSERT_GC_LOCK_HELD
|
BTREE_INSERT_GC_LOCK_HELD
|
(
!
fifo_empty
(
&
ca
->
free_inc
)
(
!
fifo_empty
(
&
ca
->
free_inc
)
?
BTREE_INSERT_NOWAIT
:
0
));
?
BTREE_INSERT_NOWAIT
:
0
));
}
bch2_btree_iter_unlock
(
&
iter
);
bch2_btree_iter_unlock
(
&
iter
);
...
...
fs/bcachefs/alloc_background.h
View file @
8fe826f9
...
@@ -6,6 +6,15 @@
...
@@ -6,6 +6,15 @@
#include "alloc_types.h"
#include "alloc_types.h"
#include "debug.h"
#include "debug.h"
struct
bkey_alloc_unpacked
{
u8
gen
;
#define x(_name, _bits) u##_bits _name;
BCH_ALLOC_FIELDS
()
#undef x
};
struct
bkey_alloc_unpacked
bch2_alloc_unpack
(
const
struct
bch_alloc
*
);
#define ALLOC_SCAN_BATCH(ca) max_t(size_t, 1, (ca)->mi.nbuckets >> 9)
#define ALLOC_SCAN_BATCH(ca) max_t(size_t, 1, (ca)->mi.nbuckets >> 9)
const
char
*
bch2_alloc_invalid
(
const
struct
bch_fs
*
,
struct
bkey_s_c
);
const
char
*
bch2_alloc_invalid
(
const
struct
bch_fs
*
,
struct
bkey_s_c
);
...
...
fs/bcachefs/bcachefs_format.h
View file @
8fe826f9
...
@@ -826,12 +826,12 @@ struct bch_alloc {
...
@@ -826,12 +826,12 @@ struct bch_alloc {
}
__attribute__
((
packed
,
aligned
(
8
)));
}
__attribute__
((
packed
,
aligned
(
8
)));
#define BCH_ALLOC_FIELDS() \
#define BCH_ALLOC_FIELDS() \
x(read_time,
2)
\
x(read_time,
16)
\
x(write_time,
2)
\
x(write_time,
16)
\
x(data_type,
1)
\
x(data_type,
8)
\
x(dirty_sectors,
2)
\
x(dirty_sectors,
16)
\
x(cached_sectors,
2)
\
x(cached_sectors,
16)
\
x(oldest_gen,
1
)
x(oldest_gen,
8
)
enum
{
enum
{
#define x(name, bytes) BCH_ALLOC_FIELD_##name,
#define x(name, bytes) BCH_ALLOC_FIELD_##name,
...
@@ -841,12 +841,12 @@ enum {
...
@@ -841,12 +841,12 @@ enum {
};
};
static
const
unsigned
BCH_ALLOC_FIELD_BYTES
[]
=
{
static
const
unsigned
BCH_ALLOC_FIELD_BYTES
[]
=
{
#define x(name, b
ytes) [BCH_ALLOC_FIELD_##name] = bytes
,
#define x(name, b
its) [BCH_ALLOC_FIELD_##name] = bits / 8
,
BCH_ALLOC_FIELDS
()
BCH_ALLOC_FIELDS
()
#undef x
#undef x
};
};
#define x(name, b
ytes) + bytes
#define x(name, b
its) + (bits / 8)
static
const
unsigned
BKEY_ALLOC_VAL_U64s_MAX
=
static
const
unsigned
BKEY_ALLOC_VAL_U64s_MAX
=
DIV_ROUND_UP
(
offsetof
(
struct
bch_alloc
,
data
)
DIV_ROUND_UP
(
offsetof
(
struct
bch_alloc
,
data
)
BCH_ALLOC_FIELDS
(),
sizeof
(
u64
));
BCH_ALLOC_FIELDS
(),
sizeof
(
u64
));
...
...
fs/bcachefs/btree_types.h
View file @
8fe826f9
...
@@ -456,6 +456,7 @@ static inline bool btree_node_is_extents(struct btree *b)
...
@@ -456,6 +456,7 @@ static inline bool btree_node_is_extents(struct btree *b)
static
inline
bool
btree_node_type_needs_gc
(
enum
btree_node_type
type
)
static
inline
bool
btree_node_type_needs_gc
(
enum
btree_node_type
type
)
{
{
switch
(
type
)
{
switch
(
type
)
{
case
BKEY_TYPE_ALLOC
:
case
BKEY_TYPE_BTREE
:
case
BKEY_TYPE_BTREE
:
case
BKEY_TYPE_EXTENTS
:
case
BKEY_TYPE_EXTENTS
:
case
BKEY_TYPE_INODES
:
case
BKEY_TYPE_INODES
:
...
...
fs/bcachefs/btree_update.h
View file @
8fe826f9
...
@@ -82,6 +82,7 @@ enum {
...
@@ -82,6 +82,7 @@ enum {
__BTREE_INSERT_USE_RESERVE
,
__BTREE_INSERT_USE_RESERVE
,
__BTREE_INSERT_USE_ALLOC_RESERVE
,
__BTREE_INSERT_USE_ALLOC_RESERVE
,
__BTREE_INSERT_JOURNAL_REPLAY
,
__BTREE_INSERT_JOURNAL_REPLAY
,
__BTREE_INSERT_NOMARK
,
__BTREE_INSERT_NOWAIT
,
__BTREE_INSERT_NOWAIT
,
__BTREE_INSERT_GC_LOCK_HELD
,
__BTREE_INSERT_GC_LOCK_HELD
,
__BCH_HASH_SET_MUST_CREATE
,
__BCH_HASH_SET_MUST_CREATE
,
...
@@ -108,12 +109,12 @@ enum {
...
@@ -108,12 +109,12 @@ enum {
#define BTREE_INSERT_USE_RESERVE (1 << __BTREE_INSERT_USE_RESERVE)
#define BTREE_INSERT_USE_RESERVE (1 << __BTREE_INSERT_USE_RESERVE)
#define BTREE_INSERT_USE_ALLOC_RESERVE (1 << __BTREE_INSERT_USE_ALLOC_RESERVE)
#define BTREE_INSERT_USE_ALLOC_RESERVE (1 << __BTREE_INSERT_USE_ALLOC_RESERVE)
/*
/* Insert is for journal replay - don't get journal reservations: */
* Insert is for journal replay: don't get journal reservations, or mark extents
* (bch_mark_key)
*/
#define BTREE_INSERT_JOURNAL_REPLAY (1 << __BTREE_INSERT_JOURNAL_REPLAY)
#define BTREE_INSERT_JOURNAL_REPLAY (1 << __BTREE_INSERT_JOURNAL_REPLAY)
/* Don't call bch2_mark_key: */
#define BTREE_INSERT_NOMARK (1 << __BTREE_INSERT_NOMARK)
/* Don't block on allocation failure (for new btree nodes: */
/* Don't block on allocation failure (for new btree nodes: */
#define BTREE_INSERT_NOWAIT (1 << __BTREE_INSERT_NOWAIT)
#define BTREE_INSERT_NOWAIT (1 << __BTREE_INSERT_NOWAIT)
#define BTREE_INSERT_GC_LOCK_HELD (1 << __BTREE_INSERT_GC_LOCK_HELD)
#define BTREE_INSERT_GC_LOCK_HELD (1 << __BTREE_INSERT_GC_LOCK_HELD)
...
...
fs/bcachefs/buckets.c
View file @
8fe826f9
...
@@ -476,6 +476,60 @@ void bch2_mark_alloc_bucket(struct bch_fs *c, struct bch_dev *ca,
...
@@ -476,6 +476,60 @@ void bch2_mark_alloc_bucket(struct bch_fs *c, struct bch_dev *ca,
ca
,
b
,
owned_by_allocator
);
ca
,
b
,
owned_by_allocator
);
}
}
static
int
bch2_mark_alloc
(
struct
bch_fs
*
c
,
struct
bkey_s_c
k
,
bool
inserting
,
struct
bch_fs_usage
*
fs_usage
,
unsigned
journal_seq
,
unsigned
flags
,
bool
gc
)
{
struct
bkey_alloc_unpacked
u
;
struct
bch_dev
*
ca
;
struct
bucket
*
g
;
struct
bucket_mark
old
,
m
;
if
(
!
inserting
)
return
0
;
/*
* alloc btree is read in by bch2_alloc_read, not gc:
*/
if
(
flags
&
BCH_BUCKET_MARK_GC
)
return
0
;
u
=
bch2_alloc_unpack
(
bkey_s_c_to_alloc
(
k
).
v
);
ca
=
bch_dev_bkey_exists
(
c
,
k
.
k
->
p
.
inode
);
g
=
__bucket
(
ca
,
k
.
k
->
p
.
offset
,
gc
);
/*
* this should currently only be getting called from the bucket
* invalidate path:
*/
BUG_ON
(
u
.
dirty_sectors
);
BUG_ON
(
u
.
cached_sectors
);
BUG_ON
(
!
g
->
mark
.
owned_by_allocator
);
old
=
bucket_data_cmpxchg
(
c
,
ca
,
fs_usage
,
g
,
m
,
({
m
.
gen
=
u
.
gen
;
m
.
data_type
=
u
.
data_type
;
m
.
dirty_sectors
=
u
.
dirty_sectors
;
m
.
cached_sectors
=
u
.
cached_sectors
;
}));
g
->
io_time
[
READ
]
=
u
.
read_time
;
g
->
io_time
[
WRITE
]
=
u
.
write_time
;
g
->
oldest_gen
=
u
.
oldest_gen
;
g
->
gen_valid
=
1
;
if
(
old
.
cached_sectors
)
{
update_cached_sectors
(
c
,
fs_usage
,
ca
->
dev_idx
,
-
old
.
cached_sectors
);
trace_invalidate
(
ca
,
bucket_to_sector
(
ca
,
k
.
k
->
p
.
offset
),
old
.
cached_sectors
);
}
return
0
;
}
#define checked_add(a, b) \
#define checked_add(a, b) \
do { \
do { \
unsigned _res = (unsigned) (a) + (b); \
unsigned _res = (unsigned) (a) + (b); \
...
@@ -840,6 +894,9 @@ static int __bch2_mark_key(struct bch_fs *c, struct bkey_s_c k,
...
@@ -840,6 +894,9 @@ static int __bch2_mark_key(struct bch_fs *c, struct bkey_s_c k,
fs_usage
=
this_cpu_ptr
(
c
->
usage
[
gc
]);
fs_usage
=
this_cpu_ptr
(
c
->
usage
[
gc
]);
switch
(
k
.
k
->
type
)
{
switch
(
k
.
k
->
type
)
{
case
KEY_TYPE_alloc
:
return
bch2_mark_alloc
(
c
,
k
,
inserting
,
fs_usage
,
journal_seq
,
flags
,
gc
);
case
KEY_TYPE_btree_ptr
:
case
KEY_TYPE_btree_ptr
:
return
bch2_mark_extent
(
c
,
k
,
inserting
return
bch2_mark_extent
(
c
,
k
,
inserting
?
c
->
opts
.
btree_node_size
?
c
->
opts
.
btree_node_size
...
@@ -922,7 +979,7 @@ void bch2_mark_update(struct btree_insert *trans,
...
@@ -922,7 +979,7 @@ void bch2_mark_update(struct btree_insert *trans,
preempt_disable
();
preempt_disable
();
fs_usage
=
bch2_fs_usage_get_scratch
(
c
);
fs_usage
=
bch2_fs_usage_get_scratch
(
c
);
if
(
!
(
trans
->
flags
&
BTREE_INSERT_
JOURNAL_REPLAY
))
if
(
!
(
trans
->
flags
&
BTREE_INSERT_
NOMARK
))
bch2_mark_key_locked
(
c
,
bkey_i_to_s_c
(
insert
->
k
),
true
,
bch2_mark_key_locked
(
c
,
bkey_i_to_s_c
(
insert
->
k
),
true
,
bpos_min
(
insert
->
k
->
k
.
p
,
b
->
key
.
k
.
p
).
offset
-
bpos_min
(
insert
->
k
->
k
.
p
,
b
->
key
.
k
.
p
).
offset
-
bkey_start_offset
(
&
insert
->
k
->
k
),
bkey_start_offset
(
&
insert
->
k
->
k
),
...
...
fs/bcachefs/buckets.h
View file @
8fe826f9
...
@@ -17,13 +17,14 @@
...
@@ -17,13 +17,14 @@
#define bucket_cmpxchg(g, new, expr) \
#define bucket_cmpxchg(g, new, expr) \
({ \
({ \
struct bucket *_g = g; \
u64 _v = atomic64_read(&(g)->_mark.v); \
u64 _v = atomic64_read(&(g)->_mark.v); \
struct bucket_mark _old; \
struct bucket_mark _old; \
\
\
do { \
do { \
(new).v.counter = _old.v.counter = _v; \
(new).v.counter = _old.v.counter = _v; \
expr; \
expr; \
} while ((_v = atomic64_cmpxchg(&(g)->_mark.v, \
} while ((_v = atomic64_cmpxchg(&(
_
g)->_mark.v, \
_old.v.counter, \
_old.v.counter, \
(new).v.counter)) != _old.v.counter);\
(new).v.counter)) != _old.v.counter);\
_old; \
_old; \
...
...
fs/bcachefs/fifo.h
View file @
8fe826f9
...
@@ -101,7 +101,7 @@ do { \
...
@@ -101,7 +101,7 @@ do { \
({ \
({ \
bool _r = !fifo_empty((fifo)); \
bool _r = !fifo_empty((fifo)); \
if (_r) \
if (_r) \
(i) = (fifo)->data[--(fifo)->back & (fifo)->mask] \
(i) = (fifo)->data[--(fifo)->back & (fifo)->mask]
;
\
_r; \
_r; \
})
})
...
...
fs/bcachefs/journal_io.c
View file @
8fe826f9
...
@@ -854,7 +854,8 @@ int bch2_journal_replay(struct bch_fs *c, struct list_head *list)
...
@@ -854,7 +854,8 @@ int bch2_journal_replay(struct bch_fs *c, struct list_head *list)
ret
=
bch2_btree_insert
(
c
,
entry
->
btree_id
,
k
,
ret
=
bch2_btree_insert
(
c
,
entry
->
btree_id
,
k
,
&
disk_res
,
NULL
,
&
disk_res
,
NULL
,
BTREE_INSERT_NOFAIL
|
BTREE_INSERT_NOFAIL
|
BTREE_INSERT_JOURNAL_REPLAY
);
BTREE_INSERT_JOURNAL_REPLAY
|
BTREE_INSERT_NOMARK
);
}
}
if
(
ret
)
{
if
(
ret
)
{
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment