Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
5663a415
Commit
5663a415
authored
Nov 27, 2018
by
Kent Overstreet
Committed by
Kent Overstreet
Oct 22, 2023
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
bcachefs: refactor bch_fs_usage
Signed-off-by:
Kent Overstreet
<
kent.overstreet@linux.dev
>
parent
73e6ab95
Changes
8
Hide whitespace changes
Inline
Side-by-side
Showing
8 changed files
with
119 additions
and
92 deletions
+119
-92
fs/bcachefs/alloc_foreground.c
fs/bcachefs/alloc_foreground.c
+1
-1
fs/bcachefs/bcachefs.h
fs/bcachefs/bcachefs.h
+8
-2
fs/bcachefs/buckets.c
fs/bcachefs/buckets.c
+70
-61
fs/bcachefs/buckets.h
fs/bcachefs/buckets.h
+25
-17
fs/bcachefs/buckets_types.h
fs/bcachefs/buckets_types.h
+9
-5
fs/bcachefs/chardev.c
fs/bcachefs/chardev.c
+1
-1
fs/bcachefs/fs.c
fs/bcachefs/fs.c
+3
-5
fs/bcachefs/super.c
fs/bcachefs/super.c
+2
-0
No files found.
fs/bcachefs/alloc_foreground.c
View file @
5663a415
...
...
@@ -721,7 +721,7 @@ static struct write_point *__writepoint_find(struct hlist_head *head,
static
inline
bool
too_many_writepoints
(
struct
bch_fs
*
c
,
unsigned
factor
)
{
u64
stranded
=
c
->
write_points_nr
*
c
->
bucket_size_max
;
u64
free
=
bch2_fs_sectors_free
(
c
,
bch2_fs_usage_read
(
c
)
);
u64
free
=
bch2_fs_sectors_free
(
c
);
return
stranded
*
factor
>
free
;
}
...
...
fs/bcachefs/bcachefs.h
View file @
5663a415
...
...
@@ -503,6 +503,10 @@ enum bch_fs_state {
BCH_FS_RW
,
};
struct
bch_fs_pcpu
{
u64
sectors_available
;
};
struct
bch_fs
{
struct
closure
cl
;
...
...
@@ -615,9 +619,11 @@ struct bch_fs {
atomic64_t
sectors_available
;
struct
bch_fs_usage
__percpu
*
usage
[
2
];
struct
bch_fs_pcpu
__percpu
*
pcpu
;
struct
bch_fs_usage
__percpu
*
usage
[
2
];
struct
percpu_rw_semaphore
mark_lock
;
struct
percpu_rw_semaphore
mark_lock
;
/*
* When we invalidate buckets, we use both the priority and the amount
...
...
fs/bcachefs/buckets.c
View file @
5663a415
...
...
@@ -272,16 +272,31 @@ static u64 avail_factor(u64 r)
return
(
r
<<
RESERVE_FACTOR
)
/
((
1
<<
RESERVE_FACTOR
)
+
1
);
}
static
inline
u64
__bch2_fs_sectors_used
(
struct
bch_fs
*
c
,
struct
bch_fs_usage
stats
)
static
inline
u64
__bch2_fs_sectors_used
(
struct
bch_fs
*
c
,
struct
bch_fs_usage
fs_usage
)
{
struct
fs_usage_sum
sum
=
__fs_usage_sum
(
stats
);
struct
fs_usage_sum
sum
=
__fs_usage_sum
(
fs_usage
);
return
sum
.
hidden
+
sum
.
data
+
reserve_factor
(
sum
.
reserved
);
}
u64
bch2_fs_sectors_used
(
struct
bch_fs
*
c
,
struct
bch_fs_usage
stats
)
u64
bch2_fs_sectors_used
(
struct
bch_fs
*
c
,
struct
bch_fs_usage
fs_usage
)
{
return
min
(
c
->
capacity
,
__bch2_fs_sectors_used
(
c
,
stats
));
return
min
(
c
->
capacity
,
__bch2_fs_sectors_used
(
c
,
fs_usage
));
}
struct
bch_fs_usage_short
bch2_fs_usage_read_short
(
struct
bch_fs
*
c
)
{
struct
bch_fs_usage
usage
=
bch2_fs_usage_read
(
c
);
struct
fs_usage_sum
sum
=
__fs_usage_sum
(
usage
);
struct
bch_fs_usage_short
ret
;
ret
.
capacity
=
READ_ONCE
(
c
->
capacity
)
-
sum
.
hidden
;
ret
.
used
=
min
(
ret
.
capacity
,
sum
.
data
+
reserve_factor
(
sum
.
reserved
));
ret
.
nr_inodes
=
usage
.
nr_inodes
;
return
ret
;
}
static
inline
int
is_unavailable_bucket
(
struct
bucket_mark
m
)
...
...
@@ -315,11 +330,11 @@ static bool bucket_became_unavailable(struct bucket_mark old,
}
void
bch2_fs_usage_apply
(
struct
bch_fs
*
c
,
struct
bch_fs_usage
*
stats
,
struct
bch_fs_usage
*
fs_usage
,
struct
disk_reservation
*
disk_res
,
struct
gc_pos
gc_pos
)
{
struct
fs_usage_sum
sum
=
__fs_usage_sum
(
*
stats
);
struct
fs_usage_sum
sum
=
__fs_usage_sum
(
*
fs_usage
);
s64
added
=
sum
.
data
+
sum
.
reserved
;
s64
should_not_have_added
;
...
...
@@ -337,24 +352,20 @@ void bch2_fs_usage_apply(struct bch_fs *c,
}
if
(
added
>
0
)
{
disk_res
->
sectors
-=
added
;
stats
->
online_reserved
-=
added
;
disk_res
->
sectors
-=
added
;
fs_usage
->
online_reserved
-=
added
;
}
preempt_disable
();
/* online_reserved not subject to gc: */
this_cpu_add
(
c
->
usage
[
0
]
->
online_reserved
,
stats
->
online_reserved
);
stats
->
online_reserved
=
0
;
bch2_usage_add
(
this_cpu_ptr
(
c
->
usage
[
0
]),
stats
);
bch2_usage_add
(
this_cpu_ptr
(
c
->
usage
[
0
]),
fs_usage
);
if
(
gc_visited
(
c
,
gc_pos
))
bch2_usage_add
(
this_cpu_ptr
(
c
->
usage
[
1
]),
stats
);
bch2_usage_add
(
this_cpu_ptr
(
c
->
usage
[
1
]),
fs_usage
);
bch2_fs_stats_verify
(
c
);
preempt_enable
();
memset
(
stats
,
0
,
sizeof
(
*
stats
));
memset
(
fs_usage
,
0
,
sizeof
(
*
fs_usage
));
}
static
void
bch2_dev_usage_update
(
struct
bch_fs
*
c
,
struct
bch_dev
*
ca
,
...
...
@@ -435,11 +446,11 @@ static void __bch2_invalidate_bucket(struct bch_fs *c, struct bch_dev *ca,
size_t
b
,
struct
bucket_mark
*
old
,
bool
gc
)
{
struct
bch_fs_usage
*
stats
=
this_cpu_ptr
(
c
->
usage
[
gc
]);
struct
bch_fs_usage
*
fs_usage
=
this_cpu_ptr
(
c
->
usage
[
gc
]);
struct
bucket
*
g
=
__bucket
(
ca
,
b
,
gc
);
struct
bucket_mark
new
;
*
old
=
bucket_data_cmpxchg
(
c
,
ca
,
stats
,
g
,
new
,
({
*
old
=
bucket_data_cmpxchg
(
c
,
ca
,
fs_usage
,
g
,
new
,
({
BUG_ON
(
!
is_available_bucket
(
new
));
new
.
owned_by_allocator
=
1
;
...
...
@@ -449,7 +460,7 @@ static void __bch2_invalidate_bucket(struct bch_fs *c, struct bch_dev *ca,
new
.
gen
++
;
}));
stats
->
replicas
[
0
].
data
[
BCH_DATA_CACHED
]
-=
old
->
cached_sectors
;
fs_usage
->
replicas
[
0
].
data
[
BCH_DATA_CACHED
]
-=
old
->
cached_sectors
;
}
void
bch2_invalidate_bucket
(
struct
bch_fs
*
c
,
struct
bch_dev
*
ca
,
...
...
@@ -468,11 +479,11 @@ static void __bch2_mark_alloc_bucket(struct bch_fs *c, struct bch_dev *ca,
size_t
b
,
bool
owned_by_allocator
,
bool
gc
)
{
struct
bch_fs_usage
*
stats
=
this_cpu_ptr
(
c
->
usage
[
gc
]);
struct
bch_fs_usage
*
fs_usage
=
this_cpu_ptr
(
c
->
usage
[
gc
]);
struct
bucket
*
g
=
__bucket
(
ca
,
b
,
gc
);
struct
bucket_mark
old
,
new
;
old
=
bucket_data_cmpxchg
(
c
,
ca
,
stats
,
g
,
new
,
({
old
=
bucket_data_cmpxchg
(
c
,
ca
,
fs_usage
,
g
,
new
,
({
new
.
owned_by_allocator
=
owned_by_allocator
;
}));
...
...
@@ -588,7 +599,7 @@ static void bch2_mark_pointer(struct bch_fs *c,
struct
extent_ptr_decoded
p
,
s64
sectors
,
enum
bch_data_type
data_type
,
struct
bch_fs_usage
*
fs_usage
,
u
64
journal_seq
,
unsigned
flags
,
u
nsigned
journal_seq
,
unsigned
flags
,
bool
gc
)
{
struct
bucket_mark
old
,
new
;
...
...
@@ -693,8 +704,8 @@ static int bch2_mark_stripe_ptr(struct bch_fs *c,
static
int
bch2_mark_extent
(
struct
bch_fs
*
c
,
struct
bkey_s_c
k
,
s64
sectors
,
enum
bch_data_type
data_type
,
struct
bch_fs_usage
*
stats
,
u
64
journal_seq
,
unsigned
flags
,
struct
bch_fs_usage
*
fs_usage
,
u
nsigned
journal_seq
,
unsigned
flags
,
bool
gc
)
{
struct
bkey_ptrs_c
ptrs
=
bch2_bkey_ptrs_c
(
k
);
...
...
@@ -717,7 +728,7 @@ static int bch2_mark_extent(struct bch_fs *c, struct bkey_s_c k,
s64
adjusted_disk_sectors
=
disk_sectors
;
bch2_mark_pointer
(
c
,
p
,
disk_sectors
,
data_type
,
stats
,
journal_seq
,
flags
,
gc
);
fs_usage
,
journal_seq
,
flags
,
gc
);
if
(
!
p
.
ptr
.
cached
)
for
(
i
=
0
;
i
<
p
.
ec_nr
;
i
++
)
{
...
...
@@ -740,13 +751,13 @@ static int bch2_mark_extent(struct bch_fs *c, struct bkey_s_c k,
}
replicas
=
clamp_t
(
unsigned
,
replicas
,
1
,
ARRAY_SIZE
(
stats
->
replicas
));
1
,
ARRAY_SIZE
(
fs_usage
->
replicas
));
ec_redundancy
=
clamp_t
(
unsigned
,
ec_redundancy
,
1
,
ARRAY_SIZE
(
stats
->
replicas
));
1
,
ARRAY_SIZE
(
fs_usage
->
replicas
));
stats
->
replicas
[
0
].
data
[
BCH_DATA_CACHED
]
+=
cached_sectors
;
stats
->
replicas
[
replicas
-
1
].
data
[
data_type
]
+=
dirty_sectors
;
stats
->
replicas
[
ec_redundancy
-
1
].
ec_data
+=
ec_sectors
;
fs_usage
->
replicas
[
0
].
data
[
BCH_DATA_CACHED
]
+=
cached_sectors
;
fs_usage
->
replicas
[
replicas
-
1
].
data
[
data_type
]
+=
dirty_sectors
;
fs_usage
->
replicas
[
ec_redundancy
-
1
].
ec_data
+=
ec_sectors
;
return
0
;
}
...
...
@@ -831,8 +842,8 @@ static int bch2_mark_stripe(struct bch_fs *c, struct bkey_s_c k,
static
int
__bch2_mark_key
(
struct
bch_fs
*
c
,
struct
bkey_s_c
k
,
bool
inserting
,
s64
sectors
,
struct
bch_fs_usage
*
stats
,
u
64
journal_seq
,
unsigned
flags
,
struct
bch_fs_usage
*
fs_usage
,
u
nsigned
journal_seq
,
unsigned
flags
,
bool
gc
)
{
int
ret
=
0
;
...
...
@@ -843,30 +854,30 @@ static int __bch2_mark_key(struct bch_fs *c, struct bkey_s_c k,
?
c
->
opts
.
btree_node_size
:
-
c
->
opts
.
btree_node_size
,
BCH_DATA_BTREE
,
stats
,
journal_seq
,
flags
,
gc
);
fs_usage
,
journal_seq
,
flags
,
gc
);
break
;
case
KEY_TYPE_extent
:
ret
=
bch2_mark_extent
(
c
,
k
,
sectors
,
BCH_DATA_USER
,
stats
,
journal_seq
,
flags
,
gc
);
fs_usage
,
journal_seq
,
flags
,
gc
);
break
;
case
KEY_TYPE_stripe
:
ret
=
bch2_mark_stripe
(
c
,
k
,
inserting
,
stats
,
journal_seq
,
flags
,
gc
);
fs_usage
,
journal_seq
,
flags
,
gc
);
break
;
case
KEY_TYPE_alloc
:
if
(
inserting
)
stats
->
nr_inodes
++
;
fs_usage
->
nr_inodes
++
;
else
stats
->
nr_inodes
--
;
fs_usage
->
nr_inodes
--
;
break
;
case
KEY_TYPE_reservation
:
{
unsigned
replicas
=
bkey_s_c_to_reservation
(
k
).
v
->
nr_replicas
;
sectors
*=
replicas
;
replicas
=
clamp_t
(
unsigned
,
replicas
,
1
,
ARRAY_SIZE
(
stats
->
replicas
));
1
,
ARRAY_SIZE
(
fs_usage
->
replicas
));
stats
->
replicas
[
replicas
-
1
].
persistent_reserved
+=
sectors
;
fs_usage
->
replicas
[
replicas
-
1
].
persistent_reserved
+=
sectors
;
break
;
}
default:
...
...
@@ -880,17 +891,15 @@ int bch2_mark_key_locked(struct bch_fs *c,
struct
bkey_s_c
k
,
bool
inserting
,
s64
sectors
,
struct
gc_pos
pos
,
struct
bch_fs_usage
*
stats
,
struct
bch_fs_usage
*
fs_usage
,
u64
journal_seq
,
unsigned
flags
)
{
int
ret
;
if
(
!
(
flags
&
BCH_BUCKET_MARK_GC
))
{
if
(
!
stats
)
stats
=
this_cpu_ptr
(
c
->
usage
[
0
]);
ret
=
__bch2_mark_key
(
c
,
k
,
inserting
,
sectors
,
stats
,
journal_seq
,
flags
,
false
);
fs_usage
?:
this_cpu_ptr
(
c
->
usage
[
0
]),
journal_seq
,
flags
,
false
);
if
(
ret
)
return
ret
;
}
...
...
@@ -910,14 +919,14 @@ int bch2_mark_key_locked(struct bch_fs *c,
int
bch2_mark_key
(
struct
bch_fs
*
c
,
struct
bkey_s_c
k
,
bool
inserting
,
s64
sectors
,
struct
gc_pos
pos
,
struct
bch_fs_usage
*
stats
,
struct
bch_fs_usage
*
fs_usage
,
u64
journal_seq
,
unsigned
flags
)
{
int
ret
;
percpu_down_read
(
&
c
->
mark_lock
);
ret
=
bch2_mark_key_locked
(
c
,
k
,
inserting
,
sectors
,
pos
,
stats
,
journal_seq
,
flags
);
pos
,
fs_usage
,
journal_seq
,
flags
);
percpu_up_read
(
&
c
->
mark_lock
);
return
ret
;
...
...
@@ -930,7 +939,7 @@ void bch2_mark_update(struct btree_insert *trans,
struct
btree_iter
*
iter
=
insert
->
iter
;
struct
btree
*
b
=
iter
->
l
[
0
].
b
;
struct
btree_node_iter
node_iter
=
iter
->
l
[
0
].
iter
;
struct
bch_fs_usage
stats
=
{
0
};
struct
bch_fs_usage
fs_usage
=
{
0
};
struct
gc_pos
pos
=
gc_pos_btree_node
(
b
);
struct
bkey_packed
*
_k
;
...
...
@@ -943,7 +952,7 @@ void bch2_mark_update(struct btree_insert *trans,
bch2_mark_key_locked
(
c
,
bkey_i_to_s_c
(
insert
->
k
),
true
,
bpos_min
(
insert
->
k
->
k
.
p
,
b
->
key
.
k
.
p
).
offset
-
bkey_start_offset
(
&
insert
->
k
->
k
),
pos
,
&
stats
,
trans
->
journal_res
.
seq
,
0
);
pos
,
&
fs_usage
,
trans
->
journal_res
.
seq
,
0
);
while
((
_k
=
bch2_btree_node_iter_peek_filter
(
&
node_iter
,
b
,
KEY_TYPE_discard
)))
{
...
...
@@ -976,7 +985,7 @@ void bch2_mark_update(struct btree_insert *trans,
BUG_ON
(
sectors
<=
0
);
bch2_mark_key_locked
(
c
,
k
,
true
,
sectors
,
pos
,
&
stats
,
trans
->
journal_res
.
seq
,
0
);
pos
,
&
fs_usage
,
trans
->
journal_res
.
seq
,
0
);
sectors
=
bkey_start_offset
(
&
insert
->
k
->
k
)
-
k
.
k
->
p
.
offset
;
...
...
@@ -987,12 +996,12 @@ void bch2_mark_update(struct btree_insert *trans,
}
bch2_mark_key_locked
(
c
,
k
,
false
,
sectors
,
pos
,
&
stats
,
trans
->
journal_res
.
seq
,
0
);
pos
,
&
fs_usage
,
trans
->
journal_res
.
seq
,
0
);
bch2_btree_node_iter_advance
(
&
node_iter
,
b
);
}
bch2_fs_usage_apply
(
c
,
&
stats
,
trans
->
disk_res
,
pos
);
bch2_fs_usage_apply
(
c
,
&
fs_usage
,
trans
->
disk_res
,
pos
);
percpu_up_read
(
&
c
->
mark_lock
);
}
...
...
@@ -1004,9 +1013,9 @@ static u64 bch2_recalc_sectors_available(struct bch_fs *c)
int
cpu
;
for_each_possible_cpu
(
cpu
)
per_cpu_ptr
(
c
->
usage
[
0
],
cpu
)
->
available_cach
e
=
0
;
per_cpu_ptr
(
c
->
pcpu
,
cpu
)
->
sectors_availabl
e
=
0
;
return
avail_factor
(
bch2_fs_sectors_free
(
c
,
bch2_fs_usage_read
(
c
)
));
return
avail_factor
(
bch2_fs_sectors_free
(
c
));
}
void
__bch2_disk_reservation_put
(
struct
bch_fs
*
c
,
struct
disk_reservation
*
res
)
...
...
@@ -1026,16 +1035,16 @@ void __bch2_disk_reservation_put(struct bch_fs *c, struct disk_reservation *res)
int
bch2_disk_reservation_add
(
struct
bch_fs
*
c
,
struct
disk_reservation
*
res
,
unsigned
sectors
,
int
flags
)
{
struct
bch_fs_
usage
*
stats
;
struct
bch_fs_
pcpu
*
pcpu
;
u64
old
,
v
,
get
;
s64
sectors_available
;
int
ret
;
percpu_down_read
(
&
c
->
mark_lock
);
preempt_disable
();
stats
=
this_cpu_ptr
(
c
->
usage
[
0
]
);
pcpu
=
this_cpu_ptr
(
c
->
pcpu
);
if
(
sectors
<=
stats
->
available_cach
e
)
if
(
sectors
<=
pcpu
->
sectors_availabl
e
)
goto
out
;
v
=
atomic64_read
(
&
c
->
sectors_available
);
...
...
@@ -1051,12 +1060,12 @@ int bch2_disk_reservation_add(struct bch_fs *c, struct disk_reservation *res,
}
while
((
v
=
atomic64_cmpxchg
(
&
c
->
sectors_available
,
old
,
old
-
get
))
!=
old
);
stats
->
available_cache
+=
get
;
pcpu
->
sectors_available
+=
get
;
out:
stats
->
available_cache
-=
sectors
;
stats
->
online_reserved
+=
sectors
;
res
->
sectors
+=
sectors
;
pcpu
->
sectors_available
-=
sectors
;
this_cpu_add
(
c
->
usage
[
0
]
->
online_reserved
,
sectors
)
;
res
->
sectors
+=
sectors
;
bch2_disk_reservations_verify
(
c
,
flags
);
bch2_fs_stats_verify
(
c
);
...
...
@@ -1089,8 +1098,8 @@ int bch2_disk_reservation_add(struct bch_fs *c, struct disk_reservation *res,
(
flags
&
BCH_DISK_RESERVATION_NOFAIL
))
{
atomic64_set
(
&
c
->
sectors_available
,
max_t
(
s64
,
0
,
sectors_available
-
sectors
));
stats
->
online_reserved
+=
sectors
;
res
->
sectors
+=
sectors
;
this_cpu_add
(
c
->
usage
[
0
]
->
online_reserved
,
sectors
)
;
res
->
sectors
+=
sectors
;
ret
=
0
;
bch2_disk_reservations_verify
(
c
,
flags
);
...
...
fs/bcachefs/buckets.h
View file @
5663a415
...
...
@@ -164,6 +164,20 @@ static inline bool bucket_unused(struct bucket_mark mark)
!
bucket_sectors_used
(
mark
);
}
static
inline
bool
is_available_bucket
(
struct
bucket_mark
mark
)
{
return
(
!
mark
.
owned_by_allocator
&&
!
mark
.
dirty_sectors
&&
!
mark
.
stripe
);
}
static
inline
bool
bucket_needs_journal_commit
(
struct
bucket_mark
m
,
u16
last_seq_ondisk
)
{
return
m
.
journal_seq_valid
&&
((
s16
)
m
.
journal_seq
-
(
s16
)
last_seq_ondisk
>
0
);
}
/* Device usage: */
struct
bch_dev_usage
__bch2_dev_usage_read
(
struct
bch_dev
*
,
bool
);
...
...
@@ -207,31 +221,21 @@ static inline u64 dev_buckets_free(struct bch_fs *c, struct bch_dev *ca)
struct
bch_fs_usage
__bch2_fs_usage_read
(
struct
bch_fs
*
,
bool
);
struct
bch_fs_usage
bch2_fs_usage_read
(
struct
bch_fs
*
);
void
bch2_fs_usage_apply
(
struct
bch_fs
*
,
struct
bch_fs_usage
*
,
struct
disk_reservation
*
,
struct
gc_pos
);
u64
bch2_fs_sectors_used
(
struct
bch_fs
*
,
struct
bch_fs_usage
);
static
inline
u64
bch2_fs_sectors_free
(
struct
bch_fs
*
c
,
struct
bch_fs_usage
stats
)
{
return
c
->
capacity
-
bch2_fs_sectors_used
(
c
,
stats
);
}
struct
bch_fs_usage_short
bch2_fs_usage_read_short
(
struct
bch_fs
*
);
static
inline
bool
is_available_bucket
(
struct
bucket_mark
mark
)
static
inline
u64
bch2_fs_sectors_free
(
struct
bch_fs
*
c
)
{
return
(
!
mark
.
owned_by_allocator
&&
!
mark
.
dirty_sectors
&&
!
mark
.
stripe
);
}
struct
bch_fs_usage_short
usage
=
bch2_fs_usage_read_short
(
c
);
static
inline
bool
bucket_needs_journal_commit
(
struct
bucket_mark
m
,
u16
last_seq_ondisk
)
{
return
m
.
journal_seq_valid
&&
((
s16
)
m
.
journal_seq
-
(
s16
)
last_seq_ondisk
>
0
);
return
usage
.
capacity
-
usage
.
used
;
}
/* key/bucket marking: */
void
bch2_bucket_seq_cleanup
(
struct
bch_fs
*
);
void
bch2_invalidate_bucket
(
struct
bch_fs
*
,
struct
bch_dev
*
,
...
...
@@ -252,6 +256,10 @@ int bch2_mark_key(struct bch_fs *, struct bkey_s_c,
bool
,
s64
,
struct
gc_pos
,
struct
bch_fs_usage
*
,
u64
,
unsigned
);
void
bch2_mark_update
(
struct
btree_insert
*
,
struct
btree_insert_entry
*
);
void
bch2_fs_usage_apply
(
struct
bch_fs
*
,
struct
bch_fs_usage
*
,
struct
disk_reservation
*
,
struct
gc_pos
);
/* disk reservations: */
void
__bch2_disk_reservation_put
(
struct
bch_fs
*
,
struct
disk_reservation
*
);
...
...
fs/bcachefs/buckets_types.h
View file @
5663a415
...
...
@@ -73,18 +73,22 @@ struct bch_fs_usage {
u64
nr_inodes
;
/* fields starting here aren't touched by gc: */
u64
online_reserved
;
u64
available_cache
;
};
struct
bch_fs_usage_short
{
u64
capacity
;
u64
used
;
u64
nr_inodes
;
};
/*
* A reservation for space on disk:
*/
struct
disk_reservation
{
u64
sectors
;
u32
gen
;
unsigned
nr_replicas
;
u64
sectors
;
u32
gen
;
unsigned
nr_replicas
;
};
struct
copygc_heap_entry
{
...
...
fs/bcachefs/chardev.c
View file @
5663a415
...
...
@@ -306,7 +306,7 @@ static ssize_t bch2_data_job_read(struct file *file, char __user *buf,
.
p
.
btree_id
=
ctx
->
stats
.
iter
.
btree_id
,
.
p
.
pos
=
ctx
->
stats
.
iter
.
pos
,
.
p
.
sectors_done
=
atomic64_read
(
&
ctx
->
stats
.
sectors_seen
),
.
p
.
sectors_total
=
bch2_fs_
sectors_used
(
c
,
bch2_fs_usage_read
(
c
))
,
.
p
.
sectors_total
=
bch2_fs_
usage_read_short
(
c
).
used
,
};
if
(
len
<
sizeof
(
e
))
...
...
fs/bcachefs/fs.c
View file @
5663a415
...
...
@@ -1423,16 +1423,14 @@ static int bch2_statfs(struct dentry *dentry, struct kstatfs *buf)
{
struct
super_block
*
sb
=
dentry
->
d_sb
;
struct
bch_fs
*
c
=
sb
->
s_fs_info
;
struct
bch_fs_usage
usage
=
bch2_fs_usage_read
(
c
);
u64
hidden_metadata
=
usage
.
buckets
[
BCH_DATA_SB
]
+
usage
.
buckets
[
BCH_DATA_JOURNAL
];
struct
bch_fs_usage_short
usage
=
bch2_fs_usage_read_short
(
c
);
unsigned
shift
=
sb
->
s_blocksize_bits
-
9
;
u64
fsid
;
buf
->
f_type
=
BCACHEFS_STATFS_MAGIC
;
buf
->
f_bsize
=
sb
->
s_blocksize
;
buf
->
f_blocks
=
(
c
->
capacity
-
hidden_metadata
)
>>
shift
;
buf
->
f_bfree
=
(
c
->
capacity
-
bch2_fs_sectors_used
(
c
,
usage
)
)
>>
shift
;
buf
->
f_blocks
=
usage
.
capacity
>>
shift
;
buf
->
f_bfree
=
(
usage
.
capacity
-
usage
.
used
)
>>
shift
;
buf
->
f_bavail
=
buf
->
f_bfree
;
buf
->
f_files
=
usage
.
nr_inodes
;
buf
->
f_ffree
=
U64_MAX
;
...
...
fs/bcachefs/super.c
View file @
5663a415
...
...
@@ -376,6 +376,7 @@ static void bch2_fs_free(struct bch_fs *c)
bch2_fs_compress_exit
(
c
);
percpu_free_rwsem
(
&
c
->
mark_lock
);
free_percpu
(
c
->
usage
[
0
]);
free_percpu
(
c
->
pcpu
);
mempool_exit
(
&
c
->
btree_iters_pool
);
mempool_exit
(
&
c
->
btree_bounce_pool
);
bioset_exit
(
&
c
->
btree_bio
);
...
...
@@ -612,6 +613,7 @@ static struct bch_fs *bch2_fs_alloc(struct bch_sb *sb, struct bch_opts opts)
offsetof
(
struct
btree_write_bio
,
wbio
.
bio
)),
BIOSET_NEED_BVECS
)
||
!
(
c
->
usage
[
0
]
=
alloc_percpu
(
struct
bch_fs_usage
))
||
!
(
c
->
pcpu
=
alloc_percpu
(
struct
bch_fs_pcpu
))
||
mempool_init_kvpmalloc_pool
(
&
c
->
btree_bounce_pool
,
1
,
btree_bytes
(
c
))
||
mempool_init_kmalloc_pool
(
&
c
->
btree_iters_pool
,
1
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment