Commit 7394e31f authored by Shaohua Li's avatar Shaohua Li Committed by Jens Axboe

blk-throttle: make bandwidth change smooth

When cgroups all reach low limit, cgroups can dispatch more IO. This
could make some cgroups dispatch more IO but others not, and even some
cgroups could dispatch less IO than their low limit. For example, cg1
low limit 10MB/s, cg2 limit 80MB/s, assume disk maximum bandwidth is
120M/s for the workload. Their bps could something like this:

cg1/cg2 bps: T1: 10/80 -> T2: 60/60 -> T3: 10/80

At T1, all cgroups reach low limit, so they can dispatch more IO later.
Then cg1 dispatch more IO and cg2 has no room to dispatch enough IO. At
T2, cg2 only dispatches 60M/s. Since We detect cg2 dispatches less IO
than its low limit 80M/s, we downgrade the queue from LIMIT_MAX to
LIMIT_LOW, then all cgroups are throttled to their low limit (T3). cg2
will have bandwidth below its low limit at most time.

The big problem here is we don't know the maximum bandwidth of the
workload, so we can't make smart decision to avoid the situation. This
patch makes cgroup bandwidth change smooth. After disk upgrades from
LIMIT_LOW to LIMIT_MAX, we don't allow cgroups use all bandwidth upto
their max limit immediately. Their bandwidth limit will be increased
gradually to avoid above situation. So above example will became
something like:

cg1/cg2 bps: 10/80 -> 15/105 -> 20/100 -> 25/95 -> 30/90 -> 35/85 -> 40/80
-> 45/75 -> 22/98

In this way cgroups bandwidth will be above their limit in majority
time, this still doesn't fully utilize disk bandwidth, but that's
something we pay for sharing.

Scale up is linear. The limit scales up 1/2 .low limit every
throtl_slice after upgrade. The scale up will stop if the adjusted limit
hits .max limit. Scale down is exponential. We cut the scale value half
if a cgroup doesn't hit its .low limit. If the scale becomes 0, we then
fully downgrade the queue to LIMIT_LOW state.

Note this doesn't completely avoid cgroup running under its low limit.
The best way to guarantee cgroup doesn't run under its limit is to set
max limit. For example, if we set cg1 max limit to 40, cg2 will never
run under its low limit.
Signed-off-by: default avatarShaohua Li <shli@fb.com>
Signed-off-by: default avatarJens Axboe <axboe@fb.com>
parent aec24246
...@@ -175,6 +175,8 @@ struct throtl_data ...@@ -175,6 +175,8 @@ struct throtl_data
unsigned long low_upgrade_time; unsigned long low_upgrade_time;
unsigned long low_downgrade_time; unsigned long low_downgrade_time;
unsigned int scale;
}; };
static void throtl_pending_timer_fn(unsigned long arg); static void throtl_pending_timer_fn(unsigned long arg);
...@@ -226,29 +228,70 @@ static struct throtl_data *sq_to_td(struct throtl_service_queue *sq) ...@@ -226,29 +228,70 @@ static struct throtl_data *sq_to_td(struct throtl_service_queue *sq)
return container_of(sq, struct throtl_data, service_queue); return container_of(sq, struct throtl_data, service_queue);
} }
/*
* cgroup's limit in LIMIT_MAX is scaled if low limit is set. This scale is to
* make the IO dispatch more smooth.
* Scale up: linearly scale up according to lapsed time since upgrade. For
* every throtl_slice, the limit scales up 1/2 .low limit till the
* limit hits .max limit
* Scale down: exponentially scale down if a cgroup doesn't hit its .low limit
*/
static uint64_t throtl_adjusted_limit(uint64_t low, struct throtl_data *td)
{
/* arbitrary value to avoid too big scale */
if (td->scale < 4096 && time_after_eq(jiffies,
td->low_upgrade_time + td->scale * td->throtl_slice))
td->scale = (jiffies - td->low_upgrade_time) / td->throtl_slice;
return low + (low >> 1) * td->scale;
}
static uint64_t tg_bps_limit(struct throtl_grp *tg, int rw) static uint64_t tg_bps_limit(struct throtl_grp *tg, int rw)
{ {
struct blkcg_gq *blkg = tg_to_blkg(tg); struct blkcg_gq *blkg = tg_to_blkg(tg);
struct throtl_data *td;
uint64_t ret; uint64_t ret;
if (cgroup_subsys_on_dfl(io_cgrp_subsys) && !blkg->parent) if (cgroup_subsys_on_dfl(io_cgrp_subsys) && !blkg->parent)
return U64_MAX; return U64_MAX;
ret = tg->bps[rw][tg->td->limit_index];
if (ret == 0 && tg->td->limit_index == LIMIT_LOW) td = tg->td;
ret = tg->bps[rw][td->limit_index];
if (ret == 0 && td->limit_index == LIMIT_LOW)
return tg->bps[rw][LIMIT_MAX]; return tg->bps[rw][LIMIT_MAX];
if (td->limit_index == LIMIT_MAX && tg->bps[rw][LIMIT_LOW] &&
tg->bps[rw][LIMIT_LOW] != tg->bps[rw][LIMIT_MAX]) {
uint64_t adjusted;
adjusted = throtl_adjusted_limit(tg->bps[rw][LIMIT_LOW], td);
ret = min(tg->bps[rw][LIMIT_MAX], adjusted);
}
return ret; return ret;
} }
static unsigned int tg_iops_limit(struct throtl_grp *tg, int rw) static unsigned int tg_iops_limit(struct throtl_grp *tg, int rw)
{ {
struct blkcg_gq *blkg = tg_to_blkg(tg); struct blkcg_gq *blkg = tg_to_blkg(tg);
struct throtl_data *td;
unsigned int ret; unsigned int ret;
if (cgroup_subsys_on_dfl(io_cgrp_subsys) && !blkg->parent) if (cgroup_subsys_on_dfl(io_cgrp_subsys) && !blkg->parent)
return UINT_MAX; return UINT_MAX;
ret = tg->iops[rw][tg->td->limit_index]; td = tg->td;
ret = tg->iops[rw][td->limit_index];
if (ret == 0 && tg->td->limit_index == LIMIT_LOW) if (ret == 0 && tg->td->limit_index == LIMIT_LOW)
return tg->iops[rw][LIMIT_MAX]; return tg->iops[rw][LIMIT_MAX];
if (td->limit_index == LIMIT_MAX && tg->iops[rw][LIMIT_LOW] &&
tg->iops[rw][LIMIT_LOW] != tg->iops[rw][LIMIT_MAX]) {
uint64_t adjusted;
adjusted = throtl_adjusted_limit(tg->iops[rw][LIMIT_LOW], td);
if (adjusted > UINT_MAX)
adjusted = UINT_MAX;
ret = min_t(unsigned int, tg->iops[rw][LIMIT_MAX], adjusted);
}
return ret; return ret;
} }
...@@ -1677,6 +1720,7 @@ static void throtl_upgrade_state(struct throtl_data *td) ...@@ -1677,6 +1720,7 @@ static void throtl_upgrade_state(struct throtl_data *td)
td->limit_index = LIMIT_MAX; td->limit_index = LIMIT_MAX;
td->low_upgrade_time = jiffies; td->low_upgrade_time = jiffies;
td->scale = 0;
rcu_read_lock(); rcu_read_lock();
blkg_for_each_descendant_post(blkg, pos_css, td->queue->root_blkg) { blkg_for_each_descendant_post(blkg, pos_css, td->queue->root_blkg) {
struct throtl_grp *tg = blkg_to_tg(blkg); struct throtl_grp *tg = blkg_to_tg(blkg);
...@@ -1694,6 +1738,13 @@ static void throtl_upgrade_state(struct throtl_data *td) ...@@ -1694,6 +1738,13 @@ static void throtl_upgrade_state(struct throtl_data *td)
static void throtl_downgrade_state(struct throtl_data *td, int new) static void throtl_downgrade_state(struct throtl_data *td, int new)
{ {
td->scale /= 2;
if (td->scale) {
td->low_upgrade_time = jiffies - td->scale * td->throtl_slice;
return;
}
td->limit_index = new; td->limit_index = new;
td->low_downgrade_time = jiffies; td->low_downgrade_time = jiffies;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment