Commit daeda1cc authored by Philipp Reisner's avatar Philipp Reisner

drbd: RCU for disk_conf

Signed-off-by: default avatarPhilipp Reisner <philipp.reisner@linbit.com>
Signed-off-by: default avatarLars Ellenberg <lars.ellenberg@linbit.com>
parent 563e4cf2
...@@ -777,7 +777,7 @@ struct drbd_backing_dev { ...@@ -777,7 +777,7 @@ struct drbd_backing_dev {
struct block_device *backing_bdev; struct block_device *backing_bdev;
struct block_device *md_bdev; struct block_device *md_bdev;
struct drbd_md md; struct drbd_md md;
struct disk_conf dc; /* The user provided config... */ struct disk_conf *disk_conf; /* RCU, for updates: mdev->tconn->conf_update */
sector_t known_size; /* last known size of that backing device */ sector_t known_size; /* last known size of that backing device */
}; };
...@@ -1644,8 +1644,13 @@ static inline union drbd_state drbd_read_state(struct drbd_conf *mdev) ...@@ -1644,8 +1644,13 @@ static inline union drbd_state drbd_read_state(struct drbd_conf *mdev)
#define __drbd_chk_io_error(m,f) __drbd_chk_io_error_(m,f, __func__) #define __drbd_chk_io_error(m,f) __drbd_chk_io_error_(m,f, __func__)
static inline void __drbd_chk_io_error_(struct drbd_conf *mdev, int forcedetach, const char *where) static inline void __drbd_chk_io_error_(struct drbd_conf *mdev, int forcedetach, const char *where)
{ {
switch (mdev->ldev->dc.on_io_error) { enum drbd_io_error_p ep;
case EP_PASS_ON:
rcu_read_lock();
ep = rcu_dereference(mdev->ldev->disk_conf)->on_io_error;
rcu_read_unlock();
switch (ep) {
case EP_PASS_ON: /* FIXME would this be better named "Ignore"? */
if (!forcedetach) { if (!forcedetach) {
if (__ratelimit(&drbd_ratelimit_state)) if (__ratelimit(&drbd_ratelimit_state))
dev_err(DEV, "Local IO failed in %s.\n", where); dev_err(DEV, "Local IO failed in %s.\n", where);
...@@ -1694,9 +1699,9 @@ static inline void drbd_chk_io_error_(struct drbd_conf *mdev, ...@@ -1694,9 +1699,9 @@ static inline void drbd_chk_io_error_(struct drbd_conf *mdev,
* BTW, for internal meta data, this happens to be the maximum capacity * BTW, for internal meta data, this happens to be the maximum capacity
* we could agree upon with our peer node. * we could agree upon with our peer node.
*/ */
static inline sector_t drbd_md_first_sector(struct drbd_backing_dev *bdev) static inline sector_t _drbd_md_first_sector(int meta_dev_idx, struct drbd_backing_dev *bdev)
{ {
switch (bdev->dc.meta_dev_idx) { switch (meta_dev_idx) {
case DRBD_MD_INDEX_INTERNAL: case DRBD_MD_INDEX_INTERNAL:
case DRBD_MD_INDEX_FLEX_INT: case DRBD_MD_INDEX_FLEX_INT:
return bdev->md.md_offset + bdev->md.bm_offset; return bdev->md.md_offset + bdev->md.bm_offset;
...@@ -1706,13 +1711,30 @@ static inline sector_t drbd_md_first_sector(struct drbd_backing_dev *bdev) ...@@ -1706,13 +1711,30 @@ static inline sector_t drbd_md_first_sector(struct drbd_backing_dev *bdev)
} }
} }
static inline sector_t drbd_md_first_sector(struct drbd_backing_dev *bdev)
{
int meta_dev_idx;
rcu_read_lock();
meta_dev_idx = rcu_dereference(bdev->disk_conf)->meta_dev_idx;
rcu_read_unlock();
return _drbd_md_first_sector(meta_dev_idx, bdev);
}
/** /**
* drbd_md_last_sector() - Return the last sector number of the meta data area * drbd_md_last_sector() - Return the last sector number of the meta data area
* @bdev: Meta data block device. * @bdev: Meta data block device.
*/ */
static inline sector_t drbd_md_last_sector(struct drbd_backing_dev *bdev) static inline sector_t drbd_md_last_sector(struct drbd_backing_dev *bdev)
{ {
switch (bdev->dc.meta_dev_idx) { int meta_dev_idx;
rcu_read_lock();
meta_dev_idx = rcu_dereference(bdev->disk_conf)->meta_dev_idx;
rcu_read_unlock();
switch (meta_dev_idx) {
case DRBD_MD_INDEX_INTERNAL: case DRBD_MD_INDEX_INTERNAL:
case DRBD_MD_INDEX_FLEX_INT: case DRBD_MD_INDEX_FLEX_INT:
return bdev->md.md_offset + MD_AL_OFFSET - 1; return bdev->md.md_offset + MD_AL_OFFSET - 1;
...@@ -1740,12 +1762,18 @@ static inline sector_t drbd_get_capacity(struct block_device *bdev) ...@@ -1740,12 +1762,18 @@ static inline sector_t drbd_get_capacity(struct block_device *bdev)
static inline sector_t drbd_get_max_capacity(struct drbd_backing_dev *bdev) static inline sector_t drbd_get_max_capacity(struct drbd_backing_dev *bdev)
{ {
sector_t s; sector_t s;
switch (bdev->dc.meta_dev_idx) { int meta_dev_idx;
rcu_read_lock();
meta_dev_idx = rcu_dereference(bdev->disk_conf)->meta_dev_idx;
rcu_read_unlock();
switch (meta_dev_idx) {
case DRBD_MD_INDEX_INTERNAL: case DRBD_MD_INDEX_INTERNAL:
case DRBD_MD_INDEX_FLEX_INT: case DRBD_MD_INDEX_FLEX_INT:
s = drbd_get_capacity(bdev->backing_bdev) s = drbd_get_capacity(bdev->backing_bdev)
? min_t(sector_t, DRBD_MAX_SECTORS_FLEX, ? min_t(sector_t, DRBD_MAX_SECTORS_FLEX,
drbd_md_first_sector(bdev)) _drbd_md_first_sector(meta_dev_idx, bdev))
: 0; : 0;
break; break;
case DRBD_MD_INDEX_FLEX_EXT: case DRBD_MD_INDEX_FLEX_EXT:
...@@ -1771,9 +1799,15 @@ static inline sector_t drbd_get_max_capacity(struct drbd_backing_dev *bdev) ...@@ -1771,9 +1799,15 @@ static inline sector_t drbd_get_max_capacity(struct drbd_backing_dev *bdev)
static inline sector_t drbd_md_ss__(struct drbd_conf *mdev, static inline sector_t drbd_md_ss__(struct drbd_conf *mdev,
struct drbd_backing_dev *bdev) struct drbd_backing_dev *bdev)
{ {
switch (bdev->dc.meta_dev_idx) { int meta_dev_idx;
rcu_read_lock();
meta_dev_idx = rcu_dereference(bdev->disk_conf)->meta_dev_idx;
rcu_read_unlock();
switch (meta_dev_idx) {
default: /* external, some index */ default: /* external, some index */
return MD_RESERVED_SECT * bdev->dc.meta_dev_idx; return MD_RESERVED_SECT * meta_dev_idx;
case DRBD_MD_INDEX_INTERNAL: case DRBD_MD_INDEX_INTERNAL:
/* with drbd08, internal meta data is always "flexible" */ /* with drbd08, internal meta data is always "flexible" */
case DRBD_MD_INDEX_FLEX_INT: case DRBD_MD_INDEX_FLEX_INT:
......
...@@ -866,6 +866,7 @@ int drbd_send_sync_param(struct drbd_conf *mdev) ...@@ -866,6 +866,7 @@ int drbd_send_sync_param(struct drbd_conf *mdev)
const int apv = mdev->tconn->agreed_pro_version; const int apv = mdev->tconn->agreed_pro_version;
enum drbd_packet cmd; enum drbd_packet cmd;
struct net_conf *nc; struct net_conf *nc;
struct disk_conf *dc;
sock = &mdev->tconn->data; sock = &mdev->tconn->data;
p = drbd_prepare_command(mdev, sock); p = drbd_prepare_command(mdev, sock);
...@@ -887,11 +888,12 @@ int drbd_send_sync_param(struct drbd_conf *mdev) ...@@ -887,11 +888,12 @@ int drbd_send_sync_param(struct drbd_conf *mdev)
memset(p->verify_alg, 0, 2 * SHARED_SECRET_MAX); memset(p->verify_alg, 0, 2 * SHARED_SECRET_MAX);
if (get_ldev(mdev)) { if (get_ldev(mdev)) {
p->rate = cpu_to_be32(mdev->ldev->dc.resync_rate); dc = rcu_dereference(mdev->ldev->disk_conf);
p->c_plan_ahead = cpu_to_be32(mdev->ldev->dc.c_plan_ahead); p->rate = cpu_to_be32(dc->resync_rate);
p->c_delay_target = cpu_to_be32(mdev->ldev->dc.c_delay_target); p->c_plan_ahead = cpu_to_be32(dc->c_plan_ahead);
p->c_fill_target = cpu_to_be32(mdev->ldev->dc.c_fill_target); p->c_delay_target = cpu_to_be32(dc->c_delay_target);
p->c_max_rate = cpu_to_be32(mdev->ldev->dc.c_max_rate); p->c_fill_target = cpu_to_be32(dc->c_fill_target);
p->c_max_rate = cpu_to_be32(dc->c_max_rate);
put_ldev(mdev); put_ldev(mdev);
} else { } else {
p->rate = cpu_to_be32(DRBD_RATE_DEF); p->rate = cpu_to_be32(DRBD_RATE_DEF);
...@@ -1056,7 +1058,9 @@ int drbd_send_sizes(struct drbd_conf *mdev, int trigger_reply, enum dds_flags fl ...@@ -1056,7 +1058,9 @@ int drbd_send_sizes(struct drbd_conf *mdev, int trigger_reply, enum dds_flags fl
if (get_ldev_if_state(mdev, D_NEGOTIATING)) { if (get_ldev_if_state(mdev, D_NEGOTIATING)) {
D_ASSERT(mdev->ldev->backing_bdev); D_ASSERT(mdev->ldev->backing_bdev);
d_size = drbd_get_max_capacity(mdev->ldev); d_size = drbd_get_max_capacity(mdev->ldev);
u_size = mdev->ldev->dc.disk_size; rcu_read_lock();
u_size = rcu_dereference(mdev->ldev->disk_conf)->disk_size;
rcu_read_unlock();
q_order_type = drbd_queue_order_type(mdev); q_order_type = drbd_queue_order_type(mdev);
max_bio_size = queue_max_hw_sectors(mdev->ldev->backing_bdev->bd_disk->queue) << 9; max_bio_size = queue_max_hw_sectors(mdev->ldev->backing_bdev->bd_disk->queue) << 9;
max_bio_size = min_t(int, max_bio_size, DRBD_MAX_BIO_SIZE); max_bio_size = min_t(int, max_bio_size, DRBD_MAX_BIO_SIZE);
...@@ -2889,7 +2893,6 @@ int drbd_md_read(struct drbd_conf *mdev, struct drbd_backing_dev *bdev) ...@@ -2889,7 +2893,6 @@ int drbd_md_read(struct drbd_conf *mdev, struct drbd_backing_dev *bdev)
for (i = UI_CURRENT; i < UI_SIZE; i++) for (i = UI_CURRENT; i < UI_SIZE; i++)
bdev->md.uuid[i] = be64_to_cpu(buffer->uuid[i]); bdev->md.uuid[i] = be64_to_cpu(buffer->uuid[i]);
bdev->md.flags = be32_to_cpu(buffer->flags); bdev->md.flags = be32_to_cpu(buffer->flags);
bdev->dc.al_extents = be32_to_cpu(buffer->al_nr_extents);
bdev->md.device_uuid = be64_to_cpu(buffer->device_uuid); bdev->md.device_uuid = be64_to_cpu(buffer->device_uuid);
spin_lock_irq(&mdev->tconn->req_lock); spin_lock_irq(&mdev->tconn->req_lock);
...@@ -2901,8 +2904,12 @@ int drbd_md_read(struct drbd_conf *mdev, struct drbd_backing_dev *bdev) ...@@ -2901,8 +2904,12 @@ int drbd_md_read(struct drbd_conf *mdev, struct drbd_backing_dev *bdev)
} }
spin_unlock_irq(&mdev->tconn->req_lock); spin_unlock_irq(&mdev->tconn->req_lock);
if (bdev->dc.al_extents < 7) mutex_lock(&mdev->tconn->conf_update);
bdev->dc.al_extents = 127; /* This blocks wants to be get removed... */
bdev->disk_conf->al_extents = be32_to_cpu(buffer->al_nr_extents);
if (bdev->disk_conf->al_extents < DRBD_AL_EXTENTS_MIN)
bdev->disk_conf->al_extents = DRBD_AL_EXTENTS_DEF;
mutex_unlock(&mdev->tconn->conf_update);
err: err:
mutex_unlock(&mdev->md_io_mutex); mutex_unlock(&mdev->md_io_mutex);
......
This diff is collapsed.
...@@ -1166,6 +1166,7 @@ static enum finish_epoch drbd_may_finish_epoch(struct drbd_conf *mdev, ...@@ -1166,6 +1166,7 @@ static enum finish_epoch drbd_may_finish_epoch(struct drbd_conf *mdev,
*/ */
void drbd_bump_write_ordering(struct drbd_conf *mdev, enum write_ordering_e wo) __must_hold(local) void drbd_bump_write_ordering(struct drbd_conf *mdev, enum write_ordering_e wo) __must_hold(local)
{ {
struct disk_conf *dc;
enum write_ordering_e pwo; enum write_ordering_e pwo;
static char *write_ordering_str[] = { static char *write_ordering_str[] = {
[WO_none] = "none", [WO_none] = "none",
...@@ -1175,10 +1176,14 @@ void drbd_bump_write_ordering(struct drbd_conf *mdev, enum write_ordering_e wo) ...@@ -1175,10 +1176,14 @@ void drbd_bump_write_ordering(struct drbd_conf *mdev, enum write_ordering_e wo)
pwo = mdev->write_ordering; pwo = mdev->write_ordering;
wo = min(pwo, wo); wo = min(pwo, wo);
if (wo == WO_bdev_flush && mdev->ldev->dc.no_disk_flush) rcu_read_lock();
dc = rcu_dereference(mdev->ldev->disk_conf);
if (wo == WO_bdev_flush && dc->no_disk_flush)
wo = WO_drain_io; wo = WO_drain_io;
if (wo == WO_drain_io && mdev->ldev->dc.no_disk_drain) if (wo == WO_drain_io && dc->no_disk_drain)
wo = WO_none; wo = WO_none;
rcu_read_unlock();
mdev->write_ordering = wo; mdev->write_ordering = wo;
if (pwo != mdev->write_ordering || wo == WO_bdev_flush) if (pwo != mdev->write_ordering || wo == WO_bdev_flush)
dev_info(DEV, "Method to ensure write ordering: %s\n", write_ordering_str[mdev->write_ordering]); dev_info(DEV, "Method to ensure write ordering: %s\n", write_ordering_str[mdev->write_ordering]);
...@@ -2190,9 +2195,14 @@ int drbd_rs_should_slow_down(struct drbd_conf *mdev, sector_t sector) ...@@ -2190,9 +2195,14 @@ int drbd_rs_should_slow_down(struct drbd_conf *mdev, sector_t sector)
struct lc_element *tmp; struct lc_element *tmp;
int curr_events; int curr_events;
int throttle = 0; int throttle = 0;
unsigned int c_min_rate;
rcu_read_lock();
c_min_rate = rcu_dereference(mdev->ldev->disk_conf)->c_min_rate;
rcu_read_unlock();
/* feature disabled? */ /* feature disabled? */
if (mdev->ldev->dc.c_min_rate == 0) if (c_min_rate == 0)
return 0; return 0;
spin_lock_irq(&mdev->al_lock); spin_lock_irq(&mdev->al_lock);
...@@ -2232,7 +2242,7 @@ int drbd_rs_should_slow_down(struct drbd_conf *mdev, sector_t sector) ...@@ -2232,7 +2242,7 @@ int drbd_rs_should_slow_down(struct drbd_conf *mdev, sector_t sector)
db = mdev->rs_mark_left[i] - rs_left; db = mdev->rs_mark_left[i] - rs_left;
dbdt = Bit2KB(db/dt); dbdt = Bit2KB(db/dt);
if (dbdt > mdev->ldev->dc.c_min_rate) if (dbdt > c_min_rate)
throttle = 1; throttle = 1;
} }
return throttle; return throttle;
...@@ -3147,6 +3157,7 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi) ...@@ -3147,6 +3157,7 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi)
struct crypto_hash *verify_tfm = NULL; struct crypto_hash *verify_tfm = NULL;
struct crypto_hash *csums_tfm = NULL; struct crypto_hash *csums_tfm = NULL;
struct net_conf *old_net_conf, *new_net_conf = NULL; struct net_conf *old_net_conf, *new_net_conf = NULL;
struct disk_conf *old_disk_conf, *new_disk_conf = NULL;
const int apv = tconn->agreed_pro_version; const int apv = tconn->agreed_pro_version;
int *rs_plan_s = NULL; int *rs_plan_s = NULL;
int fifo_size = 0; int fifo_size = 0;
...@@ -3189,24 +3200,34 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi) ...@@ -3189,24 +3200,34 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi)
if (err) if (err)
return err; return err;
if (get_ldev(mdev)) { new_disk_conf = kzalloc(sizeof(struct disk_conf), GFP_KERNEL);
mdev->ldev->dc.resync_rate = be32_to_cpu(p->rate); if (!new_disk_conf) {
put_ldev(mdev); dev_err(DEV, "Allocation of new disk_conf failed\n");
return -ENOMEM;
} }
mutex_lock(&mdev->tconn->conf_update);
old_net_conf = mdev->tconn->net_conf;
old_disk_conf = mdev->ldev->disk_conf;
*new_disk_conf = *old_disk_conf;
new_disk_conf->resync_rate = be32_to_cpu(p->rate);
if (apv >= 88) { if (apv >= 88) {
if (apv == 88) { if (apv == 88) {
if (data_size > SHARED_SECRET_MAX) { if (data_size > SHARED_SECRET_MAX) {
dev_err(DEV, "verify-alg too long, " dev_err(DEV, "verify-alg too long, "
"peer wants %u, accepting only %u byte\n", "peer wants %u, accepting only %u byte\n",
data_size, SHARED_SECRET_MAX); data_size, SHARED_SECRET_MAX);
mutex_unlock(&mdev->tconn->conf_update);
return -EIO; return -EIO;
} }
err = drbd_recv_all(mdev->tconn, p->verify_alg, data_size); err = drbd_recv_all(mdev->tconn, p->verify_alg, data_size);
if (err) if (err) {
mutex_unlock(&mdev->tconn->conf_update);
return err; return err;
}
/* we expect NUL terminated string */ /* we expect NUL terminated string */
/* but just in case someone tries to be evil */ /* but just in case someone tries to be evil */
D_ASSERT(p->verify_alg[data_size-1] == 0); D_ASSERT(p->verify_alg[data_size-1] == 0);
...@@ -3221,9 +3242,6 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi) ...@@ -3221,9 +3242,6 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi)
p->csums_alg[SHARED_SECRET_MAX-1] = 0; p->csums_alg[SHARED_SECRET_MAX-1] = 0;
} }
mutex_lock(&mdev->tconn->conf_update);
old_net_conf = mdev->tconn->net_conf;
if (strcmp(old_net_conf->verify_alg, p->verify_alg)) { if (strcmp(old_net_conf->verify_alg, p->verify_alg)) {
if (mdev->state.conn == C_WF_REPORT_PARAMS) { if (mdev->state.conn == C_WF_REPORT_PARAMS) {
dev_err(DEV, "Different verify-alg settings. me=\"%s\" peer=\"%s\"\n", dev_err(DEV, "Different verify-alg settings. me=\"%s\" peer=\"%s\"\n",
...@@ -3252,14 +3270,13 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi) ...@@ -3252,14 +3270,13 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi)
} }
} }
if (apv > 94 && get_ldev(mdev)) { if (apv > 94) {
mdev->ldev->dc.resync_rate = be32_to_cpu(p->rate); new_disk_conf->c_plan_ahead = be32_to_cpu(p->c_plan_ahead);
mdev->ldev->dc.c_plan_ahead = be32_to_cpu(p->c_plan_ahead); new_disk_conf->c_delay_target = be32_to_cpu(p->c_delay_target);
mdev->ldev->dc.c_delay_target = be32_to_cpu(p->c_delay_target); new_disk_conf->c_fill_target = be32_to_cpu(p->c_fill_target);
mdev->ldev->dc.c_fill_target = be32_to_cpu(p->c_fill_target); new_disk_conf->c_max_rate = be32_to_cpu(p->c_max_rate);
mdev->ldev->dc.c_max_rate = be32_to_cpu(p->c_max_rate);
fifo_size = (mdev->ldev->dc.c_plan_ahead * 10 * SLEEP_TIME) / HZ; fifo_size = (new_disk_conf->c_plan_ahead * 10 * SLEEP_TIME) / HZ;
if (fifo_size != mdev->rs_plan_s.size && fifo_size > 0) { if (fifo_size != mdev->rs_plan_s.size && fifo_size > 0) {
rs_plan_s = kzalloc(sizeof(int) * fifo_size, GFP_KERNEL); rs_plan_s = kzalloc(sizeof(int) * fifo_size, GFP_KERNEL);
if (!rs_plan_s) { if (!rs_plan_s) {
...@@ -3268,7 +3285,6 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi) ...@@ -3268,7 +3285,6 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi)
goto disconnect; goto disconnect;
} }
} }
put_ldev(mdev);
} }
if (verify_tfm || csums_tfm) { if (verify_tfm || csums_tfm) {
...@@ -3296,21 +3312,24 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi) ...@@ -3296,21 +3312,24 @@ static int receive_SyncParam(struct drbd_tconn *tconn, struct packet_info *pi)
} }
rcu_assign_pointer(tconn->net_conf, new_net_conf); rcu_assign_pointer(tconn->net_conf, new_net_conf);
} }
mutex_unlock(&mdev->tconn->conf_update); }
if (new_net_conf) {
synchronize_rcu();
kfree(old_net_conf);
}
spin_lock(&mdev->peer_seq_lock); rcu_assign_pointer(mdev->ldev->disk_conf, new_disk_conf);
if (fifo_size != mdev->rs_plan_s.size) { spin_lock(&mdev->peer_seq_lock);
kfree(mdev->rs_plan_s.values); if (rs_plan_s) {
mdev->rs_plan_s.values = rs_plan_s; kfree(mdev->rs_plan_s.values);
mdev->rs_plan_s.size = fifo_size; mdev->rs_plan_s.values = rs_plan_s;
mdev->rs_planed = 0; mdev->rs_plan_s.size = fifo_size;
} mdev->rs_planed = 0;
spin_unlock(&mdev->peer_seq_lock);
} }
spin_unlock(&mdev->peer_seq_lock);
mutex_unlock(&mdev->tconn->conf_update);
synchronize_rcu();
if (new_net_conf)
kfree(old_net_conf);
kfree(old_disk_conf);
return 0; return 0;
disconnect: disconnect:
...@@ -3358,37 +3377,56 @@ static int receive_sizes(struct drbd_tconn *tconn, struct packet_info *pi) ...@@ -3358,37 +3377,56 @@ static int receive_sizes(struct drbd_tconn *tconn, struct packet_info *pi)
mdev->p_size = p_size; mdev->p_size = p_size;
if (get_ldev(mdev)) { if (get_ldev(mdev)) {
rcu_read_lock();
my_usize = rcu_dereference(mdev->ldev->disk_conf)->disk_size;
rcu_read_unlock();
warn_if_differ_considerably(mdev, "lower level device sizes", warn_if_differ_considerably(mdev, "lower level device sizes",
p_size, drbd_get_max_capacity(mdev->ldev)); p_size, drbd_get_max_capacity(mdev->ldev));
warn_if_differ_considerably(mdev, "user requested size", warn_if_differ_considerably(mdev, "user requested size",
p_usize, mdev->ldev->dc.disk_size); p_usize, my_usize);
/* if this is the first connect, or an otherwise expected /* if this is the first connect, or an otherwise expected
* param exchange, choose the minimum */ * param exchange, choose the minimum */
if (mdev->state.conn == C_WF_REPORT_PARAMS) if (mdev->state.conn == C_WF_REPORT_PARAMS)
p_usize = min_not_zero((sector_t)mdev->ldev->dc.disk_size, p_usize = min_not_zero(my_usize, p_usize);
p_usize);
my_usize = mdev->ldev->dc.disk_size;
if (mdev->ldev->dc.disk_size != p_usize) {
mdev->ldev->dc.disk_size = p_usize;
dev_info(DEV, "Peer sets u_size to %lu sectors\n",
(unsigned long)mdev->ldev->dc.disk_size);
}
/* Never shrink a device with usable data during connect. /* Never shrink a device with usable data during connect.
But allow online shrinking if we are connected. */ But allow online shrinking if we are connected. */
if (drbd_new_dev_size(mdev, mdev->ldev, p_usize, 0) < if (drbd_new_dev_size(mdev, mdev->ldev, p_usize, 0) <
drbd_get_capacity(mdev->this_bdev) && drbd_get_capacity(mdev->this_bdev) &&
mdev->state.disk >= D_OUTDATED && mdev->state.disk >= D_OUTDATED &&
mdev->state.conn < C_CONNECTED) { mdev->state.conn < C_CONNECTED) {
dev_err(DEV, "The peer's disk size is too small!\n"); dev_err(DEV, "The peer's disk size is too small!\n");
conn_request_state(mdev->tconn, NS(conn, C_DISCONNECTING), CS_HARD); conn_request_state(mdev->tconn, NS(conn, C_DISCONNECTING), CS_HARD);
mdev->ldev->dc.disk_size = my_usize;
put_ldev(mdev); put_ldev(mdev);
return -EIO; return -EIO;
} }
if (my_usize != p_usize) {
struct disk_conf *old_disk_conf, *new_disk_conf = NULL;
new_disk_conf = kzalloc(sizeof(struct disk_conf), GFP_KERNEL);
if (!new_disk_conf) {
dev_err(DEV, "Allocation of new disk_conf failed\n");
put_ldev(mdev);
return -ENOMEM;
}
mutex_lock(&mdev->tconn->conf_update);
old_disk_conf = mdev->ldev->disk_conf;
*new_disk_conf = *old_disk_conf;
new_disk_conf->disk_size = p_usize;
rcu_assign_pointer(mdev->ldev->disk_conf, new_disk_conf);
mutex_unlock(&mdev->tconn->conf_update);
synchronize_rcu();
kfree(old_disk_conf);
dev_info(DEV, "Peer sets u_size to %lu sectors\n",
(unsigned long)my_usize);
}
put_ldev(mdev); put_ldev(mdev);
} }
...@@ -4268,7 +4306,9 @@ static int drbd_disconnected(int vnr, void *p, void *data) ...@@ -4268,7 +4306,9 @@ static int drbd_disconnected(int vnr, void *p, void *data)
fp = FP_DONT_CARE; fp = FP_DONT_CARE;
if (get_ldev(mdev)) { if (get_ldev(mdev)) {
fp = mdev->ldev->dc.fencing; rcu_read_lock();
fp = rcu_dereference(mdev->ldev->disk_conf)->fencing;
rcu_read_unlock();
put_ldev(mdev); put_ldev(mdev);
} }
......
...@@ -483,13 +483,13 @@ is_valid_state(struct drbd_conf *mdev, union drbd_state ns) ...@@ -483,13 +483,13 @@ is_valid_state(struct drbd_conf *mdev, union drbd_state ns)
enum drbd_state_rv rv = SS_SUCCESS; enum drbd_state_rv rv = SS_SUCCESS;
struct net_conf *nc; struct net_conf *nc;
rcu_read_lock();
fp = FP_DONT_CARE; fp = FP_DONT_CARE;
if (get_ldev(mdev)) { if (get_ldev(mdev)) {
fp = mdev->ldev->dc.fencing; fp = rcu_dereference(mdev->ldev->disk_conf)->fencing;
put_ldev(mdev); put_ldev(mdev);
} }
rcu_read_lock();
nc = rcu_dereference(mdev->tconn->net_conf); nc = rcu_dereference(mdev->tconn->net_conf);
if (nc) { if (nc) {
if (!nc->two_primaries && ns.role == R_PRIMARY) { if (!nc->two_primaries && ns.role == R_PRIMARY) {
...@@ -674,7 +674,9 @@ static union drbd_state sanitize_state(struct drbd_conf *mdev, union drbd_state ...@@ -674,7 +674,9 @@ static union drbd_state sanitize_state(struct drbd_conf *mdev, union drbd_state
fp = FP_DONT_CARE; fp = FP_DONT_CARE;
if (get_ldev(mdev)) { if (get_ldev(mdev)) {
fp = mdev->ldev->dc.fencing; rcu_read_lock();
fp = rcu_dereference(mdev->ldev->disk_conf)->fencing;
rcu_read_unlock();
put_ldev(mdev); put_ldev(mdev);
} }
...@@ -1132,7 +1134,9 @@ static void after_state_ch(struct drbd_conf *mdev, union drbd_state os, ...@@ -1132,7 +1134,9 @@ static void after_state_ch(struct drbd_conf *mdev, union drbd_state os,
fp = FP_DONT_CARE; fp = FP_DONT_CARE;
if (get_ldev(mdev)) { if (get_ldev(mdev)) {
fp = mdev->ldev->dc.fencing; rcu_read_lock();
fp = rcu_dereference(mdev->ldev->disk_conf)->fencing;
rcu_read_unlock();
put_ldev(mdev); put_ldev(mdev);
} }
...@@ -1287,7 +1291,9 @@ static void after_state_ch(struct drbd_conf *mdev, union drbd_state os, ...@@ -1287,7 +1291,9 @@ static void after_state_ch(struct drbd_conf *mdev, union drbd_state os,
/* corresponding get_ldev was in __drbd_set_state, to serialize /* corresponding get_ldev was in __drbd_set_state, to serialize
* our cleanup here with the transition to D_DISKLESS, * our cleanup here with the transition to D_DISKLESS,
* so it is safe to dreference ldev here. */ * so it is safe to dreference ldev here. */
eh = mdev->ldev->dc.on_io_error; rcu_read_lock();
eh = rcu_dereference(mdev->ldev->disk_conf)->on_io_error;
rcu_read_unlock();
was_io_error = test_and_clear_bit(WAS_IO_ERROR, &mdev->flags); was_io_error = test_and_clear_bit(WAS_IO_ERROR, &mdev->flags);
/* current state still has to be D_FAILED, /* current state still has to be D_FAILED,
......
...@@ -436,6 +436,7 @@ static void fifo_add_val(struct fifo_buffer *fb, int value) ...@@ -436,6 +436,7 @@ static void fifo_add_val(struct fifo_buffer *fb, int value)
static int drbd_rs_controller(struct drbd_conf *mdev) static int drbd_rs_controller(struct drbd_conf *mdev)
{ {
struct disk_conf *dc;
unsigned int sect_in; /* Number of sectors that came in since the last turn */ unsigned int sect_in; /* Number of sectors that came in since the last turn */
unsigned int want; /* The number of sectors we want in the proxy */ unsigned int want; /* The number of sectors we want in the proxy */
int req_sect; /* Number of sectors to request in this turn */ int req_sect; /* Number of sectors to request in this turn */
...@@ -449,14 +450,16 @@ static int drbd_rs_controller(struct drbd_conf *mdev) ...@@ -449,14 +450,16 @@ static int drbd_rs_controller(struct drbd_conf *mdev)
mdev->rs_in_flight -= sect_in; mdev->rs_in_flight -= sect_in;
spin_lock(&mdev->peer_seq_lock); /* get an atomic view on mdev->rs_plan_s */ spin_lock(&mdev->peer_seq_lock); /* get an atomic view on mdev->rs_plan_s */
rcu_read_lock();
dc = rcu_dereference(mdev->ldev->disk_conf);
steps = mdev->rs_plan_s.size; /* (mdev->ldev->dc.c_plan_ahead * 10 * SLEEP_TIME) / HZ; */ steps = mdev->rs_plan_s.size; /* (dc->c_plan_ahead * 10 * SLEEP_TIME) / HZ; */
if (mdev->rs_in_flight + sect_in == 0) { /* At start of resync */ if (mdev->rs_in_flight + sect_in == 0) { /* At start of resync */
want = ((mdev->ldev->dc.resync_rate * 2 * SLEEP_TIME) / HZ) * steps; want = ((dc->resync_rate * 2 * SLEEP_TIME) / HZ) * steps;
} else { /* normal path */ } else { /* normal path */
want = mdev->ldev->dc.c_fill_target ? mdev->ldev->dc.c_fill_target : want = dc->c_fill_target ? dc->c_fill_target :
sect_in * mdev->ldev->dc.c_delay_target * HZ / (SLEEP_TIME * 10); sect_in * dc->c_delay_target * HZ / (SLEEP_TIME * 10);
} }
correction = want - mdev->rs_in_flight - mdev->rs_planed; correction = want - mdev->rs_in_flight - mdev->rs_planed;
...@@ -468,14 +471,13 @@ static int drbd_rs_controller(struct drbd_conf *mdev) ...@@ -468,14 +471,13 @@ static int drbd_rs_controller(struct drbd_conf *mdev)
/* What we do in this step */ /* What we do in this step */
curr_corr = fifo_push(&mdev->rs_plan_s, 0); curr_corr = fifo_push(&mdev->rs_plan_s, 0);
spin_unlock(&mdev->peer_seq_lock);
mdev->rs_planed -= curr_corr; mdev->rs_planed -= curr_corr;
req_sect = sect_in + curr_corr; req_sect = sect_in + curr_corr;
if (req_sect < 0) if (req_sect < 0)
req_sect = 0; req_sect = 0;
max_sect = (mdev->ldev->dc.c_max_rate * 2 * SLEEP_TIME) / HZ; max_sect = (dc->c_max_rate * 2 * SLEEP_TIME) / HZ;
if (req_sect > max_sect) if (req_sect > max_sect)
req_sect = max_sect; req_sect = max_sect;
...@@ -484,6 +486,8 @@ static int drbd_rs_controller(struct drbd_conf *mdev) ...@@ -484,6 +486,8 @@ static int drbd_rs_controller(struct drbd_conf *mdev)
sect_in, mdev->rs_in_flight, want, correction, sect_in, mdev->rs_in_flight, want, correction,
steps, cps, mdev->rs_planed, curr_corr, req_sect); steps, cps, mdev->rs_planed, curr_corr, req_sect);
*/ */
rcu_read_unlock();
spin_unlock(&mdev->peer_seq_lock);
return req_sect; return req_sect;
} }
...@@ -491,11 +495,13 @@ static int drbd_rs_controller(struct drbd_conf *mdev) ...@@ -491,11 +495,13 @@ static int drbd_rs_controller(struct drbd_conf *mdev)
static int drbd_rs_number_requests(struct drbd_conf *mdev) static int drbd_rs_number_requests(struct drbd_conf *mdev)
{ {
int number; int number;
if (mdev->rs_plan_s.size) { /* mdev->ldev->dc.c_plan_ahead */ if (mdev->rs_plan_s.size) { /* rcu_dereference(mdev->ldev->disk_conf)->c_plan_ahead */
number = drbd_rs_controller(mdev) >> (BM_BLOCK_SHIFT - 9); number = drbd_rs_controller(mdev) >> (BM_BLOCK_SHIFT - 9);
mdev->c_sync_rate = number * HZ * (BM_BLOCK_SIZE / 1024) / SLEEP_TIME; mdev->c_sync_rate = number * HZ * (BM_BLOCK_SIZE / 1024) / SLEEP_TIME;
} else { } else {
mdev->c_sync_rate = mdev->ldev->dc.resync_rate; rcu_read_lock();
mdev->c_sync_rate = rcu_dereference(mdev->ldev->disk_conf)->resync_rate;
rcu_read_unlock();
number = SLEEP_TIME * mdev->c_sync_rate / ((BM_BLOCK_SIZE / 1024) * HZ); number = SLEEP_TIME * mdev->c_sync_rate / ((BM_BLOCK_SIZE / 1024) * HZ);
} }
...@@ -1320,13 +1326,17 @@ int w_restart_disk_io(struct drbd_work *w, int cancel) ...@@ -1320,13 +1326,17 @@ int w_restart_disk_io(struct drbd_work *w, int cancel)
static int _drbd_may_sync_now(struct drbd_conf *mdev) static int _drbd_may_sync_now(struct drbd_conf *mdev)
{ {
struct drbd_conf *odev = mdev; struct drbd_conf *odev = mdev;
int ra;
while (1) { while (1) {
if (!odev->ldev) if (!odev->ldev)
return 1; return 1;
if (odev->ldev->dc.resync_after == -1) rcu_read_lock();
ra = rcu_dereference(odev->ldev->disk_conf)->resync_after;
rcu_read_unlock();
if (ra == -1)
return 1; return 1;
odev = minor_to_mdev(odev->ldev->dc.resync_after); odev = minor_to_mdev(ra);
if (!expect(odev)) if (!expect(odev))
return 1; return 1;
if ((odev->state.conn >= C_SYNC_SOURCE && if ((odev->state.conn >= C_SYNC_SOURCE &&
...@@ -1405,6 +1415,7 @@ void suspend_other_sg(struct drbd_conf *mdev) ...@@ -1405,6 +1415,7 @@ void suspend_other_sg(struct drbd_conf *mdev)
enum drbd_ret_code drbd_sync_after_valid(struct drbd_conf *mdev, int o_minor) enum drbd_ret_code drbd_sync_after_valid(struct drbd_conf *mdev, int o_minor)
{ {
struct drbd_conf *odev; struct drbd_conf *odev;
int ra;
if (o_minor == -1) if (o_minor == -1)
return NO_ERROR; return NO_ERROR;
...@@ -1417,12 +1428,15 @@ enum drbd_ret_code drbd_sync_after_valid(struct drbd_conf *mdev, int o_minor) ...@@ -1417,12 +1428,15 @@ enum drbd_ret_code drbd_sync_after_valid(struct drbd_conf *mdev, int o_minor)
if (odev == mdev) if (odev == mdev)
return ERR_SYNC_AFTER_CYCLE; return ERR_SYNC_AFTER_CYCLE;
rcu_read_lock();
ra = rcu_dereference(odev->ldev->disk_conf)->resync_after;
rcu_read_unlock();
/* dependency chain ends here, no cycles. */ /* dependency chain ends here, no cycles. */
if (odev->ldev->dc.resync_after == -1) if (ra == -1)
return NO_ERROR; return NO_ERROR;
/* follow the dependency chain */ /* follow the dependency chain */
odev = minor_to_mdev(odev->ldev->dc.resync_after); odev = minor_to_mdev(ra);
} }
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment