Commit ad878a0d authored by Christoph Böhmwalder's avatar Christoph Böhmwalder Committed by Jens Axboe

drbd: pass drbd_peer_device to __req_mod

In preparation to support multiple connections, we need to know which
one we need to modify the request state for.

Originally-from: Lars Ellenberg <lars.ellenberg@linbit.com>
Signed-off-by: default avatarChristoph Böhmwalder <christoph.boehmwalder@linbit.com>
Link: https://lore.kernel.org/r/20230330102744.2128122-2-christoph.boehmwalder@linbit.comSigned-off-by: default avatarJens Axboe <axboe@kernel.dk>
parent db445db1
...@@ -231,9 +231,11 @@ void tl_release(struct drbd_connection *connection, unsigned int barrier_nr, ...@@ -231,9 +231,11 @@ void tl_release(struct drbd_connection *connection, unsigned int barrier_nr,
} }
req = list_prepare_entry(tmp, &connection->transfer_log, tl_requests); req = list_prepare_entry(tmp, &connection->transfer_log, tl_requests);
list_for_each_entry_safe_from(req, r, &connection->transfer_log, tl_requests) { list_for_each_entry_safe_from(req, r, &connection->transfer_log, tl_requests) {
struct drbd_peer_device *peer_device;
if (req->epoch != expect_epoch) if (req->epoch != expect_epoch)
break; break;
_req_mod(req, BARRIER_ACKED); peer_device = conn_peer_device(connection, req->device->vnr);
_req_mod(req, BARRIER_ACKED, peer_device);
} }
spin_unlock_irq(&connection->resource->req_lock); spin_unlock_irq(&connection->resource->req_lock);
...@@ -256,10 +258,13 @@ void tl_release(struct drbd_connection *connection, unsigned int barrier_nr, ...@@ -256,10 +258,13 @@ void tl_release(struct drbd_connection *connection, unsigned int barrier_nr,
/* must hold resource->req_lock */ /* must hold resource->req_lock */
void _tl_restart(struct drbd_connection *connection, enum drbd_req_event what) void _tl_restart(struct drbd_connection *connection, enum drbd_req_event what)
{ {
struct drbd_peer_device *peer_device;
struct drbd_request *req, *r; struct drbd_request *req, *r;
list_for_each_entry_safe(req, r, &connection->transfer_log, tl_requests) list_for_each_entry_safe(req, r, &connection->transfer_log, tl_requests) {
_req_mod(req, what); peer_device = conn_peer_device(connection, req->device->vnr);
_req_mod(req, what, peer_device);
}
} }
void tl_restart(struct drbd_connection *connection, enum drbd_req_event what) void tl_restart(struct drbd_connection *connection, enum drbd_req_event what)
...@@ -297,7 +302,7 @@ void tl_abort_disk_io(struct drbd_device *device) ...@@ -297,7 +302,7 @@ void tl_abort_disk_io(struct drbd_device *device)
continue; continue;
if (req->device != device) if (req->device != device)
continue; continue;
_req_mod(req, ABORT_DISK_IO); _req_mod(req, ABORT_DISK_IO, NULL);
} }
spin_unlock_irq(&connection->resource->req_lock); spin_unlock_irq(&connection->resource->req_lock);
} }
......
...@@ -2138,7 +2138,7 @@ static int receive_DataReply(struct drbd_connection *connection, struct packet_i ...@@ -2138,7 +2138,7 @@ static int receive_DataReply(struct drbd_connection *connection, struct packet_i
err = recv_dless_read(peer_device, req, sector, pi->size); err = recv_dless_read(peer_device, req, sector, pi->size);
if (!err) if (!err)
req_mod(req, DATA_RECEIVED); req_mod(req, DATA_RECEIVED, peer_device);
/* else: nothing. handled from drbd_disconnect... /* else: nothing. handled from drbd_disconnect...
* I don't think we may complete this just yet * I don't think we may complete this just yet
* in case we are "on-disconnect: freeze" */ * in case we are "on-disconnect: freeze" */
...@@ -2196,7 +2196,7 @@ static void restart_conflicting_writes(struct drbd_device *device, ...@@ -2196,7 +2196,7 @@ static void restart_conflicting_writes(struct drbd_device *device,
continue; continue;
/* as it is RQ_POSTPONED, this will cause it to /* as it is RQ_POSTPONED, this will cause it to
* be queued on the retry workqueue. */ * be queued on the retry workqueue. */
__req_mod(req, CONFLICT_RESOLVED, NULL); __req_mod(req, CONFLICT_RESOLVED, NULL, NULL);
} }
} }
...@@ -2420,6 +2420,7 @@ static blk_opf_t wire_flags_to_bio(struct drbd_connection *connection, u32 dpf) ...@@ -2420,6 +2420,7 @@ static blk_opf_t wire_flags_to_bio(struct drbd_connection *connection, u32 dpf)
static void fail_postponed_requests(struct drbd_device *device, sector_t sector, static void fail_postponed_requests(struct drbd_device *device, sector_t sector,
unsigned int size) unsigned int size)
{ {
struct drbd_peer_device *peer_device = first_peer_device(device);
struct drbd_interval *i; struct drbd_interval *i;
repeat: repeat:
...@@ -2433,7 +2434,7 @@ static void fail_postponed_requests(struct drbd_device *device, sector_t sector, ...@@ -2433,7 +2434,7 @@ static void fail_postponed_requests(struct drbd_device *device, sector_t sector,
if (!(req->rq_state & RQ_POSTPONED)) if (!(req->rq_state & RQ_POSTPONED))
continue; continue;
req->rq_state &= ~RQ_POSTPONED; req->rq_state &= ~RQ_POSTPONED;
__req_mod(req, NEG_ACKED, &m); __req_mod(req, NEG_ACKED, peer_device, &m);
spin_unlock_irq(&device->resource->req_lock); spin_unlock_irq(&device->resource->req_lock);
if (m.bio) if (m.bio)
complete_master_bio(device, &m); complete_master_bio(device, &m);
...@@ -5661,10 +5662,11 @@ static int got_IsInSync(struct drbd_connection *connection, struct packet_info * ...@@ -5661,10 +5662,11 @@ static int got_IsInSync(struct drbd_connection *connection, struct packet_info *
} }
static int static int
validate_req_change_req_state(struct drbd_device *device, u64 id, sector_t sector, validate_req_change_req_state(struct drbd_peer_device *peer_device, u64 id, sector_t sector,
struct rb_root *root, const char *func, struct rb_root *root, const char *func,
enum drbd_req_event what, bool missing_ok) enum drbd_req_event what, bool missing_ok)
{ {
struct drbd_device *device = peer_device->device;
struct drbd_request *req; struct drbd_request *req;
struct bio_and_error m; struct bio_and_error m;
...@@ -5674,7 +5676,7 @@ validate_req_change_req_state(struct drbd_device *device, u64 id, sector_t secto ...@@ -5674,7 +5676,7 @@ validate_req_change_req_state(struct drbd_device *device, u64 id, sector_t secto
spin_unlock_irq(&device->resource->req_lock); spin_unlock_irq(&device->resource->req_lock);
return -EIO; return -EIO;
} }
__req_mod(req, what, &m); __req_mod(req, what, peer_device, &m);
spin_unlock_irq(&device->resource->req_lock); spin_unlock_irq(&device->resource->req_lock);
if (m.bio) if (m.bio)
...@@ -5723,7 +5725,7 @@ static int got_BlockAck(struct drbd_connection *connection, struct packet_info * ...@@ -5723,7 +5725,7 @@ static int got_BlockAck(struct drbd_connection *connection, struct packet_info *
BUG(); BUG();
} }
return validate_req_change_req_state(device, p->block_id, sector, return validate_req_change_req_state(peer_device, p->block_id, sector,
&device->write_requests, __func__, &device->write_requests, __func__,
what, false); what, false);
} }
...@@ -5750,7 +5752,7 @@ static int got_NegAck(struct drbd_connection *connection, struct packet_info *pi ...@@ -5750,7 +5752,7 @@ static int got_NegAck(struct drbd_connection *connection, struct packet_info *pi
return 0; return 0;
} }
err = validate_req_change_req_state(device, p->block_id, sector, err = validate_req_change_req_state(peer_device, p->block_id, sector,
&device->write_requests, __func__, &device->write_requests, __func__,
NEG_ACKED, true); NEG_ACKED, true);
if (err) { if (err) {
...@@ -5781,7 +5783,7 @@ static int got_NegDReply(struct drbd_connection *connection, struct packet_info ...@@ -5781,7 +5783,7 @@ static int got_NegDReply(struct drbd_connection *connection, struct packet_info
drbd_err(device, "Got NegDReply; Sector %llus, len %u.\n", drbd_err(device, "Got NegDReply; Sector %llus, len %u.\n",
(unsigned long long)sector, be32_to_cpu(p->blksize)); (unsigned long long)sector, be32_to_cpu(p->blksize));
return validate_req_change_req_state(device, p->block_id, sector, return validate_req_change_req_state(peer_device, p->block_id, sector,
&device->read_requests, __func__, &device->read_requests, __func__,
NEG_ACKED, false); NEG_ACKED, false);
} }
......
...@@ -552,12 +552,15 @@ static inline bool is_pending_write_protocol_A(struct drbd_request *req) ...@@ -552,12 +552,15 @@ static inline bool is_pending_write_protocol_A(struct drbd_request *req)
* happen "atomically" within the req_lock, * happen "atomically" within the req_lock,
* and it enforces that we have to think in a very structured manner * and it enforces that we have to think in a very structured manner
* about the "events" that may happen to a request during its life time ... * about the "events" that may happen to a request during its life time ...
*
*
* peer_device == NULL means local disk
*/ */
int __req_mod(struct drbd_request *req, enum drbd_req_event what, int __req_mod(struct drbd_request *req, enum drbd_req_event what,
struct drbd_peer_device *peer_device,
struct bio_and_error *m) struct bio_and_error *m)
{ {
struct drbd_device *const device = req->device; struct drbd_device *const device = req->device;
struct drbd_peer_device *const peer_device = first_peer_device(device);
struct drbd_connection *const connection = peer_device ? peer_device->connection : NULL; struct drbd_connection *const connection = peer_device ? peer_device->connection : NULL;
struct net_conf *nc; struct net_conf *nc;
int p, rv = 0; int p, rv = 0;
...@@ -1100,6 +1103,7 @@ static bool drbd_should_send_out_of_sync(union drbd_dev_state s) ...@@ -1100,6 +1103,7 @@ static bool drbd_should_send_out_of_sync(union drbd_dev_state s)
static int drbd_process_write_request(struct drbd_request *req) static int drbd_process_write_request(struct drbd_request *req)
{ {
struct drbd_device *device = req->device; struct drbd_device *device = req->device;
struct drbd_peer_device *peer_device = first_peer_device(device);
int remote, send_oos; int remote, send_oos;
remote = drbd_should_do_remote(device->state); remote = drbd_should_do_remote(device->state);
...@@ -1115,7 +1119,7 @@ static int drbd_process_write_request(struct drbd_request *req) ...@@ -1115,7 +1119,7 @@ static int drbd_process_write_request(struct drbd_request *req)
/* The only size==0 bios we expect are empty flushes. */ /* The only size==0 bios we expect are empty flushes. */
D_ASSERT(device, req->master_bio->bi_opf & REQ_PREFLUSH); D_ASSERT(device, req->master_bio->bi_opf & REQ_PREFLUSH);
if (remote) if (remote)
_req_mod(req, QUEUE_AS_DRBD_BARRIER); _req_mod(req, QUEUE_AS_DRBD_BARRIER, peer_device);
return remote; return remote;
} }
...@@ -1125,10 +1129,10 @@ static int drbd_process_write_request(struct drbd_request *req) ...@@ -1125,10 +1129,10 @@ static int drbd_process_write_request(struct drbd_request *req)
D_ASSERT(device, !(remote && send_oos)); D_ASSERT(device, !(remote && send_oos));
if (remote) { if (remote) {
_req_mod(req, TO_BE_SENT); _req_mod(req, TO_BE_SENT, peer_device);
_req_mod(req, QUEUE_FOR_NET_WRITE); _req_mod(req, QUEUE_FOR_NET_WRITE, peer_device);
} else if (drbd_set_out_of_sync(device, req->i.sector, req->i.size)) } else if (drbd_set_out_of_sync(device, req->i.sector, req->i.size))
_req_mod(req, QUEUE_FOR_SEND_OOS); _req_mod(req, QUEUE_FOR_SEND_OOS, peer_device);
return remote; return remote;
} }
...@@ -1312,6 +1316,7 @@ static void drbd_update_plug(struct drbd_plug_cb *plug, struct drbd_request *req ...@@ -1312,6 +1316,7 @@ static void drbd_update_plug(struct drbd_plug_cb *plug, struct drbd_request *req
static void drbd_send_and_submit(struct drbd_device *device, struct drbd_request *req) static void drbd_send_and_submit(struct drbd_device *device, struct drbd_request *req)
{ {
struct drbd_resource *resource = device->resource; struct drbd_resource *resource = device->resource;
struct drbd_peer_device *peer_device = first_peer_device(device);
const int rw = bio_data_dir(req->master_bio); const int rw = bio_data_dir(req->master_bio);
struct bio_and_error m = { NULL, }; struct bio_and_error m = { NULL, };
bool no_remote = false; bool no_remote = false;
...@@ -1375,8 +1380,8 @@ static void drbd_send_and_submit(struct drbd_device *device, struct drbd_request ...@@ -1375,8 +1380,8 @@ static void drbd_send_and_submit(struct drbd_device *device, struct drbd_request
/* We either have a private_bio, or we can read from remote. /* We either have a private_bio, or we can read from remote.
* Otherwise we had done the goto nodata above. */ * Otherwise we had done the goto nodata above. */
if (req->private_bio == NULL) { if (req->private_bio == NULL) {
_req_mod(req, TO_BE_SENT); _req_mod(req, TO_BE_SENT, peer_device);
_req_mod(req, QUEUE_FOR_NET_READ); _req_mod(req, QUEUE_FOR_NET_READ, peer_device);
} else } else
no_remote = true; no_remote = true;
} }
...@@ -1397,7 +1402,7 @@ static void drbd_send_and_submit(struct drbd_device *device, struct drbd_request ...@@ -1397,7 +1402,7 @@ static void drbd_send_and_submit(struct drbd_device *device, struct drbd_request
req->pre_submit_jif = jiffies; req->pre_submit_jif = jiffies;
list_add_tail(&req->req_pending_local, list_add_tail(&req->req_pending_local,
&device->pending_completion[rw == WRITE]); &device->pending_completion[rw == WRITE]);
_req_mod(req, TO_BE_SUBMITTED); _req_mod(req, TO_BE_SUBMITTED, NULL);
/* but we need to give up the spinlock to submit */ /* but we need to give up the spinlock to submit */
submit_private_bio = true; submit_private_bio = true;
} else if (no_remote) { } else if (no_remote) {
......
...@@ -267,6 +267,7 @@ struct bio_and_error { ...@@ -267,6 +267,7 @@ struct bio_and_error {
extern void start_new_tl_epoch(struct drbd_connection *connection); extern void start_new_tl_epoch(struct drbd_connection *connection);
extern void drbd_req_destroy(struct kref *kref); extern void drbd_req_destroy(struct kref *kref);
extern int __req_mod(struct drbd_request *req, enum drbd_req_event what, extern int __req_mod(struct drbd_request *req, enum drbd_req_event what,
struct drbd_peer_device *peer_device,
struct bio_and_error *m); struct bio_and_error *m);
extern void complete_master_bio(struct drbd_device *device, extern void complete_master_bio(struct drbd_device *device,
struct bio_and_error *m); struct bio_and_error *m);
...@@ -280,14 +281,15 @@ extern void drbd_restart_request(struct drbd_request *req); ...@@ -280,14 +281,15 @@ extern void drbd_restart_request(struct drbd_request *req);
/* use this if you don't want to deal with calling complete_master_bio() /* use this if you don't want to deal with calling complete_master_bio()
* outside the spinlock, e.g. when walking some list on cleanup. */ * outside the spinlock, e.g. when walking some list on cleanup. */
static inline int _req_mod(struct drbd_request *req, enum drbd_req_event what) static inline int _req_mod(struct drbd_request *req, enum drbd_req_event what,
struct drbd_peer_device *peer_device)
{ {
struct drbd_device *device = req->device; struct drbd_device *device = req->device;
struct bio_and_error m; struct bio_and_error m;
int rv; int rv;
/* __req_mod possibly frees req, do not touch req after that! */ /* __req_mod possibly frees req, do not touch req after that! */
rv = __req_mod(req, what, &m); rv = __req_mod(req, what, peer_device, &m);
if (m.bio) if (m.bio)
complete_master_bio(device, &m); complete_master_bio(device, &m);
...@@ -299,7 +301,8 @@ static inline int _req_mod(struct drbd_request *req, enum drbd_req_event what) ...@@ -299,7 +301,8 @@ static inline int _req_mod(struct drbd_request *req, enum drbd_req_event what)
* of the lower level driver completion callback, so we need to * of the lower level driver completion callback, so we need to
* spin_lock_irqsave here. */ * spin_lock_irqsave here. */
static inline int req_mod(struct drbd_request *req, static inline int req_mod(struct drbd_request *req,
enum drbd_req_event what) enum drbd_req_event what,
struct drbd_peer_device *peer_device)
{ {
unsigned long flags; unsigned long flags;
struct drbd_device *device = req->device; struct drbd_device *device = req->device;
...@@ -307,7 +310,7 @@ static inline int req_mod(struct drbd_request *req, ...@@ -307,7 +310,7 @@ static inline int req_mod(struct drbd_request *req,
int rv; int rv;
spin_lock_irqsave(&device->resource->req_lock, flags); spin_lock_irqsave(&device->resource->req_lock, flags);
rv = __req_mod(req, what, &m); rv = __req_mod(req, what, peer_device, &m);
spin_unlock_irqrestore(&device->resource->req_lock, flags); spin_unlock_irqrestore(&device->resource->req_lock, flags);
if (m.bio) if (m.bio)
......
...@@ -276,7 +276,7 @@ void drbd_request_endio(struct bio *bio) ...@@ -276,7 +276,7 @@ void drbd_request_endio(struct bio *bio)
/* not req_mod(), we need irqsave here! */ /* not req_mod(), we need irqsave here! */
spin_lock_irqsave(&device->resource->req_lock, flags); spin_lock_irqsave(&device->resource->req_lock, flags);
__req_mod(req, what, &m); __req_mod(req, what, NULL, &m);
spin_unlock_irqrestore(&device->resource->req_lock, flags); spin_unlock_irqrestore(&device->resource->req_lock, flags);
put_ldev(device); put_ldev(device);
...@@ -1425,7 +1425,7 @@ int w_send_out_of_sync(struct drbd_work *w, int cancel) ...@@ -1425,7 +1425,7 @@ int w_send_out_of_sync(struct drbd_work *w, int cancel)
int err; int err;
if (unlikely(cancel)) { if (unlikely(cancel)) {
req_mod(req, SEND_CANCELED); req_mod(req, SEND_CANCELED, peer_device);
return 0; return 0;
} }
req->pre_send_jif = jiffies; req->pre_send_jif = jiffies;
...@@ -1437,7 +1437,7 @@ int w_send_out_of_sync(struct drbd_work *w, int cancel) ...@@ -1437,7 +1437,7 @@ int w_send_out_of_sync(struct drbd_work *w, int cancel)
maybe_send_barrier(connection, req->epoch); maybe_send_barrier(connection, req->epoch);
err = drbd_send_out_of_sync(peer_device, req); err = drbd_send_out_of_sync(peer_device, req);
req_mod(req, OOS_HANDED_TO_NETWORK); req_mod(req, OOS_HANDED_TO_NETWORK, peer_device);
return err; return err;
} }
...@@ -1457,7 +1457,7 @@ int w_send_dblock(struct drbd_work *w, int cancel) ...@@ -1457,7 +1457,7 @@ int w_send_dblock(struct drbd_work *w, int cancel)
int err; int err;
if (unlikely(cancel)) { if (unlikely(cancel)) {
req_mod(req, SEND_CANCELED); req_mod(req, SEND_CANCELED, peer_device);
return 0; return 0;
} }
req->pre_send_jif = jiffies; req->pre_send_jif = jiffies;
...@@ -1467,7 +1467,7 @@ int w_send_dblock(struct drbd_work *w, int cancel) ...@@ -1467,7 +1467,7 @@ int w_send_dblock(struct drbd_work *w, int cancel)
connection->send.current_epoch_writes++; connection->send.current_epoch_writes++;
err = drbd_send_dblock(peer_device, req); err = drbd_send_dblock(peer_device, req);
req_mod(req, err ? SEND_FAILED : HANDED_OVER_TO_NETWORK); req_mod(req, err ? SEND_FAILED : HANDED_OVER_TO_NETWORK, peer_device);
if (do_send_unplug && !err) if (do_send_unplug && !err)
pd_send_unplug_remote(peer_device); pd_send_unplug_remote(peer_device);
...@@ -1490,7 +1490,7 @@ int w_send_read_req(struct drbd_work *w, int cancel) ...@@ -1490,7 +1490,7 @@ int w_send_read_req(struct drbd_work *w, int cancel)
int err; int err;
if (unlikely(cancel)) { if (unlikely(cancel)) {
req_mod(req, SEND_CANCELED); req_mod(req, SEND_CANCELED, peer_device);
return 0; return 0;
} }
req->pre_send_jif = jiffies; req->pre_send_jif = jiffies;
...@@ -1502,7 +1502,7 @@ int w_send_read_req(struct drbd_work *w, int cancel) ...@@ -1502,7 +1502,7 @@ int w_send_read_req(struct drbd_work *w, int cancel)
err = drbd_send_drequest(peer_device, P_DATA_REQUEST, req->i.sector, req->i.size, err = drbd_send_drequest(peer_device, P_DATA_REQUEST, req->i.sector, req->i.size,
(unsigned long)req); (unsigned long)req);
req_mod(req, err ? SEND_FAILED : HANDED_OVER_TO_NETWORK); req_mod(req, err ? SEND_FAILED : HANDED_OVER_TO_NETWORK, peer_device);
if (do_send_unplug && !err) if (do_send_unplug && !err)
pd_send_unplug_remote(peer_device); pd_send_unplug_remote(peer_device);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment