Commit 5e2078b2 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'for-linus' of git://git.kernel.dk/linux-block

Pull misc block fixes from Jens Axboe:
 "Stuff that got collected after the merge window opened.  This
  contains:

   - NVMe:
        - Fix for non-striped transfer size setting for NVMe from
          Sathyavathi.
        - (Some) support for the weird Apple nvme controller in the
          macbooks. From Stephan Günther.

   - The error value leak for dax from Al.

   - A few minor blk-mq tweaks from me.

   - Add the new linux-block@vger.kernel.org mailing list to the
     MAINTAINERS file.

   - Discard fix for brd, from Jan.

   - A kerneldoc warning for block core from Randy.

   - An older fix from Vivek, converting a WARN_ON() to a rate limited
     printk when a device is hot removed with dirty inodes"

* 'for-linus' of git://git.kernel.dk/linux-block:
  block: don't hardcode blk_qc_t -> tag mask
  dax_io(): don't let non-error value escape via retval instead of EFAULT
  block: fix blk-core.c kernel-doc warning
  fs/block_dev.c: Remove WARN_ON() when inode writeback fails
  NVMe: add support for Apple NVMe controller
  NVMe: use split lo_hi_{read,write}q
  blk-mq: mark __blk_mq_complete_request() static
  MAINTAINERS: add reference to new linux-block list
  NVMe: Increase the max transfer size when mdts is 0
  brd: Refuse improperly aligned discard requests
parents a4d8c7c9 e3a7a3bf
...@@ -2210,6 +2210,7 @@ F: drivers/leds/leds-blinkm.c ...@@ -2210,6 +2210,7 @@ F: drivers/leds/leds-blinkm.c
BLOCK LAYER BLOCK LAYER
M: Jens Axboe <axboe@kernel.dk> M: Jens Axboe <axboe@kernel.dk>
L: linux-block@vger.kernel.org
T: git git://git.kernel.org/pub/scm/linux/kernel/git/axboe/linux-block.git T: git git://git.kernel.org/pub/scm/linux/kernel/git/axboe/linux-block.git
S: Maintained S: Maintained
F: block/ F: block/
......
...@@ -1575,6 +1575,9 @@ bool bio_attempt_front_merge(struct request_queue *q, struct request *req, ...@@ -1575,6 +1575,9 @@ bool bio_attempt_front_merge(struct request_queue *q, struct request *req,
* @q: request_queue new bio is being queued at * @q: request_queue new bio is being queued at
* @bio: new bio being queued * @bio: new bio being queued
* @request_count: out parameter for number of traversed plugged requests * @request_count: out parameter for number of traversed plugged requests
* @same_queue_rq: pointer to &struct request that gets filled in when
* another request associated with @q is found on the plug list
* (optional, may be %NULL)
* *
* Determine whether @bio being queued on @q can be merged with a request * Determine whether @bio being queued on @q can be merged with a request
* on %current's plugged list. Returns %true if merge was successful, * on %current's plugged list. Returns %true if merge was successful,
......
...@@ -358,7 +358,7 @@ static void blk_mq_ipi_complete_request(struct request *rq) ...@@ -358,7 +358,7 @@ static void blk_mq_ipi_complete_request(struct request *rq)
put_cpu(); put_cpu();
} }
void __blk_mq_complete_request(struct request *rq) static void __blk_mq_complete_request(struct request *rq)
{ {
struct request_queue *q = rq->q; struct request_queue *q = rq->q;
......
...@@ -25,7 +25,6 @@ struct blk_mq_ctx { ...@@ -25,7 +25,6 @@ struct blk_mq_ctx {
struct kobject kobj; struct kobject kobj;
} ____cacheline_aligned_in_smp; } ____cacheline_aligned_in_smp;
void __blk_mq_complete_request(struct request *rq);
void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async); void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async);
void blk_mq_freeze_queue(struct request_queue *q); void blk_mq_freeze_queue(struct request_queue *q);
void blk_mq_free_queue(struct request_queue *q); void blk_mq_free_queue(struct request_queue *q);
......
...@@ -337,6 +337,9 @@ static blk_qc_t brd_make_request(struct request_queue *q, struct bio *bio) ...@@ -337,6 +337,9 @@ static blk_qc_t brd_make_request(struct request_queue *q, struct bio *bio)
goto io_error; goto io_error;
if (unlikely(bio->bi_rw & REQ_DISCARD)) { if (unlikely(bio->bi_rw & REQ_DISCARD)) {
if (sector & ((PAGE_SIZE >> SECTOR_SHIFT) - 1) ||
bio->bi_iter.bi_size & PAGE_MASK)
goto io_error;
discard_from_brd(brd, sector, bio->bi_iter.bi_size); discard_from_brd(brd, sector, bio->bi_iter.bi_size);
goto out; goto out;
} }
......
...@@ -1725,7 +1725,7 @@ static int nvme_configure_admin_queue(struct nvme_dev *dev) ...@@ -1725,7 +1725,7 @@ static int nvme_configure_admin_queue(struct nvme_dev *dev)
{ {
int result; int result;
u32 aqa; u32 aqa;
u64 cap = readq(&dev->bar->cap); u64 cap = lo_hi_readq(&dev->bar->cap);
struct nvme_queue *nvmeq; struct nvme_queue *nvmeq;
unsigned page_shift = PAGE_SHIFT; unsigned page_shift = PAGE_SHIFT;
unsigned dev_page_min = NVME_CAP_MPSMIN(cap) + 12; unsigned dev_page_min = NVME_CAP_MPSMIN(cap) + 12;
...@@ -1774,8 +1774,8 @@ static int nvme_configure_admin_queue(struct nvme_dev *dev) ...@@ -1774,8 +1774,8 @@ static int nvme_configure_admin_queue(struct nvme_dev *dev)
dev->ctrl_config |= NVME_CC_IOSQES | NVME_CC_IOCQES; dev->ctrl_config |= NVME_CC_IOSQES | NVME_CC_IOCQES;
writel(aqa, &dev->bar->aqa); writel(aqa, &dev->bar->aqa);
writeq(nvmeq->sq_dma_addr, &dev->bar->asq); lo_hi_writeq(nvmeq->sq_dma_addr, &dev->bar->asq);
writeq(nvmeq->cq_dma_addr, &dev->bar->acq); lo_hi_writeq(nvmeq->cq_dma_addr, &dev->bar->acq);
result = nvme_enable_ctrl(dev, cap); result = nvme_enable_ctrl(dev, cap);
if (result) if (result)
...@@ -2606,7 +2606,7 @@ static int nvme_dev_add(struct nvme_dev *dev) ...@@ -2606,7 +2606,7 @@ static int nvme_dev_add(struct nvme_dev *dev)
struct pci_dev *pdev = to_pci_dev(dev->dev); struct pci_dev *pdev = to_pci_dev(dev->dev);
int res; int res;
struct nvme_id_ctrl *ctrl; struct nvme_id_ctrl *ctrl;
int shift = NVME_CAP_MPSMIN(readq(&dev->bar->cap)) + 12; int shift = NVME_CAP_MPSMIN(lo_hi_readq(&dev->bar->cap)) + 12;
res = nvme_identify_ctrl(dev, &ctrl); res = nvme_identify_ctrl(dev, &ctrl);
if (res) { if (res) {
...@@ -2622,6 +2622,8 @@ static int nvme_dev_add(struct nvme_dev *dev) ...@@ -2622,6 +2622,8 @@ static int nvme_dev_add(struct nvme_dev *dev)
memcpy(dev->firmware_rev, ctrl->fr, sizeof(ctrl->fr)); memcpy(dev->firmware_rev, ctrl->fr, sizeof(ctrl->fr));
if (ctrl->mdts) if (ctrl->mdts)
dev->max_hw_sectors = 1 << (ctrl->mdts + shift - 9); dev->max_hw_sectors = 1 << (ctrl->mdts + shift - 9);
else
dev->max_hw_sectors = UINT_MAX;
if ((pdev->vendor == PCI_VENDOR_ID_INTEL) && if ((pdev->vendor == PCI_VENDOR_ID_INTEL) &&
(pdev->device == 0x0953) && ctrl->vs[3]) { (pdev->device == 0x0953) && ctrl->vs[3]) {
unsigned int max_hw_sectors; unsigned int max_hw_sectors;
...@@ -2695,7 +2697,7 @@ static int nvme_dev_map(struct nvme_dev *dev) ...@@ -2695,7 +2697,7 @@ static int nvme_dev_map(struct nvme_dev *dev)
goto unmap; goto unmap;
} }
cap = readq(&dev->bar->cap); cap = lo_hi_readq(&dev->bar->cap);
dev->q_depth = min_t(int, NVME_CAP_MQES(cap) + 1, NVME_Q_DEPTH); dev->q_depth = min_t(int, NVME_CAP_MQES(cap) + 1, NVME_Q_DEPTH);
dev->db_stride = 1 << NVME_CAP_STRIDE(cap); dev->db_stride = 1 << NVME_CAP_STRIDE(cap);
dev->dbs = ((void __iomem *)dev->bar) + 4096; dev->dbs = ((void __iomem *)dev->bar) + 4096;
...@@ -2758,7 +2760,7 @@ static void nvme_wait_dq(struct nvme_delq_ctx *dq, struct nvme_dev *dev) ...@@ -2758,7 +2760,7 @@ static void nvme_wait_dq(struct nvme_delq_ctx *dq, struct nvme_dev *dev)
* queues than admin tags. * queues than admin tags.
*/ */
set_current_state(TASK_RUNNING); set_current_state(TASK_RUNNING);
nvme_disable_ctrl(dev, readq(&dev->bar->cap)); nvme_disable_ctrl(dev, lo_hi_readq(&dev->bar->cap));
nvme_clear_queue(dev->queues[0]); nvme_clear_queue(dev->queues[0]);
flush_kthread_worker(dq->worker); flush_kthread_worker(dq->worker);
nvme_disable_queue(dev, 0); nvme_disable_queue(dev, 0);
...@@ -3401,6 +3403,7 @@ static const struct pci_error_handlers nvme_err_handler = { ...@@ -3401,6 +3403,7 @@ static const struct pci_error_handlers nvme_err_handler = {
static const struct pci_device_id nvme_id_table[] = { static const struct pci_device_id nvme_id_table[] = {
{ PCI_DEVICE_CLASS(PCI_CLASS_STORAGE_EXPRESS, 0xffffff) }, { PCI_DEVICE_CLASS(PCI_CLASS_STORAGE_EXPRESS, 0xffffff) },
{ PCI_DEVICE(PCI_VENDOR_ID_APPLE, 0x2001) },
{ 0, } { 0, }
}; };
MODULE_DEVICE_TABLE(pci, nvme_id_table); MODULE_DEVICE_TABLE(pci, nvme_id_table);
......
...@@ -50,12 +50,21 @@ struct block_device *I_BDEV(struct inode *inode) ...@@ -50,12 +50,21 @@ struct block_device *I_BDEV(struct inode *inode)
} }
EXPORT_SYMBOL(I_BDEV); EXPORT_SYMBOL(I_BDEV);
static void bdev_write_inode(struct inode *inode) static void bdev_write_inode(struct block_device *bdev)
{ {
struct inode *inode = bdev->bd_inode;
int ret;
spin_lock(&inode->i_lock); spin_lock(&inode->i_lock);
while (inode->i_state & I_DIRTY) { while (inode->i_state & I_DIRTY) {
spin_unlock(&inode->i_lock); spin_unlock(&inode->i_lock);
WARN_ON_ONCE(write_inode_now(inode, true)); ret = write_inode_now(inode, true);
if (ret) {
char name[BDEVNAME_SIZE];
pr_warn_ratelimited("VFS: Dirty inode writeback failed "
"for block device %s (err=%d).\n",
bdevname(bdev, name), ret);
}
spin_lock(&inode->i_lock); spin_lock(&inode->i_lock);
} }
spin_unlock(&inode->i_lock); spin_unlock(&inode->i_lock);
...@@ -1504,7 +1513,7 @@ static void __blkdev_put(struct block_device *bdev, fmode_t mode, int for_part) ...@@ -1504,7 +1513,7 @@ static void __blkdev_put(struct block_device *bdev, fmode_t mode, int for_part)
* ->release can cause the queue to disappear, so flush all * ->release can cause the queue to disappear, so flush all
* dirty data before. * dirty data before.
*/ */
bdev_write_inode(bdev->bd_inode); bdev_write_inode(bdev);
} }
if (bdev->bd_contains == bdev) { if (bdev->bd_contains == bdev) {
if (disk->fops->release) if (disk->fops->release)
......
...@@ -174,8 +174,10 @@ static ssize_t dax_io(struct inode *inode, struct iov_iter *iter, ...@@ -174,8 +174,10 @@ static ssize_t dax_io(struct inode *inode, struct iov_iter *iter,
else else
len = iov_iter_zero(max - pos, iter); len = iov_iter_zero(max - pos, iter);
if (!len) if (!len) {
retval = -EFAULT;
break; break;
}
pos += len; pos += len;
addr += len; addr += len;
......
...@@ -265,7 +265,7 @@ static inline unsigned int blk_qc_t_to_queue_num(blk_qc_t cookie) ...@@ -265,7 +265,7 @@ static inline unsigned int blk_qc_t_to_queue_num(blk_qc_t cookie)
static inline unsigned int blk_qc_t_to_tag(blk_qc_t cookie) static inline unsigned int blk_qc_t_to_tag(blk_qc_t cookie)
{ {
return cookie & 0xffff; return cookie & ((1u << BLK_QC_T_SHIFT) - 1);
} }
#endif /* __LINUX_BLK_TYPES_H */ #endif /* __LINUX_BLK_TYPES_H */
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment