Commit 49bcc174 authored by Hugues Fruchet's avatar Hugues Fruchet Committed by Mauro Carvalho Chehab

media: stm32-dcmi: revisit buffer list management

Cleanup "active" field usage and enhance list management
to avoid exceptions when releasing buffers on error or
stopping streaming.
Signed-off-by: default avatarHugues Fruchet <hugues.fruchet@st.com>
Signed-off-by: default avatarHans Verkuil <hans.verkuil@cisco.com>
Signed-off-by: default avatarMauro Carvalho Chehab <mchehab+samsung@kernel.org>
parent 2d494d4a
...@@ -190,7 +190,7 @@ static inline void reg_clear(void __iomem *base, u32 reg, u32 mask) ...@@ -190,7 +190,7 @@ static inline void reg_clear(void __iomem *base, u32 reg, u32 mask)
reg_write(base, reg, reg_read(base, reg) & ~mask); reg_write(base, reg, reg_read(base, reg) & ~mask);
} }
static int dcmi_start_capture(struct stm32_dcmi *dcmi); static int dcmi_start_capture(struct stm32_dcmi *dcmi, struct dcmi_buf *buf);
static void dcmi_buffer_done(struct stm32_dcmi *dcmi, static void dcmi_buffer_done(struct stm32_dcmi *dcmi,
struct dcmi_buf *buf, struct dcmi_buf *buf,
...@@ -202,6 +202,8 @@ static void dcmi_buffer_done(struct stm32_dcmi *dcmi, ...@@ -202,6 +202,8 @@ static void dcmi_buffer_done(struct stm32_dcmi *dcmi,
if (!buf) if (!buf)
return; return;
list_del_init(&buf->list);
vbuf = &buf->vb; vbuf = &buf->vb;
vbuf->sequence = dcmi->sequence++; vbuf->sequence = dcmi->sequence++;
...@@ -219,6 +221,8 @@ static void dcmi_buffer_done(struct stm32_dcmi *dcmi, ...@@ -219,6 +221,8 @@ static void dcmi_buffer_done(struct stm32_dcmi *dcmi,
static int dcmi_restart_capture(struct stm32_dcmi *dcmi) static int dcmi_restart_capture(struct stm32_dcmi *dcmi)
{ {
struct dcmi_buf *buf;
spin_lock_irq(&dcmi->irqlock); spin_lock_irq(&dcmi->irqlock);
if (dcmi->state != RUNNING) { if (dcmi->state != RUNNING) {
...@@ -229,19 +233,16 @@ static int dcmi_restart_capture(struct stm32_dcmi *dcmi) ...@@ -229,19 +233,16 @@ static int dcmi_restart_capture(struct stm32_dcmi *dcmi)
/* Restart a new DMA transfer with next buffer */ /* Restart a new DMA transfer with next buffer */
if (list_empty(&dcmi->buffers)) { if (list_empty(&dcmi->buffers)) {
dev_dbg(dcmi->dev, "Capture restart is deferred to next buffer queueing\n"); dev_dbg(dcmi->dev, "Capture restart is deferred to next buffer queueing\n");
dcmi->active = NULL;
dcmi->state = WAIT_FOR_BUFFER; dcmi->state = WAIT_FOR_BUFFER;
spin_unlock_irq(&dcmi->irqlock); spin_unlock_irq(&dcmi->irqlock);
return 0; return 0;
} }
buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list);
dcmi->active = list_entry(dcmi->buffers.next, dcmi->active = buf;
struct dcmi_buf, list);
list_del_init(&dcmi->active->list);
spin_unlock_irq(&dcmi->irqlock); spin_unlock_irq(&dcmi->irqlock);
return dcmi_start_capture(dcmi); return dcmi_start_capture(dcmi, buf);
} }
static void dcmi_dma_callback(void *param) static void dcmi_dma_callback(void *param)
...@@ -251,6 +252,8 @@ static void dcmi_dma_callback(void *param) ...@@ -251,6 +252,8 @@ static void dcmi_dma_callback(void *param)
enum dma_status status; enum dma_status status;
struct dcmi_buf *buf = dcmi->active; struct dcmi_buf *buf = dcmi->active;
spin_lock_irq(&dcmi->irqlock);
/* Check DMA status */ /* Check DMA status */
status = dmaengine_tx_status(dcmi->dma_chan, dcmi->dma_cookie, &state); status = dmaengine_tx_status(dcmi->dma_chan, dcmi->dma_cookie, &state);
...@@ -273,15 +276,19 @@ static void dcmi_dma_callback(void *param) ...@@ -273,15 +276,19 @@ static void dcmi_dma_callback(void *param)
/* Return buffer to V4L2 */ /* Return buffer to V4L2 */
dcmi_buffer_done(dcmi, buf, buf->size, 0); dcmi_buffer_done(dcmi, buf, buf->size, 0);
spin_unlock_irq(&dcmi->irqlock);
/* Restart capture */ /* Restart capture */
if (dcmi_restart_capture(dcmi)) if (dcmi_restart_capture(dcmi))
dev_err(dcmi->dev, "%s: Cannot restart capture on DMA complete\n", dev_err(dcmi->dev, "%s: Cannot restart capture on DMA complete\n",
__func__); __func__);
break; return;
default: default:
dev_err(dcmi->dev, "%s: Received unknown status\n", __func__); dev_err(dcmi->dev, "%s: Received unknown status\n", __func__);
break; break;
} }
spin_unlock_irq(&dcmi->irqlock);
} }
static int dcmi_start_dma(struct stm32_dcmi *dcmi, static int dcmi_start_dma(struct stm32_dcmi *dcmi,
...@@ -333,10 +340,9 @@ static int dcmi_start_dma(struct stm32_dcmi *dcmi, ...@@ -333,10 +340,9 @@ static int dcmi_start_dma(struct stm32_dcmi *dcmi,
return 0; return 0;
} }
static int dcmi_start_capture(struct stm32_dcmi *dcmi) static int dcmi_start_capture(struct stm32_dcmi *dcmi, struct dcmi_buf *buf)
{ {
int ret; int ret;
struct dcmi_buf *buf = dcmi->active;
if (!buf) if (!buf)
return -EINVAL; return -EINVAL;
...@@ -490,8 +496,6 @@ static int dcmi_queue_setup(struct vb2_queue *vq, ...@@ -490,8 +496,6 @@ static int dcmi_queue_setup(struct vb2_queue *vq,
*nplanes = 1; *nplanes = 1;
sizes[0] = size; sizes[0] = size;
dcmi->active = NULL;
dev_dbg(dcmi->dev, "Setup queue, count=%d, size=%d\n", dev_dbg(dcmi->dev, "Setup queue, count=%d, size=%d\n",
*nbuffers, size); *nbuffers, size);
...@@ -549,23 +553,24 @@ static void dcmi_buf_queue(struct vb2_buffer *vb) ...@@ -549,23 +553,24 @@ static void dcmi_buf_queue(struct vb2_buffer *vb)
spin_lock_irq(&dcmi->irqlock); spin_lock_irq(&dcmi->irqlock);
dcmi->active = buf; /* Enqueue to video buffers list */
list_add_tail(&buf->list, &dcmi->buffers);
if (dcmi->state == WAIT_FOR_BUFFER) { if (dcmi->state == WAIT_FOR_BUFFER) {
dcmi->state = RUNNING; dcmi->state = RUNNING;
dcmi->active = buf;
dev_dbg(dcmi->dev, "Starting capture on buffer[%d] queued\n", dev_dbg(dcmi->dev, "Starting capture on buffer[%d] queued\n",
buf->vb.vb2_buf.index); buf->vb.vb2_buf.index);
spin_unlock_irq(&dcmi->irqlock); spin_unlock_irq(&dcmi->irqlock);
if (dcmi_start_capture(dcmi)) if (dcmi_start_capture(dcmi, buf))
dev_err(dcmi->dev, "%s: Cannot restart capture on overflow or error\n", dev_err(dcmi->dev, "%s: Cannot restart capture on overflow or error\n",
__func__); __func__);
} else { return;
/* Enqueue to video buffers list */
list_add_tail(&buf->list, &dcmi->buffers);
spin_unlock_irq(&dcmi->irqlock);
} }
spin_unlock_irq(&dcmi->irqlock);
} }
static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count) static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count)
...@@ -637,7 +642,6 @@ static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count) ...@@ -637,7 +642,6 @@ static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count)
dcmi->errors_count = 0; dcmi->errors_count = 0;
dcmi->overrun_count = 0; dcmi->overrun_count = 0;
dcmi->buffers_count = 0; dcmi->buffers_count = 0;
dcmi->active = NULL;
/* /*
* Start transfer if at least one buffer has been queued, * Start transfer if at least one buffer has been queued,
...@@ -650,15 +654,15 @@ static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count) ...@@ -650,15 +654,15 @@ static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count)
return 0; return 0;
} }
dcmi->active = list_entry(dcmi->buffers.next, struct dcmi_buf, list); buf = list_entry(dcmi->buffers.next, struct dcmi_buf, list);
list_del_init(&dcmi->active->list); dcmi->active = buf;
dev_dbg(dcmi->dev, "Start streaming, starting capture\n");
dcmi->state = RUNNING; dcmi->state = RUNNING;
dev_dbg(dcmi->dev, "Start streaming, starting capture\n");
spin_unlock_irq(&dcmi->irqlock); spin_unlock_irq(&dcmi->irqlock);
ret = dcmi_start_capture(dcmi); ret = dcmi_start_capture(dcmi, buf);
if (ret) { if (ret) {
dev_err(dcmi->dev, "%s: Start streaming failed, cannot start capture\n", dev_err(dcmi->dev, "%s: Start streaming failed, cannot start capture\n",
__func__); __func__);
...@@ -682,15 +686,11 @@ static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count) ...@@ -682,15 +686,11 @@ static int dcmi_start_streaming(struct vb2_queue *vq, unsigned int count)
* Return all buffers to vb2 in QUEUED state. * Return all buffers to vb2 in QUEUED state.
* This will give ownership back to userspace * This will give ownership back to userspace
*/ */
if (dcmi->active) {
buf = dcmi->active;
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_QUEUED);
dcmi->active = NULL;
}
list_for_each_entry_safe(buf, node, &dcmi->buffers, list) { list_for_each_entry_safe(buf, node, &dcmi->buffers, list) {
list_del_init(&buf->list); list_del_init(&buf->list);
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_QUEUED); vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_QUEUED);
} }
dcmi->active = NULL;
spin_unlock_irq(&dcmi->irqlock); spin_unlock_irq(&dcmi->irqlock);
return ret; return ret;
...@@ -732,16 +732,13 @@ static void dcmi_stop_streaming(struct vb2_queue *vq) ...@@ -732,16 +732,13 @@ static void dcmi_stop_streaming(struct vb2_queue *vq)
} }
/* Return all queued buffers to vb2 in ERROR state */ /* Return all queued buffers to vb2 in ERROR state */
if (dcmi->active) {
buf = dcmi->active;
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
dcmi->active = NULL;
}
list_for_each_entry_safe(buf, node, &dcmi->buffers, list) { list_for_each_entry_safe(buf, node, &dcmi->buffers, list) {
list_del_init(&buf->list); list_del_init(&buf->list);
vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR); vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR);
} }
dcmi->active = NULL;
spin_unlock_irq(&dcmi->irqlock); spin_unlock_irq(&dcmi->irqlock);
/* Stop all pending DMA operations */ /* Stop all pending DMA operations */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment