dmaengine: stm32-mdma: Fix incomplete Hw descriptors allocator
Only 1 Hw Descriptor is allocated. Loop over required Hw descriptor for proper allocation. Signed-off-by: Pierre-Yves MORDRET <pierre-yves.mordret@st.com> Signed-off-by: Vinod Koul <vkoul@kernel.org>
This commit is contained in:
parent
ee6de9ac52
commit
bbb5a4e1e7
@ -252,13 +252,17 @@ struct stm32_mdma_hwdesc {
|
|||||||
u32 cmdr;
|
u32 cmdr;
|
||||||
} __aligned(64);
|
} __aligned(64);
|
||||||
|
|
||||||
|
struct stm32_mdma_desc_node {
|
||||||
|
struct stm32_mdma_hwdesc *hwdesc;
|
||||||
|
dma_addr_t hwdesc_phys;
|
||||||
|
};
|
||||||
|
|
||||||
struct stm32_mdma_desc {
|
struct stm32_mdma_desc {
|
||||||
struct virt_dma_desc vdesc;
|
struct virt_dma_desc vdesc;
|
||||||
u32 ccr;
|
u32 ccr;
|
||||||
struct stm32_mdma_hwdesc *hwdesc;
|
|
||||||
dma_addr_t hwdesc_phys;
|
|
||||||
bool cyclic;
|
bool cyclic;
|
||||||
u32 count;
|
u32 count;
|
||||||
|
struct stm32_mdma_desc_node node[];
|
||||||
};
|
};
|
||||||
|
|
||||||
struct stm32_mdma_chan {
|
struct stm32_mdma_chan {
|
||||||
@ -344,30 +348,42 @@ static struct stm32_mdma_desc *stm32_mdma_alloc_desc(
|
|||||||
struct stm32_mdma_chan *chan, u32 count)
|
struct stm32_mdma_chan *chan, u32 count)
|
||||||
{
|
{
|
||||||
struct stm32_mdma_desc *desc;
|
struct stm32_mdma_desc *desc;
|
||||||
|
int i;
|
||||||
|
|
||||||
desc = kzalloc(sizeof(*desc), GFP_NOWAIT);
|
desc = kzalloc(offsetof(typeof(*desc), node[count]), GFP_NOWAIT);
|
||||||
if (!desc)
|
if (!desc)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
desc->hwdesc = dma_pool_alloc(chan->desc_pool, GFP_NOWAIT,
|
for (i = 0; i < count; i++) {
|
||||||
&desc->hwdesc_phys);
|
desc->node[i].hwdesc =
|
||||||
if (!desc->hwdesc) {
|
dma_pool_alloc(chan->desc_pool, GFP_NOWAIT,
|
||||||
dev_err(chan2dev(chan), "Failed to allocate descriptor\n");
|
&desc->node[i].hwdesc_phys);
|
||||||
kfree(desc);
|
if (!desc->node[i].hwdesc)
|
||||||
return NULL;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
desc->count = count;
|
desc->count = count;
|
||||||
|
|
||||||
return desc;
|
return desc;
|
||||||
|
|
||||||
|
err:
|
||||||
|
dev_err(chan2dev(chan), "Failed to allocate descriptor\n");
|
||||||
|
while (--i >= 0)
|
||||||
|
dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
|
||||||
|
desc->node[i].hwdesc_phys);
|
||||||
|
kfree(desc);
|
||||||
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void stm32_mdma_desc_free(struct virt_dma_desc *vdesc)
|
static void stm32_mdma_desc_free(struct virt_dma_desc *vdesc)
|
||||||
{
|
{
|
||||||
struct stm32_mdma_desc *desc = to_stm32_mdma_desc(vdesc);
|
struct stm32_mdma_desc *desc = to_stm32_mdma_desc(vdesc);
|
||||||
struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan);
|
struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan);
|
||||||
|
int i;
|
||||||
|
|
||||||
dma_pool_free(chan->desc_pool, desc->hwdesc, desc->hwdesc_phys);
|
for (i = 0; i < desc->count; i++)
|
||||||
|
dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
|
||||||
|
desc->node[i].hwdesc_phys);
|
||||||
kfree(desc);
|
kfree(desc);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -666,18 +682,18 @@ static int stm32_mdma_set_xfer_param(struct stm32_mdma_chan *chan,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void stm32_mdma_dump_hwdesc(struct stm32_mdma_chan *chan,
|
static void stm32_mdma_dump_hwdesc(struct stm32_mdma_chan *chan,
|
||||||
struct stm32_mdma_hwdesc *hwdesc)
|
struct stm32_mdma_desc_node *node)
|
||||||
{
|
{
|
||||||
dev_dbg(chan2dev(chan), "hwdesc: 0x%p\n", hwdesc);
|
dev_dbg(chan2dev(chan), "hwdesc: %pad\n", &node->hwdesc_phys);
|
||||||
dev_dbg(chan2dev(chan), "CTCR: 0x%08x\n", hwdesc->ctcr);
|
dev_dbg(chan2dev(chan), "CTCR: 0x%08x\n", node->hwdesc->ctcr);
|
||||||
dev_dbg(chan2dev(chan), "CBNDTR: 0x%08x\n", hwdesc->cbndtr);
|
dev_dbg(chan2dev(chan), "CBNDTR: 0x%08x\n", node->hwdesc->cbndtr);
|
||||||
dev_dbg(chan2dev(chan), "CSAR: 0x%08x\n", hwdesc->csar);
|
dev_dbg(chan2dev(chan), "CSAR: 0x%08x\n", node->hwdesc->csar);
|
||||||
dev_dbg(chan2dev(chan), "CDAR: 0x%08x\n", hwdesc->cdar);
|
dev_dbg(chan2dev(chan), "CDAR: 0x%08x\n", node->hwdesc->cdar);
|
||||||
dev_dbg(chan2dev(chan), "CBRUR: 0x%08x\n", hwdesc->cbrur);
|
dev_dbg(chan2dev(chan), "CBRUR: 0x%08x\n", node->hwdesc->cbrur);
|
||||||
dev_dbg(chan2dev(chan), "CLAR: 0x%08x\n", hwdesc->clar);
|
dev_dbg(chan2dev(chan), "CLAR: 0x%08x\n", node->hwdesc->clar);
|
||||||
dev_dbg(chan2dev(chan), "CTBR: 0x%08x\n", hwdesc->ctbr);
|
dev_dbg(chan2dev(chan), "CTBR: 0x%08x\n", node->hwdesc->ctbr);
|
||||||
dev_dbg(chan2dev(chan), "CMAR: 0x%08x\n", hwdesc->cmar);
|
dev_dbg(chan2dev(chan), "CMAR: 0x%08x\n", node->hwdesc->cmar);
|
||||||
dev_dbg(chan2dev(chan), "CMDR: 0x%08x\n\n", hwdesc->cmdr);
|
dev_dbg(chan2dev(chan), "CMDR: 0x%08x\n\n", node->hwdesc->cmdr);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void stm32_mdma_setup_hwdesc(struct stm32_mdma_chan *chan,
|
static void stm32_mdma_setup_hwdesc(struct stm32_mdma_chan *chan,
|
||||||
@ -691,7 +707,7 @@ static void stm32_mdma_setup_hwdesc(struct stm32_mdma_chan *chan,
|
|||||||
struct stm32_mdma_hwdesc *hwdesc;
|
struct stm32_mdma_hwdesc *hwdesc;
|
||||||
u32 next = count + 1;
|
u32 next = count + 1;
|
||||||
|
|
||||||
hwdesc = &desc->hwdesc[count];
|
hwdesc = desc->node[count].hwdesc;
|
||||||
hwdesc->ctcr = ctcr;
|
hwdesc->ctcr = ctcr;
|
||||||
hwdesc->cbndtr &= ~(STM32_MDMA_CBNDTR_BRC_MK |
|
hwdesc->cbndtr &= ~(STM32_MDMA_CBNDTR_BRC_MK |
|
||||||
STM32_MDMA_CBNDTR_BRDUM |
|
STM32_MDMA_CBNDTR_BRDUM |
|
||||||
@ -701,19 +717,20 @@ static void stm32_mdma_setup_hwdesc(struct stm32_mdma_chan *chan,
|
|||||||
hwdesc->csar = src_addr;
|
hwdesc->csar = src_addr;
|
||||||
hwdesc->cdar = dst_addr;
|
hwdesc->cdar = dst_addr;
|
||||||
hwdesc->cbrur = 0;
|
hwdesc->cbrur = 0;
|
||||||
hwdesc->clar = desc->hwdesc_phys + next * sizeof(*hwdesc);
|
|
||||||
hwdesc->ctbr = ctbr;
|
hwdesc->ctbr = ctbr;
|
||||||
hwdesc->cmar = config->mask_addr;
|
hwdesc->cmar = config->mask_addr;
|
||||||
hwdesc->cmdr = config->mask_data;
|
hwdesc->cmdr = config->mask_data;
|
||||||
|
|
||||||
if (is_last) {
|
if (is_last) {
|
||||||
if (is_cyclic)
|
if (is_cyclic)
|
||||||
hwdesc->clar = desc->hwdesc_phys;
|
hwdesc->clar = desc->node[0].hwdesc_phys;
|
||||||
else
|
else
|
||||||
hwdesc->clar = 0;
|
hwdesc->clar = 0;
|
||||||
|
} else {
|
||||||
|
hwdesc->clar = desc->node[next].hwdesc_phys;
|
||||||
}
|
}
|
||||||
|
|
||||||
stm32_mdma_dump_hwdesc(chan, hwdesc);
|
stm32_mdma_dump_hwdesc(chan, &desc->node[count]);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int stm32_mdma_setup_xfer(struct stm32_mdma_chan *chan,
|
static int stm32_mdma_setup_xfer(struct stm32_mdma_chan *chan,
|
||||||
@ -777,7 +794,7 @@ stm32_mdma_prep_slave_sg(struct dma_chan *c, struct scatterlist *sgl,
|
|||||||
{
|
{
|
||||||
struct stm32_mdma_chan *chan = to_stm32_mdma_chan(c);
|
struct stm32_mdma_chan *chan = to_stm32_mdma_chan(c);
|
||||||
struct stm32_mdma_desc *desc;
|
struct stm32_mdma_desc *desc;
|
||||||
int ret;
|
int i, ret;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Once DMA is in setup cyclic mode the channel we cannot assign this
|
* Once DMA is in setup cyclic mode the channel we cannot assign this
|
||||||
@ -803,7 +820,9 @@ stm32_mdma_prep_slave_sg(struct dma_chan *c, struct scatterlist *sgl,
|
|||||||
return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
|
return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
|
||||||
|
|
||||||
xfer_setup_err:
|
xfer_setup_err:
|
||||||
dma_pool_free(chan->desc_pool, &desc->hwdesc, desc->hwdesc_phys);
|
for (i = 0; i < desc->count; i++)
|
||||||
|
dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
|
||||||
|
desc->node[i].hwdesc_phys);
|
||||||
kfree(desc);
|
kfree(desc);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
@ -892,7 +911,9 @@ stm32_mdma_prep_dma_cyclic(struct dma_chan *c, dma_addr_t buf_addr,
|
|||||||
return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
|
return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
|
||||||
|
|
||||||
xfer_setup_err:
|
xfer_setup_err:
|
||||||
dma_pool_free(chan->desc_pool, &desc->hwdesc, desc->hwdesc_phys);
|
for (i = 0; i < desc->count; i++)
|
||||||
|
dma_pool_free(chan->desc_pool, desc->node[i].hwdesc,
|
||||||
|
desc->node[i].hwdesc_phys);
|
||||||
kfree(desc);
|
kfree(desc);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
@ -1006,7 +1027,7 @@ stm32_mdma_prep_dma_memcpy(struct dma_chan *c, dma_addr_t dest, dma_addr_t src,
|
|||||||
ctcr |= STM32_MDMA_CTCR_PKE;
|
ctcr |= STM32_MDMA_CTCR_PKE;
|
||||||
|
|
||||||
/* Prepare hardware descriptor */
|
/* Prepare hardware descriptor */
|
||||||
hwdesc = desc->hwdesc;
|
hwdesc = desc->node[0].hwdesc;
|
||||||
hwdesc->ctcr = ctcr;
|
hwdesc->ctcr = ctcr;
|
||||||
hwdesc->cbndtr = cbndtr;
|
hwdesc->cbndtr = cbndtr;
|
||||||
hwdesc->csar = src;
|
hwdesc->csar = src;
|
||||||
@ -1017,7 +1038,7 @@ stm32_mdma_prep_dma_memcpy(struct dma_chan *c, dma_addr_t dest, dma_addr_t src,
|
|||||||
hwdesc->cmar = 0;
|
hwdesc->cmar = 0;
|
||||||
hwdesc->cmdr = 0;
|
hwdesc->cmdr = 0;
|
||||||
|
|
||||||
stm32_mdma_dump_hwdesc(chan, hwdesc);
|
stm32_mdma_dump_hwdesc(chan, &desc->node[0]);
|
||||||
} else {
|
} else {
|
||||||
/* Setup a LLI transfer */
|
/* Setup a LLI transfer */
|
||||||
ctcr |= STM32_MDMA_CTCR_TRGM(STM32_MDMA_LINKED_LIST) |
|
ctcr |= STM32_MDMA_CTCR_TRGM(STM32_MDMA_LINKED_LIST) |
|
||||||
@ -1117,7 +1138,7 @@ static void stm32_mdma_start_transfer(struct stm32_mdma_chan *chan)
|
|||||||
}
|
}
|
||||||
|
|
||||||
chan->desc = to_stm32_mdma_desc(vdesc);
|
chan->desc = to_stm32_mdma_desc(vdesc);
|
||||||
hwdesc = chan->desc->hwdesc;
|
hwdesc = chan->desc->node[0].hwdesc;
|
||||||
chan->curr_hwdesc = 0;
|
chan->curr_hwdesc = 0;
|
||||||
|
|
||||||
stm32_mdma_write(dmadev, STM32_MDMA_CCR(id), chan->desc->ccr);
|
stm32_mdma_write(dmadev, STM32_MDMA_CCR(id), chan->desc->ccr);
|
||||||
@ -1195,7 +1216,7 @@ static int stm32_mdma_resume(struct dma_chan *c)
|
|||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
u32 status, reg;
|
u32 status, reg;
|
||||||
|
|
||||||
hwdesc = &chan->desc->hwdesc[chan->curr_hwdesc];
|
hwdesc = chan->desc->node[chan->curr_hwdesc].hwdesc;
|
||||||
|
|
||||||
spin_lock_irqsave(&chan->vchan.lock, flags);
|
spin_lock_irqsave(&chan->vchan.lock, flags);
|
||||||
|
|
||||||
@ -1265,13 +1286,13 @@ static size_t stm32_mdma_desc_residue(struct stm32_mdma_chan *chan,
|
|||||||
u32 curr_hwdesc)
|
u32 curr_hwdesc)
|
||||||
{
|
{
|
||||||
struct stm32_mdma_device *dmadev = stm32_mdma_get_dev(chan);
|
struct stm32_mdma_device *dmadev = stm32_mdma_get_dev(chan);
|
||||||
|
struct stm32_mdma_hwdesc *hwdesc = desc->node[0].hwdesc;
|
||||||
u32 cbndtr, residue, modulo, burst_size;
|
u32 cbndtr, residue, modulo, burst_size;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
residue = 0;
|
residue = 0;
|
||||||
for (i = curr_hwdesc + 1; i < desc->count; i++) {
|
for (i = curr_hwdesc + 1; i < desc->count; i++) {
|
||||||
struct stm32_mdma_hwdesc *hwdesc = &desc->hwdesc[i];
|
hwdesc = desc->node[i].hwdesc;
|
||||||
|
|
||||||
residue += STM32_MDMA_CBNDTR_BNDT(hwdesc->cbndtr);
|
residue += STM32_MDMA_CBNDTR_BNDT(hwdesc->cbndtr);
|
||||||
}
|
}
|
||||||
cbndtr = stm32_mdma_read(dmadev, STM32_MDMA_CBNDTR(chan->id));
|
cbndtr = stm32_mdma_read(dmadev, STM32_MDMA_CBNDTR(chan->id));
|
||||||
|
Loading…
x
Reference in New Issue
Block a user