On 4/23/21 7:46 PM, Mark Brown wrote:
On Fri, Apr 23, 2021 at 09:54:38PM +0530, Vijendar Mukunda wrote:
For CZ/StoneyRidge platforms, ACP DMA between ACP SRAM and I2S FIFO should be stopped before stopping I2S Controller DMA.
When DMA is progressing and stop request received, while DMA transfer ongoing between ACP SRAM and I2S FIFO, Stopping I2S DMA prior to ACP DMA stop resulting DMA Channel stop failure.
This again... copying in Peter for the sequencing discussion. If we need to do this I'm not convinced that bodging it in the driver is a good idea, and especially not deferring it outside of the trigger operation - for example on a suspend or pause we won't actually do a shutdown() so the trigger will end up not happening which seems like it may cause problems.
It will certainly leave the i2s running and can lead to hard to explain issues
We'd probably be better off with the core knowing what's going on and being able to reorder the callbacks although designing an interface for that seems a bit annoying.
I agree, it would be better to have some sort of flag which tells the core that there is an integration issue between the DMA and peripheral. I believe this is only affecting playback?
This issue can't be fixed in ACP DMA driver due to design constraint.
What is the design constraint here - can't we fix the design? Or is it a hardware design constraint (presumably broken signalling between the I2S and DMA blocks)?
From the description my guess is that stop on the DMA want to flush it's
FIFO (complete the in progress packet, segment). Since the peripheral is stopped it will not pull in more data -> the DMA will time out internally.
The question: how the ACP DMA driver's terminate_all is implemented? It can not really wait for the DMA to stop, we can not use terminate_all_sync() in trigger, it must just set a stop bit and make sure at synchronize() time that it has stopped, right?
What happens if the time between the DMA stop and the DAI stop is less then it would take to flush the DMA FIFO? You would have the same issue, but in a rather hard to reproducible way?
As sidenote: TI's k3-udma initially had similar issue at the design phase on the playback side which got solved by a flush bit on the channel to detach it from the peripheral and set it to free run to drain w/o peripheral. On capture however I need to push a dummy 'drain' packet to flush out the data from the DMA (if the stop happens when we did not have active descriptor on the channel).
With a flag to reorder the DMA/DAI stop sequence it might work most of the time, but imho it is going to introduce a nasty time-bomb of failure. Also your DAI will underflow instead of the DMA error.