[PATCH v5 2/3] dmaengine: amlogic: Add general DMA driver for A9

Xianwei Zhao xianwei.zhao at amlogic.com
Wed Mar 4 21:21:27 PST 2026



On 2026/3/5 12:17, Frank Li wrote:
> On Thu, Mar 05, 2026 at 11:35:15AM +0800, Xianwei Zhao wrote:
>> Hi Frank,
>>     Thanks for your review.
>>
>> On 2026/3/4 23:48, Frank Li wrote:
>>> On Wed, Mar 04, 2026 at 06:14:13AM +0000, Xianwei Zhao wrote:
>>>> Amlogic A9 SoCs include a general-purpose DMA controller that can be used
>>>> by multiple peripherals, such as I2C PIO and I3C. Each peripheral group
>>>> is associated with a dedicated DMA channel in hardware.
>>>>
>>>> Signed-off-by: Xianwei Zhao<xianwei.zhao at amlogic.com>
>>>> ---
>>>>    drivers/dma/Kconfig       |   9 +
>>>>    drivers/dma/Makefile      |   1 +
>>>>    drivers/dma/amlogic-dma.c | 585 ++++++++++++++++++++++++++++++++++++++++++++++
>>>>    3 files changed, 595 insertions(+)
>>>>
>>> ...
>>>> +
>>>> +static int aml_dma_alloc_chan_resources(struct dma_chan *chan)
>>>> +{
>>>> +     struct aml_dma_chan *aml_chan = to_aml_dma_chan(chan);
>>>> +     struct aml_dma_dev *aml_dma = aml_chan->aml_dma;
>>>> +     size_t size = size_mul(sizeof(struct aml_dma_sg_link), DMA_MAX_LINK);
>>>> +
>>>> +     aml_chan->sg_link = dma_alloc_coherent(aml_dma->dma_device.dev, size,
>>>> +                                            &aml_chan->sg_link_phys, GFP_KERNEL);
>>>> +     if (!aml_chan->sg_link)
>>>> +             return  -ENOMEM;
>>>> +
>>>> +     /* offset is the same RCH_CFG and WCH_CFG */
>>>> +     regmap_update_bits(aml_dma->regmap, aml_chan->reg_offs + RCH_CFG, CFG_CLEAR, CFG_CLEAR);
>>> regmap_set_bits()
>>>
>>>> +     aml_chan->status = DMA_COMPLETE;
>>>> +     dma_async_tx_descriptor_init(&aml_chan->desc, chan);
>>>> +     aml_chan->desc.tx_submit = aml_dma_tx_submit;
>>>> +     regmap_update_bits(aml_dma->regmap, aml_chan->reg_offs + RCH_CFG, CFG_CLEAR, 0);
>>> regmap_clear_bits();
>>>
>>>> +
>>>> +     return 0;
>>>> +}
>>>> +
>>>> +static void aml_dma_free_chan_resources(struct dma_chan *chan)
>>>> +{
>>>> +     struct aml_dma_chan *aml_chan = to_aml_dma_chan(chan);
>>>> +     struct aml_dma_dev *aml_dma = aml_chan->aml_dma;
>>>> +
>>>> +     aml_chan->status = DMA_COMPLETE;
>>>> +     dma_free_coherent(aml_dma->dma_device.dev,
>>>> +                       sizeof(struct aml_dma_sg_link) * DMA_MAX_LINK,
>>>> +                       aml_chan->sg_link, aml_chan->sg_link_phys);
>>>> +}
>>>> +
>>> ...
>>>> +
>>>> +static struct dma_async_tx_descriptor *aml_dma_prep_slave_sg
>>>> +             (struct dma_chan *chan, struct scatterlist *sgl,
>>>> +             unsigned int sg_len, enum dma_transfer_direction direction,
>>>> +             unsigned long flags, void *context)
>>>> +{
>>>> +     struct aml_dma_chan *aml_chan = to_aml_dma_chan(chan);
>>>> +     struct aml_dma_dev *aml_dma = aml_chan->aml_dma;
>>>> +     struct aml_dma_sg_link *sg_link;
>>>> +     struct scatterlist *sg;
>>>> +     int idx = 0;
>>>> +     u64 paddr;
>>>> +     u32 reg, link_count, avail, chan_id;
>>>> +     u32 i;
>>>> +
>>>> +     if (aml_chan->direction != direction) {
>>>> +             dev_err(aml_dma->dma_device.dev, "direction not support\n");
>>>> +             return NULL;
>>>> +     }
>>>> +
>>>> +     switch (aml_chan->status) {
>>>> +     case DMA_IN_PROGRESS:
>>>> +             dev_err(aml_dma->dma_device.dev, "not support multi tx_desciptor\n");
>>>> +             return NULL;
>>>> +
>>>> +     case DMA_COMPLETE:
>>>> +             aml_chan->data_len = 0;
>>>> +             chan_id = aml_chan->chan_id;
>>>> +             reg = (direction == DMA_DEV_TO_MEM) ? WCH_INT_MASK : RCH_INT_MASK;
>>>> +             regmap_update_bits(aml_dma->regmap, reg, BIT(chan_id), BIT(chan_id));
>>>> +
>>>> +             break;
>>>> +     default:
>>>> +             dev_err(aml_dma->dma_device.dev, "status error\n");
>>>> +             return NULL;
>>>> +     }
>>>> +
>>>> +     link_count = sg_nents_for_dma(sgl, sg_len, SG_MAX_LEN);
>>>> +
>>>> +     if (link_count > DMA_MAX_LINK) {
>>>> +             dev_err(aml_dma->dma_device.dev,
>>>> +                     "maximum number of sg exceeded: %d > %d\n",
>>>> +                     sg_len, DMA_MAX_LINK);
>>>> +             aml_chan->status = DMA_ERROR;
>>>> +             return NULL;
>>>> +     }
>>>> +
>>>> +     aml_chan->status = DMA_IN_PROGRESS;
>>>> +
>>>> +     for_each_sg(sgl, sg, sg_len, i) {
>>>> +             avail = sg_dma_len(sg);
>>>> +             paddr = sg->dma_address;
>>>> +             while (avail > SG_MAX_LEN) {
>>>> +                     sg_link = &aml_chan->sg_link[idx++];
>>>> +                     /* set dma address and len  to sglink*/
>>>> +                     sg_link->address = paddr;
>>>> +                     sg_link->ctl = FIELD_PREP(LINK_LEN, SG_MAX_LEN);
>>>> +                     paddr = paddr + SG_MAX_LEN;
>>>> +                     avail = avail - SG_MAX_LEN;
>>>> +             }
>>>> +             sg_link = &aml_chan->sg_link[idx++];
>>>> +             /* set dma address and len  to sglink*/
>>>> +             sg_link->address = paddr;
>>> Support here dma_wmb() to make previous write complete before update
>>> OWNER BIT.
>>>
>>> Where update OWNER bit to tall DMA engine sg_link ready?
>>>
>> This DMA hardware does not have OWNER BIT.
>>
>> DMA working steps:
>> The first step is to prepare the corresponding link memory.
>> (This is what the aml_dma_prep_slave_sg work involves.)
>>
>> The second step is to write link phy address into the control register, and
>> data length into the control register. THis will trigger DMA work.
> Is data length total transfer size?
> 

Yes.

> then DMA decrease when process one item in link?

Yes. When the link with the LINK_EOC flag is processed, the DMA  will 
stop, and rasie a interrupt.

> 
> Frank
> 
>> For the memory-to-device channel, an additional register needs to be written
>> to trigger the transfer
>> (This part is implemented in aml_enable_dma_channel function.)
>>
>> In v1 and v2 I placed dma_wmb() at the beginning of aml_enable_dma_channel.
>> You said it was okay not to use it, so I drop it.
>>
>>>> +             sg_link->ctl = FIELD_PREP(LINK_LEN, avail);
>>>> +
>>>> +             aml_chan->data_len += sg_dma_len(sg);
>>>> +     }
>>>> +     aml_chan->sg_link_cnt = idx;
>>>> +
>>>> +     return &aml_chan->desc;
>>>> +}
>>>> +
>>>> +static int aml_dma_pause_chan(struct dma_chan *chan)
>>>> +{
>>>> +     struct aml_dma_chan *aml_chan = to_aml_dma_chan(chan);
>>>> +     struct aml_dma_dev *aml_dma = aml_chan->aml_dma;
>>>> +
>>>> +     regmap_update_bits(aml_dma->regmap, aml_chan->reg_offs + RCH_CFG, CFG_PAUSE, CFG_PAUSE);
>>> regmap_set_bits(), check others
>>>
>>>> +     aml_chan->pre_status = aml_chan->status;
>>>> +     aml_chan->status = DMA_PAUSED;
>>>> +
>>>> +     return 0;
>>>> +}
>>>> +
>>> ...
>>>> +
>>>> +     dma_set_max_seg_size(dma_dev->dev, SG_MAX_LEN);
>>>> +
>>>> +     dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
>>>> +     dma_dev->device_alloc_chan_resources = aml_dma_alloc_chan_resources;
>>>> +     dma_dev->device_free_chan_resources = aml_dma_free_chan_resources;
>>>> +     dma_dev->device_tx_status = aml_dma_tx_status;
>>>> +     dma_dev->device_prep_slave_sg = aml_dma_prep_slave_sg;
>>>> +
>>>> +     dma_dev->device_pause = aml_dma_pause_chan;
>>>> +     dma_dev->device_resume = aml_dma_resume_chan;
>>> align callback name, aml_dma_chan_resume()
>>>
>>>> +     dma_dev->device_terminate_all = aml_dma_terminate_all;
>>>> +     dma_dev->device_issue_pending = aml_dma_enable_chan;
>>> aml_dma_issue_pending()
>>>
>>> Frank
>>>> +     /* PIO 4 bytes and I2C 1 byte */
>>>> +     dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) | BIT(DMA_SLAVE_BUSWIDTH_1_BYTE);
>>>> +     dma_dev->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
>>>> +     dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
>>>> +
>>> ...
>>>> --
>>>> 2.52.0



More information about the linux-amlogic mailing list