Re: [PATCH RFC 1/2] dmaengine: Introduce scheduled DMA framework
From: Ludovic Desroches
Date: Mon Mar 23 2015 - 09:26:23 EST
On Sat, Mar 21, 2015 at 12:42:06PM -0700, Maxime Ripard wrote:
> This framework aims at easing the development of dmaengine drivers by providing
> generic implementations of the functions usually required by dmaengine, while
> abstracting away most of the logic required.
>
For sure it will ease writing new dmaengine drivers! Moreover, adopting
this framework will convert the driver to use virtual channels isn't it?
> It is very relevant for controllers that have more requests than channels,
> where you need to have some scheduling that is usually very bug prone, and
> shouldn't be implemented in each and every driver.
>
> This introduces a new set of callbacks for drivers to implement the device
> specific behaviour. These new sets of callbacks aims at providing both how to
> handle channel related operations (start the transfer of a new descriptor,
> pause, resume, etc.) and the LLI related operations (iterator and various
> accessors).
>
> So far, the only transfer types supported are memcpy and slave transfers, but
> eventually should support new transfer types as new drivers get converted.
>
> Signed-off-by: Maxime Ripard <maxime.ripard@xxxxxxxxxxxxxxxxxx>
> ---
> drivers/dma/Kconfig | 4 +
> drivers/dma/Makefile | 1 +
> drivers/dma/scheduled-dma.c | 571 ++++++++++++++++++++++++++++++++++++++++++++
> drivers/dma/scheduled-dma.h | 140 +++++++++++
> 4 files changed, 716 insertions(+)
> create mode 100644 drivers/dma/scheduled-dma.c
> create mode 100644 drivers/dma/scheduled-dma.h
>
> diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
> index a874b6ec6650..032bf5fcd58b 100644
> --- a/drivers/dma/Kconfig
> +++ b/drivers/dma/Kconfig
> @@ -406,6 +406,7 @@ config DMA_SUN6I
> depends on RESET_CONTROLLER
> select DMA_ENGINE
> select DMA_VIRTUAL_CHANNELS
> + select DMA_SCHEDULED
> help
> Support for the DMA engine first found in Allwinner A31 SoCs.
>
> @@ -431,6 +432,9 @@ config DMA_ENGINE
> config DMA_VIRTUAL_CHANNELS
> tristate
>
> +config DMA_SCHEDULED
> + bool
> +
I think, it should select DMA_VIRTUAL_CHANNELS
> config DMA_ACPI
> def_bool y
> depends on ACPI
> diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
> index f915f61ec574..1db31814c749 100644
> --- a/drivers/dma/Makefile
> +++ b/drivers/dma/Makefile
> @@ -5,6 +5,7 @@ obj-$(CONFIG_DMA_ENGINE) += dmaengine.o
> obj-$(CONFIG_DMA_VIRTUAL_CHANNELS) += virt-dma.o
> obj-$(CONFIG_DMA_ACPI) += acpi-dma.o
> obj-$(CONFIG_DMA_OF) += of-dma.o
> +obj-$(CONFIG_DMA_SCHEDULED) += scheduled-dma.o
>
> obj-$(CONFIG_INTEL_MID_DMAC) += intel_mid_dma.o
> obj-$(CONFIG_DMATEST) += dmatest.o
> diff --git a/drivers/dma/scheduled-dma.c b/drivers/dma/scheduled-dma.c
> new file mode 100644
> index 000000000000..482d04f2ccbc
> --- /dev/null
> +++ b/drivers/dma/scheduled-dma.c
> @@ -0,0 +1,571 @@
> +/*
> + * Copyright (C) 2015 Maxime Ripard
> + * Maxime Ripard <maxime.ripard@xxxxxxxxxxxxxxxxxx>
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + */
> +
> +#include <linux/dmaengine.h>
> +#include <linux/dmapool.h>
> +#include <linux/list.h>
> +#include <linux/slab.h>
> +
> +#include "scheduled-dma.h"
> +#include "virt-dma.h"
> +
> +static struct sdma_request *sdma_pop_queued_transfer(struct sdma *sdma,
> + struct sdma_channel *schan)
> +{
> + struct sdma_request *sreq = NULL;
> + unsigned long flags;
> +
> + spin_lock_irqsave(&sdma->lock, flags);
> +
> + /* No requests are awaiting an available channel */
> + if (list_empty(&sdma->pend_reqs))
> + goto out;
> +
> + /*
> + * If we don't have any validate_request callback, any request
> + * can be dispatched to any channel.
> + *
> + * Remove the first entry and return it.
> + */
> + if (!sdma->ops->validate_request) {
> + sreq = list_first_entry(&sdma->pend_reqs,
> + struct sdma_request, node);
> + list_del_init(&sreq->node);
> + goto out;
> + }
> +
> + list_for_each_entry(sreq, &sdma->pend_reqs, node) {
> + /*
> + * Ask the driver to validate that the request can
> + * happen on the channel.
> + */
> + if (sdma->ops->validate_request(schan, sreq)) {
> + list_del_init(&sreq->node);
> + goto out;
> + }
> +
> + /* Otherwise, just keep looping */
> + }
> +
> +out:
> + spin_unlock_irqrestore(&sdma->lock, flags);
> +
> + return sreq;
> +}
> +
> +/*
> + * Besoin d'une fonction pour pusher un descriptor vers un pchan
> + *
> + * Flow normal:
> + * - Election d'un pchan (Framework)
> + * - Push d'un descripteur vers le pchan (Driver)
> + * - idle....
> + * - interrupt (driver)
> + * - Transfert terminé, notification vers le framework (driver)
> + * - Nouveau transfert dans la queue?
> + * + Si oui, on est reparti
> + * + Si non, on sort de l'interrupt, le pchan est libéré
> + */
> +
> +struct sdma_desc *sdma_report(struct sdma *sdma,
> + struct sdma_channel *schan,
> + enum sdma_report_status status)
> +{
> + struct sdma_desc *sdesc = NULL;
> + struct virt_dma_desc *vdesc;
> + struct sdma_request *sreq;
> +
> + switch (status) {
> + case SDMA_REPORT_TRANSFER:
> + /*
> + * We're done with the current transfer that was in this
> + * physical channel.
> + */
> + vchan_cookie_complete(&schan->desc->vdesc);
> +
> + /*
> + * Now, try to see if there's any queued transfer
> + * awaiting an available channel.
> + *
> + * If not, just bail out, and mark the pchan as
> + * available.
> + *
> + * If there's such a transfer, validate that the
> + * driver can handle it, and ask it to do the
> + * transfer.
> + */
> + sreq = sdma_pop_queued_transfer(sdma, schan);
> + if (!sreq) {
> + list_add_tail(&schan->node, &sdma->avail_chans);
> + return NULL;
> + }
> +
> + /* Mark the request as assigned to a particular channel */
> + sreq->chan = schan;
> +
> + /* Retrieve the next transfer descriptor */
> + vdesc = vchan_next_desc(&sreq->vchan);
> + schan->desc = sdesc = to_sdma_desc(&vdesc->tx);
> +
> + break;
> +
> + default:
> + break;
> + }
> +
> + return sdesc;
> +}
> +EXPORT_SYMBOL_GPL(sdma_report);
> +
> +static enum dma_status sdma_tx_status(struct dma_chan *chan,
> + dma_cookie_t cookie,
> + struct dma_tx_state *state)
> +{
> + struct sdma_request *sreq = to_sdma_request(chan);
> + struct sdma *sdma = to_sdma(chan->device);
> + struct sdma_channel *schan = sreq->chan;
> + struct virt_dma_desc *vd;
> + struct sdma_desc *desc;
> + enum dma_status ret;
> + unsigned long flags;
> + size_t bytes = 0;
> + void *lli;
> +
> + spin_lock_irqsave(&sreq->vchan.lock, flags);
> +
> + ret = dma_cookie_status(chan, cookie, state);
> + if (ret == DMA_COMPLETE)
> + goto out;
> +
> + vd = vchan_find_desc(&sreq->vchan, cookie);
> + desc = to_sdma_desc(&vd->tx);
> +
> + if (vd) {
> + lli = desc->v_lli;
> + while (true) {
> + bytes += sdma->ops->lli_size(lli);
> +
> + if (!sdma->ops->lli_has_next(lli))
> + break;
> +
> + lli = sdma->ops->lli_next(lli);
> + }
> + } else if (chan) {
> + bytes = sdma->ops->channel_residue(schan);
> + }
> +
> + dma_set_residue(state, bytes);
> +
> +out:
> + spin_unlock_irqrestore(&sreq->vchan.lock, flags);
> +
> + return ret;
> +};
> +
> +static int sdma_config(struct dma_chan *chan,
> + struct dma_slave_config *config)
> +{
> + struct sdma_request *sreq = to_sdma_request(chan);
> + unsigned long flags;
> +
> + spin_lock_irqsave(&sreq->vchan.lock, flags);
> + memcpy(&sreq->cfg, config, sizeof(*config));
> + spin_unlock_irqrestore(&sreq->vchan.lock, flags);
> +
> + return 0;
> +}
> +
> +static int sdma_pause(struct dma_chan *chan)
> +{
> + struct sdma_request *sreq = to_sdma_request(chan);
> + struct sdma_channel *schan = sreq->chan;
> + struct sdma *sdma = to_sdma(chan->device);
> + unsigned long flags;
> +
> + spin_lock_irqsave(&sreq->vchan.lock, flags);
> +
> + /*
> + * If the request is currently scheduled on a channel, just
> + * pause the channel.
> + *
> + * If not, remove the request from the pending list.
> + */
> + if (schan) {
> + sdma->ops->channel_pause(schan);
> + } else {
> + spin_lock(&sdma->lock);
> + list_del_init(&sreq->node);
> + spin_unlock(&sdma->lock);
> + }
> +
> + spin_unlock_irqrestore(&sreq->vchan.lock, flags);
> +
> + return 0;
> +}
> +
> +static int sdma_resume(struct dma_chan *chan)
> +{
> + struct sdma_request *sreq = to_sdma_request(chan);
> + struct sdma_channel *schan = sreq->chan;
> + struct sdma *sdma = to_sdma(chan->device);
> + unsigned long flags;
> +
> + spin_lock_irqsave(&sreq->vchan.lock, flags);
> +
> + /*
> + * If the request is currently scheduled on a channel, just
> + * resume the channel.
> + *
> + * If not, add the request from the pending list.
> + */
Is such a case possible? The request should be already in the pending
list, isn't it? By the way, I may have missed something but where do
you add a request to the pending list? I have seen it only in the resume
case.
> + if (schan) {
> + sdma->ops->channel_resume(schan);
> + } else {
> + spin_lock(&sdma->lock);
> + list_add_tail(&sreq->node, &sdma->pend_reqs);
> + spin_unlock(&sdma->lock);
> + }
> +
> + spin_unlock_irqrestore(&sreq->vchan.lock, flags);
> +
> + return 0;
> +}
> +
> +static int sdma_terminate(struct dma_chan *chan)
> +{
> + struct sdma_request *sreq = to_sdma_request(chan);
> + struct sdma_channel *schan = sreq->chan;
> + struct sdma *sdma = to_sdma(chan->device);
> + unsigned long flags;
> +
> + spin_lock_irqsave(&sreq->vchan.lock, flags);
> +
> + /*
> + * If the request is currently scheduled on a channel,
> + * terminate the channel.
> + *
> + * If not, prevent the request from being scheduled.
> + */
> + if (schan) {
> + sdma->ops->channel_terminate(schan);
> + } else {
> + spin_lock(&sdma->lock);
> + list_del_init(&sreq->node);
> + spin_unlock(&sdma->lock);
> + }
> +
> + spin_unlock_irqrestore(&sreq->vchan.lock, flags);
> +
> + /*
> + * Flush all the pending descriptors from our vchan
> + */
> + vchan_free_chan_resources(&sreq->vchan);
> +
> + return 0;
> +}
> +
> +
> +static struct dma_async_tx_descriptor *sdma_prep_memcpy(struct dma_chan *chan,
> + dma_addr_t dest, dma_addr_t src,
> + size_t len, unsigned long flags)
> +{
> + struct sdma_request *req = to_sdma_request(chan);
> + struct sdma *sdma = to_sdma(chan->device);
> + struct sdma_desc *desc;
> + dma_addr_t p_lli;
> + void *v_lli;
> +
> + if (!len)
> + return NULL;
> +
> + /* Allocate our representation of a descriptor */
> + desc = kzalloc(sizeof(*desc), GFP_NOWAIT);
> + if (!desc)
> + return NULL;
> +
> + v_lli = dma_pool_alloc(sdma->pool, GFP_NOWAIT, &p_lli);
> + if (!v_lli)
> + goto err_desc_free;
> +
> + /* Ask the driver to initialise its hardware descriptor */
> + if (sdma->ops->lli_init(v_lli, req->private,
> + SDMA_TRANSFER_MEMCPY,
> + DMA_MEM_TO_MEM, src, dest, len,
> + NULL))
> + goto err_lli_free;
> +
> + /* Create our single item LLI */
> + sdma->ops->lli_queue(NULL, v_lli, p_lli);
> + desc->p_lli = p_lli;
> + desc->v_lli = v_lli;
> +
> + return vchan_tx_prep(&req->vchan, &desc->vdesc, flags);
> +
> +err_lli_free:
> + dma_pool_free(sdma->pool, v_lli, p_lli);
> +err_desc_free:
> + kfree(desc);
> +
> + return NULL;
> +}
> +
> +static struct dma_async_tx_descriptor *sdma_prep_slave_sg(struct dma_chan *chan,
> + struct scatterlist *sgl,
> + unsigned int sg_len,
> + enum dma_transfer_direction dir,
> + unsigned long flags, void *context)
> +{
> + struct sdma_request *req = to_sdma_request(chan);
> + struct dma_slave_config *config = &req->cfg;
> + struct sdma *sdma = to_sdma(chan->device);
> + void *v_lli, *prev_v_lli = NULL;
> + struct scatterlist *sg;
> + struct sdma_desc *desc;
> + dma_addr_t p_lli;
> + int i;
> +
> + if (!sgl || !sg_len)
> + return NULL;
> +
> + /* Allocate our representation of a descriptor */
> + desc = kzalloc(sizeof(*desc), GFP_NOWAIT);
> + if (!desc)
> + return NULL;
> +
> + /*
> + * For each scatter list entry, build up our representation of
> + * the LLI, and ask the driver to create its hardware
> + * descriptor.
> + */
> + for_each_sg(sgl, sg, sg_len, i) {
> + v_lli = dma_pool_alloc(sdma->pool, GFP_NOWAIT, &p_lli);
> + if (!v_lli)
> + goto err_lli_free;
> +
> + /* Ask the driver to initialise its hardware descriptor */
> + if (sdma->ops->lli_init(v_lli, req->private,
> + SDMA_TRANSFER_SLAVE,
> + dir, sg_dma_address(sg),
> + config->dst_addr, sg_dma_len(sg),
> + config))
> + goto err_lli_free;
> +
> + /*
> + * If it's our first item, initialise our descriptor
> + * pointers to the lli.
> + *
> + * Otherwise, queue it to the end of the LLI.
> + */
> + if (!prev_v_lli) {
> + desc->p_lli = p_lli;
> + desc->v_lli = v_lli;
> + prev_v_lli = v_lli;
> + } else {
> + /* And to queue it at the end of its hardware LLI */
> + prev_v_lli = sdma->ops->lli_queue(prev_v_lli, v_lli, p_lli);
> + }
> + }
> +
> + return vchan_tx_prep(&req->vchan, &desc->vdesc, flags);
> +
> +err_lli_free:
> +#warning "Free the LLI"
> +
> + kfree(desc);
> + return NULL;
> +}
> +
> +static void sdma_issue_pending(struct dma_chan *chan)
> +{
> + struct sdma_request *sreq = to_sdma_request(chan);
> + struct sdma *sdma = to_sdma(chan->device);
> + struct virt_dma_desc *vdesc;
> + struct sdma_channel *schan;
> + struct sdma_desc *sdesc;
> + unsigned long flags;
> +
> + spin_lock_irqsave(&sreq->vchan.lock, flags);
> +
> + /* See if we have anything to do */
> + if (!vchan_issue_pending(&sreq->vchan))
> + goto out_chan_unlock;
> +
> + /* Is some work in progress already? */
> + if (sreq->chan)
> + goto out_chan_unlock;
> +
> + spin_lock(&sdma->lock);
> +
> + /* Is there an available channel */
> + if (list_empty(&sdma->avail_chans))
> + goto out_main_unlock;
> +
> + /*
> + * If there's no validate_request callback, it means that all
> + * channels can transfer any request. Pick the first available
> + * channel.
> + *
> + * Otherwise, iterate over all the pending channels and call
> + * validate_request.
> + */
> + if (!sdma->ops->validate_request) {
> + schan = list_first_entry(&sdma->avail_chans,
> + struct sdma_channel, node);
> + } else {
> + list_for_each_entry(schan, &sdma->avail_chans, node) {
> + if (sdma->ops->validate_request(schan, sreq)) {
> + list_del_init(&schan->node);
> + break;
> + }
> + }
> + }
> +
> + if (!schan)
> + goto out_main_unlock;
> +
> + sreq->chan = schan;
> +
> + /* Retrieve the next transfer descriptor */
> + vdesc = vchan_next_desc(&sreq->vchan);
> + schan->desc = sdesc = to_sdma_desc(&vdesc->tx);
> +
> + sdma->ops->channel_start(schan, sdesc);
> +
> +out_main_unlock:
> + spin_unlock(&sdma->lock);
> +out_chan_unlock:
> + spin_unlock_irqrestore(&sreq->vchan.lock, flags);
> +}
> +
> +static void sdma_free_chan_resources(struct dma_chan *chan)
> +{
> + struct sdma_request *sreq = to_sdma_request(chan);
> + unsigned long flags;
> +
> + spin_lock_irqsave(&sreq->vchan.lock, flags);
> + list_del_init(&sreq->node);
> + spin_unlock_irqrestore(&sreq->vchan.lock, flags);
> +
> + vchan_free_chan_resources(&sreq->vchan);
> +}
> +
> +static void sdma_free_desc(struct virt_dma_desc *vdesc)
> +{
> +#warning "Free the descriptors"
> +}
> +
> +struct sdma *sdma_alloc(struct device *dev,
> + unsigned int channels,
> + unsigned int requests,
> + ssize_t lli_size,
> + ssize_t priv_size)
> +{
> + struct sdma *sdma;
> + int ret, i;
> +
> + sdma = devm_kzalloc(dev, sizeof(*sdma) + priv_size, GFP_KERNEL);
> + if (!sdma) {
> + ret = -ENOMEM;
> + goto out;
> + }
> + INIT_LIST_HEAD(&sdma->pend_reqs);
> + INIT_LIST_HEAD(&sdma->avail_chans);
> + spin_lock_init(&sdma->lock);
> +
> + sdma->pool = dmam_pool_create(dev_name(dev), dev, lli_size, 4, 0);
> + if (!sdma->pool) {
> + ret = -ENOMEM;
> + goto out;
> + }
> +
> + sdma->channels_nr = channels;
> + sdma->channels = devm_kcalloc(dev, channels, sizeof(*sdma->channels), GFP_KERNEL);
> + if (!sdma->channels) {
> + ret = -ENOMEM;
> + goto out;
> + }
> +
> + for (i = 0; i < channels; i++) {
> + struct sdma_channel *schan = &sdma->channels[i];
> +
> + list_add_tail(&schan->node, &sdma->avail_chans);
> + schan->index = i;
> + }
> +
> + sdma->requests_nr = requests;
> + sdma->requests = devm_kcalloc(dev, requests, sizeof(*sdma->requests), GFP_KERNEL);
> + if (!sdma->channels) {
> + ret = -ENOMEM;
> + goto out;
> + }
> +
> + INIT_LIST_HEAD(&sdma->ddev.channels);
> +
> + for (i = 0; i < requests; i++) {
> + struct sdma_request *sreq = &sdma->requests[i];
> +
> + INIT_LIST_HEAD(&sreq->node);
> + sreq->vchan.desc_free = sdma_free_desc;
> + vchan_init(&sreq->vchan, &sdma->ddev);
> + }
> +
> + return sdma;
> +
> +out:
> + return ERR_PTR(ret);
> +}
> +EXPORT_SYMBOL(sdma_alloc);
> +
> +void sdma_free(struct sdma *sdma)
> +{
> + return;
> +}
> +EXPORT_SYMBOL(sdma_free);
> +
> +int sdma_register(struct sdma *sdma,
> + struct sdma_ops *ops)
> +{
> + struct dma_device *ddev = &sdma->ddev;
> +
> + sdma->ops = ops;
> +
> + ddev->device_config = sdma_config;
> + ddev->device_tx_status = sdma_tx_status;
> + ddev->device_issue_pending = sdma_issue_pending;
> + ddev->device_free_chan_resources = sdma_free_chan_resources;
> +
> + if (ops->channel_pause)
> + ddev->device_pause = sdma_pause;
> +
> + if (ops->channel_resume)
> + ddev->device_resume = sdma_resume;
> +
> + if (ops->channel_terminate)
> + ddev->device_terminate_all = sdma_terminate;
> +
> + if (dma_has_cap(DMA_SLAVE, ddev->cap_mask))
> + ddev->device_prep_slave_sg = sdma_prep_slave_sg;
> +
> + if (dma_has_cap(DMA_MEMCPY, ddev->cap_mask))
> + ddev->device_prep_dma_memcpy = sdma_prep_memcpy;
> +
> + dma_async_device_register(ddev);
> +
> + return 0;
> +}
> +EXPORT_SYMBOL(sdma_register);
> +
> +int sdma_unregister(struct sdma *sdma)
> +{
> + dma_async_device_unregister(&sdma->ddev);
> +
> + return 0;
> +}
> +EXPORT_SYMBOL(sdma_unregister);
> diff --git a/drivers/dma/scheduled-dma.h b/drivers/dma/scheduled-dma.h
> new file mode 100644
> index 000000000000..d24c8143b2b6
> --- /dev/null
> +++ b/drivers/dma/scheduled-dma.h
> @@ -0,0 +1,140 @@
> +/*
> + * Copyright (C) 2015 Maxime Ripard
> + * Maxime Ripard <maxime.ripard@xxxxxxxxxxxxxxxxxx>
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + */
> +
> +#include "virt-dma.h"
> +
> +#ifndef _SCHEDULED_DMA_H_
> +#define _SCHEDULED_DMA_H_
> +
> +enum sdma_transfer_type {
> + SDMA_TRANSFER_MEMCPY,
> + SDMA_TRANSFER_SLAVE,
> +};
> +
> +enum sdma_report_status {
> + SDMA_REPORT_CHUNK,
> + SDMA_REPORT_TRANSFER,
> +};
> +
> +struct sdma_desc {
> + struct virt_dma_desc vdesc;
> +
> + /* Entry point to our LLI */
> + dma_addr_t p_lli;
> + void *v_lli;
> +};
> +
> +struct sdma_channel {
> + struct sdma_desc *desc;
> + unsigned int index;
> + struct list_head node;
> + void *private;
> +};
> +
> +struct sdma_request {
> + struct dma_slave_config cfg;
> + struct list_head node;
> + struct virt_dma_chan vchan;
> +
> + struct sdma_channel *chan;
> + void *private;
> +};
> +
> +struct sdma_ops {
> + /* LLI management operations */
> + bool (*lli_has_next)(void *v_lli);
> + void *(*lli_next)(void *v_lli);
> + int (*lli_init)(void *v_lli, void *sreq_priv,
> + enum sdma_transfer_type type,
> + enum dma_transfer_direction dir,
> + dma_addr_t src,
> + dma_addr_t dst, u32 len,
> + struct dma_slave_config *config);
> + void *(*lli_queue)(void *prev_v_lli, void *v_lli, dma_addr_t p_lli);
> + size_t (*lli_size)(void *v_lli);
> +
> + /* Scheduler helper */
> + struct sdma_request *(*validate_request)(struct sdma_channel *chan,
> + struct sdma_request *req);
> +
> + /* Transfer Management Functions */
> + int (*channel_pause)(struct sdma_channel *chan);
> + int (*channel_resume)(struct sdma_channel *chan);
> + int (*channel_start)(struct sdma_channel *chan, struct sdma_desc *sdesc);
> + int (*channel_terminate)(struct sdma_channel *chan);
> + size_t (*channel_residue)(struct sdma_channel *chan);
> +};
> +
> +struct sdma {
> + struct dma_device ddev;
> + struct sdma_ops *ops;
> +
> + struct dma_pool *pool;
> +
> + struct sdma_channel *channels;
> + int channels_nr;
> + struct sdma_request *requests;
> + int requests_nr;
> +
> + struct list_head avail_chans;
> + struct list_head pend_reqs;
> +
> + spinlock_t lock;
> +
> + unsigned long private[];
> +};
> +
> +static inline struct sdma *to_sdma(struct dma_device *d)
> +{
> + return container_of(d, struct sdma, ddev);
> +}
> +
> +static inline struct sdma_request *to_sdma_request(struct dma_chan *chan)
> +{
> + return container_of(chan, struct sdma_request, vchan.chan);
> +}
> +
> +static inline struct sdma_desc *to_sdma_desc(struct dma_async_tx_descriptor *tx)
> +{
> + return container_of(tx, struct sdma_desc, vdesc.tx);
> +}
> +
> +static inline void *sdma_priv(struct sdma *sdma)
> +{
> + return (void*)sdma->private;
> +}
> +
> +static inline void sdma_set_chan_private(struct sdma *sdma, void *ptr)
> +{
> + int i;
> +
> + for (i = 0; i < sdma->channels_nr; i++) {
> + struct sdma_channel *schan = &sdma->channels[i];
> +
> + schan->private = ptr;
> + }
> +}
> +
> +struct sdma_desc *sdma_report(struct sdma *sdma,
> + struct sdma_channel *chan,
> + enum sdma_report_status status);
> +
> +struct sdma *sdma_alloc(struct device *dev,
> + unsigned int channels,
> + unsigned int requests,
> + ssize_t lli_size,
> + ssize_t priv_size);
> +void sdma_free(struct sdma *sdma);
> +
> +int sdma_register(struct sdma *sdma,
> + struct sdma_ops *ops);
> +int sdma_unregister(struct sdma *sdma);
> +
> +#endif /* _SCHEDULED_DMA_H_ */
> --
> 2.3.3
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/