Re: [PATCH 1/6] dma: bcm-sba-raid: Improve memory allocation in SBA RAID driver
From: Vinod Koul
Date: Wed Jul 26 2017 - 13:06:47 EST
On Wed, Jul 26, 2017 at 11:06:39AM +0530, Anup Patel wrote:
> This patch improves memory allocation in SBA RAID driver in
> following ways:
> 1. Simplify struct sba_request to reduce memory consumption
what is the simplification?? You need to document that
> 2. Allocate sba resources before registering dma device
what is the motivation for that
So, reading this log doesnt help me to know what to expect in this patch
>
> Signed-off-by: Anup Patel <anup.patel@xxxxxxxxxxxx>
> Reviewed-by: Scott Branden <scott.branden@xxxxxxxxxxxx>
> Reviewed-by: Ray Jui <ray.jui@xxxxxxxxxxxx>
> Reviewed-by: Vikram Prakash <vikram.prakash@xxxxxxxxxxxx>
> ---
> drivers/dma/bcm-sba-raid.c | 439 +++++++++++++++++++++++----------------------
> 1 file changed, 226 insertions(+), 213 deletions(-)
>
> diff --git a/drivers/dma/bcm-sba-raid.c b/drivers/dma/bcm-sba-raid.c
> index e41bbc7..6d15fed 100644
> --- a/drivers/dma/bcm-sba-raid.c
> +++ b/drivers/dma/bcm-sba-raid.c
> @@ -48,7 +48,8 @@
>
> #include "dmaengine.h"
>
> -/* SBA command related defines */
> +/* ====== Driver macros and defines ===== */
why this noise, seems unrelated to the change!
> +
> #define SBA_TYPE_SHIFT 48
> #define SBA_TYPE_MASK GENMASK(1, 0)
> #define SBA_TYPE_A 0x0
> @@ -82,39 +83,41 @@
> #define SBA_CMD_WRITE_BUFFER 0xc
> #define SBA_CMD_GALOIS 0xe
>
> -/* Driver helper macros */
> +#define SBA_MAX_REQ_PER_MBOX_CHANNEL 8192
> +
> #define to_sba_request(tx) \
> container_of(tx, struct sba_request, tx)
> #define to_sba_device(dchan) \
> container_of(dchan, struct sba_device, dma_chan)
>
> -enum sba_request_state {
> - SBA_REQUEST_STATE_FREE = 1,
> - SBA_REQUEST_STATE_ALLOCED = 2,
> - SBA_REQUEST_STATE_PENDING = 3,
> - SBA_REQUEST_STATE_ACTIVE = 4,
> - SBA_REQUEST_STATE_RECEIVED = 5,
> - SBA_REQUEST_STATE_COMPLETED = 6,
> - SBA_REQUEST_STATE_ABORTED = 7,
> +/* ===== Driver data structures ===== */
> +
> +enum sba_request_flags {
> + SBA_REQUEST_STATE_FREE = 0x001,
> + SBA_REQUEST_STATE_ALLOCED = 0x002,
> + SBA_REQUEST_STATE_PENDING = 0x004,
> + SBA_REQUEST_STATE_ACTIVE = 0x008,
> + SBA_REQUEST_STATE_RECEIVED = 0x010,
> + SBA_REQUEST_STATE_COMPLETED = 0x020,
> + SBA_REQUEST_STATE_ABORTED = 0x040,
> + SBA_REQUEST_STATE_MASK = 0x0ff,
> + SBA_REQUEST_FENCE = 0x100,
how does this help in mem alloctn?
> };
>
> struct sba_request {
> /* Global state */
> struct list_head node;
> struct sba_device *sba;
> - enum sba_request_state state;
> - bool fence;
> + u32 flags;
> /* Chained requests management */
> struct sba_request *first;
> struct list_head next;
> - unsigned int next_count;
> atomic_t next_pending_count;
> /* BRCM message data */
> - void *resp;
> - dma_addr_t resp_dma;
> - struct brcm_sba_command *cmds;
> struct brcm_message msg;
> struct dma_async_tx_descriptor tx;
> + /* SBA commands */
> + struct brcm_sba_command cmds[0];
> };
>
> enum sba_version {
> @@ -128,11 +131,11 @@ struct sba_device {
> /* DT configuration parameters */
> enum sba_version ver;
> /* Derived configuration parameters */
> - u32 max_req;
> u32 hw_buf_size;
> u32 hw_resp_size;
> u32 max_pq_coefs;
> u32 max_pq_srcs;
> + u32 max_req;
> u32 max_cmd_per_req;
> u32 max_xor_srcs;
> u32 max_resp_pool_size;
> @@ -152,7 +155,6 @@ struct sba_device {
> void *cmds_base;
> dma_addr_t cmds_dma_base;
> spinlock_t reqs_lock;
> - struct sba_request *reqs;
> bool reqs_fence;
> struct list_head reqs_alloc_list;
> struct list_head reqs_pending_list;
> @@ -161,10 +163,9 @@ struct sba_device {
> struct list_head reqs_completed_list;
> struct list_head reqs_aborted_list;
> struct list_head reqs_free_list;
> - int reqs_free_count;
> };
>
> -/* ====== SBA command helper routines ===== */
> +/* ====== Command helper routines ===== */
more noise..
>
> static inline u64 __pure sba_cmd_enc(u64 cmd, u32 val, u32 shift, u32 mask)
> {
> @@ -196,7 +197,7 @@ static inline u32 __pure sba_cmd_pq_c_mdata(u32 d, u32 b1, u32 b0)
> ((d & SBA_C_MDATA_DNUM_MASK) << SBA_C_MDATA_DNUM_SHIFT);
> }
>
> -/* ====== Channel resource management routines ===== */
> +/* ====== General helper routines ===== */
and it keeps getting more interesting, sigh!!!
>
> static struct sba_request *sba_alloc_request(struct sba_device *sba)
> {
> @@ -204,24 +205,20 @@ static struct sba_request *sba_alloc_request(struct sba_device *sba)
> struct sba_request *req = NULL;
>
> spin_lock_irqsave(&sba->reqs_lock, flags);
> -
> req = list_first_entry_or_null(&sba->reqs_free_list,
> struct sba_request, node);
> - if (req) {
> + if (req)
> list_move_tail(&req->node, &sba->reqs_alloc_list);
> - req->state = SBA_REQUEST_STATE_ALLOCED;
> - req->fence = false;
> - req->first = req;
> - INIT_LIST_HEAD(&req->next);
> - req->next_count = 1;
> - atomic_set(&req->next_pending_count, 1);
> -
> - sba->reqs_free_count--;
> + spin_unlock_irqrestore(&sba->reqs_lock, flags);
> + if (!req)
> + return NULL;
>
> - dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
> - }
> + req->flags = SBA_REQUEST_STATE_ALLOCED;
> + req->first = req;
> + INIT_LIST_HEAD(&req->next);
> + atomic_set(&req->next_pending_count, 1);
Cant fathom how this helps w/ mem allocation
>
> - spin_unlock_irqrestore(&sba->reqs_lock, flags);
> + dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
>
> return req;
> }
> @@ -231,7 +228,8 @@ static void _sba_pending_request(struct sba_device *sba,
> struct sba_request *req)
> {
> lockdep_assert_held(&sba->reqs_lock);
> - req->state = SBA_REQUEST_STATE_PENDING;
> + req->flags &= ~SBA_REQUEST_STATE_MASK;
> + req->flags |= SBA_REQUEST_STATE_PENDING;
> list_move_tail(&req->node, &sba->reqs_pending_list);
> if (list_empty(&sba->reqs_active_list))
> sba->reqs_fence = false;
> @@ -246,9 +244,10 @@ static bool _sba_active_request(struct sba_device *sba,
> sba->reqs_fence = false;
> if (sba->reqs_fence)
> return false;
> - req->state = SBA_REQUEST_STATE_ACTIVE;
> + req->flags &= ~SBA_REQUEST_STATE_MASK;
> + req->flags |= SBA_REQUEST_STATE_ACTIVE;
> list_move_tail(&req->node, &sba->reqs_active_list);
> - if (req->fence)
> + if (req->flags & SBA_REQUEST_FENCE)
> sba->reqs_fence = true;
> return true;
> }
> @@ -258,7 +257,8 @@ static void _sba_abort_request(struct sba_device *sba,
> struct sba_request *req)
> {
> lockdep_assert_held(&sba->reqs_lock);
> - req->state = SBA_REQUEST_STATE_ABORTED;
> + req->flags &= ~SBA_REQUEST_STATE_MASK;
> + req->flags |= SBA_REQUEST_STATE_ABORTED;
> list_move_tail(&req->node, &sba->reqs_aborted_list);
> if (list_empty(&sba->reqs_active_list))
> sba->reqs_fence = false;
> @@ -269,42 +269,34 @@ static void _sba_free_request(struct sba_device *sba,
> struct sba_request *req)
> {
> lockdep_assert_held(&sba->reqs_lock);
> - req->state = SBA_REQUEST_STATE_FREE;
> + req->flags &= ~SBA_REQUEST_STATE_MASK;
> + req->flags |= SBA_REQUEST_STATE_FREE;
> list_move_tail(&req->node, &sba->reqs_free_list);
> if (list_empty(&sba->reqs_active_list))
> sba->reqs_fence = false;
> - sba->reqs_free_count++;
> }
>
> -static void sba_received_request(struct sba_request *req)
> +/* Note: Must be called with sba->reqs_lock held */
> +static void _sba_complete_request(struct sba_device *sba,
> + struct sba_request *req)
> {
> - unsigned long flags;
> - struct sba_device *sba = req->sba;
> -
> - spin_lock_irqsave(&sba->reqs_lock, flags);
> - req->state = SBA_REQUEST_STATE_RECEIVED;
> - list_move_tail(&req->node, &sba->reqs_received_list);
> - spin_unlock_irqrestore(&sba->reqs_lock, flags);
> + lockdep_assert_held(&sba->reqs_lock);
> + req->flags &= ~SBA_REQUEST_STATE_MASK;
> + req->flags |= SBA_REQUEST_STATE_COMPLETED;
> + list_move_tail(&req->node, &sba->reqs_completed_list);
> + if (list_empty(&sba->reqs_active_list))
> + sba->reqs_fence = false;
Ok am going to stop here, sorry can't review it further.
Please split stuff up, make logical incremental patchsets and resubmit...
--
~Vinod