Re: [PATCH v16] platform/mellanox: Add TmFifo driver for Mellanox BlueField Soc

From: Andy Shevchenko
Date: Mon May 06 2019 - 05:14:26 EST


On Fri, May 3, 2019 at 4:49 PM Liming Sun <lsun@xxxxxxxxxxxx> wrote:
>
> This commit adds the TmFifo platform driver for Mellanox BlueField
> Soc. TmFifo is a shared FIFO which enables external host machine
> to exchange data with the SoC via USB or PCIe. The driver is based
> on virtio framework and has console and network access enabled.
>

Pushed to my review and testing queue, thanks!

> Reviewed-by: Vadim Pasternak <vadimp@xxxxxxxxxxxx>
> Signed-off-by: Liming Sun <lsun@xxxxxxxxxxxx>
> ---
> v15->v16:
> Rebase and resubmit (no new changes).
> v14->v15:
> Fixes for comments from Andy:
> - Remove the 'union' definition of mlxbf_tmfifo_msg_hdr and use
> on-the-fly conversion when sending the 8-byte message header
> into the FIFO;
> - Update comment of mlxbf_tmfifo_msg_hdr explaining why '__be16'
> is needed for the 'len' field. The SoC sends data stream into
> the FIFO and the other side reads it. The byte order of the data
> stream (byte-stream) stays the same. The 'len' field is encoded
> into network byte order so upper-level applications in external
> host machine with different endianness could decode it. This
> implementation was verified over USB with an external PPC host
> machine running in big-endian mode.
> - Move the 'dev_err()' line to the end of the block in function
> mlxbf_tmfifo_alloc_vrings();
> - Remove the 'irq_info->index < MLXBF_TM_MAX_IRQ' check in
> mlxbf_tmfifo_irq_handler() since it's unnecessary;
> - Remove the 'if (desc_head)' check in
> mlxbf_tmfifo_release_pending_pkt() since function
> mlxbf_tmfifo_get_pkt_len() is already NULL-aware;
> - Adjust the testing order of 'if (!(vring->index & BIT(0)))'
> in bool mlxbf_tmfifo_virtio_notify() to test the positive case
> 'if (vring->index & BIT(0))' first;
> - Add '(u64)offset' conversion in mlxbf_tmfifo_virtio_get() to
> avoid 32-bit length addition overflow;
> - Update the 'efi.get_variable' statement into single line in
> mlxbf_tmfifo_get_cfg_mac();
> - Use new helper devm_platform_ioremap_resource() to replace
> 'platform_get_resource() + devm_ioremap_resource()' in
> mlxbf_tmfifo_probe();
> v13->v14:
> Fixes for comments from Andy:
> - Add a blank line to separate the virtio header files;
> - Update the comment for 'union mlxbf_tmfifo_msg_hdr' to be
> more clear how this union is used;
> - Update the 'mlxbf_tmfifo_net_default_mac[ETH_ALEN]' definition
> to be two lines;
> - Reformat macro MLXBF_TMFIFO_NET_FEATURES to put the definition
> in a seperate line;
> - Update all 'fifo' to 'FIFO' in the comments;
> - Update mlxbf_tmfifo_alloc_vrings() to specifically release the
> allocated entries in case of failures, so the logic looks more
> clear. In the caller function the mlxbf_tmfifo_free_vrings()
> might be called again in case of other failures, which is ok
> since the 'va' pointer will be set to NULL once released;
> - Update mlxbf_tmfifo_timer() to change the first statement to
> one line;
> - Update one memcpy() to ether_addr_copy() in
> mlxbf_tmfifo_get_cfg_mac();
> - Remove 'fifo->pdev' since it is really not needed;
> - Define temporary variable to update the mlxbf_tmfifo_create_vdev()
> statement into single line.
> New changes by Liming:
> - Reorder the logic a little bit in mlxbf_tmfifo_timer(). Previously
> it has logic like "!a || !b" while the '!b' will not be evaluated
> if '!a' is true. It was changed to this way during review, but is
> actually not the desired behavior since both bits need to be
> tested/set in fifo->pend_events. This issue was found during
> verification which caused extra delays for Tx packets.
> v12->v13:
> Rebase and resubmit (no new changes).
> v11->v12:
> Fixed the two unsolved comments from v11.
> - "Change macro mlxbf_vdev_to_tmfifo() to one line"
> Done. Seems not hard.
> - "Is it appropriate use of devm_* for 'tm_vdev = devm_kzalloc'"
> Yes, understand the comment now. The tmfifo is fixed, but the
> vdev is dynamic. Use kzalloc() instead, and free the device
> in the release callback which is the right place for it.
> v10->v11:
> Fixes for comments from Andy:
> - Use GENMASK_ULL() instead of GENMASK() in mlxbf-tmfifo-regs.h
> - Removed the cpu_to_le64()/le64_to_cpu() conversion since
> readq()/writeq() already takes care of it.
> - Remove the "if (irq)" check in mlxbf_tmfifo_disable_irqs().
> - Add "u32 count" temp variable in mlxbf_tmfifo_get_tx_avail().
> - Clean up mlxbf_tmfifo_get_cfg_mac(), use ETH_ALEN instead of
> value 6.
> - Change the tx_buf to use Linux existing 'struct circ_buf'.
> Comment not applied:
> - "Change macro mlxbf_vdev_to_tmfifo() to one line"
> Couldn't fit in one line with 80 chracters
> - "Is it appropriate use of devm_* for 'tm_vdev = devm_kzalloc'"
> This is SoC, the device won't be closed or detached.
> The only case is when the driver is unloaded. So it appears
> ok to use devm_kzalloc() since it's allocated during probe()
> and released during module unload.
> Comments from Vadim: OK
> v9->v10:
> Fixes for comments from Andy:
> - Use devm_ioremap_resource() instead of devm_ioremap().
> - Use kernel-doc comments.
> - Keep Makefile contents sorted.
> - Use same fixed format for offsets.
> - Use SZ_1K/SZ_32K instead of 1024/23*1024.
> - Remove unnecessary comments.
> - Use one style for max numbers.
> - More comments for mlxbf_tmfifo_vdev and mlxbf_tmfifo_data_64bit.
> - Use globally defined MTU instead of new definition.
> - Remove forward declaration of mlxbf_tmfifo_remove().
> - Remove PAGE_ALIGN() for dma_alloc_coherent)().
> - Remove the cast of "struct vring *".
> - Check return result of test_and_set_bit().
> - Add a macro mlxbt_vdev_to_tmfifo().
> - Several other minor coding style comments.
> Comment not applied:
> - "Shouldn't be rather helper in EFI lib in kernel"
> Looks like efi.get_variable() is the way I found in the kernel
> tree.
> - "this one is not protected anyhow? Potential race condition"
> In mlxbf_tmfifo_console_tx(), the spin-lock is used to protect the
> 'tx_buf' only, not the FIFO writes. So there is no race condition.
> - "Is __packed needed in mlxbf_tmfifo_msg_hdr".
> Yes, it is needed to make sure the structure is 8 bytes.
> Fixes for comments from Vadim:
> - Use tab in mlxbf-tmfifo-regs.h
> - Use kernel-doc comments for struct mlxbf_tmfifo_msg_hdr and
> mlxbf_tmfifo_irq_info as well.
> - Use _MAX instead of _CNT in the macro definition to be consistent.
> - Fix the MODULE_LICENSE.
> - Use BIT_ULL() instead of BIT().
> - Remove argument of 'avail' for mlxbf_tmfifo_rxtx_header() and
> mlxbf_tmfifo_rxtx_word()
> - Revise logic in mlxbf_tmfifo_rxtx_one_desc() to remove the
> WARN_ON().
> - Change "union mlxbf_tmfifo_u64 u" to "union mlxbf_tmfifo_u64 buf"
> in mlxbf_tmfifo_rxtx_word().
> - Change date type of vring_change from 'int' to 'bool'.
> - Remove the blank lines after Signed-off.
> - Donât use declaration in the middle.
> - Make the network header initialization in some more elegant way.
> - Change label done to mlxbf_tmfifo_desc_done.
> - Remove some unnecessary comments, and several other misc coding
> style comments.
> - Simplify code logic in mlxbf_tmfifo_virtio_notify()
> New changes by Liming:
> - Simplify the Rx/Tx function arguments to make it more readable.
> v8->v9:
> Fixes for comments from Andy:
> - Use modern devm_xxx() API instead.
> Fixes for comments from Vadim:
> - Split the Rx/Tx function into smaller funcitons.
> - File name, copyright information.
> - Function and variable name conversion.
> - Local variable and indent coding styles.
> - Remove unnecessary 'inline' declarations.
> - Use devm_xxx() APIs.
> - Move the efi_char16_t MAC address definition to global.
> - Fix warnings reported by 'checkpatch --strict'.
> - Fix warnings reported by 'make CF="-D__CHECK_ENDIAN__"'.
> - Change select VIRTIO_xxx to depends on VIRTIO_ in Kconfig.
> - Merge mlxbf_tmfifo_vdev_tx_buf_push() and
> mlxbf_tmfifo_vdev_tx_buf_pop().
> - Add union to avoid casting between __le64 and u64.
> - Several other misc coding style comments.
> New changes by Liming:
> - Removed the DT binding documentation since only ACPI is
> supported for now by UEFI on the SoC.
> v8: Re-submit under drivers/platform/mellanox for the target-side
> platform driver only.
> v7: Added host side drivers into the same patch set.
> v5~v6: Coding style fix.
> v1~v4: Initial version for directory drivers/soc/mellanox.
> ---
> drivers/platform/mellanox/Kconfig | 12 +-
> drivers/platform/mellanox/Makefile | 1 +
> drivers/platform/mellanox/mlxbf-tmfifo-regs.h | 63 ++
> drivers/platform/mellanox/mlxbf-tmfifo.c | 1281 +++++++++++++++++++++++++
> 4 files changed, 1356 insertions(+), 1 deletion(-)
> create mode 100644 drivers/platform/mellanox/mlxbf-tmfifo-regs.h
> create mode 100644 drivers/platform/mellanox/mlxbf-tmfifo.c
>
> diff --git a/drivers/platform/mellanox/Kconfig b/drivers/platform/mellanox/Kconfig
> index cd8a908..530fe7e 100644
> --- a/drivers/platform/mellanox/Kconfig
> +++ b/drivers/platform/mellanox/Kconfig
> @@ -5,7 +5,7 @@
>
> menuconfig MELLANOX_PLATFORM
> bool "Platform support for Mellanox hardware"
> - depends on X86 || ARM || COMPILE_TEST
> + depends on X86 || ARM || ARM64 || COMPILE_TEST
> ---help---
> Say Y here to get to see options for platform support for
> Mellanox systems. This option alone does not add any kernel code.
> @@ -34,4 +34,14 @@ config MLXREG_IO
> to system resets operation, system reset causes monitoring and some
> kinds of mux selection.
>
> +config MLXBF_TMFIFO
> + tristate "Mellanox BlueField SoC TmFifo platform driver"
> + depends on ARM64
> + depends on ACPI
> + depends on VIRTIO_CONSOLE && VIRTIO_NET
> + help
> + Say y here to enable TmFifo support. The TmFifo driver provides
> + platform driver support for the TmFifo which supports console
> + and networking based on the virtio framework.
> +
> endif # MELLANOX_PLATFORM
> diff --git a/drivers/platform/mellanox/Makefile b/drivers/platform/mellanox/Makefile
> index 57074d9c..a229bda1 100644
> --- a/drivers/platform/mellanox/Makefile
> +++ b/drivers/platform/mellanox/Makefile
> @@ -3,5 +3,6 @@
> # Makefile for linux/drivers/platform/mellanox
> # Mellanox Platform-Specific Drivers
> #
> +obj-$(CONFIG_MLXBF_TMFIFO) += mlxbf-tmfifo.o
> obj-$(CONFIG_MLXREG_HOTPLUG) += mlxreg-hotplug.o
> obj-$(CONFIG_MLXREG_IO) += mlxreg-io.o
> diff --git a/drivers/platform/mellanox/mlxbf-tmfifo-regs.h b/drivers/platform/mellanox/mlxbf-tmfifo-regs.h
> new file mode 100644
> index 0000000..e4f0d2e
> --- /dev/null
> +++ b/drivers/platform/mellanox/mlxbf-tmfifo-regs.h
> @@ -0,0 +1,63 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (c) 2019, Mellanox Technologies. All rights reserved.
> + */
> +
> +#ifndef __MLXBF_TMFIFO_REGS_H__
> +#define __MLXBF_TMFIFO_REGS_H__
> +
> +#include <linux/types.h>
> +#include <linux/bits.h>
> +
> +#define MLXBF_TMFIFO_TX_DATA 0x00
> +#define MLXBF_TMFIFO_TX_STS 0x08
> +#define MLXBF_TMFIFO_TX_STS__LENGTH 0x0001
> +#define MLXBF_TMFIFO_TX_STS__COUNT_SHIFT 0
> +#define MLXBF_TMFIFO_TX_STS__COUNT_WIDTH 9
> +#define MLXBF_TMFIFO_TX_STS__COUNT_RESET_VAL 0
> +#define MLXBF_TMFIFO_TX_STS__COUNT_RMASK GENMASK_ULL(8, 0)
> +#define MLXBF_TMFIFO_TX_STS__COUNT_MASK GENMASK_ULL(8, 0)
> +#define MLXBF_TMFIFO_TX_CTL 0x10
> +#define MLXBF_TMFIFO_TX_CTL__LENGTH 0x0001
> +#define MLXBF_TMFIFO_TX_CTL__LWM_SHIFT 0
> +#define MLXBF_TMFIFO_TX_CTL__LWM_WIDTH 8
> +#define MLXBF_TMFIFO_TX_CTL__LWM_RESET_VAL 128
> +#define MLXBF_TMFIFO_TX_CTL__LWM_RMASK GENMASK_ULL(7, 0)
> +#define MLXBF_TMFIFO_TX_CTL__LWM_MASK GENMASK_ULL(7, 0)
> +#define MLXBF_TMFIFO_TX_CTL__HWM_SHIFT 8
> +#define MLXBF_TMFIFO_TX_CTL__HWM_WIDTH 8
> +#define MLXBF_TMFIFO_TX_CTL__HWM_RESET_VAL 128
> +#define MLXBF_TMFIFO_TX_CTL__HWM_RMASK GENMASK_ULL(7, 0)
> +#define MLXBF_TMFIFO_TX_CTL__HWM_MASK GENMASK_ULL(15, 8)
> +#define MLXBF_TMFIFO_TX_CTL__MAX_ENTRIES_SHIFT 32
> +#define MLXBF_TMFIFO_TX_CTL__MAX_ENTRIES_WIDTH 9
> +#define MLXBF_TMFIFO_TX_CTL__MAX_ENTRIES_RESET_VAL 256
> +#define MLXBF_TMFIFO_TX_CTL__MAX_ENTRIES_RMASK GENMASK_ULL(8, 0)
> +#define MLXBF_TMFIFO_TX_CTL__MAX_ENTRIES_MASK GENMASK_ULL(40, 32)
> +#define MLXBF_TMFIFO_RX_DATA 0x00
> +#define MLXBF_TMFIFO_RX_STS 0x08
> +#define MLXBF_TMFIFO_RX_STS__LENGTH 0x0001
> +#define MLXBF_TMFIFO_RX_STS__COUNT_SHIFT 0
> +#define MLXBF_TMFIFO_RX_STS__COUNT_WIDTH 9
> +#define MLXBF_TMFIFO_RX_STS__COUNT_RESET_VAL 0
> +#define MLXBF_TMFIFO_RX_STS__COUNT_RMASK GENMASK_ULL(8, 0)
> +#define MLXBF_TMFIFO_RX_STS__COUNT_MASK GENMASK_ULL(8, 0)
> +#define MLXBF_TMFIFO_RX_CTL 0x10
> +#define MLXBF_TMFIFO_RX_CTL__LENGTH 0x0001
> +#define MLXBF_TMFIFO_RX_CTL__LWM_SHIFT 0
> +#define MLXBF_TMFIFO_RX_CTL__LWM_WIDTH 8
> +#define MLXBF_TMFIFO_RX_CTL__LWM_RESET_VAL 128
> +#define MLXBF_TMFIFO_RX_CTL__LWM_RMASK GENMASK_ULL(7, 0)
> +#define MLXBF_TMFIFO_RX_CTL__LWM_MASK GENMASK_ULL(7, 0)
> +#define MLXBF_TMFIFO_RX_CTL__HWM_SHIFT 8
> +#define MLXBF_TMFIFO_RX_CTL__HWM_WIDTH 8
> +#define MLXBF_TMFIFO_RX_CTL__HWM_RESET_VAL 128
> +#define MLXBF_TMFIFO_RX_CTL__HWM_RMASK GENMASK_ULL(7, 0)
> +#define MLXBF_TMFIFO_RX_CTL__HWM_MASK GENMASK_ULL(15, 8)
> +#define MLXBF_TMFIFO_RX_CTL__MAX_ENTRIES_SHIFT 32
> +#define MLXBF_TMFIFO_RX_CTL__MAX_ENTRIES_WIDTH 9
> +#define MLXBF_TMFIFO_RX_CTL__MAX_ENTRIES_RESET_VAL 256
> +#define MLXBF_TMFIFO_RX_CTL__MAX_ENTRIES_RMASK GENMASK_ULL(8, 0)
> +#define MLXBF_TMFIFO_RX_CTL__MAX_ENTRIES_MASK GENMASK_ULL(40, 32)
> +
> +#endif /* !defined(__MLXBF_TMFIFO_REGS_H__) */
> diff --git a/drivers/platform/mellanox/mlxbf-tmfifo.c b/drivers/platform/mellanox/mlxbf-tmfifo.c
> new file mode 100644
> index 0000000..9a5c9fd
> --- /dev/null
> +++ b/drivers/platform/mellanox/mlxbf-tmfifo.c
> @@ -0,0 +1,1281 @@
> +// SPDX-License-Identifier: GPL-2.0+
> +/*
> + * Mellanox BlueField SoC TmFifo driver
> + *
> + * Copyright (C) 2019 Mellanox Technologies
> + */
> +
> +#include <linux/acpi.h>
> +#include <linux/bitfield.h>
> +#include <linux/circ_buf.h>
> +#include <linux/efi.h>
> +#include <linux/irq.h>
> +#include <linux/module.h>
> +#include <linux/mutex.h>
> +#include <linux/platform_device.h>
> +#include <linux/types.h>
> +
> +#include <linux/virtio_config.h>
> +#include <linux/virtio_console.h>
> +#include <linux/virtio_ids.h>
> +#include <linux/virtio_net.h>
> +#include <linux/virtio_ring.h>
> +
> +#include "mlxbf-tmfifo-regs.h"
> +
> +/* Vring size. */
> +#define MLXBF_TMFIFO_VRING_SIZE SZ_1K
> +
> +/* Console Tx buffer size. */
> +#define MLXBF_TMFIFO_CON_TX_BUF_SIZE SZ_32K
> +
> +/* Console Tx buffer reserved space. */
> +#define MLXBF_TMFIFO_CON_TX_BUF_RSV_SIZE 8
> +
> +/* House-keeping timer interval. */
> +#define MLXBF_TMFIFO_TIMER_INTERVAL (HZ / 10)
> +
> +/* Virtual devices sharing the TM FIFO. */
> +#define MLXBF_TMFIFO_VDEV_MAX (VIRTIO_ID_CONSOLE + 1)
> +
> +/*
> + * Reserve 1/16 of TmFifo space, so console messages are not starved by
> + * the networking traffic.
> + */
> +#define MLXBF_TMFIFO_RESERVE_RATIO 16
> +
> +/* Message with data needs at least two words (for header & data). */
> +#define MLXBF_TMFIFO_DATA_MIN_WORDS 2
> +
> +struct mlxbf_tmfifo;
> +
> +/**
> + * mlxbf_tmfifo_vring - Structure of the TmFifo virtual ring
> + * @va: virtual address of the ring
> + * @dma: dma address of the ring
> + * @vq: pointer to the virtio virtqueue
> + * @desc: current descriptor of the pending packet
> + * @desc_head: head descriptor of the pending packet
> + * @cur_len: processed length of the current descriptor
> + * @rem_len: remaining length of the pending packet
> + * @pkt_len: total length of the pending packet
> + * @next_avail: next avail descriptor id
> + * @num: vring size (number of descriptors)
> + * @align: vring alignment size
> + * @index: vring index
> + * @vdev_id: vring virtio id (VIRTIO_ID_xxx)
> + * @fifo: pointer to the tmfifo structure
> + */
> +struct mlxbf_tmfifo_vring {
> + void *va;
> + dma_addr_t dma;
> + struct virtqueue *vq;
> + struct vring_desc *desc;
> + struct vring_desc *desc_head;
> + int cur_len;
> + int rem_len;
> + u32 pkt_len;
> + u16 next_avail;
> + int num;
> + int align;
> + int index;
> + int vdev_id;
> + struct mlxbf_tmfifo *fifo;
> +};
> +
> +/* Interrupt types. */
> +enum {
> + MLXBF_TM_RX_LWM_IRQ,
> + MLXBF_TM_RX_HWM_IRQ,
> + MLXBF_TM_TX_LWM_IRQ,
> + MLXBF_TM_TX_HWM_IRQ,
> + MLXBF_TM_MAX_IRQ
> +};
> +
> +/* Ring types (Rx & Tx). */
> +enum {
> + MLXBF_TMFIFO_VRING_RX,
> + MLXBF_TMFIFO_VRING_TX,
> + MLXBF_TMFIFO_VRING_MAX
> +};
> +
> +/**
> + * mlxbf_tmfifo_vdev - Structure of the TmFifo virtual device
> + * @vdev: virtio device, in which the vdev.id.device field has the
> + * VIRTIO_ID_xxx id to distinguish the virtual device.
> + * @status: status of the device
> + * @features: supported features of the device
> + * @vrings: array of tmfifo vrings of this device
> + * @config.cons: virtual console config -
> + * select if vdev.id.device is VIRTIO_ID_CONSOLE
> + * @config.net: virtual network config -
> + * select if vdev.id.device is VIRTIO_ID_NET
> + * @tx_buf: tx buffer used to buffer data before writing into the FIFO
> + */
> +struct mlxbf_tmfifo_vdev {
> + struct virtio_device vdev;
> + u8 status;
> + u64 features;
> + struct mlxbf_tmfifo_vring vrings[MLXBF_TMFIFO_VRING_MAX];
> + union {
> + struct virtio_console_config cons;
> + struct virtio_net_config net;
> + } config;
> + struct circ_buf tx_buf;
> +};
> +
> +/**
> + * mlxbf_tmfifo_irq_info - Structure of the interrupt information
> + * @fifo: pointer to the tmfifo structure
> + * @irq: interrupt number
> + * @index: index into the interrupt array
> + */
> +struct mlxbf_tmfifo_irq_info {
> + struct mlxbf_tmfifo *fifo;
> + int irq;
> + int index;
> +};
> +
> +/**
> + * mlxbf_tmfifo - Structure of the TmFifo
> + * @vdev: array of the virtual devices running over the TmFifo
> + * @lock: lock to protect the TmFifo access
> + * @rx_base: mapped register base address for the Rx FIFO
> + * @tx_base: mapped register base address for the Tx FIFO
> + * @rx_fifo_size: number of entries of the Rx FIFO
> + * @tx_fifo_size: number of entries of the Tx FIFO
> + * @pend_events: pending bits for deferred events
> + * @irq_info: interrupt information
> + * @work: work struct for deferred process
> + * @timer: background timer
> + * @vring: Tx/Rx ring
> + * @spin_lock: spin lock
> + * @is_ready: ready flag
> + */
> +struct mlxbf_tmfifo {
> + struct mlxbf_tmfifo_vdev *vdev[MLXBF_TMFIFO_VDEV_MAX];
> + struct mutex lock; /* TmFifo lock */
> + void __iomem *rx_base;
> + void __iomem *tx_base;
> + int rx_fifo_size;
> + int tx_fifo_size;
> + unsigned long pend_events;
> + struct mlxbf_tmfifo_irq_info irq_info[MLXBF_TM_MAX_IRQ];
> + struct work_struct work;
> + struct timer_list timer;
> + struct mlxbf_tmfifo_vring *vring[2];
> + spinlock_t spin_lock; /* spin lock */
> + bool is_ready;
> +};
> +
> +/**
> + * mlxbf_tmfifo_msg_hdr - Structure of the TmFifo message header
> + * @type: message type
> + * @len: payload length in network byte order. Messages sent into the FIFO
> + * will be read by the other side as data stream in the same byte order.
> + * The length needs to be encoded into network order so both sides
> + * could understand it.
> + */
> +struct mlxbf_tmfifo_msg_hdr {
> + u8 type;
> + __be16 len;
> + u8 unused[5];
> +} __packed __aligned(sizeof(u64));
> +
> +/*
> + * Default MAC.
> + * This MAC address will be read from EFI persistent variable if configured.
> + * It can also be reconfigured with standard Linux tools.
> + */
> +static u8 mlxbf_tmfifo_net_default_mac[ETH_ALEN] = {
> + 0x00, 0x1A, 0xCA, 0xFF, 0xFF, 0x01
> +};
> +
> +/* EFI variable name of the MAC address. */
> +static efi_char16_t mlxbf_tmfifo_efi_name[] = L"RshimMacAddr";
> +
> +/* Maximum L2 header length. */
> +#define MLXBF_TMFIFO_NET_L2_OVERHEAD 36
> +
> +/* Supported virtio-net features. */
> +#define MLXBF_TMFIFO_NET_FEATURES \
> + (BIT_ULL(VIRTIO_NET_F_MTU) | BIT_ULL(VIRTIO_NET_F_STATUS) | \
> + BIT_ULL(VIRTIO_NET_F_MAC))
> +
> +#define mlxbf_vdev_to_tmfifo(d) container_of(d, struct mlxbf_tmfifo_vdev, vdev)
> +
> +/* Free vrings of the FIFO device. */
> +static void mlxbf_tmfifo_free_vrings(struct mlxbf_tmfifo *fifo,
> + struct mlxbf_tmfifo_vdev *tm_vdev)
> +{
> + struct mlxbf_tmfifo_vring *vring;
> + int i, size;
> +
> + for (i = 0; i < ARRAY_SIZE(tm_vdev->vrings); i++) {
> + vring = &tm_vdev->vrings[i];
> + if (vring->va) {
> + size = vring_size(vring->num, vring->align);
> + dma_free_coherent(tm_vdev->vdev.dev.parent, size,
> + vring->va, vring->dma);
> + vring->va = NULL;
> + if (vring->vq) {
> + vring_del_virtqueue(vring->vq);
> + vring->vq = NULL;
> + }
> + }
> + }
> +}
> +
> +/* Allocate vrings for the FIFO. */
> +static int mlxbf_tmfifo_alloc_vrings(struct mlxbf_tmfifo *fifo,
> + struct mlxbf_tmfifo_vdev *tm_vdev)
> +{
> + struct mlxbf_tmfifo_vring *vring;
> + struct device *dev;
> + dma_addr_t dma;
> + int i, size;
> + void *va;
> +
> + for (i = 0; i < ARRAY_SIZE(tm_vdev->vrings); i++) {
> + vring = &tm_vdev->vrings[i];
> + vring->fifo = fifo;
> + vring->num = MLXBF_TMFIFO_VRING_SIZE;
> + vring->align = SMP_CACHE_BYTES;
> + vring->index = i;
> + vring->vdev_id = tm_vdev->vdev.id.device;
> + dev = &tm_vdev->vdev.dev;
> +
> + size = vring_size(vring->num, vring->align);
> + va = dma_alloc_coherent(dev->parent, size, &dma, GFP_KERNEL);
> + if (!va) {
> + mlxbf_tmfifo_free_vrings(fifo, tm_vdev);
> + dev_err(dev->parent, "dma_alloc_coherent failed\n");
> + return -ENOMEM;
> + }
> +
> + vring->va = va;
> + vring->dma = dma;
> + }
> +
> + return 0;
> +}
> +
> +/* Disable interrupts of the FIFO device. */
> +static void mlxbf_tmfifo_disable_irqs(struct mlxbf_tmfifo *fifo)
> +{
> + int i, irq;
> +
> + for (i = 0; i < MLXBF_TM_MAX_IRQ; i++) {
> + irq = fifo->irq_info[i].irq;
> + fifo->irq_info[i].irq = 0;
> + disable_irq(irq);
> + }
> +}
> +
> +/* Interrupt handler. */
> +static irqreturn_t mlxbf_tmfifo_irq_handler(int irq, void *arg)
> +{
> + struct mlxbf_tmfifo_irq_info *irq_info = arg;
> +
> + if (!test_and_set_bit(irq_info->index, &irq_info->fifo->pend_events))
> + schedule_work(&irq_info->fifo->work);
> +
> + return IRQ_HANDLED;
> +}
> +
> +/* Get the next packet descriptor from the vring. */
> +static struct vring_desc *
> +mlxbf_tmfifo_get_next_desc(struct mlxbf_tmfifo_vring *vring)
> +{
> + const struct vring *vr = virtqueue_get_vring(vring->vq);
> + struct virtio_device *vdev = vring->vq->vdev;
> + unsigned int idx, head;
> +
> + if (vring->next_avail == virtio16_to_cpu(vdev, vr->avail->idx))
> + return NULL;
> +
> + idx = vring->next_avail % vr->num;
> + head = virtio16_to_cpu(vdev, vr->avail->ring[idx]);
> + if (WARN_ON(head >= vr->num))
> + return NULL;
> +
> + vring->next_avail++;
> +
> + return &vr->desc[head];
> +}
> +
> +/* Release virtio descriptor. */
> +static void mlxbf_tmfifo_release_desc(struct mlxbf_tmfifo_vring *vring,
> + struct vring_desc *desc, u32 len)
> +{
> + const struct vring *vr = virtqueue_get_vring(vring->vq);
> + struct virtio_device *vdev = vring->vq->vdev;
> + u16 idx, vr_idx;
> +
> + vr_idx = virtio16_to_cpu(vdev, vr->used->idx);
> + idx = vr_idx % vr->num;
> + vr->used->ring[idx].id = cpu_to_virtio32(vdev, desc - vr->desc);
> + vr->used->ring[idx].len = cpu_to_virtio32(vdev, len);
> +
> + /*
> + * Virtio could poll and check the 'idx' to decide whether the desc is
> + * done or not. Add a memory barrier here to make sure the update above
> + * completes before updating the idx.
> + */
> + mb();
> + vr->used->idx = cpu_to_virtio16(vdev, vr_idx + 1);
> +}
> +
> +/* Get the total length of the descriptor chain. */
> +static u32 mlxbf_tmfifo_get_pkt_len(struct mlxbf_tmfifo_vring *vring,
> + struct vring_desc *desc)
> +{
> + const struct vring *vr = virtqueue_get_vring(vring->vq);
> + struct virtio_device *vdev = vring->vq->vdev;
> + u32 len = 0, idx;
> +
> + while (desc) {
> + len += virtio32_to_cpu(vdev, desc->len);
> + if (!(virtio16_to_cpu(vdev, desc->flags) & VRING_DESC_F_NEXT))
> + break;
> + idx = virtio16_to_cpu(vdev, desc->next);
> + desc = &vr->desc[idx];
> + }
> +
> + return len;
> +}
> +
> +static void mlxbf_tmfifo_release_pending_pkt(struct mlxbf_tmfifo_vring *vring)
> +{
> + struct vring_desc *desc_head;
> + u32 len = 0;
> +
> + if (vring->desc_head) {
> + desc_head = vring->desc_head;
> + len = vring->pkt_len;
> + } else {
> + desc_head = mlxbf_tmfifo_get_next_desc(vring);
> + len = mlxbf_tmfifo_get_pkt_len(vring, desc_head);
> + }
> +
> + if (desc_head)
> + mlxbf_tmfifo_release_desc(vring, desc_head, len);
> +
> + vring->pkt_len = 0;
> + vring->desc = NULL;
> + vring->desc_head = NULL;
> +}
> +
> +static void mlxbf_tmfifo_init_net_desc(struct mlxbf_tmfifo_vring *vring,
> + struct vring_desc *desc, bool is_rx)
> +{
> + struct virtio_device *vdev = vring->vq->vdev;
> + struct virtio_net_hdr *net_hdr;
> +
> + net_hdr = phys_to_virt(virtio64_to_cpu(vdev, desc->addr));
> + memset(net_hdr, 0, sizeof(*net_hdr));
> +}
> +
> +/* Get and initialize the next packet. */
> +static struct vring_desc *
> +mlxbf_tmfifo_get_next_pkt(struct mlxbf_tmfifo_vring *vring, bool is_rx)
> +{
> + struct vring_desc *desc;
> +
> + desc = mlxbf_tmfifo_get_next_desc(vring);
> + if (desc && is_rx && vring->vdev_id == VIRTIO_ID_NET)
> + mlxbf_tmfifo_init_net_desc(vring, desc, is_rx);
> +
> + vring->desc_head = desc;
> + vring->desc = desc;
> +
> + return desc;
> +}
> +
> +/* House-keeping timer. */
> +static void mlxbf_tmfifo_timer(struct timer_list *t)
> +{
> + struct mlxbf_tmfifo *fifo = container_of(t, struct mlxbf_tmfifo, timer);
> + int rx, tx;
> +
> + rx = !test_and_set_bit(MLXBF_TM_RX_HWM_IRQ, &fifo->pend_events);
> + tx = !test_and_set_bit(MLXBF_TM_TX_LWM_IRQ, &fifo->pend_events);
> +
> + if (rx || tx)
> + schedule_work(&fifo->work);
> +
> + mod_timer(&fifo->timer, jiffies + MLXBF_TMFIFO_TIMER_INTERVAL);
> +}
> +
> +/* Copy one console packet into the output buffer. */
> +static void mlxbf_tmfifo_console_output_one(struct mlxbf_tmfifo_vdev *cons,
> + struct mlxbf_tmfifo_vring *vring,
> + struct vring_desc *desc)
> +{
> + const struct vring *vr = virtqueue_get_vring(vring->vq);
> + struct virtio_device *vdev = &cons->vdev;
> + u32 len, idx, seg;
> + void *addr;
> +
> + while (desc) {
> + addr = phys_to_virt(virtio64_to_cpu(vdev, desc->addr));
> + len = virtio32_to_cpu(vdev, desc->len);
> +
> + seg = CIRC_SPACE_TO_END(cons->tx_buf.head, cons->tx_buf.tail,
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE);
> + if (len <= seg) {
> + memcpy(cons->tx_buf.buf + cons->tx_buf.head, addr, len);
> + } else {
> + memcpy(cons->tx_buf.buf + cons->tx_buf.head, addr, seg);
> + addr += seg;
> + memcpy(cons->tx_buf.buf, addr, len - seg);
> + }
> + cons->tx_buf.head = (cons->tx_buf.head + len) %
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE;
> +
> + if (!(virtio16_to_cpu(vdev, desc->flags) & VRING_DESC_F_NEXT))
> + break;
> + idx = virtio16_to_cpu(vdev, desc->next);
> + desc = &vr->desc[idx];
> + }
> +}
> +
> +/* Copy console data into the output buffer. */
> +static void mlxbf_tmfifo_console_output(struct mlxbf_tmfifo_vdev *cons,
> + struct mlxbf_tmfifo_vring *vring)
> +{
> + struct vring_desc *desc;
> + u32 len, avail;
> +
> + desc = mlxbf_tmfifo_get_next_desc(vring);
> + while (desc) {
> + /* Release the packet if not enough space. */
> + len = mlxbf_tmfifo_get_pkt_len(vring, desc);
> + avail = CIRC_SPACE(cons->tx_buf.head, cons->tx_buf.tail,
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE);
> + if (len + MLXBF_TMFIFO_CON_TX_BUF_RSV_SIZE > avail) {
> + mlxbf_tmfifo_release_desc(vring, desc, len);
> + break;
> + }
> +
> + mlxbf_tmfifo_console_output_one(cons, vring, desc);
> + mlxbf_tmfifo_release_desc(vring, desc, len);
> + desc = mlxbf_tmfifo_get_next_desc(vring);
> + }
> +}
> +
> +/* Get the number of available words in Rx FIFO for receiving. */
> +static int mlxbf_tmfifo_get_rx_avail(struct mlxbf_tmfifo *fifo)
> +{
> + u64 sts;
> +
> + sts = readq(fifo->rx_base + MLXBF_TMFIFO_RX_STS);
> + return FIELD_GET(MLXBF_TMFIFO_RX_STS__COUNT_MASK, sts);
> +}
> +
> +/* Get the number of available words in the TmFifo for sending. */
> +static int mlxbf_tmfifo_get_tx_avail(struct mlxbf_tmfifo *fifo, int vdev_id)
> +{
> + int tx_reserve;
> + u32 count;
> + u64 sts;
> +
> + /* Reserve some room in FIFO for console messages. */
> + if (vdev_id == VIRTIO_ID_NET)
> + tx_reserve = fifo->tx_fifo_size / MLXBF_TMFIFO_RESERVE_RATIO;
> + else
> + tx_reserve = 1;
> +
> + sts = readq(fifo->tx_base + MLXBF_TMFIFO_TX_STS);
> + count = FIELD_GET(MLXBF_TMFIFO_TX_STS__COUNT_MASK, sts);
> + return fifo->tx_fifo_size - tx_reserve - count;
> +}
> +
> +/* Console Tx (move data from the output buffer into the TmFifo). */
> +static void mlxbf_tmfifo_console_tx(struct mlxbf_tmfifo *fifo, int avail)
> +{
> + struct mlxbf_tmfifo_msg_hdr hdr;
> + struct mlxbf_tmfifo_vdev *cons;
> + unsigned long flags;
> + int size, seg;
> + void *addr;
> + u64 data;
> +
> + /* Return if not enough space available. */
> + if (avail < MLXBF_TMFIFO_DATA_MIN_WORDS)
> + return;
> +
> + cons = fifo->vdev[VIRTIO_ID_CONSOLE];
> + if (!cons || !cons->tx_buf.buf)
> + return;
> +
> + /* Return if no data to send. */
> + size = CIRC_CNT(cons->tx_buf.head, cons->tx_buf.tail,
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE);
> + if (size == 0)
> + return;
> +
> + /* Adjust the size to available space. */
> + if (size + sizeof(hdr) > avail * sizeof(u64))
> + size = avail * sizeof(u64) - sizeof(hdr);
> +
> + /* Write header. */
> + hdr.type = VIRTIO_ID_CONSOLE;
> + hdr.len = htons(size);
> + writeq(*(u64 *)&hdr, fifo->tx_base + MLXBF_TMFIFO_TX_DATA);
> +
> + /* Use spin-lock to protect the 'cons->tx_buf'. */
> + spin_lock_irqsave(&fifo->spin_lock, flags);
> +
> + while (size > 0) {
> + addr = cons->tx_buf.buf + cons->tx_buf.tail;
> +
> + seg = CIRC_CNT_TO_END(cons->tx_buf.head, cons->tx_buf.tail,
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE);
> + if (seg >= sizeof(u64)) {
> + memcpy(&data, addr, sizeof(u64));
> + } else {
> + memcpy(&data, addr, seg);
> + memcpy((u8 *)&data + seg, cons->tx_buf.buf,
> + sizeof(u64) - seg);
> + }
> + writeq(data, fifo->tx_base + MLXBF_TMFIFO_TX_DATA);
> +
> + if (size >= sizeof(u64)) {
> + cons->tx_buf.tail = (cons->tx_buf.tail + sizeof(u64)) %
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE;
> + size -= sizeof(u64);
> + } else {
> + cons->tx_buf.tail = (cons->tx_buf.tail + size) %
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE;
> + size = 0;
> + }
> + }
> +
> + spin_unlock_irqrestore(&fifo->spin_lock, flags);
> +}
> +
> +/* Rx/Tx one word in the descriptor buffer. */
> +static void mlxbf_tmfifo_rxtx_word(struct mlxbf_tmfifo_vring *vring,
> + struct vring_desc *desc,
> + bool is_rx, int len)
> +{
> + struct virtio_device *vdev = vring->vq->vdev;
> + struct mlxbf_tmfifo *fifo = vring->fifo;
> + void *addr;
> + u64 data;
> +
> + /* Get the buffer address of this desc. */
> + addr = phys_to_virt(virtio64_to_cpu(vdev, desc->addr));
> +
> + /* Read a word from FIFO for Rx. */
> + if (is_rx)
> + data = readq(fifo->rx_base + MLXBF_TMFIFO_RX_DATA);
> +
> + if (vring->cur_len + sizeof(u64) <= len) {
> + /* The whole word. */
> + if (is_rx)
> + memcpy(addr + vring->cur_len, &data, sizeof(u64));
> + else
> + memcpy(&data, addr + vring->cur_len, sizeof(u64));
> + vring->cur_len += sizeof(u64);
> + } else {
> + /* Leftover bytes. */
> + if (is_rx)
> + memcpy(addr + vring->cur_len, &data,
> + len - vring->cur_len);
> + else
> + memcpy(&data, addr + vring->cur_len,
> + len - vring->cur_len);
> + vring->cur_len = len;
> + }
> +
> + /* Write the word into FIFO for Tx. */
> + if (!is_rx)
> + writeq(data, fifo->tx_base + MLXBF_TMFIFO_TX_DATA);
> +}
> +
> +/*
> + * Rx/Tx packet header.
> + *
> + * In Rx case, the packet might be found to belong to a different vring since
> + * the TmFifo is shared by different services. In such case, the 'vring_change'
> + * flag is set.
> + */
> +static void mlxbf_tmfifo_rxtx_header(struct mlxbf_tmfifo_vring *vring,
> + struct vring_desc *desc,
> + bool is_rx, bool *vring_change)
> +{
> + struct mlxbf_tmfifo *fifo = vring->fifo;
> + struct virtio_net_config *config;
> + struct mlxbf_tmfifo_msg_hdr hdr;
> + int vdev_id, hdr_len;
> +
> + /* Read/Write packet header. */
> + if (is_rx) {
> + /* Drain one word from the FIFO. */
> + *(u64 *)&hdr = readq(fifo->rx_base + MLXBF_TMFIFO_RX_DATA);
> +
> + /* Skip the length 0 packets (keepalive). */
> + if (hdr.len == 0)
> + return;
> +
> + /* Check packet type. */
> + if (hdr.type == VIRTIO_ID_NET) {
> + vdev_id = VIRTIO_ID_NET;
> + hdr_len = sizeof(struct virtio_net_hdr);
> + config = &fifo->vdev[vdev_id]->config.net;
> + if (ntohs(hdr.len) > config->mtu +
> + MLXBF_TMFIFO_NET_L2_OVERHEAD)
> + return;
> + } else {
> + vdev_id = VIRTIO_ID_CONSOLE;
> + hdr_len = 0;
> + }
> +
> + /*
> + * Check whether the new packet still belongs to this vring.
> + * If not, update the pkt_len of the new vring.
> + */
> + if (vdev_id != vring->vdev_id) {
> + struct mlxbf_tmfifo_vdev *tm_dev2 = fifo->vdev[vdev_id];
> +
> + if (!tm_dev2)
> + return;
> + vring->desc = desc;
> + vring = &tm_dev2->vrings[MLXBF_TMFIFO_VRING_RX];
> + *vring_change = true;
> + }
> + vring->pkt_len = ntohs(hdr.len) + hdr_len;
> + } else {
> + /* Network virtio has an extra header. */
> + hdr_len = (vring->vdev_id == VIRTIO_ID_NET) ?
> + sizeof(struct virtio_net_hdr) : 0;
> + vring->pkt_len = mlxbf_tmfifo_get_pkt_len(vring, desc);
> + hdr.type = (vring->vdev_id == VIRTIO_ID_NET) ?
> + VIRTIO_ID_NET : VIRTIO_ID_CONSOLE;
> + hdr.len = htons(vring->pkt_len - hdr_len);
> + writeq(*(u64 *)&hdr, fifo->tx_base + MLXBF_TMFIFO_TX_DATA);
> + }
> +
> + vring->cur_len = hdr_len;
> + vring->rem_len = vring->pkt_len;
> + fifo->vring[is_rx] = vring;
> +}
> +
> +/*
> + * Rx/Tx one descriptor.
> + *
> + * Return true to indicate more data available.
> + */
> +static bool mlxbf_tmfifo_rxtx_one_desc(struct mlxbf_tmfifo_vring *vring,
> + bool is_rx, int *avail)
> +{
> + const struct vring *vr = virtqueue_get_vring(vring->vq);
> + struct mlxbf_tmfifo *fifo = vring->fifo;
> + struct virtio_device *vdev;
> + bool vring_change = false;
> + struct vring_desc *desc;
> + unsigned long flags;
> + u32 len, idx;
> +
> + vdev = &fifo->vdev[vring->vdev_id]->vdev;
> +
> + /* Get the descriptor of the next packet. */
> + if (!vring->desc) {
> + desc = mlxbf_tmfifo_get_next_pkt(vring, is_rx);
> + if (!desc)
> + return false;
> + } else {
> + desc = vring->desc;
> + }
> +
> + /* Beginning of a packet. Start to Rx/Tx packet header. */
> + if (vring->pkt_len == 0) {
> + mlxbf_tmfifo_rxtx_header(vring, desc, is_rx, &vring_change);
> + (*avail)--;
> +
> + /* Return if new packet is for another ring. */
> + if (vring_change)
> + return false;
> + goto mlxbf_tmfifo_desc_done;
> + }
> +
> + /* Get the length of this desc. */
> + len = virtio32_to_cpu(vdev, desc->len);
> + if (len > vring->rem_len)
> + len = vring->rem_len;
> +
> + /* Rx/Tx one word (8 bytes) if not done. */
> + if (vring->cur_len < len) {
> + mlxbf_tmfifo_rxtx_word(vring, desc, is_rx, len);
> + (*avail)--;
> + }
> +
> + /* Check again whether it's done. */
> + if (vring->cur_len == len) {
> + vring->cur_len = 0;
> + vring->rem_len -= len;
> +
> + /* Get the next desc on the chain. */
> + if (vring->rem_len > 0 &&
> + (virtio16_to_cpu(vdev, desc->flags) & VRING_DESC_F_NEXT)) {
> + idx = virtio16_to_cpu(vdev, desc->next);
> + desc = &vr->desc[idx];
> + goto mlxbf_tmfifo_desc_done;
> + }
> +
> + /* Done and release the pending packet. */
> + mlxbf_tmfifo_release_pending_pkt(vring);
> + desc = NULL;
> + fifo->vring[is_rx] = NULL;
> +
> + /* Notify upper layer that packet is done. */
> + spin_lock_irqsave(&fifo->spin_lock, flags);
> + vring_interrupt(0, vring->vq);
> + spin_unlock_irqrestore(&fifo->spin_lock, flags);
> + }
> +
> +mlxbf_tmfifo_desc_done:
> + /* Save the current desc. */
> + vring->desc = desc;
> +
> + return true;
> +}
> +
> +/* Rx & Tx processing of a queue. */
> +static void mlxbf_tmfifo_rxtx(struct mlxbf_tmfifo_vring *vring, bool is_rx)
> +{
> + int avail = 0, devid = vring->vdev_id;
> + struct mlxbf_tmfifo *fifo;
> + bool more;
> +
> + fifo = vring->fifo;
> +
> + /* Return if vdev is not ready. */
> + if (!fifo->vdev[devid])
> + return;
> +
> + /* Return if another vring is running. */
> + if (fifo->vring[is_rx] && fifo->vring[is_rx] != vring)
> + return;
> +
> + /* Only handle console and network for now. */
> + if (WARN_ON(devid != VIRTIO_ID_NET && devid != VIRTIO_ID_CONSOLE))
> + return;
> +
> + do {
> + /* Get available FIFO space. */
> + if (avail == 0) {
> + if (is_rx)
> + avail = mlxbf_tmfifo_get_rx_avail(fifo);
> + else
> + avail = mlxbf_tmfifo_get_tx_avail(fifo, devid);
> + if (avail <= 0)
> + break;
> + }
> +
> + /* Console output always comes from the Tx buffer. */
> + if (!is_rx && devid == VIRTIO_ID_CONSOLE) {
> + mlxbf_tmfifo_console_tx(fifo, avail);
> + break;
> + }
> +
> + /* Handle one descriptor. */
> + more = mlxbf_tmfifo_rxtx_one_desc(vring, is_rx, &avail);
> + } while (more);
> +}
> +
> +/* Handle Rx or Tx queues. */
> +static void mlxbf_tmfifo_work_rxtx(struct mlxbf_tmfifo *fifo, int queue_id,
> + int irq_id, bool is_rx)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev;
> + struct mlxbf_tmfifo_vring *vring;
> + int i;
> +
> + if (!test_and_clear_bit(irq_id, &fifo->pend_events) ||
> + !fifo->irq_info[irq_id].irq)
> + return;
> +
> + for (i = 0; i < MLXBF_TMFIFO_VDEV_MAX; i++) {
> + tm_vdev = fifo->vdev[i];
> + if (tm_vdev) {
> + vring = &tm_vdev->vrings[queue_id];
> + if (vring->vq)
> + mlxbf_tmfifo_rxtx(vring, is_rx);
> + }
> + }
> +}
> +
> +/* Work handler for Rx and Tx case. */
> +static void mlxbf_tmfifo_work_handler(struct work_struct *work)
> +{
> + struct mlxbf_tmfifo *fifo;
> +
> + fifo = container_of(work, struct mlxbf_tmfifo, work);
> + if (!fifo->is_ready)
> + return;
> +
> + mutex_lock(&fifo->lock);
> +
> + /* Tx (Send data to the TmFifo). */
> + mlxbf_tmfifo_work_rxtx(fifo, MLXBF_TMFIFO_VRING_TX,
> + MLXBF_TM_TX_LWM_IRQ, false);
> +
> + /* Rx (Receive data from the TmFifo). */
> + mlxbf_tmfifo_work_rxtx(fifo, MLXBF_TMFIFO_VRING_RX,
> + MLXBF_TM_RX_HWM_IRQ, true);
> +
> + mutex_unlock(&fifo->lock);
> +}
> +
> +/* The notify function is called when new buffers are posted. */
> +static bool mlxbf_tmfifo_virtio_notify(struct virtqueue *vq)
> +{
> + struct mlxbf_tmfifo_vring *vring = vq->priv;
> + struct mlxbf_tmfifo_vdev *tm_vdev;
> + struct mlxbf_tmfifo *fifo;
> + unsigned long flags;
> +
> + fifo = vring->fifo;
> +
> + /*
> + * Virtio maintains vrings in pairs, even number ring for Rx
> + * and odd number ring for Tx.
> + */
> + if (vring->index & BIT(0)) {
> + /*
> + * Console could make blocking call with interrupts disabled.
> + * In such case, the vring needs to be served right away. For
> + * other cases, just set the TX LWM bit to start Tx in the
> + * worker handler.
> + */
> + if (vring->vdev_id == VIRTIO_ID_CONSOLE) {
> + spin_lock_irqsave(&fifo->spin_lock, flags);
> + tm_vdev = fifo->vdev[VIRTIO_ID_CONSOLE];
> + mlxbf_tmfifo_console_output(tm_vdev, vring);
> + spin_unlock_irqrestore(&fifo->spin_lock, flags);
> + } else if (test_and_set_bit(MLXBF_TM_TX_LWM_IRQ,
> + &fifo->pend_events)) {
> + return true;
> + }
> + } else {
> + if (test_and_set_bit(MLXBF_TM_RX_HWM_IRQ, &fifo->pend_events))
> + return true;
> + }
> +
> + schedule_work(&fifo->work);
> +
> + return true;
> +}
> +
> +/* Get the array of feature bits for this device. */
> +static u64 mlxbf_tmfifo_virtio_get_features(struct virtio_device *vdev)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + return tm_vdev->features;
> +}
> +
> +/* Confirm device features to use. */
> +static int mlxbf_tmfifo_virtio_finalize_features(struct virtio_device *vdev)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + tm_vdev->features = vdev->features;
> +
> + return 0;
> +}
> +
> +/* Free virtqueues found by find_vqs(). */
> +static void mlxbf_tmfifo_virtio_del_vqs(struct virtio_device *vdev)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> + struct mlxbf_tmfifo_vring *vring;
> + struct virtqueue *vq;
> + int i;
> +
> + for (i = 0; i < ARRAY_SIZE(tm_vdev->vrings); i++) {
> + vring = &tm_vdev->vrings[i];
> +
> + /* Release the pending packet. */
> + if (vring->desc)
> + mlxbf_tmfifo_release_pending_pkt(vring);
> + vq = vring->vq;
> + if (vq) {
> + vring->vq = NULL;
> + vring_del_virtqueue(vq);
> + }
> + }
> +}
> +
> +/* Create and initialize the virtual queues. */
> +static int mlxbf_tmfifo_virtio_find_vqs(struct virtio_device *vdev,
> + unsigned int nvqs,
> + struct virtqueue *vqs[],
> + vq_callback_t *callbacks[],
> + const char * const names[],
> + const bool *ctx,
> + struct irq_affinity *desc)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> + struct mlxbf_tmfifo_vring *vring;
> + struct virtqueue *vq;
> + int i, ret, size;
> +
> + if (nvqs > ARRAY_SIZE(tm_vdev->vrings))
> + return -EINVAL;
> +
> + for (i = 0; i < nvqs; ++i) {
> + if (!names[i]) {
> + ret = -EINVAL;
> + goto error;
> + }
> + vring = &tm_vdev->vrings[i];
> +
> + /* zero vring */
> + size = vring_size(vring->num, vring->align);
> + memset(vring->va, 0, size);
> + vq = vring_new_virtqueue(i, vring->num, vring->align, vdev,
> + false, false, vring->va,
> + mlxbf_tmfifo_virtio_notify,
> + callbacks[i], names[i]);
> + if (!vq) {
> + dev_err(&vdev->dev, "vring_new_virtqueue failed\n");
> + ret = -ENOMEM;
> + goto error;
> + }
> +
> + vqs[i] = vq;
> + vring->vq = vq;
> + vq->priv = vring;
> + }
> +
> + return 0;
> +
> +error:
> + mlxbf_tmfifo_virtio_del_vqs(vdev);
> + return ret;
> +}
> +
> +/* Read the status byte. */
> +static u8 mlxbf_tmfifo_virtio_get_status(struct virtio_device *vdev)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + return tm_vdev->status;
> +}
> +
> +/* Write the status byte. */
> +static void mlxbf_tmfifo_virtio_set_status(struct virtio_device *vdev,
> + u8 status)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + tm_vdev->status = status;
> +}
> +
> +/* Reset the device. Not much here for now. */
> +static void mlxbf_tmfifo_virtio_reset(struct virtio_device *vdev)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + tm_vdev->status = 0;
> +}
> +
> +/* Read the value of a configuration field. */
> +static void mlxbf_tmfifo_virtio_get(struct virtio_device *vdev,
> + unsigned int offset,
> + void *buf,
> + unsigned int len)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + if ((u64)offset + len > sizeof(tm_vdev->config))
> + return;
> +
> + memcpy(buf, (u8 *)&tm_vdev->config + offset, len);
> +}
> +
> +/* Write the value of a configuration field. */
> +static void mlxbf_tmfifo_virtio_set(struct virtio_device *vdev,
> + unsigned int offset,
> + const void *buf,
> + unsigned int len)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + if ((u64)offset + len > sizeof(tm_vdev->config))
> + return;
> +
> + memcpy((u8 *)&tm_vdev->config + offset, buf, len);
> +}
> +
> +static void tmfifo_virtio_dev_release(struct device *device)
> +{
> + struct virtio_device *vdev =
> + container_of(device, struct virtio_device, dev);
> + struct mlxbf_tmfifo_vdev *tm_vdev = mlxbf_vdev_to_tmfifo(vdev);
> +
> + kfree(tm_vdev);
> +}
> +
> +/* Virtio config operations. */
> +static const struct virtio_config_ops mlxbf_tmfifo_virtio_config_ops = {
> + .get_features = mlxbf_tmfifo_virtio_get_features,
> + .finalize_features = mlxbf_tmfifo_virtio_finalize_features,
> + .find_vqs = mlxbf_tmfifo_virtio_find_vqs,
> + .del_vqs = mlxbf_tmfifo_virtio_del_vqs,
> + .reset = mlxbf_tmfifo_virtio_reset,
> + .set_status = mlxbf_tmfifo_virtio_set_status,
> + .get_status = mlxbf_tmfifo_virtio_get_status,
> + .get = mlxbf_tmfifo_virtio_get,
> + .set = mlxbf_tmfifo_virtio_set,
> +};
> +
> +/* Create vdev for the FIFO. */
> +static int mlxbf_tmfifo_create_vdev(struct device *dev,
> + struct mlxbf_tmfifo *fifo,
> + int vdev_id, u64 features,
> + void *config, u32 size)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev, *reg_dev = NULL;
> + int ret;
> +
> + mutex_lock(&fifo->lock);
> +
> + tm_vdev = fifo->vdev[vdev_id];
> + if (tm_vdev) {
> + dev_err(dev, "vdev %d already exists\n", vdev_id);
> + ret = -EEXIST;
> + goto fail;
> + }
> +
> + tm_vdev = kzalloc(sizeof(*tm_vdev), GFP_KERNEL);
> + if (!tm_vdev) {
> + ret = -ENOMEM;
> + goto fail;
> + }
> +
> + tm_vdev->vdev.id.device = vdev_id;
> + tm_vdev->vdev.config = &mlxbf_tmfifo_virtio_config_ops;
> + tm_vdev->vdev.dev.parent = dev;
> + tm_vdev->vdev.dev.release = tmfifo_virtio_dev_release;
> + tm_vdev->features = features;
> + if (config)
> + memcpy(&tm_vdev->config, config, size);
> +
> + if (mlxbf_tmfifo_alloc_vrings(fifo, tm_vdev)) {
> + dev_err(dev, "unable to allocate vring\n");
> + ret = -ENOMEM;
> + goto vdev_fail;
> + }
> +
> + /* Allocate an output buffer for the console device. */
> + if (vdev_id == VIRTIO_ID_CONSOLE)
> + tm_vdev->tx_buf.buf = devm_kmalloc(dev,
> + MLXBF_TMFIFO_CON_TX_BUF_SIZE,
> + GFP_KERNEL);
> + fifo->vdev[vdev_id] = tm_vdev;
> +
> + /* Register the virtio device. */
> + ret = register_virtio_device(&tm_vdev->vdev);
> + reg_dev = tm_vdev;
> + if (ret) {
> + dev_err(dev, "register_virtio_device failed\n");
> + goto vdev_fail;
> + }
> +
> + mutex_unlock(&fifo->lock);
> + return 0;
> +
> +vdev_fail:
> + mlxbf_tmfifo_free_vrings(fifo, tm_vdev);
> + fifo->vdev[vdev_id] = NULL;
> + if (reg_dev)
> + put_device(&tm_vdev->vdev.dev);
> + else
> + kfree(tm_vdev);
> +fail:
> + mutex_unlock(&fifo->lock);
> + return ret;
> +}
> +
> +/* Delete vdev for the FIFO. */
> +static int mlxbf_tmfifo_delete_vdev(struct mlxbf_tmfifo *fifo, int vdev_id)
> +{
> + struct mlxbf_tmfifo_vdev *tm_vdev;
> +
> + mutex_lock(&fifo->lock);
> +
> + /* Unregister vdev. */
> + tm_vdev = fifo->vdev[vdev_id];
> + if (tm_vdev) {
> + unregister_virtio_device(&tm_vdev->vdev);
> + mlxbf_tmfifo_free_vrings(fifo, tm_vdev);
> + fifo->vdev[vdev_id] = NULL;
> + }
> +
> + mutex_unlock(&fifo->lock);
> +
> + return 0;
> +}
> +
> +/* Read the configured network MAC address from efi variable. */
> +static void mlxbf_tmfifo_get_cfg_mac(u8 *mac)
> +{
> + efi_guid_t guid = EFI_GLOBAL_VARIABLE_GUID;
> + unsigned long size = ETH_ALEN;
> + u8 buf[ETH_ALEN];
> + efi_status_t rc;
> +
> + rc = efi.get_variable(mlxbf_tmfifo_efi_name, &guid, NULL, &size, buf);
> + if (rc == EFI_SUCCESS && size == ETH_ALEN)
> + ether_addr_copy(mac, buf);
> + else
> + ether_addr_copy(mac, mlxbf_tmfifo_net_default_mac);
> +}
> +
> +/* Set TmFifo thresolds which is used to trigger interrupts. */
> +static void mlxbf_tmfifo_set_threshold(struct mlxbf_tmfifo *fifo)
> +{
> + u64 ctl;
> +
> + /* Get Tx FIFO size and set the low/high watermark. */
> + ctl = readq(fifo->tx_base + MLXBF_TMFIFO_TX_CTL);
> + fifo->tx_fifo_size =
> + FIELD_GET(MLXBF_TMFIFO_TX_CTL__MAX_ENTRIES_MASK, ctl);
> + ctl = (ctl & ~MLXBF_TMFIFO_TX_CTL__LWM_MASK) |
> + FIELD_PREP(MLXBF_TMFIFO_TX_CTL__LWM_MASK,
> + fifo->tx_fifo_size / 2);
> + ctl = (ctl & ~MLXBF_TMFIFO_TX_CTL__HWM_MASK) |
> + FIELD_PREP(MLXBF_TMFIFO_TX_CTL__HWM_MASK,
> + fifo->tx_fifo_size - 1);
> + writeq(ctl, fifo->tx_base + MLXBF_TMFIFO_TX_CTL);
> +
> + /* Get Rx FIFO size and set the low/high watermark. */
> + ctl = readq(fifo->rx_base + MLXBF_TMFIFO_RX_CTL);
> + fifo->rx_fifo_size =
> + FIELD_GET(MLXBF_TMFIFO_RX_CTL__MAX_ENTRIES_MASK, ctl);
> + ctl = (ctl & ~MLXBF_TMFIFO_RX_CTL__LWM_MASK) |
> + FIELD_PREP(MLXBF_TMFIFO_RX_CTL__LWM_MASK, 0);
> + ctl = (ctl & ~MLXBF_TMFIFO_RX_CTL__HWM_MASK) |
> + FIELD_PREP(MLXBF_TMFIFO_RX_CTL__HWM_MASK, 1);
> + writeq(ctl, fifo->rx_base + MLXBF_TMFIFO_RX_CTL);
> +}
> +
> +static void mlxbf_tmfifo_cleanup(struct mlxbf_tmfifo *fifo)
> +{
> + int i;
> +
> + fifo->is_ready = false;
> + del_timer_sync(&fifo->timer);
> + mlxbf_tmfifo_disable_irqs(fifo);
> + cancel_work_sync(&fifo->work);
> + for (i = 0; i < MLXBF_TMFIFO_VDEV_MAX; i++)
> + mlxbf_tmfifo_delete_vdev(fifo, i);
> +}
> +
> +/* Probe the TMFIFO. */
> +static int mlxbf_tmfifo_probe(struct platform_device *pdev)
> +{
> + struct virtio_net_config net_config;
> + struct device *dev = &pdev->dev;
> + struct mlxbf_tmfifo *fifo;
> + int i, rc;
> +
> + fifo = devm_kzalloc(dev, sizeof(*fifo), GFP_KERNEL);
> + if (!fifo)
> + return -ENOMEM;
> +
> + spin_lock_init(&fifo->spin_lock);
> + INIT_WORK(&fifo->work, mlxbf_tmfifo_work_handler);
> + mutex_init(&fifo->lock);
> +
> + /* Get the resource of the Rx FIFO. */
> + fifo->rx_base = devm_platform_ioremap_resource(pdev, 0);
> + if (IS_ERR(fifo->rx_base))
> + return PTR_ERR(fifo->rx_base);
> +
> + /* Get the resource of the Tx FIFO. */
> + fifo->tx_base = devm_platform_ioremap_resource(pdev, 1);
> + if (IS_ERR(fifo->tx_base))
> + return PTR_ERR(fifo->tx_base);
> +
> + platform_set_drvdata(pdev, fifo);
> +
> + timer_setup(&fifo->timer, mlxbf_tmfifo_timer, 0);
> +
> + for (i = 0; i < MLXBF_TM_MAX_IRQ; i++) {
> + fifo->irq_info[i].index = i;
> + fifo->irq_info[i].fifo = fifo;
> + fifo->irq_info[i].irq = platform_get_irq(pdev, i);
> + rc = devm_request_irq(dev, fifo->irq_info[i].irq,
> + mlxbf_tmfifo_irq_handler, 0,
> + "tmfifo", &fifo->irq_info[i]);
> + if (rc) {
> + dev_err(dev, "devm_request_irq failed\n");
> + fifo->irq_info[i].irq = 0;
> + return rc;
> + }
> + }
> +
> + mlxbf_tmfifo_set_threshold(fifo);
> +
> + /* Create the console vdev. */
> + rc = mlxbf_tmfifo_create_vdev(dev, fifo, VIRTIO_ID_CONSOLE, 0, NULL, 0);
> + if (rc)
> + goto fail;
> +
> + /* Create the network vdev. */
> + memset(&net_config, 0, sizeof(net_config));
> + net_config.mtu = ETH_DATA_LEN;
> + net_config.status = VIRTIO_NET_S_LINK_UP;
> + mlxbf_tmfifo_get_cfg_mac(net_config.mac);
> + rc = mlxbf_tmfifo_create_vdev(dev, fifo, VIRTIO_ID_NET,
> + MLXBF_TMFIFO_NET_FEATURES, &net_config,
> + sizeof(net_config));
> + if (rc)
> + goto fail;
> +
> + mod_timer(&fifo->timer, jiffies + MLXBF_TMFIFO_TIMER_INTERVAL);
> +
> + fifo->is_ready = true;
> + return 0;
> +
> +fail:
> + mlxbf_tmfifo_cleanup(fifo);
> + return rc;
> +}
> +
> +/* Device remove function. */
> +static int mlxbf_tmfifo_remove(struct platform_device *pdev)
> +{
> + struct mlxbf_tmfifo *fifo = platform_get_drvdata(pdev);
> +
> + mlxbf_tmfifo_cleanup(fifo);
> +
> + return 0;
> +}
> +
> +static const struct acpi_device_id mlxbf_tmfifo_acpi_match[] = {
> + { "MLNXBF01", 0 },
> + {}
> +};
> +MODULE_DEVICE_TABLE(acpi, mlxbf_tmfifo_acpi_match);
> +
> +static struct platform_driver mlxbf_tmfifo_driver = {
> + .probe = mlxbf_tmfifo_probe,
> + .remove = mlxbf_tmfifo_remove,
> + .driver = {
> + .name = "bf-tmfifo",
> + .acpi_match_table = mlxbf_tmfifo_acpi_match,
> + },
> +};
> +
> +module_platform_driver(mlxbf_tmfifo_driver);
> +
> +MODULE_DESCRIPTION("Mellanox BlueField SoC TmFifo Driver");
> +MODULE_LICENSE("GPL v2");
> +MODULE_AUTHOR("Mellanox Technologies");
> --
> 1.8.3.1
>


--
With Best Regards,
Andy Shevchenko