root/drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
// SPDX-License-Identifier: GPL-2.0-or-later
/*
        Copyright (C) 2004 - 2009 Ivo van Doorn <IvDoorn@gmail.com>
        <http://rt2x00.serialmonkey.com>

 */

/*
        Module: rt2x00mmio
        Abstract: rt2x00 generic mmio device routines.
 */

#include <linux/dma-mapping.h>
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/slab.h>

#include "rt2x00.h"
#include "rt2x00mmio.h"

/*
 * Register access.
 */
int rt2x00mmio_regbusy_read(struct rt2x00_dev *rt2x00dev,
                            const unsigned int offset,
                            const struct rt2x00_field32 field,
                            u32 *reg)
{
        unsigned int i;

        if (!test_bit(DEVICE_STATE_PRESENT, &rt2x00dev->flags))
                return 0;

        for (i = 0; i < REGISTER_BUSY_COUNT; i++) {
                *reg = rt2x00mmio_register_read(rt2x00dev, offset);
                if (!rt2x00_get_field32(*reg, field))
                        return 1;
                udelay(REGISTER_BUSY_DELAY);
        }

        printk_once(KERN_ERR "%s() Indirect register access failed: "
              "offset=0x%.08x, value=0x%.08x\n", __func__, offset, *reg);
        *reg = ~0;

        return 0;
}
EXPORT_SYMBOL_GPL(rt2x00mmio_regbusy_read);

bool rt2x00mmio_rxdone(struct rt2x00_dev *rt2x00dev)
{
        struct data_queue *queue = rt2x00dev->rx;
        struct queue_entry *entry;
        struct queue_entry_priv_mmio *entry_priv;
        struct skb_frame_desc *skbdesc;
        int max_rx = 16;

        while (--max_rx) {
                entry = rt2x00queue_get_entry(queue, Q_INDEX);
                entry_priv = entry->priv_data;

                if (rt2x00dev->ops->lib->get_entry_state(entry))
                        break;

                /*
                 * Fill in desc fields of the skb descriptor
                 */
                skbdesc = get_skb_frame_desc(entry->skb);
                skbdesc->desc = entry_priv->desc;
                skbdesc->desc_len = entry->queue->desc_size;

                /*
                 * DMA is already done, notify rt2x00lib that
                 * it finished successfully.
                 */
                rt2x00lib_dmastart(entry);
                rt2x00lib_dmadone(entry);

                /*
                 * Send the frame to rt2x00lib for further processing.
                 */
                rt2x00lib_rxdone(entry, GFP_ATOMIC);
        }

        return !max_rx;
}
EXPORT_SYMBOL_GPL(rt2x00mmio_rxdone);

void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop)
{
        unsigned int i;

        for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++)
                msleep(50);
}
EXPORT_SYMBOL_GPL(rt2x00mmio_flush_queue);

/*
 * Device initialization handlers.
 */
static int rt2x00mmio_alloc_queue_dma(struct rt2x00_dev *rt2x00dev,
                                      struct data_queue *queue)
{
        struct queue_entry_priv_mmio *entry_priv;
        void *addr;
        dma_addr_t dma;
        unsigned int i;

        /*
         * Allocate DMA memory for descriptor and buffer.
         */
        addr = dma_alloc_coherent(rt2x00dev->dev,
                                  queue->limit * queue->desc_size, &dma,
                                  GFP_KERNEL);
        if (!addr)
                return -ENOMEM;

        /*
         * Initialize all queue entries to contain valid addresses.
         */
        for (i = 0; i < queue->limit; i++) {
                entry_priv = queue->entries[i].priv_data;
                entry_priv->desc = addr + i * queue->desc_size;
                entry_priv->desc_dma = dma + i * queue->desc_size;
        }

        return 0;
}

static void rt2x00mmio_free_queue_dma(struct rt2x00_dev *rt2x00dev,
                                      struct data_queue *queue)
{
        struct queue_entry_priv_mmio *entry_priv =
            queue->entries[0].priv_data;

        if (entry_priv->desc)
                dma_free_coherent(rt2x00dev->dev,
                                  queue->limit * queue->desc_size,
                                  entry_priv->desc, entry_priv->desc_dma);
        entry_priv->desc = NULL;
}

int rt2x00mmio_initialize(struct rt2x00_dev *rt2x00dev)
{
        struct data_queue *queue;
        int status;

        /*
         * Allocate DMA
         */
        queue_for_each(rt2x00dev, queue) {
                status = rt2x00mmio_alloc_queue_dma(rt2x00dev, queue);
                if (status)
                        goto exit;
        }

        /*
         * Register interrupt handler.
         */
        status = request_irq(rt2x00dev->irq,
                             rt2x00dev->ops->lib->irq_handler,
                             IRQF_SHARED, rt2x00dev->name, rt2x00dev);
        if (status) {
                rt2x00_err(rt2x00dev, "IRQ %d allocation failed (error %d)\n",
                           rt2x00dev->irq, status);
                goto exit;
        }

        return 0;

exit:
        queue_for_each(rt2x00dev, queue)
                rt2x00mmio_free_queue_dma(rt2x00dev, queue);

        return status;
}
EXPORT_SYMBOL_GPL(rt2x00mmio_initialize);

void rt2x00mmio_uninitialize(struct rt2x00_dev *rt2x00dev)
{
        struct data_queue *queue;

        /*
         * Free irq line.
         */
        free_irq(rt2x00dev->irq, rt2x00dev);

        /*
         * Free DMA
         */
        queue_for_each(rt2x00dev, queue)
                rt2x00mmio_free_queue_dma(rt2x00dev, queue);
}
EXPORT_SYMBOL_GPL(rt2x00mmio_uninitialize);

/*
 * rt2x00mmio module information.
 */
MODULE_AUTHOR(DRV_PROJECT);
MODULE_VERSION(DRV_VERSION);
MODULE_DESCRIPTION("rt2x00 mmio library");
MODULE_LICENSE("GPL");