root/drivers/net/ethernet/intel/igc/igc_tsn.c
// SPDX-License-Identifier: GPL-2.0
/* Copyright (c)  2019 Intel Corporation */

#include "igc.h"
#include "igc_base.h"
#include "igc_hw.h"
#include "igc_tsn.h"

#define MIN_MULTPLIER_TX_MIN_FRAG       0
#define MAX_MULTPLIER_TX_MIN_FRAG       3
/* Frag size is based on the Section 8.12.2 of the SW User Manual */
#define TX_MIN_FRAG_SIZE                64
#define TX_MAX_FRAG_SIZE        (TX_MIN_FRAG_SIZE * \
                                 (MAX_MULTPLIER_TX_MIN_FRAG + 1))

enum tx_queue {
        TX_QUEUE_0 = 0,
        TX_QUEUE_1,
        TX_QUEUE_2,
        TX_QUEUE_3,
};

DEFINE_STATIC_KEY_FALSE(igc_fpe_enabled);

static int igc_fpe_init_smd_frame(struct igc_ring *ring,
                                  struct igc_tx_buffer *buffer,
                                  struct sk_buff *skb)
{
        dma_addr_t dma = dma_map_single(ring->dev, skb->data, skb->len,
                                        DMA_TO_DEVICE);

        if (dma_mapping_error(ring->dev, dma)) {
                netdev_err_once(ring->netdev, "Failed to map DMA for TX\n");
                return -ENOMEM;
        }

        buffer->skb = skb;
        buffer->protocol = 0;
        buffer->bytecount = skb->len;
        buffer->gso_segs = 1;
        buffer->time_stamp = jiffies;
        dma_unmap_len_set(buffer, len, skb->len);
        dma_unmap_addr_set(buffer, dma, dma);

        return 0;
}

static int igc_fpe_init_tx_descriptor(struct igc_ring *ring,
                                      struct sk_buff *skb,
                                      enum igc_txd_popts_type type)
{
        u32 cmd_type, olinfo_status = 0;
        struct igc_tx_buffer *buffer;
        union igc_adv_tx_desc *desc;
        int err;

        if (!igc_desc_unused(ring))
                return -EBUSY;

        buffer = &ring->tx_buffer_info[ring->next_to_use];
        err = igc_fpe_init_smd_frame(ring, buffer, skb);
        if (err)
                return err;

        cmd_type = IGC_ADVTXD_DTYP_DATA | IGC_ADVTXD_DCMD_DEXT |
                   IGC_ADVTXD_DCMD_IFCS | IGC_TXD_DCMD |
                   buffer->bytecount;

        olinfo_status |= FIELD_PREP(IGC_ADVTXD_PAYLEN_MASK, buffer->bytecount);

        switch (type) {
        case SMD_V:
        case SMD_R:
                olinfo_status |= FIELD_PREP(IGC_TXD_POPTS_SMD_MASK, type);
                break;
        }

        desc = IGC_TX_DESC(ring, ring->next_to_use);
        desc->read.cmd_type_len = cpu_to_le32(cmd_type);
        desc->read.olinfo_status = cpu_to_le32(olinfo_status);
        desc->read.buffer_addr = cpu_to_le64(dma_unmap_addr(buffer, dma));

        netdev_tx_sent_queue(txring_txq(ring), skb->len);

        buffer->next_to_watch = desc;
        ring->next_to_use = (ring->next_to_use + 1) % ring->count;

        return 0;
}

static int igc_fpe_xmit_smd_frame(struct igc_adapter *adapter,
                                  enum igc_txd_popts_type type)
{
        int cpu = smp_processor_id();
        struct netdev_queue *nq;
        struct igc_ring *ring;
        struct sk_buff *skb;
        int err;

        ring = igc_get_tx_ring(adapter, cpu);
        nq = txring_txq(ring);

        skb = alloc_skb(SMD_FRAME_SIZE, GFP_ATOMIC);
        if (!skb)
                return -ENOMEM;

        skb_put_zero(skb, SMD_FRAME_SIZE);

        __netif_tx_lock(nq, cpu);

        err = igc_fpe_init_tx_descriptor(ring, skb, type);
        igc_flush_tx_descriptors(ring);

        __netif_tx_unlock(nq);

        return err;
}

static void igc_fpe_configure_tx(struct ethtool_mmsv *mmsv, bool tx_enable)
{
        struct igc_fpe_t *fpe = container_of(mmsv, struct igc_fpe_t, mmsv);
        struct igc_adapter *adapter;

        adapter = container_of(fpe, struct igc_adapter, fpe);
        adapter->fpe.tx_enabled = tx_enable;

        /* Update config since tx_enabled affects preemptible queue configuration */
        igc_tsn_offload_apply(adapter);
}

static void igc_fpe_send_mpacket(struct ethtool_mmsv *mmsv,
                                 enum ethtool_mpacket type)
{
        struct igc_fpe_t *fpe = container_of(mmsv, struct igc_fpe_t, mmsv);
        struct igc_adapter *adapter;
        int err;

        adapter = container_of(fpe, struct igc_adapter, fpe);

        if (type == ETHTOOL_MPACKET_VERIFY) {
                err = igc_fpe_xmit_smd_frame(adapter, SMD_V);
                if (err && net_ratelimit())
                        netdev_err(adapter->netdev, "Error sending SMD-V\n");
        } else if (type == ETHTOOL_MPACKET_RESPONSE) {
                err = igc_fpe_xmit_smd_frame(adapter, SMD_R);
                if (err && net_ratelimit())
                        netdev_err(adapter->netdev, "Error sending SMD-R frame\n");
        }
}

static const struct ethtool_mmsv_ops igc_mmsv_ops = {
        .configure_tx = igc_fpe_configure_tx,
        .send_mpacket = igc_fpe_send_mpacket,
};

void igc_fpe_init(struct igc_adapter *adapter)
{
        adapter->fpe.tx_min_frag_size = TX_MIN_FRAG_SIZE;
        adapter->fpe.tx_enabled = false;
        ethtool_mmsv_init(&adapter->fpe.mmsv, adapter->netdev, &igc_mmsv_ops);
}

void igc_fpe_clear_preempt_queue(struct igc_adapter *adapter)
{
        for (int i = 0; i < adapter->num_tx_queues; i++) {
                struct igc_ring *tx_ring = adapter->tx_ring[i];

                tx_ring->preemptible = false;
        }
}

static u32 igc_fpe_map_preempt_tc_to_queue(const struct igc_adapter *adapter,
                                           unsigned long preemptible_tcs)
{
        struct net_device *dev = adapter->netdev;
        u32 i, queue = 0;

        for (i = 0; i < dev->num_tc; i++) {
                u32 offset, count;

                if (!(preemptible_tcs & BIT(i)))
                        continue;

                offset = dev->tc_to_txq[i].offset;
                count = dev->tc_to_txq[i].count;
                queue |= GENMASK(offset + count - 1, offset);
        }

        return queue;
}

void igc_fpe_save_preempt_queue(struct igc_adapter *adapter,
                                const struct tc_mqprio_qopt_offload *mqprio)
{
        u32 preemptible_queue = igc_fpe_map_preempt_tc_to_queue(adapter,
                                                                mqprio->preemptible_tcs);

        for (int i = 0; i < adapter->num_tx_queues; i++) {
                struct igc_ring *tx_ring = adapter->tx_ring[i];

                tx_ring->preemptible = !!(preemptible_queue & BIT(i));
        }
}

static bool is_any_launchtime(struct igc_adapter *adapter)
{
        int i;

        for (i = 0; i < adapter->num_tx_queues; i++) {
                struct igc_ring *ring = adapter->tx_ring[i];

                if (ring->launchtime_enable)
                        return true;
        }

        return false;
}

static bool is_cbs_enabled(struct igc_adapter *adapter)
{
        int i;

        for (i = 0; i < adapter->num_tx_queues; i++) {
                struct igc_ring *ring = adapter->tx_ring[i];

                if (ring->cbs_enable)
                        return true;
        }

        return false;
}

static unsigned int igc_tsn_new_flags(struct igc_adapter *adapter)
{
        unsigned int new_flags = adapter->flags & ~IGC_FLAG_TSN_ANY_ENABLED;


        if (adapter->taprio_offload_enable || is_any_launchtime(adapter) ||
            adapter->strict_priority_enable)
                new_flags |= IGC_FLAG_TSN_QBV_ENABLED;

        if (is_cbs_enabled(adapter))
                new_flags |= IGC_FLAG_TSN_QAV_ENABLED;

        if (adapter->fpe.mmsv.pmac_enabled)
                new_flags |= IGC_FLAG_TSN_PREEMPT_ENABLED;

        return new_flags;
}

static bool igc_tsn_is_tx_mode_in_tsn(struct igc_adapter *adapter)
{
        struct igc_hw *hw = &adapter->hw;

        return !!(rd32(IGC_TQAVCTRL) & IGC_TQAVCTRL_TRANSMIT_MODE_TSN);
}

void igc_tsn_adjust_txtime_offset(struct igc_adapter *adapter)
{
        struct igc_hw *hw = &adapter->hw;
        u16 txoffset;

        if (!igc_tsn_is_tx_mode_in_tsn(adapter))
                return;

        switch (adapter->link_speed) {
        case SPEED_10:
                txoffset = IGC_TXOFFSET_SPEED_10;
                break;
        case SPEED_100:
                txoffset = IGC_TXOFFSET_SPEED_100;
                break;
        case SPEED_1000:
                txoffset = IGC_TXOFFSET_SPEED_1000;
                break;
        case SPEED_2500:
                txoffset = IGC_TXOFFSET_SPEED_2500;
                break;
        default:
                txoffset = 0;
                break;
        }

        wr32(IGC_GTXOFFSET, txoffset);
}

static void igc_tsn_restore_retx_default(struct igc_adapter *adapter)
{
        struct igc_hw *hw = &adapter->hw;
        u32 retxctl;

        retxctl = rd32(IGC_RETX_CTL) & IGC_RETX_CTL_WATERMARK_MASK;
        wr32(IGC_RETX_CTL, retxctl);
}

bool igc_tsn_is_taprio_activated_by_user(struct igc_adapter *adapter)
{
        struct igc_hw *hw = &adapter->hw;

        return (rd32(IGC_BASET_H) || rd32(IGC_BASET_L)) &&
                adapter->taprio_offload_enable;
}

static void igc_tsn_tx_arb(struct igc_adapter *adapter, bool reverse_prio)
{
        struct igc_hw *hw = &adapter->hw;
        u32 txarb;

        txarb = rd32(IGC_TXARB);

        txarb &= ~(IGC_TXARB_TXQ_PRIO_0_MASK |
                   IGC_TXARB_TXQ_PRIO_1_MASK |
                   IGC_TXARB_TXQ_PRIO_2_MASK |
                   IGC_TXARB_TXQ_PRIO_3_MASK);

        if (reverse_prio) {
                txarb |= IGC_TXARB_TXQ_PRIO_0(TX_QUEUE_3);
                txarb |= IGC_TXARB_TXQ_PRIO_1(TX_QUEUE_2);
                txarb |= IGC_TXARB_TXQ_PRIO_2(TX_QUEUE_1);
                txarb |= IGC_TXARB_TXQ_PRIO_3(TX_QUEUE_0);
        } else {
                txarb |= IGC_TXARB_TXQ_PRIO_0(TX_QUEUE_0);
                txarb |= IGC_TXARB_TXQ_PRIO_1(TX_QUEUE_1);
                txarb |= IGC_TXARB_TXQ_PRIO_2(TX_QUEUE_2);
                txarb |= IGC_TXARB_TXQ_PRIO_3(TX_QUEUE_3);
        }

        wr32(IGC_TXARB, txarb);
}

/**
 * igc_tsn_set_rxpbsize - Set the receive packet buffer size
 * @adapter: Pointer to the igc_adapter structure
 * @rxpbs_exp_bmc_be: Value to set the receive packet buffer size, including
 *                    express buffer, BMC buffer, and Best Effort buffer
 *
 * The IGC_RXPBS register value may include allocations for the Express buffer,
 * BMC buffer, Best Effort buffer, and the timestamp descriptor buffer
 * (IGC_RXPBS_CFG_TS_EN).
 */
static void igc_tsn_set_rxpbsize(struct igc_adapter *adapter,
                                 u32 rxpbs_exp_bmc_be)
{
        struct igc_hw *hw = &adapter->hw;
        u32 rxpbs = rd32(IGC_RXPBS);

        rxpbs &= ~(IGC_RXPBSIZE_EXP_MASK | IGC_BMC2OSPBSIZE_MASK |
                   IGC_RXPBSIZE_BE_MASK);
        rxpbs |= rxpbs_exp_bmc_be;

        wr32(IGC_RXPBS, rxpbs);
}

/* Returns the TSN specific registers to their default values after
 * the adapter is reset.
 */
static int igc_tsn_disable_offload(struct igc_adapter *adapter)
{
        struct igc_hw *hw = &adapter->hw;
        u32 tqavctrl;
        int i;

        wr32(IGC_GTXOFFSET, 0);
        wr32(IGC_TXPBS, IGC_TXPBSIZE_DEFAULT);
        wr32(IGC_DTXMXPKTSZ, IGC_DTXMXPKTSZ_DEFAULT);

        igc_tsn_set_rxpbsize(adapter, IGC_RXPBSIZE_EXP_BMC_DEFAULT);

        if (igc_is_device_id_i226(hw))
                igc_tsn_restore_retx_default(adapter);

        tqavctrl = rd32(IGC_TQAVCTRL);
        tqavctrl &= ~(IGC_TQAVCTRL_TRANSMIT_MODE_TSN |
                      IGC_TQAVCTRL_ENHANCED_QAV | IGC_TQAVCTRL_FUTSCDDIS |
                      IGC_TQAVCTRL_PREEMPT_ENA | IGC_TQAVCTRL_MIN_FRAG_MASK);

        wr32(IGC_TQAVCTRL, tqavctrl);

        for (i = 0; i < adapter->num_tx_queues; i++) {
                int reg_idx = adapter->tx_ring[i]->reg_idx;
                u32 txdctl;

                wr32(IGC_TXQCTL(i), 0);
                wr32(IGC_STQT(i), 0);
                wr32(IGC_ENDQT(i), NSEC_PER_SEC);

                txdctl = rd32(IGC_TXDCTL(reg_idx));
                txdctl &= ~IGC_TXDCTL_PRIORITY_HIGH;
                wr32(IGC_TXDCTL(reg_idx), txdctl);
        }

        wr32(IGC_QBVCYCLET_S, 0);
        wr32(IGC_QBVCYCLET, NSEC_PER_SEC);

        /* Restore the default Tx arbitration: Priority 0 has the highest
         * priority and is assigned to queue 0 and so on and so forth.
         */
        igc_tsn_tx_arb(adapter, false);

        adapter->flags &= ~IGC_FLAG_TSN_QBV_ENABLED;

        return 0;
}

/* To partially fix i226 HW errata, reduce MAC internal buffering from 192 Bytes
 * to 88 Bytes by setting RETX_CTL register using the recommendation from:
 * a) Ethernet Controller I225/I226 Specification Update Rev 2.1
 *    Item 9: TSN: Packet Transmission Might Cross the Qbv Window
 * b) I225/6 SW User Manual Rev 1.2.4: Section 8.11.5 Retry Buffer Control
 */
static void igc_tsn_set_retx_qbvfullthreshold(struct igc_adapter *adapter)
{
        struct igc_hw *hw = &adapter->hw;
        u32 retxctl, watermark;

        retxctl = rd32(IGC_RETX_CTL);
        watermark = retxctl & IGC_RETX_CTL_WATERMARK_MASK;
        /* Set QBVFULLTH value using watermark and set QBVFULLEN */
        retxctl |= (watermark << IGC_RETX_CTL_QBVFULLTH_SHIFT) |
                   IGC_RETX_CTL_QBVFULLEN;
        wr32(IGC_RETX_CTL, retxctl);
}

static u8 igc_fpe_get_frag_size_mult(const struct igc_fpe_t *fpe)
{
        u8 mult = (fpe->tx_min_frag_size / TX_MIN_FRAG_SIZE) - 1;

        return clamp_t(u8, mult, MIN_MULTPLIER_TX_MIN_FRAG,
                       MAX_MULTPLIER_TX_MIN_FRAG);
}

u32 igc_fpe_get_supported_frag_size(u32 frag_size)
{
        static const u32 supported_sizes[] = { 64, 128, 192, 256 };

        /* Find the smallest supported size that is >= frag_size */
        for (int i = 0; i < ARRAY_SIZE(supported_sizes); i++) {
                if (frag_size <= supported_sizes[i])
                        return supported_sizes[i];
        }

        /* Should not happen */
        return TX_MAX_FRAG_SIZE;
}

static int igc_tsn_enable_offload(struct igc_adapter *adapter)
{
        struct igc_hw *hw = &adapter->hw;
        u32 tqavctrl, baset_l, baset_h;
        u32 sec, nsec, cycle;
        ktime_t base_time, systim;
        u32 frag_size_mult;
        int i;

        wr32(IGC_TSAUXC, 0);
        wr32(IGC_DTXMXPKTSZ, IGC_DTXMXPKTSZ_TSN);
        wr32(IGC_TXPBS, IGC_TXPBSIZE_TSN);

        igc_tsn_set_rxpbsize(adapter, IGC_RXPBSIZE_EXP_BMC_BE_TSN);

        if (igc_is_device_id_i226(hw))
                igc_tsn_set_retx_qbvfullthreshold(adapter);

        if (adapter->strict_priority_enable ||
            adapter->flags & IGC_FLAG_TSN_REVERSE_TXQ_PRIO)
                igc_tsn_tx_arb(adapter, true);

        for (i = 0; i < adapter->num_tx_queues; i++) {
                struct igc_ring *ring = adapter->tx_ring[i];
                u32 txdctl = rd32(IGC_TXDCTL(ring->reg_idx));
                u32 txqctl = 0;
                u16 cbs_value;
                u32 tqavcc;

                wr32(IGC_STQT(i), ring->start_time);
                wr32(IGC_ENDQT(i), ring->end_time);

                if (adapter->taprio_offload_enable) {
                        /* If taprio_offload_enable is set we are in "taprio"
                         * mode and we need to be strict about the
                         * cycles: only transmit a packet if it can be
                         * completed during that cycle.
                         *
                         * If taprio_offload_enable is NOT true when
                         * enabling TSN offload, the cycle should have
                         * no external effects, but is only used internally
                         * to adapt the base time register after a second
                         * has passed.
                         *
                         * Enabling strict mode in this case would
                         * unnecessarily prevent the transmission of
                         * certain packets (i.e. at the boundary of a
                         * second) and thus interfere with the launchtime
                         * feature that promises transmission at a
                         * certain point in time.
                         */
                        txqctl |= IGC_TXQCTL_STRICT_CYCLE |
                                IGC_TXQCTL_STRICT_END;
                }

                if (ring->launchtime_enable)
                        txqctl |= IGC_TXQCTL_QUEUE_MODE_LAUNCHT;

                if (!adapter->fpe.tx_enabled) {
                        /* fpe inactive: clear both flags */
                        txqctl &= ~IGC_TXQCTL_PREEMPTIBLE;
                        txdctl &= ~IGC_TXDCTL_PRIORITY_HIGH;
                } else if (ring->preemptible) {
                        /* fpe active + preemptible: enable preemptible queue + set low priority */
                        txqctl |= IGC_TXQCTL_PREEMPTIBLE;
                        txdctl &= ~IGC_TXDCTL_PRIORITY_HIGH;
                } else {
                        /* fpe active + express: enable express queue + set high priority */
                        txqctl &= ~IGC_TXQCTL_PREEMPTIBLE;
                        txdctl |= IGC_TXDCTL_PRIORITY_HIGH;
                }

                wr32(IGC_TXDCTL(ring->reg_idx), txdctl);

                /* Skip configuring CBS for Q2 and Q3 */
                if (i > 1)
                        goto skip_cbs;

                if (ring->cbs_enable) {
                        if (i == 0)
                                txqctl |= IGC_TXQCTL_QAV_SEL_CBS0;
                        else
                                txqctl |= IGC_TXQCTL_QAV_SEL_CBS1;

                        /* According to i225 datasheet section 7.5.2.7, we
                         * should set the 'idleSlope' field from TQAVCC
                         * register following the equation:
                         *
                         * value = link-speed   0x7736 * BW * 0.2
                         *         ---------- *  -----------------         (E1)
                         *          100Mbps            2.5
                         *
                         * Note that 'link-speed' is in Mbps.
                         *
                         * 'BW' is the percentage bandwidth out of full
                         * link speed which can be found with the
                         * following equation. Note that idleSlope here
                         * is the parameter from this function
                         * which is in kbps.
                         *
                         *     BW =     idleSlope
                         *          -----------------                      (E2)
                         *          link-speed * 1000
                         *
                         * That said, we can come up with a generic
                         * equation to calculate the value we should set
                         * it TQAVCC register by replacing 'BW' in E1 by E2.
                         * The resulting equation is:
                         *
                         * value = link-speed * 0x7736 * idleSlope * 0.2
                         *         -------------------------------------   (E3)
                         *             100 * 2.5 * link-speed * 1000
                         *
                         * 'link-speed' is present in both sides of the
                         * fraction so it is canceled out. The final
                         * equation is the following:
                         *
                         *     value = idleSlope * 61036
                         *             -----------------                   (E4)
                         *                  2500000
                         *
                         * NOTE: For i225, given the above, we can see
                         *       that idleslope is represented in
                         *       40.959433 kbps units by the value at
                         *       the TQAVCC register (2.5Gbps / 61036),
                         *       which reduces the granularity for
                         *       idleslope increments.
                         *
                         * In i225 controller, the sendSlope and loCredit
                         * parameters from CBS are not configurable
                         * by software so we don't do any
                         * 'controller configuration' in respect to
                         * these parameters.
                         */
                        cbs_value = DIV_ROUND_UP_ULL(ring->idleslope
                                                     * 61036ULL, 2500000);

                        tqavcc = rd32(IGC_TQAVCC(i));
                        tqavcc &= ~IGC_TQAVCC_IDLESLOPE_MASK;
                        tqavcc |= cbs_value | IGC_TQAVCC_KEEP_CREDITS;
                        wr32(IGC_TQAVCC(i), tqavcc);

                        wr32(IGC_TQAVHC(i),
                             0x80000000 + ring->hicredit * 0x7736);
                } else {
                        /* Disable any CBS for the queue */
                        txqctl &= ~(IGC_TXQCTL_QAV_SEL_MASK);

                        /* Set idleSlope to zero. */
                        tqavcc = rd32(IGC_TQAVCC(i));
                        tqavcc &= ~(IGC_TQAVCC_IDLESLOPE_MASK |
                                    IGC_TQAVCC_KEEP_CREDITS);
                        wr32(IGC_TQAVCC(i), tqavcc);

                        /* Set hiCredit to zero. */
                        wr32(IGC_TQAVHC(i), 0);
                }
skip_cbs:
                wr32(IGC_TXQCTL(i), txqctl);
        }

        tqavctrl = rd32(IGC_TQAVCTRL) & ~(IGC_TQAVCTRL_FUTSCDDIS |
                   IGC_TQAVCTRL_PREEMPT_ENA | IGC_TQAVCTRL_MIN_FRAG_MASK);
        tqavctrl |= IGC_TQAVCTRL_TRANSMIT_MODE_TSN | IGC_TQAVCTRL_ENHANCED_QAV;

        if (adapter->fpe.mmsv.pmac_enabled)
                tqavctrl |= IGC_TQAVCTRL_PREEMPT_ENA;

        frag_size_mult = igc_fpe_get_frag_size_mult(&adapter->fpe);
        tqavctrl |= FIELD_PREP(IGC_TQAVCTRL_MIN_FRAG_MASK, frag_size_mult);

        adapter->qbv_count++;

        cycle = adapter->cycle_time;
        base_time = adapter->base_time;

        nsec = rd32(IGC_SYSTIML);
        sec = rd32(IGC_SYSTIMH);

        systim = ktime_set(sec, nsec);
        if (ktime_compare(systim, base_time) > 0) {
                s64 n = div64_s64(ktime_sub_ns(systim, base_time), cycle);

                base_time = ktime_add_ns(base_time, (n + 1) * cycle);
        } else {
                if (igc_is_device_id_i226(hw)) {
                        ktime_t adjust_time, expires_time;

                       /* According to datasheet section 7.5.2.9.3.3, FutScdDis bit
                        * has to be configured before the cycle time and base time.
                        * Tx won't hang if a GCL is already running,
                        * so in this case we don't need to set FutScdDis.
                        */
                        if (!(rd32(IGC_BASET_H) || rd32(IGC_BASET_L)))
                                tqavctrl |= IGC_TQAVCTRL_FUTSCDDIS;

                        nsec = rd32(IGC_SYSTIML);
                        sec = rd32(IGC_SYSTIMH);
                        systim = ktime_set(sec, nsec);

                        adjust_time = adapter->base_time;
                        expires_time = ktime_sub_ns(adjust_time, systim);
                        hrtimer_start(&adapter->hrtimer, expires_time, HRTIMER_MODE_REL);
                }
        }

        wr32(IGC_TQAVCTRL, tqavctrl);

        wr32(IGC_QBVCYCLET_S, cycle);
        wr32(IGC_QBVCYCLET, cycle);

        baset_h = div_s64_rem(base_time, NSEC_PER_SEC, &baset_l);
        wr32(IGC_BASET_H, baset_h);

        /* In i226, Future base time is only supported when FutScdDis bit
         * is enabled and only active for re-configuration.
         * In this case, initialize the base time with zero to create
         * "re-configuration" scenario then only set the desired base time.
         */
        if (tqavctrl & IGC_TQAVCTRL_FUTSCDDIS)
                wr32(IGC_BASET_L, 0);
        wr32(IGC_BASET_L, baset_l);

        return 0;
}

int igc_tsn_reset(struct igc_adapter *adapter)
{
        unsigned int new_flags;
        int err = 0;

        if (adapter->fpe.mmsv.pmac_enabled) {
                err = igc_enable_empty_addr_recv(adapter);
                if (err && net_ratelimit())
                        netdev_err(adapter->netdev, "Error adding empty address to MAC filter\n");
        } else {
                igc_disable_empty_addr_recv(adapter);
        }

        new_flags = igc_tsn_new_flags(adapter);

        if (!(new_flags & IGC_FLAG_TSN_ANY_ENABLED))
                return igc_tsn_disable_offload(adapter);

        err = igc_tsn_enable_offload(adapter);
        if (err < 0)
                return err;

        adapter->flags = new_flags;

        return err;
}

static bool igc_tsn_will_tx_mode_change(struct igc_adapter *adapter)
{
        bool any_tsn_enabled = !!(igc_tsn_new_flags(adapter) &
                                  IGC_FLAG_TSN_ANY_ENABLED);

        return (any_tsn_enabled && !igc_tsn_is_tx_mode_in_tsn(adapter)) ||
               (!any_tsn_enabled && igc_tsn_is_tx_mode_in_tsn(adapter));
}

int igc_tsn_offload_apply(struct igc_adapter *adapter)
{
        /* Per I225/6 HW Design Section 7.5.2.1 guideline, if tx mode change
         * from legacy->tsn or tsn->legacy, then reset adapter is needed.
         */
        if (netif_running(adapter->netdev) &&
            igc_tsn_will_tx_mode_change(adapter)) {
                schedule_work(&adapter->reset_task);
                return 0;
        }

        igc_tsn_reset(adapter);

        return 0;
}