--- /dev/null
+/* $OpenBSD: if_aq_pci.c,v 1.1 2021/09/02 10:11:21 mlarkin Exp $ */
+/* $NetBSD: if_aq.c,v 1.27 2021/06/16 00:21:18 riastradh Exp $ */
+
+/*
+ * Copyright (c) 2021 Jonathan Matthew <jonathan@d14n.org>
+ * Copyright (c) 2021 Mike Larkin <mlarkin@openbsd.org>
+ *
+ * Permission to use, copy, modify, and distribute this software for any
+ * purpose with or without fee is hereby granted, provided that the above
+ * copyright notice and this permission notice appear in all copies.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
+ * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
+ * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
+ * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
+ * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
+ * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
+ * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
+ */
+
+/**
+ * aQuantia Corporation Network Driver
+ * Copyright (C) 2014-2017 aQuantia Corporation. All rights reserved
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ *
+ * (1) Redistributions of source code must retain the above
+ * copyright notice, this list of conditions and the following
+ * disclaimer.
+ *
+ * (2) Redistributions in binary form must reproduce the above
+ * copyright notice, this list of conditions and the following
+ * disclaimer in the documentation and/or other materials provided
+ * with the distribution.
+ *
+ * (3) The name of the author may not be used to endorse or promote
+ * products derived from this software without specific prior
+ * written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
+ * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
+ * GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
+ * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
+ * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ *
+ */
+
+/*-
+ * Copyright (c) 2020 Ryo Shimizu <ryo@nerv.org>
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
+ * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
+ * DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT,
+ * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
+ * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
+ * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
+ * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+#include <sys/types.h>
+#include <sys/device.h>
+#include <sys/param.h>
+#include <sys/kernel.h>
+#include <sys/sockio.h>
+#include <sys/systm.h>
+
+#include <net/if.h>
+#include <net/if_media.h>
+
+#include <netinet/in.h>
+#include <netinet/if_ether.h>
+
+#include <dev/pci/pcireg.h>
+#include <dev/pci/pcivar.h>
+#include <dev/pci/pcidevs.h>
+
+/* #define AQ_DEBUG 1 */
+#ifdef AQ_DEBUG
+#define DPRINTF(x) printf x
+#else
+#define DPRINTF(x)
+#endif /* AQ_DEBUG */
+
+#define DEVNAME(_s) ((_s)->sc_dev.dv_xname)
+
+#define AQ_BAR0 0x10
+#define AQ_MAXQ 8
+
+#define AQ_TXD_NUM 2048
+#define AQ_RXD_NUM 2048
+
+#define AQ_TX_MAX_SEGMENTS 1 /* XXX */
+
+#define AQ_LINKSTAT_IRQ 31
+
+#define RPF_ACTION_HOST 1
+
+#define AQ_FW_SOFTRESET_REG 0x0000
+#define AQ_FW_SOFTRESET_DIS (1 << 14)
+#define AQ_FW_SOFTRESET_RESET (1 << 15)
+#define AQ_FW_VERSION_REG 0x0018
+#define AQ_HW_REVISION_REG 0x001c
+#define AQ_GLB_NVR_INTERFACE1_REG 0x0100
+#define AQ_FW_MBOX_CMD_REG 0x0200
+#define AQ_FW_MBOX_CMD_EXECUTE 0x00008000
+#define AQ_FW_MBOX_CMD_BUSY 0x00000100
+#define AQ_FW_MBOX_ADDR_REG 0x0208
+#define AQ_FW_MBOX_VAL_REG 0x020C
+#define AQ_FW_GLB_CPU_SEM_REG(i) (0x03a0 + (i) * 4)
+#define AQ_FW_SEM_RAM_REG AQ_FW_GLB_CPU_SEM_REG(2)
+#define AQ_FW_GLB_CTL2_REG 0x0404
+#define AQ_GLB_GENERAL_PROVISIONING9_REG 0x0520
+#define AQ_GLB_NVR_PROVISIONING2_REG 0x0534
+#define AQ_INTR_STATUS_REG 0x2000 /* intr status */
+#define AQ_INTR_STATUS_CLR_REG 0x2050 /* intr status clear */
+#define AQ_INTR_MASK_REG 0x2060 /* intr mask set */
+#define AQ_INTR_MASK_CLR_REG 0x2070 /* intr mask clear */
+#define AQ_INTR_AUTOMASK_REG 0x2090
+
+/* AQ_INTR_IRQ_MAP_TXRX_REG 0x2100-0x2140 */
+#define AQ_INTR_IRQ_MAP_TXRX_REG(i) (0x2100 + ((i) / 2) * 4)
+#define AQ_INTR_IRQ_MAP_TX_REG(i) AQ_INTR_IRQ_MAP_TXRX_REG(i)
+#define AQ_INTR_IRQ_MAP_TX_IRQMAP(i) (0x1FU << (((i) & 1) ? 16 : 24))
+#define AQ_INTR_IRQ_MAP_TX_EN(i) (1U << (((i) & 1) ? 23 : 31))
+#define AQ_INTR_IRQ_MAP_RX_REG(i) AQ_INTR_IRQ_MAP_TXRX_REG(i)
+#define AQ_INTR_IRQ_MAP_RX_IRQMAP(i) (0x1FU << (((i) & 1) ? 0 : 8))
+#define AQ_INTR_IRQ_MAP_RX_EN(i) (1U << (((i) & 1) ? 7 : 15))
+
+/* AQ_GEN_INTR_MAP_REG[AQ_RINGS_NUM] 0x2180-0x2200 */
+#define AQ_GEN_INTR_MAP_REG(i) (0x2180 + (i) * 4)
+#define AQ_B0_ERR_INT 8U
+
+#define AQ_INTR_CTRL_REG 0x2300
+#define AQ_INTR_CTRL_IRQMODE ((1 << 0) | (1 << 1))
+#define AQ_INTR_CTRL_IRQMODE_LEGACY 0
+#define AQ_INTR_CTRL_IRQMODE_MSI 1
+#define AQ_INTR_CTRL_IRQMODE_MSIX 2
+#define AQ_INTR_CTRL_MULTIVEC (1 << 2)
+#define AQ_INTR_CTRL_RESET_DIS (1 << 29)
+#define AQ_INTR_CTRL_RESET_IRQ (1 << 31)
+#define AQ_MBOXIF_POWER_GATING_CONTROL_REG 0x32a8
+
+#define FW_MPI_MBOX_ADDR_REG 0x0360
+#define FW1X_MPI_INIT1_REG 0x0364
+#define FW1X_MPI_INIT2_REG 0x0370
+#define FW1X_MPI_EFUSEADDR_REG 0x0374
+
+#define FW2X_MPI_EFUSEADDR_REG 0x0364
+#define FW2X_MPI_CONTROL_REG 0x0368 /* 64bit */
+#define FW2X_MPI_STATE_REG 0x0370 /* 64bit */
+#define FW_BOOT_EXIT_CODE_REG 0x0388
+
+#define FW_BOOT_EXIT_CODE_REG 0x0388
+#define RBL_STATUS_DEAD 0x0000dead
+#define RBL_STATUS_SUCCESS 0x0000abba
+#define RBL_STATUS_FAILURE 0x00000bad
+#define RBL_STATUS_HOST_BOOT 0x0000f1a7
+#define FW_MPI_DAISY_CHAIN_STATUS_REG 0x0704
+#define AQ_PCI_REG_CONTROL_6_REG 0x1014
+
+#define FW_MPI_RESETCTRL_REG 0x4000
+#define FW_MPI_RESETCTRL_RESET_DIS (1 << 29)
+
+#define RX_SYSCONTROL_REG 0x5000
+#define RX_SYSCONTROL_RESET_DIS (1 << 29)
+
+#define RX_TCP_RSS_HASH_REG 0x5040
+
+#define RPF_L2BC_REG 0x5100
+#define RPF_L2BC_EN (1 << 0)
+#define RPF_L2BC_PROMISC (1 << 3)
+#define RPF_L2BC_ACTION 0x7000
+#define RPF_L2BC_THRESHOLD 0xFFFF0000
+
+#define AQ_HW_MAC_OWN 0
+
+/* RPF_L2UC_*_REG[34] (actual [38]?) */
+#define RPF_L2UC_LSW_REG(i) (0x5110 + (i) * 8)
+#define RPF_L2UC_MSW_REG(i) (0x5114 + (i) * 8)
+#define RPF_L2UC_MSW_MACADDR_HI 0xFFFF
+#define RPF_L2UC_MSW_ACTION 0x70000
+#define RPF_L2UC_MSW_EN (1 << 31)
+#define AQ_HW_MAC_NUM 34
+
+/* RPF_MCAST_FILTER_REG[8] 0x5250-0x5270 */
+#define RPF_MCAST_FILTER_REG(i) (0x5250 + (i) * 4)
+#define RPF_MCAST_FILTER_EN (1 << 31)
+#define RPF_MCAST_FILTER_MASK_REG 0x5270
+#define RPF_MCAST_FILTER_MASK_ALLMULTI (1 << 14)
+
+#define RPF_VLAN_MODE_REG 0x5280
+#define RPF_VLAN_MODE_PROMISC (1 << 1)
+#define RPF_VLAN_MODE_ACCEPT_UNTAGGED (1 << 2)
+#define RPF_VLAN_MODE_UNTAGGED_ACTION 0x38
+
+#define RPF_VLAN_TPID_REG 0x5284
+#define RPF_VLAN_TPID_OUTER 0xFFFF0000
+#define RPF_VLAN_TPID_INNER 0xFFFF
+
+/* RPF_ETHERTYPE_FILTER_REG[AQ_RINGS_NUM] 0x5300-0x5380 */
+#define RPF_ETHERTYPE_FILTER_REG(i) (0x5300 + (i) * 4)
+#define RPF_ETHERTYPE_FILTER_EN (1 << 31)
+
+/* RPF_L3_FILTER_REG[8] 0x5380-0x53a0 */
+#define RPF_L3_FILTER_REG(i) (0x5380 + (i) * 4)
+#define RPF_L3_FILTER_L4_EN (1 << 31)
+
+#define RX_FLR_RSS_CONTROL1_REG 0x54c0
+#define RX_FLR_RSS_CONTROL1_EN (1 << 31)
+
+#define RPF_RPB_RX_TC_UPT_REG 0x54c4
+#define RPF_RPB_RX_TC_UPT_MASK(i) (0x00000007 << ((i) * 4))
+
+#define RPB_RPF_RX_REG 0x5700
+#define RPB_RPF_RX_TC_MODE (1 << 8)
+#define RPB_RPF_RX_FC_MODE 0x30
+#define RPB_RPF_RX_BUF_EN (1 << 0)
+
+/* RPB_RXB_BUFSIZE_REG[AQ_TRAFFICCLASS_NUM] 0x5710-0x5790 */
+#define RPB_RXB_BUFSIZE_REG(i) (0x5710 + (i) * 0x10)
+#define RPB_RXB_BUFSIZE 0x1FF
+#define RPB_RXB_XOFF_REG(i) (0x5714 + (i) * 0x10)
+#define RPB_RXB_XOFF_EN (1 << 31)
+#define RPB_RXB_XOFF_THRESH_HI 0x3FFF0000
+#define RPB_RXB_XOFF_THRESH_LO 0x3FFF
+
+#define RX_DMA_INT_DESC_WRWB_EN_REG 0x5a30
+#define RX_DMA_INT_DESC_WRWB_EN (1 << 2)
+#define RX_DMA_INT_DESC_MODERATE_EN (1 << 3)
+
+#define RX_INTR_MODERATION_CTL_REG(i) (0x5a40 + (i) * 4)
+#define RX_INTR_MODERATION_CTL_EN (1 << 1)
+
+#define RX_DMA_DESC_BASE_ADDRLSW_REG(i) (0x5b00 + (i) * 0x20)
+#define RX_DMA_DESC_BASE_ADDRMSW_REG(i) (0x5b04 + (i) * 0x20)
+#define RX_DMA_DESC_REG(i) (0x5b08 + (i) * 0x20)
+#define RX_DMA_DESC_LEN (0x3FF << 3)
+#define RX_DMA_DESC_RESET (1 << 25)
+#define RX_DMA_DESC_HEADER_SPLIT (1 << 28)
+#define RX_DMA_DESC_VLAN_STRIP (1 << 29)
+#define RX_DMA_DESC_EN (1 << 31)
+#define RX_DMA_DESC_HEAD_PTR_REG(i) (0x5b0c + (i) * 0x20)
+#define RX_DMA_DESC_HEAD_PTR 0xFFF
+#define RX_DMA_DESC_TAIL_PTR_REG(i) (0x5b10 + (i) * 0x20)
+#define RX_DMA_DESC_BUFSIZE_REG(i) (0x5b18 + (i) * 0x20)
+#define RX_DMA_DESC_BUFSIZE_DATA 0x000F
+#define RX_DMA_DESC_BUFSIZE_HDR 0x0FF0
+
+#define RX_DMA_DCAD_REG(i) (0x6100 + (i) * 4)
+#define RX_DMA_DCAD_CPUID 0xFF
+#define RX_DMA_DCAD_PAYLOAD_EN (1 << 29)
+#define RX_DMA_DCAD_HEADER_EN (1 << 30)
+#define RX_DMA_DCAD_DESC_EN (1 << 31)
+
+#define RX_DMA_DCA_REG 0x6180
+#define RX_DMA_DCA_EN (1 << 31)
+#define RX_DMA_DCA_MODE 0xF
+
+#define TX_SYSCONTROL_REG 0x7000
+#define TX_SYSCONTROL_RESET_DIS (1 << 29)
+
+#define TX_TPO2_REG 0x7040
+#define TX_TPO2_EN (1 << 16)
+
+#define TPS_DESC_VM_ARB_MODE_REG 0x7300
+#define TPS_DESC_VM_ARB_MODE (1 << 0)
+#define TPS_DESC_RATE_REG 0x7310
+#define TPS_DESC_RATE_TA_RST (1 << 31)
+#define TPS_DESC_RATE_LIM 0x7FF
+#define TPS_DESC_TC_ARB_MODE_REG 0x7200
+#define TPS_DESC_TC_ARB_MODE 0x3
+#define TPS_DATA_TC_ARB_MODE_REG 0x7100
+#define TPS_DATA_TC_ARB_MODE (1 << 0)
+
+/* TPS_DATA_TCT_REG[AQ_TRAFFICCLASS_NUM] 0x7110-0x7130 */
+#define TPS_DATA_TCT_REG(i) (0x7110 + (i) * 4)
+#define TPS_DATA_TCT_CREDIT_MAX 0xFFF0000
+#define TPS_DATA_TCT_WEIGHT 0x1FF
+/* TPS_DATA_TCT_REG[AQ_TRAFFICCLASS_NUM] 0x7210-0x7230 */
+#define TPS_DESC_TCT_REG(i) (0x7210 + (i) * 4)
+#define TPS_DESC_TCT_CREDIT_MAX 0xFFF0000
+#define TPS_DESC_TCT_WEIGHT 0x1FF
+
+#define AQ_HW_TXBUF_MAX 160
+#define AQ_HW_RXBUF_MAX 320
+
+#define THM_LSO_TCP_FLAG1_REG 0x7820
+#define THM_LSO_TCP_FLAG1_FIRST 0xFFF
+#define THM_LSO_TCP_FLAG1_MID 0xFFF0000
+#define THM_LSO_TCP_FLAG2_REG 0x7824
+#define THM_LSO_TCP_FLAG2_LAST 0xFFF
+
+#define TPB_TX_BUF_REG 0x7900
+#define TPB_TX_BUF_EN (1 << 0)
+#define TPB_TX_BUF_SCP_INS_EN (1 << 2)
+#define TPB_TX_BUF_TC_MODE_EN (1 << 8)
+
+/* TPB_TXB_BUFSIZE_REG[AQ_TRAFFICCLASS_NUM] 0x7910-7990 */
+#define TPB_TXB_BUFSIZE_REG(i) (0x7910 + (i) * 0x10)
+#define TPB_TXB_BUFSIZE (0xFF)
+#define TPB_TXB_THRESH_REG(i) (0x7914 + (i) * 0x10)
+#define TPB_TXB_THRESH_HI 0x1FFF0000
+#define TPB_TXB_THRESH_LO 0x1FFF
+
+#define AQ_HW_TX_DMA_TOTAL_REQ_LIMIT_REG 0x7b20
+
+#define TX_DMA_INT_DESC_WRWB_EN_REG 0x7b40
+#define TX_DMA_INT_DESC_WRWB_EN (1 << 1)
+#define TX_DMA_INT_DESC_MODERATE_EN (1 << 4)
+
+#define TX_DMA_DESC_BASE_ADDRLSW_REG(i) (0x7c00 + (i) * 0x40)
+#define TX_DMA_DESC_BASE_ADDRMSW_REG(i) (0x7c04 + (i) * 0x40)
+#define TX_DMA_DESC_REG(i) (0x7c08 + (i) * 0x40)
+#define TX_DMA_DESC_LEN 0x00000FF8
+#define TX_DMA_DESC_EN 0x80000000
+#define TX_DMA_DESC_HEAD_PTR_REG(i) (0x7c0c + (i) * 0x40)
+#define TX_DMA_DESC_HEAD_PTR 0x00000FFF
+#define TX_DMA_DESC_TAIL_PTR_REG(i) (0x7c10 + (i) * 0x40)
+#define TX_DMA_DESC_WRWB_THRESH_REG(i) (0x7c18 + (i) * 0x40)
+#define TX_DMA_DESC_WRWB_THRESH 0x00003F00
+
+#define TDM_DCAD_REG(i) (0x8400 + (i) * 4)
+#define TDM_DCAD_CPUID 0x7F
+#define TDM_DCAD_CPUID_EN 0x80000000
+
+#define TDM_DCA_REG 0x8480
+#define TDM_DCA_EN (1 << 31)
+#define TDM_DCA_MODE 0xF
+
+#define TX_INTR_MODERATION_CTL_REG(i) (0x8980 + (i) * 4)
+#define TX_INTR_MODERATION_CTL_EN (1 << 1)
+
+#define __LOWEST_SET_BIT(__mask) (((((uint32_t)__mask) - 1) & ((uint32_t)__mask)) ^ ((uint32_t)__mask))
+#define __SHIFTIN(__x, __mask) ((__x) * __LOWEST_SET_BIT(__mask))
+
+#if 0
+#define AQ_READ_REG(sc, reg) \
+ bus_space_read_4((sc)->sc_iot, (sc)->sc_ioh, (reg))
+
+#endif
+#define AQ_WRITE_REG(sc, reg, val) \
+ bus_space_write_4((sc)->sc_iot, (sc)->sc_ioh, (reg), (val))
+
+#define AQ_WRITE_REG_BIT(sc, reg, mask, val) \
+ do { \
+ uint32_t _v; \
+ _v = AQ_READ_REG((sc), (reg)); \
+ _v &= ~(mask); \
+ if ((val) != 0) \
+ _v |= __SHIFTIN((val), (mask)); \
+ AQ_WRITE_REG((sc), (reg), _v); \
+ } while (/* CONSTCOND */ 0)
+
+#define AQ_READ64_REG(sc, reg) \
+ ((uint64_t)AQ_READ_REG(sc, reg) | \
+ (((uint64_t)AQ_READ_REG(sc, (reg) + 4)) << 32))
+
+#define AQ_WRITE64_REG(sc, reg, val) \
+ do { \
+ AQ_WRITE_REG(sc, reg, (uint32_t)val); \
+ AQ_WRITE_REG(sc, reg + 4, (uint32_t)(val >> 32)); \
+ } while (/* CONSTCOND */0)
+
+#define WAIT_FOR(expr, us, n, errp) \
+ do { \
+ unsigned int _n; \
+ for (_n = n; (!(expr)) && _n != 0; --_n) { \
+ delay((us)); \
+ } \
+ if ((errp != NULL)) { \
+ if (_n == 0) \
+ *(errp) = ETIMEDOUT; \
+ else \
+ *(errp) = 0; \
+ } \
+ } while (/* CONSTCOND */ 0)
+
+#define FW_VERSION_MAJOR(sc) (((sc)->sc_fw_version >> 24) & 0xff)
+#define FW_VERSION_MINOR(sc) (((sc)->sc_fw_version >> 16) & 0xff)
+#define FW_VERSION_BUILD(sc) ((sc)->sc_fw_version & 0xffff)
+
+#define FEATURES_MIPS 0x00000001
+#define FEATURES_TPO2 0x00000002
+#define FEATURES_RPF2 0x00000004
+#define FEATURES_MPI_AQ 0x00000008
+#define FEATURES_REV_A0 0x10000000
+#define FEATURES_REV_A (FEATURES_REV_A0)
+#define FEATURES_REV_B0 0x20000000
+#define FEATURES_REV_B1 0x40000000
+#define FEATURES_REV_B (FEATURES_REV_B0|FEATURES_REV_B1)
+
+/* lock for FW2X_MPI_{CONTROL,STATE]_REG read-modify-write */
+#define AQ_MPI_LOCK(sc) mtx_enter(&(sc)->sc_mpi_mutex);
+#define AQ_MPI_UNLOCK(sc) mtx_leave(&(sc)->sc_mpi_mutex);
+
+#define FW2X_CTRL_10BASET_HD (1 << 0)
+#define FW2X_CTRL_10BASET_FD (1 << 1)
+#define FW2X_CTRL_100BASETX_HD (1 << 2)
+#define FW2X_CTRL_100BASET4_HD (1 << 3)
+#define FW2X_CTRL_100BASET2_HD (1 << 4)
+#define FW2X_CTRL_100BASETX_FD (1 << 5)
+#define FW2X_CTRL_100BASET2_FD (1 << 6)
+#define FW2X_CTRL_1000BASET_HD (1 << 7)
+#define FW2X_CTRL_1000BASET_FD (1 << 8)
+#define FW2X_CTRL_2P5GBASET_FD (1 << 9)
+#define FW2X_CTRL_5GBASET_FD (1 << 10)
+#define FW2X_CTRL_10GBASET_FD (1 << 11)
+#define FW2X_CTRL_RESERVED1 (1ULL << 32)
+#define FW2X_CTRL_10BASET_EEE (1ULL << 33)
+#define FW2X_CTRL_RESERVED2 (1ULL << 34)
+#define FW2X_CTRL_PAUSE (1ULL << 35)
+#define FW2X_CTRL_ASYMMETRIC_PAUSE (1ULL << 36)
+#define FW2X_CTRL_100BASETX_EEE (1ULL << 37)
+#define FW2X_CTRL_RESERVED3 (1ULL << 38)
+#define FW2X_CTRL_RESERVED4 (1ULL << 39)
+#define FW2X_CTRL_1000BASET_FD_EEE (1ULL << 40)
+#define FW2X_CTRL_2P5GBASET_FD_EEE (1ULL << 41)
+#define FW2X_CTRL_5GBASET_FD_EEE (1ULL << 42)
+#define FW2X_CTRL_10GBASET_FD_EEE (1ULL << 43)
+#define FW2X_CTRL_RESERVED5 (1ULL << 44)
+#define FW2X_CTRL_RESERVED6 (1ULL << 45)
+#define FW2X_CTRL_RESERVED7 (1ULL << 46)
+#define FW2X_CTRL_RESERVED8 (1ULL << 47)
+#define FW2X_CTRL_RESERVED9 (1ULL << 48)
+#define FW2X_CTRL_CABLE_DIAG (1ULL << 49)
+#define FW2X_CTRL_TEMPERATURE (1ULL << 50)
+#define FW2X_CTRL_DOWNSHIFT (1ULL << 51)
+#define FW2X_CTRL_PTP_AVB_EN (1ULL << 52)
+#define FW2X_CTRL_MEDIA_DETECT (1ULL << 53)
+#define FW2X_CTRL_LINK_DROP (1ULL << 54)
+#define FW2X_CTRL_SLEEP_PROXY (1ULL << 55)
+#define FW2X_CTRL_WOL (1ULL << 56)
+#define FW2X_CTRL_MAC_STOP (1ULL << 57)
+#define FW2X_CTRL_EXT_LOOPBACK (1ULL << 58)
+#define FW2X_CTRL_INT_LOOPBACK (1ULL << 59)
+#define FW2X_CTRL_EFUSE_AGENT (1ULL << 60)
+#define FW2X_CTRL_WOL_TIMER (1ULL << 61)
+#define FW2X_CTRL_STATISTICS (1ULL << 62)
+#define FW2X_CTRL_TRANSACTION_ID (1ULL << 63)
+
+#define FW2X_CTRL_RATE_100M FW2X_CTRL_100BASETX_FD
+#define FW2X_CTRL_RATE_1G FW2X_CTRL_1000BASET_FD
+#define FW2X_CTRL_RATE_2G5 FW2X_CTRL_2P5GBASET_FD
+#define FW2X_CTRL_RATE_5G FW2X_CTRL_5GBASET_FD
+#define FW2X_CTRL_RATE_10G FW2X_CTRL_10GBASET_FD
+#define FW2X_CTRL_RATE_MASK \
+ (FW2X_CTRL_RATE_100M | \
+ FW2X_CTRL_RATE_1G | \
+ FW2X_CTRL_RATE_2G5 | \
+ FW2X_CTRL_RATE_5G | \
+ FW2X_CTRL_RATE_10G)
+#define FW2X_CTRL_EEE_MASK \
+ (FW2X_CTRL_10BASET_EEE | \
+ FW2X_CTRL_100BASETX_EEE | \
+ FW2X_CTRL_1000BASET_FD_EEE | \
+ FW2X_CTRL_2P5GBASET_FD_EEE | \
+ FW2X_CTRL_5GBASET_FD_EEE | \
+ FW2X_CTRL_10GBASET_FD_EEE)
+
+enum aq_fw_bootloader_mode {
+ FW_BOOT_MODE_UNKNOWN = 0,
+ FW_BOOT_MODE_FLB,
+ FW_BOOT_MODE_RBL_FLASH,
+ FW_BOOT_MODE_RBL_HOST_BOOTLOAD
+};
+
+enum aq_media_type {
+ AQ_MEDIA_TYPE_UNKNOWN = 0,
+ AQ_MEDIA_TYPE_FIBRE,
+ AQ_MEDIA_TYPE_TP
+};
+
+enum aq_link_speed {
+ AQ_LINK_NONE = 0,
+ AQ_LINK_100M = (1 << 0),
+ AQ_LINK_1G = (1 << 1),
+ AQ_LINK_2G5 = (1 << 2),
+ AQ_LINK_5G = (1 << 3),
+ AQ_LINK_10G = (1 << 4)
+};
+
+#define AQ_LINK_ALL (AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | \
+ AQ_LINK_5G | AQ_LINK_10G )
+#define AQ_LINK_AUTO AQ_LINK_ALL
+
+enum aq_link_eee {
+ AQ_EEE_DISABLE = 0,
+ AQ_EEE_ENABLE = 1
+};
+
+enum aq_hw_fw_mpi_state {
+ MPI_DEINIT = 0,
+ MPI_RESET = 1,
+ MPI_INIT = 2,
+ MPI_POWER = 4
+};
+
+enum aq_link_fc {
+ AQ_FC_NONE = 0,
+ AQ_FC_RX = (1 << 0),
+ AQ_FC_TX = (1 << 1),
+ AQ_FC_ALL = (AQ_FC_RX | AQ_FC_TX)
+};
+
+struct aq_dmamem {
+ bus_dmamap_t aqm_map;
+ bus_dma_segment_t aqm_seg;
+ int aqm_nsegs;
+ size_t aqm_size;
+ caddr_t aqm_kva;
+};
+
+#define AQ_DMA_MAP(_aqm) ((_aqm)->aqm_map)
+#define AQ_DMA_DVA(_aqm) ((_aqm)->aqm_map->dm_segs[0].ds_addr)
+#define AQ_DMA_KVA(_aqm) ((void *)(_aqm)->aqm_kva)
+#define AQ_DMA_LEN(_aqm) ((_aqm)->aqm_size)
+
+
+struct aq_mailbox_header {
+ uint32_t version;
+ uint32_t transaction_id;
+ int32_t error;
+} __packed __aligned(4);
+
+struct aq_hw_stats_s {
+ uint32_t uprc;
+ uint32_t mprc;
+ uint32_t bprc;
+ uint32_t erpt;
+ uint32_t uptc;
+ uint32_t mptc;
+ uint32_t bptc;
+ uint32_t erpr;
+ uint32_t mbtc;
+ uint32_t bbtc;
+ uint32_t mbrc;
+ uint32_t bbrc;
+ uint32_t ubrc;
+ uint32_t ubtc;
+ uint32_t ptc;
+ uint32_t prc;
+ uint32_t dpc; /* not exists in fw2x_msm_statistics */
+ uint32_t cprc; /* not exists in fw2x_msm_statistics */
+} __packed __aligned(4);
+
+struct aq_fw2x_capabilities {
+ uint32_t caps_lo;
+ uint32_t caps_hi;
+} __packed __aligned(4);
+
+struct aq_fw2x_msm_statistics {
+ uint32_t uprc;
+ uint32_t mprc;
+ uint32_t bprc;
+ uint32_t erpt;
+ uint32_t uptc;
+ uint32_t mptc;
+ uint32_t bptc;
+ uint32_t erpr;
+ uint32_t mbtc;
+ uint32_t bbtc;
+ uint32_t mbrc;
+ uint32_t bbrc;
+ uint32_t ubrc;
+ uint32_t ubtc;
+ uint32_t ptc;
+ uint32_t prc;
+} __packed __aligned(4);
+
+struct aq_fw2x_phy_cable_diag_data {
+ uint32_t lane_data[4];
+} __packed __aligned(4);
+
+struct aq_fw2x_mailbox { /* struct fwHostInterface */
+ struct aq_mailbox_header header;
+ struct aq_fw2x_msm_statistics msm; /* msmStatistics_t msm; */
+
+ uint32_t phy_info1;
+#define PHYINFO1_FAULT_CODE __BITS(31,16)
+#define PHYINFO1_PHY_H_BIT __BITS(0,15)
+ uint32_t phy_info2;
+#define PHYINFO2_TEMPERATURE __BITS(15,0)
+#define PHYINFO2_CABLE_LEN __BITS(23,16)
+
+ struct aq_fw2x_phy_cable_diag_data diag_data;
+ uint32_t reserved[8];
+
+ struct aq_fw2x_capabilities caps;
+
+ /* ... */
+} __packed __aligned(4);
+
+struct aq_rx_desc_read {
+ uint64_t buf_addr;
+ uint64_t hdr_addr;
+} __packed;
+
+struct aq_rx_desc_wb {
+ uint32_t type;
+#define AQ_RXDESC_TYPE_RSSTYPE 0x000f
+#define AQ_RXDESC_TYPE_ETHER 0x0030
+#define AQ_RXDESC_TYPE_PROTO 0x01c0
+#define AQ_RXDESC_TYPE_VLAN (1 << 9)
+#define AQ_RXDESC_TYPE_VLAN2 (1 << 10)
+#define AQ_RXDESC_TYPE_DMA_ERR (1 << 12)
+#define AQ_RXDESC_TYPE_V4_SUM (1 << 19)
+#define AQ_RXDESC_TYPE_TCP_SUM (1 << 20)
+ uint32_t rss_hash;
+ uint16_t status;
+#define AQ_RXDESC_STATUS_DD (1 << 0)
+#define AQ_RXDESC_STATUS_EOP (1 << 1)
+#define AQ_RXDESC_STATUS_MACERR (1 << 2)
+#define AQ_RXDESC_STATUS_V4_SUM (1 << 3)
+#define AQ_RXDESC_STATUS_L4_SUM_ERR (1 << 4)
+#define AQ_RXDESC_STATUS_L4_SUM_OK (1 << 5)
+ uint16_t pkt_len;
+ uint16_t next_desc_ptr;
+ uint16_t vlan;
+} __packed;
+
+struct aq_tx_desc {
+ uint64_t buf_addr;
+ uint32_t ctl1;
+#define AQ_TXDESC_CTL1_TYPE_TXD 0x00000001
+#define AQ_TXDESC_CTL1_TYPE_TXC 0x00000002
+#define AQ_TXDESC_CTL1_BLEN_SHIFT 4
+#define AQ_TXDESC_CTL1_DD (1 << 20)
+#define AQ_TXDESC_CTL1_CMD_EOP (1 << 21)
+#define AQ_TXDESC_CTL1_CMD_VLAN (1 << 22)
+#define AQ_TXDESC_CTL1_CMD_FCS (1 << 23)
+#define AQ_TXDESC_CTL1_CMD_IP4CSUM (1 << 24)
+#define AQ_TXDESC_CTL1_CMD_L4CSUM (1 << 25)
+#define AQ_TXDESC_CTL1_CMD_WB (1 << 27)
+
+#define AQ_TXDESC_CTL1_VID_SHIFT 4
+ uint32_t ctl2;
+#define AQ_TXDESC_CTL2_LEN_SHIFT 14
+#define AQ_TXDESC_CTL2_CTX_EN (1 << 13)
+} __packed;
+
+struct aq_slot {
+ bus_dmamap_t as_map;
+ struct mbuf *as_m;
+};
+
+struct aq_rxring {
+ struct ifiqueue *rx_ifiq;
+ struct aq_dmamem rx_mem;
+ struct aq_slot *rx_slots;
+ int rx_q;
+ int rx_irq;
+
+ struct timeout rx_refill;
+ struct if_rxring rx_rxr;
+ uint32_t rx_prod;
+ uint32_t rx_cons;
+};
+
+struct aq_txring {
+ struct ifqueue *tx_ifq;
+ struct aq_dmamem tx_mem;
+ struct aq_slot *tx_slots;
+ int tx_q;
+ int tx_irq;
+ uint32_t tx_prod;
+ uint32_t tx_cons;
+};
+
+struct aq_queues {
+ char q_name[16];
+ void *q_ihc;
+ struct aq_softc *q_sc;
+ int q_index;
+ struct aq_rxring q_rx;
+ struct aq_txring q_tx;
+};
+
+
+struct aq_softc;
+struct aq_firmware_ops {
+ int (*reset)(struct aq_softc *);
+ int (*set_mode)(struct aq_softc *, enum aq_hw_fw_mpi_state,
+ enum aq_link_speed, enum aq_link_fc, enum aq_link_eee);
+ int (*get_mode)(struct aq_softc *, enum aq_hw_fw_mpi_state *,
+ enum aq_link_speed *, enum aq_link_fc *, enum aq_link_eee *);
+ int (*get_stats)(struct aq_softc *, struct aq_hw_stats_s *);
+};
+
+struct aq_softc {
+ struct device sc_dev;
+ uint16_t sc_product;
+ uint16_t sc_revision;
+ bus_dma_tag_t sc_dmat;
+ pci_chipset_tag_t sc_pc;
+ pcitag_t sc_pcitag;
+ int sc_nqueues;
+ struct aq_queues sc_queues[AQ_MAXQ];
+ struct intrmap *sc_intrmap;
+ void *sc_ih;
+ bus_space_handle_t sc_ioh;
+ bus_space_tag_t sc_iot;
+
+ uint32_t sc_mbox_addr;
+ int sc_rbl_enabled;
+ int sc_fast_start_enabled;
+ int sc_flash_present;
+ uint32_t sc_fw_version;
+ const struct aq_firmware_ops *sc_fw_ops;
+ uint64_t sc_fw_caps;
+ enum aq_media_type sc_media_type;
+ enum aq_link_speed sc_available_rates;
+ uint32_t sc_features;
+ int sc_linkstat_irq;
+ struct arpcom sc_arpcom;
+ struct ifmedia sc_media;
+
+ struct ether_addr sc_enaddr;
+ struct mutex sc_mpi_mutex;
+};
+
+const struct pci_matchid aq_devices[] = {
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100S },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107S },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108S },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109S },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111S },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112S },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D100 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D107 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D108 },
+ { PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D109 },
+};
+
+const struct aq_product {
+ pci_vendor_id_t aq_vendor;
+ pci_product_id_t aq_product;
+ enum aq_media_type aq_media_type;
+ enum aq_link_speed aq_available_rates;
+} aq_products[] = {
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100,
+ AQ_MEDIA_TYPE_FIBRE, AQ_LINK_ALL
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_ALL
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC100S,
+ AQ_MEDIA_TYPE_FIBRE, AQ_LINK_ALL
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC107S,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_ALL
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC108S,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC109S,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC111S,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_AQC112S,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D100,
+ AQ_MEDIA_TYPE_FIBRE, AQ_LINK_ALL
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D107,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_ALL
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D108,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5 | AQ_LINK_5G
+},
+{ PCI_VENDOR_AQUANTIA, PCI_PRODUCT_AQUANTIA_D109,
+ AQ_MEDIA_TYPE_TP, AQ_LINK_100M | AQ_LINK_1G | AQ_LINK_2G5
+}
+};
+
+int aq_match(struct device *, void *, void *);
+void aq_attach(struct device *, struct device *, void *);
+int aq_detach(struct device *, int);
+int aq_activate(struct device *, int);
+int aq_intr(void *);
+void aq_global_software_reset(struct aq_softc *);
+int aq_fw_reset(struct aq_softc *);
+int aq_mac_soft_reset(struct aq_softc *, enum aq_fw_bootloader_mode *);
+int aq_mac_soft_reset_rbl(struct aq_softc *, enum aq_fw_bootloader_mode *);
+int aq_mac_soft_reset_flb(struct aq_softc *);
+int aq_fw_read_version(struct aq_softc *);
+int aq_fw_version_init(struct aq_softc *);
+int aq_hw_init_ucp(struct aq_softc *);
+int aq_fw_downld_dwords(struct aq_softc *, uint32_t, uint32_t *, uint32_t);
+int aq_get_mac_addr(struct aq_softc *);
+int aq_hw_reset(struct aq_softc *);
+int aq_hw_init(struct aq_softc *, int);
+void aq_hw_qos_set(struct aq_softc *);
+void aq_l3_filter_set(struct aq_softc *);
+void aq_hw_init_tx_path(struct aq_softc *);
+void aq_hw_init_rx_path(struct aq_softc *);
+int aq_set_mac_addr(struct aq_softc *, int, uint8_t *);
+int aq_set_linkmode(struct aq_softc *, enum aq_link_speed,
+ enum aq_link_fc, enum aq_link_eee);
+void aq_watchdog(struct ifnet *);
+void aq_enable_intr(struct aq_softc *, int, int);
+int aq_ioctl(struct ifnet *, u_long, caddr_t);
+int aq_up(struct aq_softc *);
+void aq_down(struct aq_softc *);
+void aq_iff(struct aq_softc *);
+void aq_start(struct ifqueue *);
+void aq_ifmedia_status(struct ifnet *, struct ifmediareq *);
+int aq_ifmedia_change(struct ifnet *);
+void aq_update_link_status(struct aq_softc *);
+
+void aq_refill(void *);
+int aq_rx_fill(struct aq_softc *, struct aq_rxring *);
+static inline unsigned int aq_rx_fill_slots(struct aq_softc *,
+ struct aq_rxring *, uint);
+
+int aq_dmamem_alloc(struct aq_softc *, struct aq_dmamem *,
+ bus_size_t, u_int);
+void aq_dmamem_zero(struct aq_dmamem *);
+void aq_dmamem_free(struct aq_softc *, struct aq_dmamem *);
+
+int aq_fw1x_reset(struct aq_softc *);
+int aq_fw1x_get_mode(struct aq_softc *, enum aq_hw_fw_mpi_state *,
+ enum aq_link_speed *, enum aq_link_fc *, enum aq_link_eee *);
+int aq_fw1x_set_mode(struct aq_softc *, enum aq_hw_fw_mpi_state,
+ enum aq_link_speed, enum aq_link_fc, enum aq_link_eee);
+int aq_fw1x_get_stats(struct aq_softc *, struct aq_hw_stats_s *);
+
+int aq_fw2x_reset(struct aq_softc *);
+int aq_fw2x_get_mode(struct aq_softc *, enum aq_hw_fw_mpi_state *,
+ enum aq_link_speed *, enum aq_link_fc *, enum aq_link_eee *);
+int aq_fw2x_set_mode(struct aq_softc *, enum aq_hw_fw_mpi_state,
+ enum aq_link_speed, enum aq_link_fc, enum aq_link_eee);
+int aq_fw2x_get_stats(struct aq_softc *, struct aq_hw_stats_s *);
+
+const struct aq_firmware_ops aq_fw1x_ops = {
+ .reset = aq_fw1x_reset,
+ .set_mode = aq_fw1x_set_mode,
+ .get_mode = aq_fw1x_get_mode,
+ .get_stats = aq_fw1x_get_stats,
+};
+
+const struct aq_firmware_ops aq_fw2x_ops = {
+ .reset = aq_fw2x_reset,
+ .set_mode = aq_fw2x_set_mode,
+ .get_mode = aq_fw2x_get_mode,
+ .get_stats = aq_fw2x_get_stats,
+};
+
+struct cfattach aq_ca = {
+ sizeof(struct aq_softc), aq_match, aq_attach, NULL,
+ aq_activate
+};
+
+struct cfdriver aq_cd = {
+ NULL, "aq", DV_IFNET
+};
+
+uint32_t
+AQ_READ_REG(struct aq_softc *sc, uint32_t reg)
+{
+ uint32_t res;
+
+ res = bus_space_read_4(sc->sc_iot, sc->sc_ioh, reg);
+
+ return res;
+}
+
+
+int
+aq_match(struct device *dev, void *match, void *aux)
+{
+ return pci_matchbyid((struct pci_attach_args *)aux, aq_devices,
+ sizeof(aq_devices) / sizeof(aq_devices[0]));
+}
+
+const struct aq_product *
+aq_lookup(const struct pci_attach_args *pa)
+{
+ unsigned int i;
+
+ for (i = 0; i < sizeof(aq_products) / sizeof(aq_products[0]); i++) {
+ if (PCI_VENDOR(pa->pa_id) == aq_products[i].aq_vendor &&
+ PCI_PRODUCT(pa->pa_id) == aq_products[i].aq_product) {
+ return &aq_products[i];
+ }
+ }
+
+ return NULL;
+}
+
+void
+aq_attach(struct device *parent, struct device *self, void *aux)
+{
+ struct aq_softc *sc = (struct aq_softc *)self;
+ struct pci_attach_args *pa = aux;
+ const struct aq_product *aqp;
+ pcireg_t command, bar, memtype;
+ pci_chipset_tag_t pc;
+ pci_intr_handle_t ih;
+ int (*isr)(void *);
+ const char *intrstr;
+ pcitag_t tag;
+ struct ifnet *ifp = &sc->sc_arpcom.ac_if;
+ int irqmode;
+ int i;
+
+ mtx_init(&sc->sc_mpi_mutex, IPL_NET);
+
+ sc->sc_dmat = pa->pa_dmat;
+ sc->sc_pc = pc = pa->pa_pc;
+ sc->sc_pcitag = tag = pa->pa_tag;
+
+ command = pci_conf_read(pa->pa_pc, pa->pa_tag, PCI_COMMAND_STATUS_REG);
+ command |= PCI_COMMAND_MASTER_ENABLE;
+ pci_conf_write(pa->pa_pc, pa->pa_tag, PCI_COMMAND_STATUS_REG, command);
+
+ sc->sc_product = PCI_PRODUCT(pa->pa_id);
+ sc->sc_revision = PCI_REVISION(pa->pa_class);
+
+ aqp = aq_lookup(pa);
+
+ bar = pci_conf_read(pc, tag, AQ_BAR0);
+ if ((PCI_MAPREG_MEM_ADDR(bar) == 0) ||
+ (PCI_MAPREG_TYPE(bar) != PCI_MAPREG_TYPE_MEM)) {
+ printf("%s: wrong BAR type\n", DEVNAME(sc));
+ return;
+ }
+
+ memtype = pci_mapreg_type(pc, tag, AQ_BAR0);
+ if (pci_mapreg_map(pa, AQ_BAR0, memtype, 0, &sc->sc_iot, &sc->sc_ioh,
+ NULL, NULL, 0)) {
+ printf(": failed to map BAR0\n");
+ return;
+ }
+
+ sc->sc_nqueues = 1;
+
+ if (pci_intr_map_msix(pa, 0, &ih) == 0) {
+ irqmode = AQ_INTR_CTRL_IRQMODE_MSIX;
+ } else if (pci_intr_map_msi(pa, &ih) == 0) {
+ irqmode = AQ_INTR_CTRL_IRQMODE_MSI;
+ } else if (pci_intr_map(pa, &ih) == 0) {
+ irqmode = AQ_INTR_CTRL_IRQMODE_LEGACY;
+ } else {
+ printf(": failed to map interrupt\n");
+ return;
+ }
+
+ isr = aq_intr;
+ sc->sc_ih = pci_intr_establish(pa->pa_pc, ih,
+ IPL_NET | IPL_MPSAFE, isr, sc, self->dv_xname);
+ intrstr = pci_intr_string(pa->pa_pc, ih);
+ if (intrstr)
+ printf(": %s", intrstr);
+
+ if (aq_fw_reset(sc))
+ return;
+
+ DPRINTF((", FW version 0x%x", sc->sc_fw_version));
+
+ if (aq_fw_version_init(sc))
+ return;
+
+ if (aq_hw_init_ucp(sc))
+ return;
+
+ if (aq_hw_reset(sc))
+ return;
+
+ if (aq_get_mac_addr(sc))
+ return;
+
+ if (aq_hw_init(sc, irqmode))
+ return;
+
+ sc->sc_media_type = aqp->aq_media_type;
+ sc->sc_available_rates = aqp->aq_available_rates;
+
+ ifmedia_init(&sc->sc_media, IFM_IMASK, aq_ifmedia_change,
+ aq_ifmedia_status);
+
+ bcopy(sc->sc_enaddr.ether_addr_octet, sc->sc_arpcom.ac_enaddr, 6);
+ strlcpy(ifp->if_xname, self->dv_xname, IFNAMSIZ);
+ ifp->if_softc = sc;
+ ifp->if_flags = IFF_BROADCAST | IFF_MULTICAST | IFF_SIMPLEX;
+ ifp->if_xflags = IFXF_MPSAFE;
+ ifp->if_ioctl = aq_ioctl;
+ ifp->if_qstart = aq_start;
+ ifp->if_watchdog = aq_watchdog;
+ ifp->if_hardmtu = 9000;
+ ifp->if_capabilities = IFCAP_VLAN_MTU;
+ ifq_set_maxlen(&ifp->if_snd, AQ_TXD_NUM);
+
+ ifmedia_init(&sc->sc_media, IFM_IMASK, aq_ifmedia_change,
+ aq_ifmedia_status);
+ if (sc->sc_available_rates & AQ_LINK_100M) {
+ ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_100_TX, 0, NULL);
+ ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_100_TX|IFM_FDX, 0,
+ NULL);
+ }
+
+ if (sc->sc_available_rates & AQ_LINK_1G) {
+ ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_1000_T, 0, NULL);
+ ifmedia_add(&sc->sc_media, IFM_ETHER|IFM_1000_T|IFM_FDX, 0,
+ NULL);
+ }
+
+ if (sc->sc_available_rates & AQ_LINK_2G5) {
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_2500_T, 0, NULL);
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_2500_T | IFM_FDX,
+ 0, NULL);
+ }
+
+ if (sc->sc_available_rates & AQ_LINK_5G) {
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_5000_T, 0, NULL);
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_5000_T | IFM_FDX,
+ 0, NULL);
+ }
+
+ if (sc->sc_available_rates & AQ_LINK_10G) {
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10G_T, 0, NULL);
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10G_T | IFM_FDX,
+ 0, NULL);
+ }
+
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO, 0, NULL);
+ ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO | IFM_FDX, 0, NULL);
+ ifmedia_set(&sc->sc_media, IFM_ETHER | IFM_AUTO);
+ aq_set_linkmode(sc, AQ_LINK_AUTO, AQ_FC_NONE, AQ_EEE_DISABLE);
+
+ if_attach(ifp);
+ ether_ifattach(ifp);
+
+ if_attach_iqueues(ifp, sc->sc_nqueues);
+ if_attach_queues(ifp, sc->sc_nqueues);
+
+ for (i = 0; i < sc->sc_nqueues; i++) {
+ struct aq_queues *aq = &sc->sc_queues[i];
+ struct aq_rxring *rx = &aq->q_rx;
+ struct aq_txring *tx = &aq->q_tx;
+
+ aq->q_sc = sc;
+ aq->q_index = i;
+ rx->rx_q = i;
+ rx->rx_irq = i * 2;
+ rx->rx_ifiq = ifp->if_iqs[i];
+ ifp->if_iqs[i]->ifiq_softc = aq;
+ timeout_set(&rx->rx_refill, aq_refill, rx);
+
+ tx->tx_q = i;
+ tx->tx_irq = rx->rx_irq + 1;
+ tx->tx_ifq = ifp->if_ifqs[i];
+ ifp->if_ifqs[i]->ifq_softc = aq;
+
+ if (sc->sc_nqueues > 1) {
+ /* map msix */
+ }
+
+ AQ_WRITE_REG(sc, TX_INTR_MODERATION_CTL_REG(i), 0);
+ AQ_WRITE_REG(sc, RX_INTR_MODERATION_CTL_REG(i), 0);
+ }
+
+ AQ_WRITE_REG_BIT(sc, TX_DMA_INT_DESC_WRWB_EN_REG,
+ TX_DMA_INT_DESC_WRWB_EN, 1);
+ AQ_WRITE_REG_BIT(sc, TX_DMA_INT_DESC_WRWB_EN_REG,
+ TX_DMA_INT_DESC_MODERATE_EN, 0);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_INT_DESC_WRWB_EN_REG,
+ RX_DMA_INT_DESC_WRWB_EN, 1);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_INT_DESC_WRWB_EN_REG,
+ RX_DMA_INT_DESC_MODERATE_EN, 0);
+
+ aq_enable_intr(sc, 1, 0);
+ printf("\n");
+}
+
+int
+aq_fw_reset(struct aq_softc *sc)
+{
+ uint32_t ver, v, boot_exit_code;
+ int i, error;
+ enum aq_fw_bootloader_mode mode;
+
+ mode = FW_BOOT_MODE_UNKNOWN;
+
+ ver = AQ_READ_REG(sc, AQ_FW_VERSION_REG);
+
+ for (i = 1000; i > 0; i--) {
+ v = AQ_READ_REG(sc, FW_MPI_DAISY_CHAIN_STATUS_REG);
+ boot_exit_code = AQ_READ_REG(sc, FW_BOOT_EXIT_CODE_REG);
+ if (v != 0x06000000 || boot_exit_code != 0)
+ break;
+ }
+
+ if (i <= 0) {
+ printf("%s: F/W reset failed. Neither RBL nor FLB started",
+ DEVNAME(sc));
+ return ETIMEDOUT;
+ }
+
+ sc->sc_rbl_enabled = (boot_exit_code != 0);
+
+ /*
+ * Having FW version 0 is an indicator that cold start
+ * is in progress. This means two things:
+ * 1) Driver have to wait for FW/HW to finish boot (500ms giveup)
+ * 2) Driver may skip reset sequence and save time.
+ */
+ if (sc->sc_fast_start_enabled && (ver != 0)) {
+ error = aq_fw_read_version(sc);
+ /* Skip reset as it just completed */
+ if (error == 0)
+ return 0;
+ }
+
+ error = aq_mac_soft_reset(sc, &mode);
+ if (error != 0) {
+ printf("%s: MAC reset failed: %d\n", DEVNAME(sc), error);
+ return error;
+ }
+
+ switch (mode) {
+ case FW_BOOT_MODE_FLB:
+ DPRINTF(("%s: FLB> F/W successfully loaded from flash.",
+ DEVNAME(sc)));
+ sc->sc_flash_present = 1;
+ return aq_fw_read_version(sc);
+ case FW_BOOT_MODE_RBL_FLASH:
+ DPRINTF(("%s: RBL> F/W loaded from flash. Host Bootload "
+ "disabled.", DEVNAME(sc)));
+ sc->sc_flash_present = 1;
+ return aq_fw_read_version(sc);
+ case FW_BOOT_MODE_UNKNOWN:
+ printf("%s: F/W bootload error: unknown bootloader type",
+ DEVNAME(sc));
+ return ENOTSUP;
+ case FW_BOOT_MODE_RBL_HOST_BOOTLOAD:
+ printf("%s: RBL> F/W Host Bootload not implemented", DEVNAME(sc));
+ return ENOTSUP;
+ }
+
+ return ENOTSUP;
+}
+
+int
+aq_mac_soft_reset_rbl(struct aq_softc *sc, enum aq_fw_bootloader_mode *mode)
+{
+ int timo;
+
+ DPRINTF(("%s: RBL> MAC reset STARTED!\n", DEVNAME(sc)));
+
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x40e1);
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CPU_SEM_REG(0), 1);
+ AQ_WRITE_REG(sc, AQ_MBOXIF_POWER_GATING_CONTROL_REG, 0);
+
+ /* MAC FW will reload PHY FW if 1E.1000.3 was cleaned - #undone */
+ AQ_WRITE_REG(sc, FW_BOOT_EXIT_CODE_REG, RBL_STATUS_DEAD);
+
+ aq_global_software_reset(sc);
+
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x40e0);
+
+ /* Wait for RBL to finish boot process. */
+#define RBL_TIMEOUT_MS 10000
+ uint16_t rbl_status;
+ for (timo = RBL_TIMEOUT_MS; timo > 0; timo--) {
+ rbl_status = AQ_READ_REG(sc, FW_BOOT_EXIT_CODE_REG) & 0xffff;
+ if (rbl_status != 0 && rbl_status != RBL_STATUS_DEAD)
+ break;
+ delay(1000);
+ }
+
+ if (timo <= 0) {
+ printf("%s: RBL> RBL restart failed: timeout\n", DEVNAME(sc));
+ return EBUSY;
+ }
+
+ switch (rbl_status) {
+ case RBL_STATUS_SUCCESS:
+ if (mode != NULL)
+ *mode = FW_BOOT_MODE_RBL_FLASH;
+ DPRINTF(("%s: RBL> reset complete! [Flash]\n", DEVNAME(sc)));
+ break;
+ case RBL_STATUS_HOST_BOOT:
+ if (mode != NULL)
+ *mode = FW_BOOT_MODE_RBL_HOST_BOOTLOAD;
+ DPRINTF(("%s: RBL> reset complete! [Host Bootload]\n",
+ DEVNAME(sc)));
+ break;
+ case RBL_STATUS_FAILURE:
+ default:
+ printf("%s: unknown RBL status 0x%x\n", DEVNAME(sc),
+ rbl_status);
+ return EBUSY;
+ }
+
+ return 0;
+}
+
+int
+aq_mac_soft_reset_flb(struct aq_softc *sc)
+{
+ uint32_t v;
+ int timo;
+
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x40e1);
+ /*
+ * Let Felicity hardware to complete SMBUS transaction before
+ * Global software reset.
+ */
+ delay(50000);
+
+ /*
+ * If SPI burst transaction was interrupted(before running the script),
+ * global software reset may not clear SPI interface.
+ * Clean it up manually before global reset.
+ */
+ AQ_WRITE_REG(sc, AQ_GLB_NVR_PROVISIONING2_REG, 0x00a0);
+ AQ_WRITE_REG(sc, AQ_GLB_NVR_INTERFACE1_REG, 0x009f);
+ AQ_WRITE_REG(sc, AQ_GLB_NVR_INTERFACE1_REG, 0x809f);
+ delay(50000);
+
+ v = AQ_READ_REG(sc, AQ_FW_SOFTRESET_REG);
+ v &= ~AQ_FW_SOFTRESET_DIS;
+ v |= AQ_FW_SOFTRESET_RESET;
+ AQ_WRITE_REG(sc, AQ_FW_SOFTRESET_REG, v);
+
+ /* Kickstart. */
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x80e0);
+ AQ_WRITE_REG(sc, AQ_MBOXIF_POWER_GATING_CONTROL_REG, 0);
+ if (!sc->sc_fast_start_enabled)
+ AQ_WRITE_REG(sc, AQ_GLB_GENERAL_PROVISIONING9_REG, 1);
+
+ /*
+ * For the case SPI burst transaction was interrupted (by MCP reset
+ * above), wait until it is completed by hardware.
+ */
+ delay(50000);
+
+ /* MAC Kickstart */
+ if (!sc->sc_fast_start_enabled) {
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x180e0);
+
+ uint32_t flb_status;
+ for (timo = 0; timo < 1000; timo++) {
+ flb_status = AQ_READ_REG(sc,
+ FW_MPI_DAISY_CHAIN_STATUS_REG) & 0x10;
+ if (flb_status != 0)
+ break;
+ delay(1000);
+ }
+ if (flb_status == 0) {
+ printf("%s: FLB> MAC kickstart failed: timed out\n",
+ DEVNAME(sc));
+ return ETIMEDOUT;
+ }
+ DPRINTF(("%s: FLB> MAC kickstart done, %d ms\n", DEVNAME(sc),
+ timo));
+ /* FW reset */
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CTL2_REG, 0x80e0);
+ /*
+ * Let Felicity hardware complete SMBUS transaction before
+ * Global software reset.
+ */
+ delay(50000);
+ sc->sc_fast_start_enabled = true;
+ }
+ AQ_WRITE_REG(sc, AQ_FW_GLB_CPU_SEM_REG(0), 1);
+
+ /* PHY Kickstart: #undone */
+ aq_global_software_reset(sc);
+
+ for (timo = 0; timo < 1000; timo++) {
+ if (AQ_READ_REG(sc, AQ_FW_VERSION_REG) != 0)
+ break;
+ delay(10000);
+ }
+ if (timo >= 1000) {
+ printf("%s: FLB> Global Soft Reset failed\n", DEVNAME(sc));
+ return ETIMEDOUT;
+ }
+ DPRINTF(("%s: FLB> F/W restart: %d ms\n", DEVNAME(sc), timo * 10));
+
+ return 0;
+
+}
+
+int
+aq_mac_soft_reset(struct aq_softc *sc, enum aq_fw_bootloader_mode *mode)
+{
+ if (sc->sc_rbl_enabled)
+ return aq_mac_soft_reset_rbl(sc, mode);
+
+ if (mode != NULL)
+ *mode = FW_BOOT_MODE_FLB;
+ return aq_mac_soft_reset_flb(sc);
+}
+
+void
+aq_global_software_reset(struct aq_softc *sc)
+{
+ uint32_t v;
+
+ AQ_WRITE_REG_BIT(sc, RX_SYSCONTROL_REG, RX_SYSCONTROL_RESET_DIS, 0);
+ AQ_WRITE_REG_BIT(sc, TX_SYSCONTROL_REG, TX_SYSCONTROL_RESET_DIS, 0);
+ AQ_WRITE_REG_BIT(sc, FW_MPI_RESETCTRL_REG,
+ FW_MPI_RESETCTRL_RESET_DIS, 0);
+
+ v = AQ_READ_REG(sc, AQ_FW_SOFTRESET_REG);
+ v &= ~AQ_FW_SOFTRESET_DIS;
+ v |= AQ_FW_SOFTRESET_RESET;
+ AQ_WRITE_REG(sc, AQ_FW_SOFTRESET_REG, v);
+}
+
+int
+aq_fw_read_version(struct aq_softc *sc)
+{
+ int i, error = EBUSY;
+#define MAC_FW_START_TIMEOUT_MS 10000
+ for (i = 0; i < MAC_FW_START_TIMEOUT_MS; i++) {
+ sc->sc_fw_version = AQ_READ_REG(sc, AQ_FW_VERSION_REG);
+ if (sc->sc_fw_version != 0) {
+ error = 0;
+ break;
+ }
+ delay(1000);
+ }
+ return error;
+}
+
+int
+aq_fw_version_init(struct aq_softc *sc)
+{
+ int error = 0;
+ char fw_vers[sizeof("F/W version xxxxx.xxxxx.xxxxx")];
+
+ if (FW_VERSION_MAJOR(sc) == 1) {
+ sc->sc_fw_ops = &aq_fw1x_ops;
+ } else if ((FW_VERSION_MAJOR(sc) == 2) || (FW_VERSION_MAJOR(sc) == 3)) {
+ sc->sc_fw_ops = &aq_fw2x_ops;
+ } else {
+ printf("%s: Unsupported F/W version %d.%d.%d\n",
+ DEVNAME(sc),
+ FW_VERSION_MAJOR(sc), FW_VERSION_MINOR(sc),
+ FW_VERSION_BUILD(sc));
+ return ENOTSUP;
+ }
+ snprintf(fw_vers, sizeof(fw_vers), "F/W version %d.%d.%d",
+ FW_VERSION_MAJOR(sc), FW_VERSION_MINOR(sc), FW_VERSION_BUILD(sc));
+
+ /* detect revision */
+ uint32_t hwrev = AQ_READ_REG(sc, AQ_HW_REVISION_REG);
+ switch (hwrev & 0x0000000f) {
+ case 0x01:
+ printf(", revision A0, %s", fw_vers);
+ sc->sc_features |= FEATURES_REV_A0 |
+ FEATURES_MPI_AQ | FEATURES_MIPS;
+ break;
+ case 0x02:
+ printf(", revision B0, %s", fw_vers);
+ sc->sc_features |= FEATURES_REV_B0 |
+ FEATURES_MPI_AQ | FEATURES_MIPS |
+ FEATURES_TPO2 | FEATURES_RPF2;
+ break;
+ case 0x0A:
+ printf(", revision B1, %s", fw_vers);
+ sc->sc_features |= FEATURES_REV_B1 |
+ FEATURES_MPI_AQ | FEATURES_MIPS |
+ FEATURES_TPO2 | FEATURES_RPF2;
+ break;
+ default:
+ printf(", Unknown revision (0x%08x)", hwrev);
+ error = ENOTSUP;
+ break;
+ }
+ return error;
+}
+
+int
+aq_hw_init_ucp(struct aq_softc *sc)
+{
+ int timo;
+
+ if (FW_VERSION_MAJOR(sc) == 1) {
+ if (AQ_READ_REG(sc, FW1X_MPI_INIT2_REG) == 0) {
+ uint32_t data;
+ arc4random_buf(&data, sizeof(data));
+ data &= 0xfefefefe;
+ data |= 0x02020202;
+ AQ_WRITE_REG(sc, FW1X_MPI_INIT2_REG, data);
+ }
+ AQ_WRITE_REG(sc, FW1X_MPI_INIT1_REG, 0);
+ }
+
+ for (timo = 100; timo > 0; timo--) {
+ sc->sc_mbox_addr = AQ_READ_REG(sc, FW_MPI_MBOX_ADDR_REG);
+ if (sc->sc_mbox_addr != 0)
+ break;
+ delay(1000);
+ }
+
+#define AQ_FW_MIN_VERSION 0x01050006
+#define AQ_FW_MIN_VERSION_STR "1.5.6"
+ if (sc->sc_fw_version < AQ_FW_MIN_VERSION) {
+ printf("%s: atlantic: wrong FW version: " AQ_FW_MIN_VERSION_STR
+ " or later required, this is %d.%d.%d\n",
+ DEVNAME(sc),
+ FW_VERSION_MAJOR(sc),
+ FW_VERSION_MINOR(sc),
+ FW_VERSION_BUILD(sc));
+ return ENOTSUP;
+ }
+
+ if (sc->sc_mbox_addr == 0)
+ printf("%s: NULL MBOX!!\n", DEVNAME(sc));
+
+ return 0;
+}
+
+int
+aq_hw_reset(struct aq_softc *sc)
+{
+ int error;
+
+ /* disable irq */
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_RESET_DIS, 0);
+
+ /* apply */
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_RESET_IRQ, 1);
+
+ /* wait ack 10 times by 1ms */
+ WAIT_FOR(
+ (AQ_READ_REG(sc, AQ_INTR_CTRL_REG) & AQ_INTR_CTRL_RESET_IRQ) == 0,
+ 1000, 10, &error);
+ if (error != 0) {
+ printf("%s: atlantic: IRQ reset failed: %d\n", DEVNAME(sc),
+ error);
+ return error;
+ }
+
+ return sc->sc_fw_ops->reset(sc);
+}
+
+int
+aq_get_mac_addr(struct aq_softc *sc)
+{
+ uint32_t mac_addr[2];
+ uint32_t efuse_shadow_addr;
+ int err;
+
+ efuse_shadow_addr = 0;
+ if (FW_VERSION_MAJOR(sc) >= 2)
+ efuse_shadow_addr = AQ_READ_REG(sc, FW2X_MPI_EFUSEADDR_REG);
+ else
+ efuse_shadow_addr = AQ_READ_REG(sc, FW1X_MPI_EFUSEADDR_REG);
+
+ if (efuse_shadow_addr == 0) {
+ printf("%s: cannot get efuse addr", DEVNAME(sc));
+ return ENXIO;
+ }
+
+ DPRINTF(("%s: efuse_shadow_addr = %x\n", DEVNAME(sc), efuse_shadow_addr));
+
+ memset(mac_addr, 0, sizeof(mac_addr));
+ err = aq_fw_downld_dwords(sc, efuse_shadow_addr + (40 * 4),
+ mac_addr, 2);
+ if (err < 0)
+ return err;
+
+ if (mac_addr[0] == 0 && mac_addr[1] == 0) {
+ printf("%s: mac address not found", DEVNAME(sc));
+ return ENXIO;
+ }
+
+ DPRINTF(("%s: mac0 %x mac1 %x\n", DEVNAME(sc), mac_addr[0],
+ mac_addr[1]));
+
+ mac_addr[0] = htobe32(mac_addr[0]);
+ mac_addr[1] = htobe32(mac_addr[1]);
+
+ DPRINTF(("%s: mac0 %x mac1 %x\n", DEVNAME(sc), mac_addr[0],
+ mac_addr[1]));
+
+ memcpy(sc->sc_enaddr.ether_addr_octet,
+ (uint8_t *)mac_addr, ETHER_ADDR_LEN);
+ DPRINTF((": %s", ether_sprintf(sc->sc_enaddr.ether_addr_octet)));
+
+ return 0;
+}
+
+int
+aq_activate(struct device *self, int act)
+{
+ return 0;
+}
+
+int
+aq_fw_downld_dwords(struct aq_softc *sc, uint32_t addr, uint32_t *p,
+ uint32_t cnt)
+{
+ uint32_t v;
+ int error = 0;
+
+ WAIT_FOR(AQ_READ_REG(sc, AQ_FW_SEM_RAM_REG) == 1, 1, 10000, &error);
+ if (error != 0) {
+ AQ_WRITE_REG(sc, AQ_FW_SEM_RAM_REG, 1);
+ v = AQ_READ_REG(sc, AQ_FW_SEM_RAM_REG);
+ if (v == 0) {
+ printf("%s: %s:%d: timeout\n",
+ DEVNAME(sc), __func__, __LINE__);
+ return ETIMEDOUT;
+ }
+ }
+
+ AQ_WRITE_REG(sc, AQ_FW_MBOX_ADDR_REG, addr);
+
+ error = 0;
+ for (; cnt > 0 && error == 0; cnt--) {
+ /* execute mailbox interface */
+ AQ_WRITE_REG_BIT(sc, AQ_FW_MBOX_CMD_REG,
+ AQ_FW_MBOX_CMD_EXECUTE, 1);
+ if (sc->sc_features & FEATURES_REV_B1) {
+ WAIT_FOR(AQ_READ_REG(sc, AQ_FW_MBOX_ADDR_REG) != addr,
+ 1, 1000, &error);
+ } else {
+ WAIT_FOR((AQ_READ_REG(sc, AQ_FW_MBOX_CMD_REG) &
+ AQ_FW_MBOX_CMD_BUSY) == 0,
+ 1, 1000, &error);
+ }
+ *p++ = AQ_READ_REG(sc, AQ_FW_MBOX_VAL_REG);
+ addr += sizeof(uint32_t);
+ }
+ AQ_WRITE_REG(sc, AQ_FW_SEM_RAM_REG, 1);
+
+ if (error != 0)
+ printf("%s: %s:%d: timeout\n",
+ DEVNAME(sc), __func__, __LINE__);
+
+ return error;
+}
+
+int
+aq_fw2x_reset(struct aq_softc *sc)
+{
+ struct aq_fw2x_capabilities caps = { 0 };
+ int error;
+
+ error = aq_fw_downld_dwords(sc,
+ sc->sc_mbox_addr + offsetof(struct aq_fw2x_mailbox, caps),
+ (uint32_t *)&caps, sizeof caps / sizeof(uint32_t));
+ if (error != 0) {
+ printf("%s: fw2x> can't get F/W capabilities mask, error %d\n",
+ DEVNAME(sc), error);
+ return error;
+ }
+ sc->sc_fw_caps = caps.caps_lo | ((uint64_t)caps.caps_hi << 32);
+
+ DPRINTF(("%s: fw2x> F/W capabilities=0x%llx\n", DEVNAME(sc),
+ sc->sc_fw_caps));
+
+ return 0;
+}
+
+int
+aq_fw1x_reset(struct aq_softc *sc)
+{
+ printf("%s: unimplemented %s\n", DEVNAME(sc), __func__);
+ return 0;
+}
+
+int
+aq_fw1x_set_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state w,
+ enum aq_link_speed x, enum aq_link_fc y, enum aq_link_eee z)
+{
+ return 0;
+}
+
+int
+aq_fw1x_get_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state *w,
+ enum aq_link_speed *x, enum aq_link_fc *y, enum aq_link_eee *z)
+{
+ return 0;
+}
+
+int
+aq_fw1x_get_stats(struct aq_softc *sc, struct aq_hw_stats_s *w)
+{
+ return 0;
+}
+
+
+int
+aq_fw2x_get_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state *modep,
+ enum aq_link_speed *speedp, enum aq_link_fc *fcp, enum aq_link_eee *eeep)
+{
+ uint64_t mpi_state, mpi_ctrl;
+ enum aq_link_speed speed;
+ enum aq_link_fc fc;
+
+ AQ_MPI_LOCK(sc);
+
+ mpi_state = AQ_READ64_REG(sc, FW2X_MPI_STATE_REG);
+ if (modep != NULL) {
+ mpi_ctrl = AQ_READ64_REG(sc, FW2X_MPI_CONTROL_REG);
+ if (mpi_ctrl & FW2X_CTRL_RATE_MASK)
+ *modep = MPI_INIT;
+ else
+ *modep = MPI_DEINIT;
+ }
+
+ AQ_MPI_UNLOCK(sc);
+
+ if (mpi_state & FW2X_CTRL_RATE_10G)
+ speed = AQ_LINK_10G;
+ else if (mpi_state & FW2X_CTRL_RATE_5G)
+ speed = AQ_LINK_5G;
+ else if (mpi_state & FW2X_CTRL_RATE_2G5)
+ speed = AQ_LINK_2G5;
+ else if (mpi_state & FW2X_CTRL_RATE_1G)
+ speed = AQ_LINK_1G;
+ else if (mpi_state & FW2X_CTRL_RATE_100M)
+ speed = AQ_LINK_100M;
+ else
+ speed = AQ_LINK_NONE;
+ if (speedp != NULL)
+ *speedp = speed;
+
+ fc = AQ_FC_NONE;
+ if (mpi_state & FW2X_CTRL_PAUSE)
+ fc |= AQ_FC_RX;
+ if (mpi_state & FW2X_CTRL_ASYMMETRIC_PAUSE)
+ fc |= AQ_FC_TX;
+ if (fcp != NULL)
+ *fcp = fc;
+
+ if (eeep != NULL)
+ *eeep = AQ_EEE_DISABLE;
+
+ return 0;
+}
+
+int
+aq_fw2x_get_stats(struct aq_softc *sc, struct aq_hw_stats_s *w)
+{
+ return 0;
+}
+
+void
+aq_hw_l3_filter_set(struct aq_softc *sc)
+{
+ int i;
+
+ /* clear all filter */
+ for (i = 0; i < 8; i++) {
+ AQ_WRITE_REG_BIT(sc, RPF_L3_FILTER_REG(i),
+ RPF_L3_FILTER_L4_EN, 0);
+ }
+}
+
+int
+aq_hw_init(struct aq_softc *sc, int irqmode)
+{
+ uint32_t v;
+
+ /* Force limit MRRS on RDM/TDM to 2K */
+ v = AQ_READ_REG(sc, AQ_PCI_REG_CONTROL_6_REG);
+ AQ_WRITE_REG(sc, AQ_PCI_REG_CONTROL_6_REG, (v & ~0x0707) | 0x0404);
+
+ /*
+ * TX DMA total request limit. B0 hardware is not capable to
+ * handle more than (8K-MRRS) incoming DMA data.
+ * Value 24 in 256byte units
+ */
+ AQ_WRITE_REG(sc, AQ_HW_TX_DMA_TOTAL_REQ_LIMIT_REG, 24);
+
+ aq_hw_init_tx_path(sc);
+ aq_hw_init_rx_path(sc);
+
+ if (aq_set_mac_addr(sc, AQ_HW_MAC_OWN, sc->sc_enaddr.ether_addr_octet))
+ return EINVAL;
+
+ aq_set_linkmode(sc, AQ_LINK_NONE, AQ_FC_NONE, AQ_EEE_DISABLE);
+
+ aq_hw_qos_set(sc);
+
+ /* Enable interrupt */
+ AQ_WRITE_REG(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_RESET_DIS);
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_MULTIVEC, 0);
+
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_CTRL_REG, AQ_INTR_CTRL_IRQMODE, irqmode);
+
+ AQ_WRITE_REG(sc, AQ_INTR_AUTOMASK_REG, 0xffffffff);
+
+ AQ_WRITE_REG(sc, AQ_GEN_INTR_MAP_REG(0),
+ ((AQ_B0_ERR_INT << 24) | (1U << 31)) |
+ ((AQ_B0_ERR_INT << 16) | (1 << 23))
+ );
+
+ /* link interrupt */
+ sc->sc_linkstat_irq = AQ_LINKSTAT_IRQ;
+ AQ_WRITE_REG(sc, AQ_GEN_INTR_MAP_REG(3),
+ (1 << 7) | sc->sc_linkstat_irq);
+
+ return 0;
+}
+
+void
+aq_hw_init_tx_path(struct aq_softc *sc)
+{
+ /* Tx TC/RSS number config */
+ AQ_WRITE_REG_BIT(sc, TPB_TX_BUF_REG, TPB_TX_BUF_TC_MODE_EN, 1);
+
+ AQ_WRITE_REG_BIT(sc, THM_LSO_TCP_FLAG1_REG,
+ THM_LSO_TCP_FLAG1_FIRST, 0x0ff6);
+ AQ_WRITE_REG_BIT(sc, THM_LSO_TCP_FLAG1_REG,
+ THM_LSO_TCP_FLAG1_MID, 0x0ff6);
+ AQ_WRITE_REG_BIT(sc, THM_LSO_TCP_FLAG2_REG,
+ THM_LSO_TCP_FLAG2_LAST, 0x0f7f);
+
+ /* misc */
+ AQ_WRITE_REG(sc, TX_TPO2_REG,
+ (sc->sc_features & FEATURES_TPO2) ? TX_TPO2_EN : 0);
+ AQ_WRITE_REG_BIT(sc, TDM_DCA_REG, TDM_DCA_EN, 0);
+ AQ_WRITE_REG_BIT(sc, TDM_DCA_REG, TDM_DCA_MODE, 0);
+
+ AQ_WRITE_REG_BIT(sc, TPB_TX_BUF_REG, TPB_TX_BUF_SCP_INS_EN, 1);
+}
+
+void
+aq_hw_init_rx_path(struct aq_softc *sc)
+{
+ int i;
+
+ /* clear setting */
+ AQ_WRITE_REG_BIT(sc, RPB_RPF_RX_REG, RPB_RPF_RX_TC_MODE, 0);
+ AQ_WRITE_REG_BIT(sc, RPB_RPF_RX_REG, RPB_RPF_RX_FC_MODE, 0);
+ AQ_WRITE_REG(sc, RX_FLR_RSS_CONTROL1_REG, 0);
+ for (i = 0; i < 32; i++) {
+ AQ_WRITE_REG_BIT(sc, RPF_ETHERTYPE_FILTER_REG(i),
+ RPF_ETHERTYPE_FILTER_EN, 0);
+ }
+
+ /* L2 and Multicast filters */
+ for (i = 0; i < AQ_HW_MAC_NUM; i++) {
+ AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(i), RPF_L2UC_MSW_EN, 0);
+ AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(i), RPF_L2UC_MSW_ACTION,
+ RPF_ACTION_HOST);
+ }
+ AQ_WRITE_REG(sc, RPF_MCAST_FILTER_MASK_REG, 0);
+ AQ_WRITE_REG(sc, RPF_MCAST_FILTER_REG(0), 0x00010fff);
+
+ /* Vlan filters */
+ AQ_WRITE_REG_BIT(sc, RPF_VLAN_TPID_REG, RPF_VLAN_TPID_OUTER,
+ ETHERTYPE_QINQ);
+ AQ_WRITE_REG_BIT(sc, RPF_VLAN_TPID_REG, RPF_VLAN_TPID_INNER,
+ ETHERTYPE_VLAN);
+ AQ_WRITE_REG_BIT(sc, RPF_VLAN_MODE_REG, RPF_VLAN_MODE_PROMISC, 0);
+
+ if (sc->sc_features & FEATURES_REV_B) {
+ AQ_WRITE_REG_BIT(sc, RPF_VLAN_MODE_REG,
+ RPF_VLAN_MODE_ACCEPT_UNTAGGED, 1);
+ AQ_WRITE_REG_BIT(sc, RPF_VLAN_MODE_REG,
+ RPF_VLAN_MODE_UNTAGGED_ACTION, RPF_ACTION_HOST);
+ }
+
+ AQ_WRITE_REG(sc, RX_TCP_RSS_HASH_REG, 0);
+
+ AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_EN, 1);
+ AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_ACTION, RPF_ACTION_HOST);
+ AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_THRESHOLD, 0xffff);
+
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DCA_REG, RX_DMA_DCA_EN, 0);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DCA_REG, RX_DMA_DCA_MODE, 0);
+}
+
+/* set multicast filter. index 0 for own address */
+int
+aq_set_mac_addr(struct aq_softc *sc, int index, uint8_t *enaddr)
+{
+ uint32_t h, l;
+
+ if (index >= AQ_HW_MAC_NUM)
+ return EINVAL;
+
+ if (enaddr == NULL) {
+ /* disable */
+ AQ_WRITE_REG_BIT(sc,
+ RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_EN, 0);
+ return 0;
+ }
+
+ h = (enaddr[0] << 8) | (enaddr[1]);
+ l = ((uint32_t)enaddr[2] << 24) | (enaddr[3] << 16) |
+ (enaddr[4] << 8) | (enaddr[5]);
+
+ /* disable, set, and enable */
+ AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_EN, 0);
+ AQ_WRITE_REG(sc, RPF_L2UC_LSW_REG(index), l);
+ AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index),
+ RPF_L2UC_MSW_MACADDR_HI, h);
+ AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_ACTION, 1);
+ AQ_WRITE_REG_BIT(sc, RPF_L2UC_MSW_REG(index), RPF_L2UC_MSW_EN, 1);
+
+ return 0;
+}
+
+int
+aq_get_linkmode(struct aq_softc *sc, enum aq_link_speed *speed,
+ enum aq_link_fc *fc, enum aq_link_eee *eee)
+{
+ enum aq_hw_fw_mpi_state mode;
+ int error;
+
+ error = sc->sc_fw_ops->get_mode(sc, &mode, speed, fc, eee);
+ if (error != 0)
+ return error;
+ if (mode != MPI_INIT)
+ return ENXIO;
+
+ return 0;
+}
+
+int
+aq_set_linkmode(struct aq_softc *sc, enum aq_link_speed speed,
+ enum aq_link_fc fc, enum aq_link_eee eee)
+{
+ return sc->sc_fw_ops->set_mode(sc, MPI_INIT, speed, fc, eee);
+}
+
+int
+aq_fw2x_set_mode(struct aq_softc *sc, enum aq_hw_fw_mpi_state mode,
+ enum aq_link_speed speed, enum aq_link_fc fc, enum aq_link_eee eee)
+{
+ uint64_t mpi_ctrl;
+ int error = 0;
+
+ AQ_MPI_LOCK(sc);
+
+ mpi_ctrl = AQ_READ64_REG(sc, FW2X_MPI_CONTROL_REG);
+
+ switch (mode) {
+ case MPI_INIT:
+ mpi_ctrl &= ~FW2X_CTRL_RATE_MASK;
+ if (speed & AQ_LINK_10G)
+ mpi_ctrl |= FW2X_CTRL_RATE_10G;
+ if (speed & AQ_LINK_5G)
+ mpi_ctrl |= FW2X_CTRL_RATE_5G;
+ if (speed & AQ_LINK_2G5)
+ mpi_ctrl |= FW2X_CTRL_RATE_2G5;
+ if (speed & AQ_LINK_1G)
+ mpi_ctrl |= FW2X_CTRL_RATE_1G;
+ if (speed & AQ_LINK_100M)
+ mpi_ctrl |= FW2X_CTRL_RATE_100M;
+
+ mpi_ctrl &= ~FW2X_CTRL_LINK_DROP;
+
+ mpi_ctrl &= ~FW2X_CTRL_EEE_MASK;
+ if (eee == AQ_EEE_ENABLE)
+ mpi_ctrl |= FW2X_CTRL_EEE_MASK;
+
+ mpi_ctrl &= ~(FW2X_CTRL_PAUSE | FW2X_CTRL_ASYMMETRIC_PAUSE);
+ if (fc & AQ_FC_RX)
+ mpi_ctrl |= FW2X_CTRL_PAUSE;
+ if (fc & AQ_FC_TX)
+ mpi_ctrl |= FW2X_CTRL_ASYMMETRIC_PAUSE;
+ break;
+ case MPI_DEINIT:
+ mpi_ctrl &= ~(FW2X_CTRL_RATE_MASK | FW2X_CTRL_EEE_MASK);
+ mpi_ctrl &= ~(FW2X_CTRL_PAUSE | FW2X_CTRL_ASYMMETRIC_PAUSE);
+ break;
+ default:
+ printf("%s: fw2x> unknown MPI state %d\n", DEVNAME(sc), mode);
+ error = EINVAL;
+ goto failure;
+ }
+ AQ_WRITE64_REG(sc, FW2X_MPI_CONTROL_REG, mpi_ctrl);
+
+ failure:
+ AQ_MPI_UNLOCK(sc);
+ return error;
+}
+
+void
+aq_hw_qos_set(struct aq_softc *sc)
+{
+ uint32_t tc = 0;
+ uint32_t buff_size;
+
+ /* TPS Descriptor rate init */
+ AQ_WRITE_REG_BIT(sc, TPS_DESC_RATE_REG, TPS_DESC_RATE_TA_RST, 0);
+ AQ_WRITE_REG_BIT(sc, TPS_DESC_RATE_REG, TPS_DESC_RATE_LIM, 0xa);
+
+ /* TPS VM init */
+ AQ_WRITE_REG_BIT(sc, TPS_DESC_VM_ARB_MODE_REG, TPS_DESC_VM_ARB_MODE, 0);
+
+ /* TPS TC credits init */
+ AQ_WRITE_REG_BIT(sc, TPS_DESC_TC_ARB_MODE_REG, TPS_DESC_TC_ARB_MODE, 0);
+ AQ_WRITE_REG_BIT(sc, TPS_DATA_TC_ARB_MODE_REG, TPS_DATA_TC_ARB_MODE, 0);
+
+ AQ_WRITE_REG_BIT(sc, TPS_DATA_TCT_REG(tc),
+ TPS_DATA_TCT_CREDIT_MAX, 0xfff);
+ AQ_WRITE_REG_BIT(sc, TPS_DATA_TCT_REG(tc),
+ TPS_DATA_TCT_WEIGHT, 0x64);
+ AQ_WRITE_REG_BIT(sc, TPS_DESC_TCT_REG(tc),
+ TPS_DESC_TCT_CREDIT_MAX, 0x50);
+ AQ_WRITE_REG_BIT(sc, TPS_DESC_TCT_REG(tc),
+ TPS_DESC_TCT_WEIGHT, 0x1e);
+
+ /* Tx buf size */
+ tc = 0;
+ buff_size = AQ_HW_TXBUF_MAX;
+ AQ_WRITE_REG_BIT(sc, TPB_TXB_BUFSIZE_REG(tc), TPB_TXB_BUFSIZE,
+ buff_size);
+ AQ_WRITE_REG_BIT(sc, TPB_TXB_THRESH_REG(tc), TPB_TXB_THRESH_HI,
+ (buff_size * (1024 / 32) * 66) / 100);
+ AQ_WRITE_REG_BIT(sc, TPB_TXB_THRESH_REG(tc), TPB_TXB_THRESH_LO,
+ (buff_size * (1024 / 32) * 50) / 100);
+
+ /* QoS Rx buf size per TC */
+ tc = 0;
+ buff_size = AQ_HW_RXBUF_MAX;
+ AQ_WRITE_REG_BIT(sc, RPB_RXB_BUFSIZE_REG(tc), RPB_RXB_BUFSIZE,
+ buff_size);
+ AQ_WRITE_REG_BIT(sc, RPB_RXB_XOFF_REG(tc), RPB_RXB_XOFF_EN, 0);
+ AQ_WRITE_REG_BIT(sc, RPB_RXB_XOFF_REG(tc), RPB_RXB_XOFF_THRESH_HI,
+ (buff_size * (1024 / 32) * 66) / 100);
+ AQ_WRITE_REG_BIT(sc, RPB_RXB_XOFF_REG(tc), RPB_RXB_XOFF_THRESH_LO,
+ (buff_size * (1024 / 32) * 50) / 100);
+
+ /* QoS 802.1p priority -> TC mapping */
+ int i_priority;
+ for (i_priority = 0; i_priority < 8; i_priority++) {
+ AQ_WRITE_REG_BIT(sc, RPF_RPB_RX_TC_UPT_REG,
+ RPF_RPB_RX_TC_UPT_MASK(i_priority), 0);
+ }
+}
+
+void
+aq_txring_reset(struct aq_softc *sc, struct aq_txring *tx, int start)
+{
+ daddr_t paddr;
+
+ tx->tx_prod = 0;
+ tx->tx_cons = 0;
+
+ /* empty slots? */
+
+ AQ_WRITE_REG_BIT(sc, TX_DMA_DESC_REG(tx->tx_q), TX_DMA_DESC_EN, 0);
+
+ if (start == 0)
+ return;
+
+ paddr = AQ_DMA_DVA(&tx->tx_mem);
+ AQ_WRITE_REG(sc, TX_DMA_DESC_BASE_ADDRLSW_REG(tx->tx_q), paddr);
+ AQ_WRITE_REG(sc, TX_DMA_DESC_BASE_ADDRMSW_REG(tx->tx_q),
+ paddr >> 32);
+
+ AQ_WRITE_REG_BIT(sc, TX_DMA_DESC_REG(tx->tx_q), TX_DMA_DESC_LEN,
+ AQ_TXD_NUM / 8);
+
+ tx->tx_prod = AQ_READ_REG(sc, TX_DMA_DESC_TAIL_PTR_REG(tx->tx_q));
+ tx->tx_cons = tx->tx_prod;
+ AQ_WRITE_REG(sc, TX_DMA_DESC_WRWB_THRESH_REG(tx->tx_q), 0);
+
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_TX_REG(tx->tx_q),
+ AQ_INTR_IRQ_MAP_TX_IRQMAP(tx->tx_q), tx->tx_irq);
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_TX_REG(tx->tx_q),
+ AQ_INTR_IRQ_MAP_TX_EN(tx->tx_q), 1);
+
+ AQ_WRITE_REG_BIT(sc, TX_DMA_DESC_REG(tx->tx_q), TX_DMA_DESC_EN, 1);
+
+ AQ_WRITE_REG_BIT(sc, TDM_DCAD_REG(tx->tx_q), TDM_DCAD_CPUID, 0);
+ AQ_WRITE_REG_BIT(sc, TDM_DCAD_REG(tx->tx_q), TDM_DCAD_CPUID_EN, 0);
+}
+
+void
+aq_rxring_reset(struct aq_softc *sc, struct aq_rxring *rx, int start)
+{
+ daddr_t paddr;
+
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), RX_DMA_DESC_EN, 0);
+ /* drain */
+
+ if (start == 0)
+ return;
+
+ paddr = AQ_DMA_DVA(&rx->rx_mem);
+ AQ_WRITE_REG(sc, RX_DMA_DESC_BASE_ADDRLSW_REG(rx->rx_q), paddr);
+ AQ_WRITE_REG(sc, RX_DMA_DESC_BASE_ADDRMSW_REG(rx->rx_q),
+ paddr >> 32);
+
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), RX_DMA_DESC_LEN,
+ AQ_RXD_NUM / 8);
+
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_BUFSIZE_REG(rx->rx_q),
+ RX_DMA_DESC_BUFSIZE_DATA, MCLBYTES / 1024);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_BUFSIZE_REG(rx->rx_q),
+ RX_DMA_DESC_BUFSIZE_HDR, 0);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q),
+ RX_DMA_DESC_HEADER_SPLIT, 0);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q),
+ RX_DMA_DESC_VLAN_STRIP, 0);
+
+ rx->rx_cons = AQ_READ_REG(sc, RX_DMA_DESC_HEAD_PTR_REG(rx->rx_q)) &
+ RX_DMA_DESC_HEAD_PTR;
+ AQ_WRITE_REG(sc, RX_DMA_DESC_TAIL_PTR_REG(rx->rx_q), rx->rx_cons);
+ rx->rx_prod = rx->rx_cons;
+
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_RX_REG(rx->rx_q),
+ AQ_INTR_IRQ_MAP_RX_IRQMAP(rx->rx_q), rx->rx_irq);
+ AQ_WRITE_REG_BIT(sc, AQ_INTR_IRQ_MAP_RX_REG(rx->rx_q),
+ AQ_INTR_IRQ_MAP_RX_EN(rx->rx_q), 1);
+
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q),
+ RX_DMA_DCAD_CPUID, 0);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q),
+ RX_DMA_DCAD_DESC_EN, 0);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q),
+ RX_DMA_DCAD_HEADER_EN, 0);
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DCAD_REG(rx->rx_q),
+ RX_DMA_DCAD_PAYLOAD_EN, 0);
+
+ AQ_WRITE_REG_BIT(sc, RX_DMA_DESC_REG(rx->rx_q), RX_DMA_DESC_EN, 1);
+}
+
+static inline unsigned int
+aq_rx_fill_slots(struct aq_softc *sc, struct aq_rxring *rx, uint nslots)
+{
+ struct aq_rx_desc_read *ring, *rd;
+ struct aq_slot *as;
+ struct mbuf *m;
+ uint p, fills;
+
+ ring = AQ_DMA_KVA(&rx->rx_mem);
+ p = rx->rx_prod;
+
+ bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&rx->rx_mem), 0,
+ AQ_DMA_LEN(&rx->rx_mem), BUS_DMASYNC_POSTWRITE);
+
+ for (fills = 0; fills < nslots; fills++) {
+ as = &rx->rx_slots[p];
+ rd = &ring[p];
+
+ m = MCLGETL(NULL, M_DONTWAIT, MCLBYTES + ETHER_ALIGN);
+ if (m == NULL)
+ break;
+
+ m->m_data += (m->m_ext.ext_size - MCLBYTES);
+ m->m_data += ETHER_ALIGN;
+ m->m_len = m->m_pkthdr.len = MCLBYTES;
+
+ if (bus_dmamap_load_mbuf(sc->sc_dmat, as->as_map, m,
+ BUS_DMA_NOWAIT) != 0) {
+ m_freem(m);
+ break;
+ }
+ as->as_m = m;
+
+ htolem64(&rd->buf_addr, as->as_map->dm_segs[0].ds_addr);
+ rd->hdr_addr = 0;
+ p++;
+ if (p == AQ_RXD_NUM)
+ p = 0;
+ }
+
+ bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&rx->rx_mem), 0,
+ AQ_DMA_LEN(&rx->rx_mem), BUS_DMASYNC_PREWRITE);
+
+ rx->rx_prod = p;
+ AQ_WRITE_REG(sc, RX_DMA_DESC_TAIL_PTR_REG(rx->rx_q), rx->rx_prod);
+ return (nslots - fills);
+}
+
+int
+aq_rx_fill(struct aq_softc *sc, struct aq_rxring *rx)
+{
+ u_int slots;
+
+ slots = if_rxr_get(&rx->rx_rxr, AQ_RXD_NUM);
+ if (slots == 0)
+ return 1;
+
+ slots = aq_rx_fill_slots(sc, rx, slots);
+ if_rxr_put(&rx->rx_rxr, slots);
+ return 0;
+}
+
+void
+aq_refill(void *xq)
+{
+ struct aq_queues *q = xq;
+ struct aq_softc *sc = q->q_sc;
+
+ aq_rx_fill(sc, &q->q_rx);
+
+ if (if_rxr_inuse(&q->q_rx.rx_rxr) == 0)
+ timeout_add(&q->q_rx.rx_refill, 1);
+}
+
+void
+aq_rxeof(struct aq_softc *sc, struct aq_rxring *rx)
+{
+ struct ifnet *ifp = &sc->sc_arpcom.ac_if;
+ struct aq_rx_desc_wb *rxd;
+ struct aq_rx_desc_wb *ring;
+ struct aq_slot *as;
+ uint32_t end, idx;
+ uint16_t pktlen, status;
+ uint32_t rxd_type;
+ struct mbuf *m;
+ struct mbuf_list ml = MBUF_LIST_INITIALIZER();
+ int rxfree;
+
+ if (!ISSET(ifp->if_flags, IFF_RUNNING))
+ return;
+
+ end = AQ_READ_REG(sc, RX_DMA_DESC_HEAD_PTR_REG(rx->rx_q)) &
+ RX_DMA_DESC_HEAD_PTR;
+
+ bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&rx->rx_mem), 0,
+ AQ_DMA_LEN(&rx->rx_mem), BUS_DMASYNC_POSTREAD);
+
+ rxfree = 0;
+ idx = rx->rx_cons;
+ ring = AQ_DMA_KVA(&rx->rx_mem);
+ while (idx != end) {
+ rxd = &ring[idx];
+ as = &rx->rx_slots[idx];
+
+ bus_dmamap_sync(sc->sc_dmat, as->as_map, 0,
+ as->as_map->dm_mapsize, BUS_DMASYNC_POSTREAD);
+ bus_dmamap_unload(sc->sc_dmat, as->as_map);
+
+ status = lemtoh16(&rxd->status);
+ if ((status & AQ_RXDESC_STATUS_DD) == 0)
+ break;
+
+ rxfree++;
+ m = as->as_m;
+ as->as_m = NULL;
+
+ pktlen = lemtoh16(&rxd->pkt_len);
+ rxd_type = lemtoh32(&rxd->type);
+ /* rss hash, vlan */
+
+ if ((status & AQ_RXDESC_STATUS_MACERR) ||
+ (rxd_type & AQ_RXDESC_TYPE_DMA_ERR)) {
+ printf("%s:rx: rx error (status %x type %x)\n",
+ DEVNAME(sc), status, rxd_type);
+ m_freem(m);
+ } else {
+ m->m_pkthdr.len = m->m_len = pktlen;
+ ml_enqueue(&ml, m);
+ }
+
+ idx++;
+ if (idx == AQ_RXD_NUM)
+ idx = 0;
+ }
+ rx->rx_cons = idx;
+
+ if (rxfree > 0) {
+ if_rxr_put(&rx->rx_rxr, rxfree);
+ if (ifiq_input(rx->rx_ifiq, &ml))
+ if_rxr_livelocked(&rx->rx_rxr);
+
+ aq_rx_fill(sc, rx);
+ if (if_rxr_inuse(&rx->rx_rxr) == 0)
+ timeout_add(&rx->rx_refill, 1);
+ }
+}
+
+void
+aq_txeof(struct aq_softc *sc, struct aq_txring *tx)
+{
+ struct ifnet *ifp = &sc->sc_arpcom.ac_if;
+ struct aq_slot *as;
+ uint32_t idx, end, free;
+
+ if (!ISSET(ifp->if_flags, IFF_RUNNING))
+ return;
+
+ idx = tx->tx_cons;
+ end = AQ_READ_REG(sc, TX_DMA_DESC_HEAD_PTR_REG(tx->tx_q)) &
+ TX_DMA_DESC_HEAD_PTR;
+ free = 0;
+
+ bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&tx->tx_mem), 0,
+ AQ_DMA_LEN(&tx->tx_mem), BUS_DMASYNC_POSTREAD);
+
+ while (idx != end) {
+ as = &tx->tx_slots[idx];
+ bus_dmamap_unload(sc->sc_dmat, as->as_map);
+
+ m_freem(as->as_m);
+ as->as_m = NULL;
+
+ idx++;
+ if (idx == AQ_TXD_NUM)
+ idx = 0;
+ free++;
+ }
+
+ tx->tx_cons = idx;
+
+ if (free != 0) {
+ if (ifq_is_oactive(tx->tx_ifq))
+ ifq_restart(tx->tx_ifq);
+ }
+}
+
+void
+aq_start(struct ifqueue *ifq)
+{
+ struct aq_queues *aq = ifq->ifq_softc;
+ struct aq_softc *sc = aq->q_sc;
+ struct aq_txring *tx = &aq->q_tx;
+ struct aq_tx_desc *ring, *txd;
+ struct aq_slot *as;
+ struct mbuf *m;
+ uint32_t idx, free, used, ctl1, ctl2;
+
+ idx = tx->tx_prod;
+ free = tx->tx_cons + AQ_TXD_NUM - tx->tx_prod;
+ used = 0;
+
+ bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&tx->tx_mem), 0,
+ AQ_DMA_LEN(&tx->tx_mem), BUS_DMASYNC_POSTWRITE);
+ ring = (struct aq_tx_desc *)AQ_DMA_KVA(&tx->tx_mem);
+
+ for (;;) {
+ if (used + AQ_TX_MAX_SEGMENTS >= free) {
+ ifq_set_oactive(ifq);
+ break;
+ }
+
+ m = ifq_dequeue(ifq);
+ if (m == NULL)
+ break;
+
+ txd = ring + idx;
+ as = &tx->tx_slots[idx];
+
+ if (m_defrag(m, M_DONTWAIT) != 0) {
+ m_freem(m);
+ break;
+ }
+
+ if (bus_dmamap_load_mbuf(sc->sc_dmat, as->as_map, m,
+ BUS_DMA_STREAMING | BUS_DMA_NOWAIT) != 0) {
+ m_freem(m);
+ break;
+ }
+
+ as->as_m = m;
+
+#if NBPFILTER > 0
+ if (ifp->if_bpf)
+ bpf_mtap_ether(ifp->if_bpf, m, BPF_DIRECTION_OUT);
+#endif
+ bus_dmamap_sync(sc->sc_dmat, as->as_map, 0,
+ as->as_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
+
+ ctl1 = AQ_TXDESC_CTL1_TYPE_TXD | (as->as_map->dm_segs[0].ds_len <<
+ AQ_TXDESC_CTL1_BLEN_SHIFT) | AQ_TXDESC_CTL1_CMD_FCS |
+ AQ_TXDESC_CTL1_CMD_EOP | AQ_TXDESC_CTL1_CMD_WB;
+ ctl2 = m->m_pkthdr.len << AQ_TXDESC_CTL2_LEN_SHIFT;
+
+ txd->buf_addr = htole64(as->as_map->dm_segs[0].ds_addr);
+ txd->ctl1 = htole32(ctl1);
+ txd->ctl2 = htole32(ctl2);
+
+ idx++;
+ if (idx == AQ_TXD_NUM)
+ idx = 0;
+ used++;
+ }
+
+ bus_dmamap_sync(sc->sc_dmat, AQ_DMA_MAP(&tx->tx_mem), 0,
+ AQ_DMA_LEN(&tx->tx_mem), BUS_DMASYNC_PREWRITE);
+
+ if (used != 0) {
+ tx->tx_prod = idx;
+ AQ_WRITE_REG(sc, TX_DMA_DESC_TAIL_PTR_REG(tx->tx_q),
+ tx->tx_prod);
+ }
+}
+
+int
+aq_intr(void *arg)
+{
+ struct aq_softc *sc = arg;
+ struct aq_queues *aq = &sc->sc_queues[0];
+ uint32_t status;
+
+ status = AQ_READ_REG(sc, AQ_INTR_STATUS_REG);
+ AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG, 0xffffffff);
+
+ if (status & (1 << sc->sc_linkstat_irq))
+ aq_update_link_status(sc);
+
+ if (status & (1 << aq->q_tx.tx_irq)) {
+ aq_txeof(sc, &aq->q_tx);
+ AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG,
+ (1 << aq->q_tx.tx_irq));
+ }
+ if (status & (1 << aq->q_rx.rx_irq)) {
+ aq_rxeof(sc, &aq->q_rx);
+ AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG,
+ (1 << aq->q_rx.rx_irq));
+ }
+
+ return 1;
+}
+
+void
+aq_watchdog(struct ifnet *ifp)
+{
+
+}
+
+void
+aq_free_slots(struct aq_softc *sc, struct aq_slot *slots, int allocated,
+ int total)
+{
+ struct aq_slot *as;
+
+ int i = allocated;
+ while (i-- > 0) {
+ as = &slots[i];
+ bus_dmamap_destroy(sc->sc_dmat, as->as_map);
+ if (as->as_m != NULL)
+ m_freem(as->as_m);
+ }
+ free(slots, M_DEVBUF, total * sizeof(*as));
+}
+
+int
+aq_queue_up(struct aq_softc *sc, struct aq_queues *aq)
+{
+ struct aq_rxring *rx;
+ struct aq_txring *tx;
+ struct aq_slot *as;
+ int i;
+
+ rx = &aq->q_rx;
+ rx->rx_slots = mallocarray(sizeof(*as), AQ_RXD_NUM, M_DEVBUF,
+ M_WAITOK | M_ZERO);
+ if (rx->rx_slots == NULL) {
+ printf("%s: failed to allocate rx slots %d\n", DEVNAME(sc),
+ aq->q_index);
+ return ENOMEM;
+ }
+
+ for (i = 0; i < AQ_RXD_NUM; i++) {
+ as = &rx->rx_slots[i];
+ if (bus_dmamap_create(sc->sc_dmat, MCLBYTES, 1, MCLBYTES, 0,
+ BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW | BUS_DMA_64BIT,
+ &as->as_map) != 0) {
+ printf("%s: failed to allocate rx dma maps %d\n",
+ DEVNAME(sc), aq->q_index);
+ goto destroy_rx_slots;
+ }
+ }
+
+ if (aq_dmamem_alloc(sc, &rx->rx_mem, AQ_RXD_NUM *
+ sizeof(struct aq_rx_desc_read), PAGE_SIZE) != 0) {
+ printf("%s: unable to allocate rx ring %d\n", DEVNAME(sc),
+ aq->q_index);
+ goto destroy_rx_slots;
+ }
+
+ tx = &aq->q_tx;
+ tx->tx_slots = mallocarray(sizeof(*as), AQ_TXD_NUM, M_DEVBUF,
+ M_WAITOK | M_ZERO);
+ if (tx->tx_slots == NULL) {
+ printf("%s: failed to allocate tx slots %d\n", DEVNAME(sc),
+ aq->q_index);
+ goto destroy_rx_ring;
+ }
+
+ for (i = 0; i < AQ_TXD_NUM; i++) {
+ as = &tx->tx_slots[i];
+ if (bus_dmamap_create(sc->sc_dmat, MCLBYTES,
+ AQ_TX_MAX_SEGMENTS, MCLBYTES, 0,
+ BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW | BUS_DMA_64BIT,
+ &as->as_map) != 0) {
+ printf("%s: failed to allocated tx dma maps %d\n",
+ DEVNAME(sc), aq->q_index);
+ goto destroy_tx_slots;
+ }
+ }
+
+ if (aq_dmamem_alloc(sc, &tx->tx_mem, AQ_TXD_NUM *
+ sizeof(struct aq_tx_desc), PAGE_SIZE) != 0) {
+ printf("%s: unable to allocate tx ring %d\n", DEVNAME(sc),
+ aq->q_index);
+ goto destroy_tx_slots;
+ }
+
+ aq_txring_reset(sc, tx, 1);
+ aq_rxring_reset(sc, rx, 1);
+ return 0;
+
+destroy_tx_slots:
+ aq_free_slots(sc, tx->tx_slots, i, AQ_TXD_NUM);
+ tx->tx_slots = NULL;
+ i = AQ_RXD_NUM;
+
+destroy_rx_ring:
+ aq_dmamem_free(sc, &rx->rx_mem);
+destroy_rx_slots:
+ aq_free_slots(sc, rx->rx_slots, i, AQ_RXD_NUM);
+ rx->rx_slots = NULL;
+ return ENOMEM;
+}
+
+void
+aq_queue_down(struct aq_softc *sc, struct aq_queues *aq)
+{
+ struct aq_txring *tx;
+ struct aq_rxring *rx;
+
+ tx = &aq->q_tx;
+ aq_txring_reset(sc, &aq->q_tx, 0);
+ if (tx->tx_slots != NULL) {
+ aq_free_slots(sc, tx->tx_slots, AQ_TXD_NUM, AQ_TXD_NUM);
+ tx->tx_slots = NULL;
+ }
+
+ aq_dmamem_free(sc, &tx->tx_mem);
+
+ rx = &aq->q_rx;
+ aq_rxring_reset(sc, &aq->q_rx, 0);
+ if (rx->rx_slots != NULL) {
+ aq_free_slots(sc, rx->rx_slots, AQ_RXD_NUM, AQ_RXD_NUM);
+ rx->rx_slots = NULL;
+ }
+
+ aq_dmamem_free(sc, &rx->rx_mem);
+}
+
+int
+aq_up(struct aq_softc *sc)
+{
+ struct ifnet *ifp = &sc->sc_arpcom.ac_if;
+ int i;
+
+ for (i = 0; i < sc->sc_nqueues; i++) {
+ if (aq_queue_up(sc, &sc->sc_queues[i]) != 0)
+ goto downqueues;
+ }
+
+ /* filters? */
+ /* enable checksum offload */
+
+ SET(ifp->if_flags, IFF_RUNNING);
+ aq_enable_intr(sc, 1, 1);
+ AQ_WRITE_REG_BIT(sc, TPB_TX_BUF_REG, TPB_TX_BUF_EN, 1);
+ AQ_WRITE_REG_BIT(sc, RPB_RPF_RX_REG, RPB_RPF_RX_BUF_EN, 1);
+
+ for (i = 0; i < sc->sc_nqueues; i++) {
+ struct aq_queues *aq = &sc->sc_queues[i];
+
+ if_rxr_init(&aq->q_rx.rx_rxr, 1, AQ_RXD_NUM - 1);
+ aq_rx_fill(sc, &aq->q_rx);
+
+ ifq_clr_oactive(aq->q_tx.tx_ifq);
+ }
+
+ return ENETRESET;
+
+downqueues:
+ for (i = 0; i < sc->sc_nqueues; i++)
+ aq_queue_down(sc, &sc->sc_queues[i]);
+ return ENOMEM;
+}
+
+void
+aq_down(struct aq_softc *sc)
+{
+ struct ifnet *ifp = &sc->sc_arpcom.ac_if;
+ int i;
+
+ CLR(ifp->if_flags, IFF_RUNNING);
+
+ aq_enable_intr(sc, 1, 0);
+ intr_barrier(sc->sc_ih);
+
+ for (i = 0; i < sc->sc_nqueues; i++) {
+ /* queue intr barrier? */
+ aq_queue_down(sc, &sc->sc_queues[i]);
+ }
+}
+
+void
+aq_enable_intr(struct aq_softc *sc, int link, int txrx)
+{
+ uint32_t imask = 0;
+ int i;
+
+ if (txrx) {
+ for (i = 0; i < sc->sc_nqueues; i++) {
+ imask |= (1 << sc->sc_queues[i].q_tx.tx_irq);
+ imask |= (1 << sc->sc_queues[i].q_rx.rx_irq);
+ }
+ }
+
+ if (link)
+ imask |= (1 << sc->sc_linkstat_irq);
+
+ AQ_WRITE_REG(sc, AQ_INTR_MASK_REG, imask);
+ AQ_WRITE_REG(sc, AQ_INTR_STATUS_CLR_REG, 0xffffffff);
+}
+
+void
+aq_ifmedia_status(struct ifnet *ifp, struct ifmediareq *ifmr)
+{
+ struct aq_softc *aq = ifp->if_softc;
+ enum aq_link_speed speed;
+ enum aq_link_fc fc;
+ int media;
+ int flow;
+
+ if (aq_get_linkmode(aq, &speed, &fc, NULL) != 0)
+ return;
+
+ switch (speed) {
+ case AQ_LINK_10G:
+ media = IFM_10G_T;
+ break;
+ case AQ_LINK_5G:
+ media = IFM_5000_T;
+ break;
+ case AQ_LINK_2G5:
+ media = IFM_2500_T;
+ break;
+ case AQ_LINK_1G:
+ media = IFM_1000_T;
+ break;
+ case AQ_LINK_100M:
+ media = IFM_100_TX;
+ break;
+ case AQ_LINK_NONE:
+ media = 0;
+ break;
+ }
+
+ flow = 0;
+ if (fc & AQ_FC_RX)
+ flow |= IFM_ETH_RXPAUSE;
+ if (fc & AQ_FC_TX)
+ flow |= IFM_ETH_TXPAUSE;
+
+ ifmr->ifm_status = IFM_AVALID;
+ if (speed != AQ_LINK_NONE) {
+ ifmr->ifm_status |= IFM_ACTIVE;
+ ifmr->ifm_active = IFM_ETHER | IFM_AUTO | media | flow;
+ }
+}
+
+int
+aq_ifmedia_change(struct ifnet *ifp)
+{
+ struct aq_softc *sc = ifp->if_softc;
+ enum aq_link_speed rate = AQ_LINK_NONE;
+ enum aq_link_fc fc = AQ_FC_NONE;
+
+ if (IFM_TYPE(sc->sc_media.ifm_media) != IFM_ETHER)
+ return EINVAL;
+
+ switch (IFM_SUBTYPE(sc->sc_media.ifm_media)) {
+ case IFM_AUTO:
+ rate = AQ_LINK_AUTO;
+ break;
+ case IFM_NONE:
+ rate = AQ_LINK_NONE;
+ break;
+ case IFM_100_TX:
+ rate = AQ_LINK_100M;
+ break;
+ case IFM_1000_T:
+ rate = AQ_LINK_1G;
+ break;
+ case IFM_2500_T:
+ rate = AQ_LINK_2G5;
+ break;
+ case IFM_5000_T:
+ rate = AQ_LINK_5G;
+ break;
+ case IFM_10G_T:
+ rate = AQ_LINK_10G;
+ break;
+ default:
+ return ENODEV;
+ }
+
+ if (sc->sc_media.ifm_media & IFM_FLOW)
+ fc = AQ_FC_ALL;
+
+ return aq_set_linkmode(sc, rate, fc, AQ_EEE_DISABLE);
+}
+
+void
+aq_update_link_status(struct aq_softc *sc)
+{
+ struct ifnet *ifp = &sc->sc_arpcom.ac_if;
+ enum aq_link_speed speed;
+ enum aq_link_fc fc;
+
+ if (aq_get_linkmode(sc, &speed, &fc, NULL) != 0)
+ return;
+
+ if (speed == AQ_LINK_NONE) {
+ if (ifp->if_link_state != LINK_STATE_DOWN) {
+ ifp->if_link_state = LINK_STATE_DOWN;
+ if_link_state_change(ifp);
+ }
+ } else {
+ if (ifp->if_link_state != LINK_STATE_FULL_DUPLEX) {
+ ifp->if_link_state = LINK_STATE_FULL_DUPLEX;
+ if_link_state_change(ifp);
+ }
+ }
+}
+
+
+int
+aq_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
+{
+ struct aq_softc *sc = ifp->if_softc;
+ struct ifreq *ifr = (struct ifreq *)data;
+ int error = 0, s;
+
+ s = splnet();
+
+ switch (cmd) {
+ case SIOCSIFADDR:
+ ifp->if_flags |= IFF_UP;
+ if ((ifp->if_flags & IFF_RUNNING) == 0)
+ error = aq_up(sc);
+ break;
+ case SIOCSIFFLAGS:
+ if (ifp->if_flags & IFF_UP) {
+ if (ifp->if_flags & IFF_RUNNING)
+ error = ENETRESET;
+ else
+ error = aq_up(sc);
+ } else {
+ if (ifp->if_flags & IFF_RUNNING)
+ aq_down(sc);
+ }
+ break;
+ case SIOCSIFMEDIA:
+ case SIOCGIFMEDIA:
+ error = ifmedia_ioctl(ifp, ifr, &sc->sc_media, cmd);
+ break;
+ default:
+ error = ether_ioctl(ifp, &sc->sc_arpcom, cmd, data);
+ }
+
+ if (error == ENETRESET) {
+ if (ifp->if_flags & IFF_RUNNING)
+ aq_iff(sc);
+ error = 0;
+ }
+
+ splx(s);
+ return error;
+}
+
+void
+aq_iff(struct aq_softc *sc)
+{
+ struct ifnet *ifp = &sc->sc_arpcom.ac_if;
+ struct arpcom *ac = &sc->sc_arpcom;
+ struct ether_multi *enm;
+ struct ether_multistep step;
+ int idx;
+
+ if (ifp->if_flags & IFF_PROMISC || ac->ac_multirangecnt > 0) {
+ ifp->if_flags |= IFF_ALLMULTI;
+ AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 1);
+ } else if (ac->ac_multicnt >= AQ_HW_MAC_NUM ||
+ ISSET(ifp->if_flags, IFF_ALLMULTI)) {
+ AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 0);
+ AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_MASK_REG,
+ RPF_MCAST_FILTER_MASK_ALLMULTI, 1);
+ AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_REG(0),
+ RPF_MCAST_FILTER_EN, 1);
+ } else if (ac->ac_multicnt == 0) {
+ AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 0);
+ AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_REG(0),
+ RPF_MCAST_FILTER_EN, 0);
+ } else {
+ idx = AQ_HW_MAC_OWN + 1;
+
+ /* turn on allmulti while we're rewriting? */
+ AQ_WRITE_REG_BIT(sc, RPF_L2BC_REG, RPF_L2BC_PROMISC, 0);
+
+ ETHER_FIRST_MULTI(step, ac, enm);
+ while (enm != NULL) {
+ aq_set_mac_addr(sc, idx++, enm->enm_addrlo);
+ ETHER_NEXT_MULTI(step, enm);
+ }
+
+ for (; idx < AQ_HW_MAC_NUM; idx++)
+ aq_set_mac_addr(sc, idx, NULL);
+
+ AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_MASK_REG,
+ RPF_MCAST_FILTER_MASK_ALLMULTI, 0);
+ AQ_WRITE_REG_BIT(sc, RPF_MCAST_FILTER_REG(0),
+ RPF_MCAST_FILTER_EN, 1);
+ }
+}
+
+int
+aq_dmamem_alloc(struct aq_softc *sc, struct aq_dmamem *aqm,
+ bus_size_t size, u_int align)
+{
+ aqm->aqm_size = size;
+
+ if (bus_dmamap_create(sc->sc_dmat, aqm->aqm_size, 1,
+ aqm->aqm_size, 0,
+ BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW | BUS_DMA_64BIT,
+ &aqm->aqm_map) != 0)
+ return (1);
+ if (bus_dmamem_alloc(sc->sc_dmat, aqm->aqm_size,
+ align, 0, &aqm->aqm_seg, 1, &aqm->aqm_nsegs,
+ BUS_DMA_WAITOK | BUS_DMA_ZERO) != 0)
+ goto destroy;
+ if (bus_dmamem_map(sc->sc_dmat, &aqm->aqm_seg, aqm->aqm_nsegs,
+ aqm->aqm_size, &aqm->aqm_kva, BUS_DMA_WAITOK) != 0)
+ goto free;
+ if (bus_dmamap_load(sc->sc_dmat, aqm->aqm_map, aqm->aqm_kva,
+ aqm->aqm_size, NULL, BUS_DMA_WAITOK) != 0)
+ goto unmap;
+
+ return (0);
+unmap:
+ bus_dmamem_unmap(sc->sc_dmat, aqm->aqm_kva, aqm->aqm_size);
+free:
+ bus_dmamem_free(sc->sc_dmat, &aqm->aqm_seg, 1);
+destroy:
+ bus_dmamap_destroy(sc->sc_dmat, aqm->aqm_map);
+ return (1);
+}
+
+void
+aq_dmamem_free(struct aq_softc *sc, struct aq_dmamem *aqm)
+{
+ bus_dmamap_unload(sc->sc_dmat, aqm->aqm_map);
+ bus_dmamem_unmap(sc->sc_dmat, aqm->aqm_kva, aqm->aqm_size);
+ bus_dmamem_free(sc->sc_dmat, &aqm->aqm_seg, 1);
+ bus_dmamap_destroy(sc->sc_dmat, aqm->aqm_map);
+}