#include "bpfilter.h"
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/mbuf.h>
#include <sys/socket.h>
#include <sys/ioctl.h>
#include <sys/errno.h>
#include <sys/malloc.h>
#include <sys/kernel.h>
#include <sys/device.h>
#include <net/if.h>
#include <net/if_media.h>
#include <netinet/in.h>
#include <netinet/if_ether.h>
#include <dev/mii/miivar.h>
#include <machine/bus.h>
#if NBPFILTER > 0
#include <net/bpf.h>
#endif
#include <dev/ic/xlreg.h>
#define XL905B_TXCSUM_BROKEN 1
int xl_newbuf(struct xl_softc *, struct xl_chain_onefrag *);
void xl_stats_update(void *);
int xl_encap(struct xl_softc *, struct xl_chain *,
struct mbuf * );
void xl_rxeof(struct xl_softc *);
void xl_txeof(struct xl_softc *);
void xl_txeof_90xB(struct xl_softc *);
void xl_txeoc(struct xl_softc *);
int xl_intr(void *);
void xl_start(struct ifnet *);
void xl_start_90xB(struct ifnet *);
int xl_ioctl(struct ifnet *, u_long, caddr_t);
void xl_freetxrx(struct xl_softc *);
void xl_watchdog(struct ifnet *);
int xl_ifmedia_upd(struct ifnet *);
void xl_ifmedia_sts(struct ifnet *, struct ifmediareq *);
int xl_eeprom_wait(struct xl_softc *);
int xl_read_eeprom(struct xl_softc *, caddr_t, int, int, int);
void xl_mii_sync(struct xl_softc *);
void xl_mii_send(struct xl_softc *, u_int32_t, int);
int xl_mii_readreg(struct xl_softc *, struct xl_mii_frame *);
int xl_mii_writereg(struct xl_softc *, struct xl_mii_frame *);
void xl_setcfg(struct xl_softc *);
void xl_setmode(struct xl_softc *, uint64_t);
void xl_iff(struct xl_softc *);
void xl_iff_90x(struct xl_softc *);
void xl_iff_905b(struct xl_softc *);
int xl_list_rx_init(struct xl_softc *);
void xl_fill_rx_ring(struct xl_softc *);
int xl_list_tx_init(struct xl_softc *);
int xl_list_tx_init_90xB(struct xl_softc *);
void xl_wait(struct xl_softc *);
void xl_mediacheck(struct xl_softc *);
void xl_choose_xcvr(struct xl_softc *, int);
int xl_miibus_readreg(struct device *, int, int);
void xl_miibus_writereg(struct device *, int, int, int);
void xl_miibus_statchg(struct device *);
#ifndef SMALL_KERNEL
int xl_wol(struct ifnet *, int);
void xl_wol_power(struct xl_softc *);
#endif
int
xl_activate(struct device *self, int act)
{
struct xl_softc *sc = (struct xl_softc *)self;
struct ifnet *ifp = &sc->sc_arpcom.ac_if;
switch (act) {
case DVACT_SUSPEND:
if (ifp->if_flags & IFF_RUNNING)
xl_stop(sc);
break;
case DVACT_RESUME:
if (ifp->if_flags & IFF_UP)
xl_init(sc);
break;
case DVACT_POWERDOWN:
#ifndef SMALL_KERNEL
xl_wol_power(sc);
#endif
break;
}
return (0);
}
void
xl_wait(struct xl_softc *sc)
{
int i;
for (i = 0; i < XL_TIMEOUT; i++) {
if (!(CSR_READ_2(sc, XL_STATUS) & XL_STAT_CMDBUSY))
break;
}
if (i == XL_TIMEOUT)
printf("%s: command never completed!\n", sc->sc_dev.dv_xname);
}
#define MII_SET(x) \
CSR_WRITE_2(sc, XL_W4_PHY_MGMT, \
CSR_READ_2(sc, XL_W4_PHY_MGMT) | (x))
#define MII_CLR(x) \
CSR_WRITE_2(sc, XL_W4_PHY_MGMT, \
CSR_READ_2(sc, XL_W4_PHY_MGMT) & ~(x))
void
xl_mii_sync(struct xl_softc *sc)
{
int i;
XL_SEL_WIN(4);
MII_SET(XL_MII_DIR|XL_MII_DATA);
for (i = 0; i < 32; i++) {
MII_SET(XL_MII_CLK);
MII_SET(XL_MII_DATA);
MII_SET(XL_MII_DATA);
MII_CLR(XL_MII_CLK);
MII_SET(XL_MII_DATA);
MII_SET(XL_MII_DATA);
}
}
void
xl_mii_send(struct xl_softc *sc, u_int32_t bits, int cnt)
{
int i;
XL_SEL_WIN(4);
MII_CLR(XL_MII_CLK);
for (i = (0x1 << (cnt - 1)); i; i >>= 1) {
if (bits & i) {
MII_SET(XL_MII_DATA);
} else {
MII_CLR(XL_MII_DATA);
}
MII_CLR(XL_MII_CLK);
MII_SET(XL_MII_CLK);
}
}
int
xl_mii_readreg(struct xl_softc *sc, struct xl_mii_frame *frame)
{
int i, ack, s;
s = splnet();
frame->mii_stdelim = XL_MII_STARTDELIM;
frame->mii_opcode = XL_MII_READOP;
frame->mii_turnaround = 0;
frame->mii_data = 0;
XL_SEL_WIN(4);
CSR_WRITE_2(sc, XL_W4_PHY_MGMT, 0);
MII_SET(XL_MII_DIR);
xl_mii_sync(sc);
xl_mii_send(sc, frame->mii_stdelim, 2);
xl_mii_send(sc, frame->mii_opcode, 2);
xl_mii_send(sc, frame->mii_phyaddr, 5);
xl_mii_send(sc, frame->mii_regaddr, 5);
MII_CLR((XL_MII_CLK|XL_MII_DATA));
MII_SET(XL_MII_CLK);
MII_CLR(XL_MII_DIR);
MII_CLR(XL_MII_CLK);
ack = CSR_READ_2(sc, XL_W4_PHY_MGMT) & XL_MII_DATA;
MII_SET(XL_MII_CLK);
if (ack) {
for(i = 0; i < 16; i++) {
MII_CLR(XL_MII_CLK);
MII_SET(XL_MII_CLK);
}
goto fail;
}
for (i = 0x8000; i; i >>= 1) {
MII_CLR(XL_MII_CLK);
if (!ack) {
if (CSR_READ_2(sc, XL_W4_PHY_MGMT) & XL_MII_DATA)
frame->mii_data |= i;
}
MII_SET(XL_MII_CLK);
}
fail:
MII_CLR(XL_MII_CLK);
MII_SET(XL_MII_CLK);
splx(s);
if (ack)
return (1);
return (0);
}
int
xl_mii_writereg(struct xl_softc *sc, struct xl_mii_frame *frame)
{
int s;
s = splnet();
frame->mii_stdelim = XL_MII_STARTDELIM;
frame->mii_opcode = XL_MII_WRITEOP;
frame->mii_turnaround = XL_MII_TURNAROUND;
XL_SEL_WIN(4);
MII_SET(XL_MII_DIR);
xl_mii_sync(sc);
xl_mii_send(sc, frame->mii_stdelim, 2);
xl_mii_send(sc, frame->mii_opcode, 2);
xl_mii_send(sc, frame->mii_phyaddr, 5);
xl_mii_send(sc, frame->mii_regaddr, 5);
xl_mii_send(sc, frame->mii_turnaround, 2);
xl_mii_send(sc, frame->mii_data, 16);
MII_SET(XL_MII_CLK);
MII_CLR(XL_MII_CLK);
MII_CLR(XL_MII_DIR);
splx(s);
return (0);
}
int
xl_miibus_readreg(struct device *self, int phy, int reg)
{
struct xl_softc *sc = (struct xl_softc *)self;
struct xl_mii_frame frame;
if (!(sc->xl_flags & XL_FLAG_PHYOK) && phy != 24)
return (0);
bzero(&frame, sizeof(frame));
frame.mii_phyaddr = phy;
frame.mii_regaddr = reg;
xl_mii_readreg(sc, &frame);
return (frame.mii_data);
}
void
xl_miibus_writereg(struct device *self, int phy, int reg, int data)
{
struct xl_softc *sc = (struct xl_softc *)self;
struct xl_mii_frame frame;
if (!(sc->xl_flags & XL_FLAG_PHYOK) && phy != 24)
return;
bzero(&frame, sizeof(frame));
frame.mii_phyaddr = phy;
frame.mii_regaddr = reg;
frame.mii_data = data;
xl_mii_writereg(sc, &frame);
}
void
xl_miibus_statchg(struct device *self)
{
struct xl_softc *sc = (struct xl_softc *)self;
xl_setcfg(sc);
XL_SEL_WIN(3);
if ((sc->sc_mii.mii_media_active & IFM_GMASK) == IFM_FDX)
CSR_WRITE_1(sc, XL_W3_MAC_CTRL, XL_MACCTRL_DUPLEX);
else
CSR_WRITE_1(sc, XL_W3_MAC_CTRL,
(CSR_READ_1(sc, XL_W3_MAC_CTRL) & ~XL_MACCTRL_DUPLEX));
}
int
xl_eeprom_wait(struct xl_softc *sc)
{
int i;
for (i = 0; i < 100; i++) {
if (CSR_READ_2(sc, XL_W0_EE_CMD) & XL_EE_BUSY)
DELAY(162);
else
break;
}
if (i == 100) {
printf("%s: eeprom failed to come ready\n", sc->sc_dev.dv_xname);
return (1);
}
return (0);
}
int
xl_read_eeprom(struct xl_softc *sc, caddr_t dest, int off, int cnt, int swap)
{
int err = 0, i;
u_int16_t word = 0, *ptr;
#define EEPROM_5BIT_OFFSET(A) ((((A) << 2) & 0x7F00) | ((A) & 0x003F))
#define EEPROM_8BIT_OFFSET(A) ((A) & 0x003F)
XL_SEL_WIN(0);
if (xl_eeprom_wait(sc))
return (1);
if (sc->xl_flags & XL_FLAG_EEPROM_OFFSET_30)
off += 0x30;
for (i = 0; i < cnt; i++) {
if (sc->xl_flags & XL_FLAG_8BITROM)
CSR_WRITE_2(sc, XL_W0_EE_CMD,
XL_EE_8BIT_READ | EEPROM_8BIT_OFFSET(off + i));
else
CSR_WRITE_2(sc, XL_W0_EE_CMD,
XL_EE_READ | EEPROM_5BIT_OFFSET(off + i));
err = xl_eeprom_wait(sc);
if (err)
break;
word = CSR_READ_2(sc, XL_W0_EE_DATA);
ptr = (u_int16_t *)(dest + (i * 2));
if (swap)
*ptr = ntohs(word);
else
*ptr = word;
}
return (err ? 1 : 0);
}
void
xl_iff(struct xl_softc *sc)
{
if (sc->xl_type == XL_TYPE_905B)
xl_iff_905b(sc);
else
xl_iff_90x(sc);
}
void
xl_iff_90x(struct xl_softc *sc)
{
struct ifnet *ifp = &sc->sc_arpcom.ac_if;
struct arpcom *ac = &sc->sc_arpcom;
u_int8_t rxfilt;
XL_SEL_WIN(5);
rxfilt = CSR_READ_1(sc, XL_W5_RX_FILTER);
rxfilt &= ~(XL_RXFILTER_ALLFRAMES | XL_RXFILTER_ALLMULTI |
XL_RXFILTER_BROADCAST | XL_RXFILTER_INDIVIDUAL);
ifp->if_flags &= ~IFF_ALLMULTI;
rxfilt |= XL_RXFILTER_BROADCAST | XL_RXFILTER_INDIVIDUAL;
if (ifp->if_flags & IFF_PROMISC || ac->ac_multicnt > 0) {
ifp->if_flags |= IFF_ALLMULTI;
if (ifp->if_flags & IFF_PROMISC)
rxfilt |= XL_RXFILTER_ALLFRAMES;
else
rxfilt |= XL_RXFILTER_ALLMULTI;
}
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_SET_FILT | rxfilt);
XL_SEL_WIN(7);
}
void
xl_iff_905b(struct xl_softc *sc)
{
struct ifnet *ifp = &sc->sc_arpcom.ac_if;
struct arpcom *ac = &sc->sc_arpcom;
int h = 0, i;
struct ether_multi *enm;
struct ether_multistep step;
u_int8_t rxfilt;
XL_SEL_WIN(5);
rxfilt = CSR_READ_1(sc, XL_W5_RX_FILTER);
rxfilt &= ~(XL_RXFILTER_ALLFRAMES | XL_RXFILTER_ALLMULTI |
XL_RXFILTER_BROADCAST | XL_RXFILTER_INDIVIDUAL |
XL_RXFILTER_MULTIHASH);
ifp->if_flags &= ~IFF_ALLMULTI;
rxfilt |= XL_RXFILTER_BROADCAST | XL_RXFILTER_INDIVIDUAL;
if (ifp->if_flags & IFF_PROMISC || ac->ac_multirangecnt > 0) {
ifp->if_flags |= IFF_ALLMULTI;
if (ifp->if_flags & IFF_PROMISC)
rxfilt |= XL_RXFILTER_ALLFRAMES;
else
rxfilt |= XL_RXFILTER_ALLMULTI;
} else {
rxfilt |= XL_RXFILTER_MULTIHASH;
for (i = 0; i < XL_HASHFILT_SIZE; i++)
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_SET_HASH|i);
ETHER_FIRST_MULTI(step, ac, enm);
while (enm != NULL) {
h = ether_crc32_be(enm->enm_addrlo, ETHER_ADDR_LEN) &
0x000000FF;
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_SET_HASH |
XL_HASH_SET | h);
ETHER_NEXT_MULTI(step, enm);
}
}
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_SET_FILT | rxfilt);
XL_SEL_WIN(7);
}
void
xl_setcfg(struct xl_softc *sc)
{
u_int32_t icfg;
XL_SEL_WIN(3);
icfg = CSR_READ_4(sc, XL_W3_INTERNAL_CFG);
icfg &= ~XL_ICFG_CONNECTOR_MASK;
if (sc->xl_media & XL_MEDIAOPT_MII ||
sc->xl_media & XL_MEDIAOPT_BT4)
icfg |= (XL_XCVR_MII << XL_ICFG_CONNECTOR_BITS);
if (sc->xl_media & XL_MEDIAOPT_BTX)
icfg |= (XL_XCVR_AUTO << XL_ICFG_CONNECTOR_BITS);
CSR_WRITE_4(sc, XL_W3_INTERNAL_CFG, icfg);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_COAX_STOP);
}
void
xl_setmode(struct xl_softc *sc, uint64_t media)
{
struct ifnet *ifp = &sc->sc_arpcom.ac_if;
u_int32_t icfg;
u_int16_t mediastat;
XL_SEL_WIN(4);
mediastat = CSR_READ_2(sc, XL_W4_MEDIA_STATUS);
XL_SEL_WIN(3);
icfg = CSR_READ_4(sc, XL_W3_INTERNAL_CFG);
if (sc->xl_media & XL_MEDIAOPT_BT) {
if (IFM_SUBTYPE(media) == IFM_10_T) {
ifp->if_baudrate = IF_Mbps(10);
sc->xl_xcvr = XL_XCVR_10BT;
icfg &= ~XL_ICFG_CONNECTOR_MASK;
icfg |= (XL_XCVR_10BT << XL_ICFG_CONNECTOR_BITS);
mediastat |= XL_MEDIASTAT_LINKBEAT|
XL_MEDIASTAT_JABGUARD;
mediastat &= ~XL_MEDIASTAT_SQEENB;
}
}
if (sc->xl_media & XL_MEDIAOPT_BFX) {
if (IFM_SUBTYPE(media) == IFM_100_FX) {
ifp->if_baudrate = IF_Mbps(100);
sc->xl_xcvr = XL_XCVR_100BFX;
icfg &= ~XL_ICFG_CONNECTOR_MASK;
icfg |= (XL_XCVR_100BFX << XL_ICFG_CONNECTOR_BITS);
mediastat |= XL_MEDIASTAT_LINKBEAT;
mediastat &= ~XL_MEDIASTAT_SQEENB;
}
}
if (sc->xl_media & (XL_MEDIAOPT_AUI|XL_MEDIAOPT_10FL)) {
if (IFM_SUBTYPE(media) == IFM_10_5) {
ifp->if_baudrate = IF_Mbps(10);
sc->xl_xcvr = XL_XCVR_AUI;
icfg &= ~XL_ICFG_CONNECTOR_MASK;
icfg |= (XL_XCVR_AUI << XL_ICFG_CONNECTOR_BITS);
mediastat &= ~(XL_MEDIASTAT_LINKBEAT|
XL_MEDIASTAT_JABGUARD);
mediastat |= ~XL_MEDIASTAT_SQEENB;
}
if (IFM_SUBTYPE(media) == IFM_10_FL) {
ifp->if_baudrate = IF_Mbps(10);
sc->xl_xcvr = XL_XCVR_AUI;
icfg &= ~XL_ICFG_CONNECTOR_MASK;
icfg |= (XL_XCVR_AUI << XL_ICFG_CONNECTOR_BITS);
mediastat &= ~(XL_MEDIASTAT_LINKBEAT|
XL_MEDIASTAT_JABGUARD);
mediastat |= ~XL_MEDIASTAT_SQEENB;
}
}
if (sc->xl_media & XL_MEDIAOPT_BNC) {
if (IFM_SUBTYPE(media) == IFM_10_2) {
ifp->if_baudrate = IF_Mbps(10);
sc->xl_xcvr = XL_XCVR_COAX;
icfg &= ~XL_ICFG_CONNECTOR_MASK;
icfg |= (XL_XCVR_COAX << XL_ICFG_CONNECTOR_BITS);
mediastat &= ~(XL_MEDIASTAT_LINKBEAT|
XL_MEDIASTAT_JABGUARD|
XL_MEDIASTAT_SQEENB);
}
}
if ((media & IFM_GMASK) == IFM_FDX ||
IFM_SUBTYPE(media) == IFM_100_FX) {
XL_SEL_WIN(3);
CSR_WRITE_1(sc, XL_W3_MAC_CTRL, XL_MACCTRL_DUPLEX);
} else {
XL_SEL_WIN(3);
CSR_WRITE_1(sc, XL_W3_MAC_CTRL,
(CSR_READ_1(sc, XL_W3_MAC_CTRL) & ~XL_MACCTRL_DUPLEX));
}
if (IFM_SUBTYPE(media) == IFM_10_2)
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_COAX_START);
else
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_COAX_STOP);
CSR_WRITE_4(sc, XL_W3_INTERNAL_CFG, icfg);
XL_SEL_WIN(4);
CSR_WRITE_2(sc, XL_W4_MEDIA_STATUS, mediastat);
DELAY(800);
XL_SEL_WIN(7);
}
void
xl_reset(struct xl_softc *sc)
{
int i;
XL_SEL_WIN(0);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RESET |
((sc->xl_flags & XL_FLAG_WEIRDRESET) ?
XL_RESETOPT_DISADVFD:0));
DELAY(100000);
for (i = 0; i < XL_TIMEOUT; i++) {
DELAY(10);
if (!(CSR_READ_2(sc, XL_STATUS) & XL_STAT_CMDBUSY))
break;
}
if (i == XL_TIMEOUT)
printf("%s: reset didn't complete\n", sc->sc_dev.dv_xname);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_RESET);
DELAY(100000);
xl_wait(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_RESET);
xl_wait(sc);
if (sc->xl_flags & XL_FLAG_INVERT_LED_PWR ||
sc->xl_flags & XL_FLAG_INVERT_MII_PWR) {
XL_SEL_WIN(2);
CSR_WRITE_2(sc, XL_W2_RESET_OPTIONS, CSR_READ_2(sc,
XL_W2_RESET_OPTIONS)
| ((sc->xl_flags & XL_FLAG_INVERT_LED_PWR)?XL_RESETOPT_INVERT_LED:0)
| ((sc->xl_flags & XL_FLAG_INVERT_MII_PWR)?XL_RESETOPT_INVERT_MII:0)
);
}
DELAY(100000);
}
void
xl_mediacheck(struct xl_softc *sc)
{
if (sc->xl_media & (XL_MEDIAOPT_MASK & ~XL_MEDIAOPT_VCO)) {
if (sc->xl_xcvr <= XL_XCVR_AUTO)
return;
else {
printf("%s: bogus xcvr value "
"in EEPROM (%x)\n", sc->sc_dev.dv_xname, sc->xl_xcvr);
printf("%s: choosing new default based "
"on card type\n", sc->sc_dev.dv_xname);
}
} else {
if (sc->xl_type == XL_TYPE_905B &&
sc->xl_media & XL_MEDIAOPT_10FL)
return;
printf("%s: WARNING: no media options bits set in "
"the media options register!!\n", sc->sc_dev.dv_xname);
printf("%s: this could be a manufacturing defect in "
"your adapter or system\n", sc->sc_dev.dv_xname);
printf("%s: attempting to guess media type; you "
"should probably consult your vendor\n", sc->sc_dev.dv_xname);
}
xl_choose_xcvr(sc, 1);
}
void
xl_choose_xcvr(struct xl_softc *sc, int verbose)
{
u_int16_t devid;
xl_read_eeprom(sc, (caddr_t)&devid, XL_EE_PRODID, 1, 0);
switch(devid) {
case TC_DEVICEID_BOOMERANG_10BT:
case TC_DEVICEID_KRAKATOA_10BT:
sc->xl_media = XL_MEDIAOPT_BT;
sc->xl_xcvr = XL_XCVR_10BT;
if (verbose)
printf("%s: guessing 10BaseT transceiver\n",
sc->sc_dev.dv_xname);
break;
case TC_DEVICEID_BOOMERANG_10BT_COMBO:
case TC_DEVICEID_KRAKATOA_10BT_COMBO:
sc->xl_media = XL_MEDIAOPT_BT|XL_MEDIAOPT_BNC|XL_MEDIAOPT_AUI;
sc->xl_xcvr = XL_XCVR_10BT;
if (verbose)
printf("%s: guessing COMBO (AUI/BNC/TP)\n",
sc->sc_dev.dv_xname);
break;
case TC_DEVICEID_KRAKATOA_10BT_TPC:
sc->xl_media = XL_MEDIAOPT_BT|XL_MEDIAOPT_BNC;
sc->xl_xcvr = XL_XCVR_10BT;
if (verbose)
printf("%s: guessing TPC (BNC/TP)\n", sc->sc_dev.dv_xname);
break;
case TC_DEVICEID_CYCLONE_10FL:
sc->xl_media = XL_MEDIAOPT_10FL;
sc->xl_xcvr = XL_XCVR_AUI;
if (verbose)
printf("%s: guessing 10baseFL\n", sc->sc_dev.dv_xname);
break;
case TC_DEVICEID_BOOMERANG_10_100BT:
case TC_DEVICEID_HURRICANE_555:
case TC_DEVICEID_HURRICANE_556:
case TC_DEVICEID_HURRICANE_556B:
case TC_DEVICEID_HURRICANE_575A:
case TC_DEVICEID_HURRICANE_575B:
case TC_DEVICEID_HURRICANE_575C:
case TC_DEVICEID_HURRICANE_656:
case TC_DEVICEID_HURRICANE_656B:
case TC_DEVICEID_TORNADO_656C:
case TC_DEVICEID_TORNADO_10_100BT_920B:
sc->xl_media = XL_MEDIAOPT_MII;
sc->xl_xcvr = XL_XCVR_MII;
if (verbose)
printf("%s: guessing MII\n", sc->sc_dev.dv_xname);
break;
case TC_DEVICEID_BOOMERANG_100BT4:
case TC_DEVICEID_CYCLONE_10_100BT4:
sc->xl_media = XL_MEDIAOPT_BT4;
sc->xl_xcvr = XL_XCVR_MII;
if (verbose)
printf("%s: guessing 100BaseT4/MII\n", sc->sc_dev.dv_xname);
break;
case TC_DEVICEID_HURRICANE_10_100BT:
case TC_DEVICEID_HURRICANE_10_100BT_SERV:
case TC_DEVICEID_TORNADO_10_100BT_SERV:
case TC_DEVICEID_HURRICANE_SOHO100TX:
case TC_DEVICEID_TORNADO_10_100BT:
case TC_DEVICEID_TORNADO_HOMECONNECT:
sc->xl_media = XL_MEDIAOPT_BTX;
sc->xl_xcvr = XL_XCVR_AUTO;
if (verbose)
printf("%s: guessing 10/100 internal\n",
sc->sc_dev.dv_xname);
break;
case TC_DEVICEID_CYCLONE_10_100_COMBO:
sc->xl_media = XL_MEDIAOPT_BTX|XL_MEDIAOPT_BNC|XL_MEDIAOPT_AUI;
sc->xl_xcvr = XL_XCVR_AUTO;
if (verbose)
printf("%s: guessing 10/100 plus BNC/AUI\n",
sc->sc_dev.dv_xname);
break;
default:
printf("%s: unknown device ID: %x -- "
"defaulting to 10baseT\n", sc->sc_dev.dv_xname, devid);
sc->xl_media = XL_MEDIAOPT_BT;
break;
}
}
int
xl_list_tx_init(struct xl_softc *sc)
{
struct xl_chain_data *cd;
struct xl_list_data *ld;
int i;
cd = &sc->xl_cdata;
ld = sc->xl_ldata;
for (i = 0; i < XL_TX_LIST_CNT; i++) {
cd->xl_tx_chain[i].xl_ptr = &ld->xl_tx_list[i];
if (i == (XL_TX_LIST_CNT - 1))
cd->xl_tx_chain[i].xl_next = NULL;
else
cd->xl_tx_chain[i].xl_next = &cd->xl_tx_chain[i + 1];
}
cd->xl_tx_free = &cd->xl_tx_chain[0];
cd->xl_tx_tail = cd->xl_tx_head = NULL;
return (0);
}
int
xl_list_tx_init_90xB(struct xl_softc *sc)
{
struct xl_chain_data *cd;
struct xl_list_data *ld;
int i, next, prev;
cd = &sc->xl_cdata;
ld = sc->xl_ldata;
for (i = 0; i < XL_TX_LIST_CNT; i++) {
if (i == (XL_TX_LIST_CNT - 1))
next = 0;
else
next = i + 1;
if (i == 0)
prev = XL_TX_LIST_CNT - 1;
else
prev = i - 1;
cd->xl_tx_chain[i].xl_ptr = &ld->xl_tx_list[i];
cd->xl_tx_chain[i].xl_phys =
sc->sc_listmap->dm_segs[0].ds_addr +
offsetof(struct xl_list_data, xl_tx_list[i]);
cd->xl_tx_chain[i].xl_next = &cd->xl_tx_chain[next];
cd->xl_tx_chain[i].xl_prev = &cd->xl_tx_chain[prev];
}
bzero(ld->xl_tx_list, sizeof(struct xl_list) * XL_TX_LIST_CNT);
ld->xl_tx_list[0].xl_status = htole32(XL_TXSTAT_EMPTY);
cd->xl_tx_prod = 1;
cd->xl_tx_cons = 1;
cd->xl_tx_cnt = 0;
return (0);
}
int
xl_list_rx_init(struct xl_softc *sc)
{
struct xl_chain_data *cd;
struct xl_list_data *ld;
int i, n;
bus_addr_t next;
cd = &sc->xl_cdata;
ld = sc->xl_ldata;
for (i = 0; i < XL_RX_LIST_CNT; i++) {
cd->xl_rx_chain[i].xl_ptr =
(struct xl_list_onefrag *)&ld->xl_rx_list[i];
if (i == (XL_RX_LIST_CNT - 1))
n = 0;
else
n = i + 1;
cd->xl_rx_chain[i].xl_next = &cd->xl_rx_chain[n];
next = sc->sc_listmap->dm_segs[0].ds_addr +
offsetof(struct xl_list_data, xl_rx_list[n]);
ld->xl_rx_list[i].xl_next = htole32(next);
}
cd->xl_rx_prod = cd->xl_rx_cons = &cd->xl_rx_chain[0];
if_rxr_init(&cd->xl_rx_ring, 2, XL_RX_LIST_CNT - 1);
xl_fill_rx_ring(sc);
return (0);
}
void
xl_fill_rx_ring(struct xl_softc *sc)
{
struct xl_chain_data *cd;
u_int slots;
cd = &sc->xl_cdata;
for (slots = if_rxr_get(&cd->xl_rx_ring, XL_RX_LIST_CNT);
slots > 0; slots--) {
if (xl_newbuf(sc, cd->xl_rx_prod) == ENOBUFS)
break;
cd->xl_rx_prod = cd->xl_rx_prod->xl_next;
}
if_rxr_put(&cd->xl_rx_ring, slots);
}
int
xl_newbuf(struct xl_softc *sc, struct xl_chain_onefrag *c)
{
struct mbuf *m_new = NULL;
bus_dmamap_t map;
m_new = MCLGETL(NULL, M_DONTWAIT, MCLBYTES);
if (!m_new)
return (ENOBUFS);
m_new->m_len = m_new->m_pkthdr.len = MCLBYTES;
if (bus_dmamap_load(sc->sc_dmat, sc->sc_rx_sparemap,
mtod(m_new, caddr_t), MCLBYTES, NULL, BUS_DMA_NOWAIT) != 0) {
m_freem(m_new);
return (ENOBUFS);
}
if (c->map->dm_nsegs != 0) {
bus_dmamap_sync(sc->sc_dmat, c->map,
0, c->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
bus_dmamap_unload(sc->sc_dmat, c->map);
}
map = c->map;
c->map = sc->sc_rx_sparemap;
sc->sc_rx_sparemap = map;
m_adj(m_new, ETHER_ALIGN);
bus_dmamap_sync(sc->sc_dmat, c->map, 0, c->map->dm_mapsize,
BUS_DMASYNC_PREREAD);
c->xl_mbuf = m_new;
c->xl_ptr->xl_frag.xl_addr =
htole32(c->map->dm_segs[0].ds_addr + ETHER_ALIGN);
c->xl_ptr->xl_frag.xl_len =
htole32(c->map->dm_segs[0].ds_len | XL_LAST_FRAG);
c->xl_ptr->xl_status = htole32(0);
bus_dmamap_sync(sc->sc_dmat, sc->sc_listmap,
((caddr_t)c->xl_ptr - sc->sc_listkva), sizeof(struct xl_list),
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
return (0);
}
void
xl_rxeof(struct xl_softc *sc)
{
struct mbuf_list ml = MBUF_LIST_INITIALIZER();
struct mbuf *m;
struct ifnet *ifp;
struct xl_chain_onefrag *cur_rx;
int total_len = 0;
u_int32_t rxstat;
u_int16_t sumflags = 0;
ifp = &sc->sc_arpcom.ac_if;
again:
while (if_rxr_inuse(&sc->xl_cdata.xl_rx_ring) > 0) {
cur_rx = sc->xl_cdata.xl_rx_cons;
bus_dmamap_sync(sc->sc_dmat, sc->sc_listmap,
((caddr_t)cur_rx->xl_ptr - sc->sc_listkva),
sizeof(struct xl_list),
BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
if ((rxstat = letoh32(sc->xl_cdata.xl_rx_cons->xl_ptr->xl_status)) == 0)
break;
m = cur_rx->xl_mbuf;
cur_rx->xl_mbuf = NULL;
sc->xl_cdata.xl_rx_cons = cur_rx->xl_next;
if_rxr_put(&sc->xl_cdata.xl_rx_ring, 1);
total_len = rxstat & XL_RXSTAT_LENMASK;
if (total_len > XL_MAX_FRAMELEN)
rxstat |= (XL_RXSTAT_UP_ERROR|XL_RXSTAT_OVERSIZE);
if (rxstat & XL_RXSTAT_UP_ERROR) {
ifp->if_ierrors++;
cur_rx->xl_ptr->xl_status = htole32(0);
m_freem(m);
continue;
}
if (!(rxstat & XL_RXSTAT_UP_CMPLT)) {
printf("%s: bad receive status -- "
"packet dropped\n", sc->sc_dev.dv_xname);
ifp->if_ierrors++;
cur_rx->xl_ptr->xl_status = htole32(0);
m_freem(m);
continue;
}
m->m_pkthdr.len = m->m_len = total_len;
if (sc->xl_type == XL_TYPE_905B) {
if (!(rxstat & XL_RXSTAT_IPCKERR) &&
(rxstat & XL_RXSTAT_IPCKOK))
sumflags |= M_IPV4_CSUM_IN_OK;
if (!(rxstat & XL_RXSTAT_TCPCKERR) &&
(rxstat & XL_RXSTAT_TCPCKOK))
sumflags |= M_TCP_CSUM_IN_OK;
if (!(rxstat & XL_RXSTAT_UDPCKERR) &&
(rxstat & XL_RXSTAT_UDPCKOK))
sumflags |= M_UDP_CSUM_IN_OK;
m->m_pkthdr.csum_flags = sumflags;
}
ml_enqueue(&ml, m);
}
if (ifiq_input(&ifp->if_rcv, &ml))
if_rxr_livelocked(&sc->xl_cdata.xl_rx_ring);
xl_fill_rx_ring(sc);
if (CSR_READ_4(sc, XL_UPLIST_PTR) == 0 ||
CSR_READ_4(sc, XL_UPLIST_STATUS) & XL_PKTSTAT_UP_STALLED) {
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_UP_STALL);
xl_wait(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_UP_UNSTALL);
xl_fill_rx_ring(sc);
goto again;
}
}
void
xl_txeof(struct xl_softc *sc)
{
struct xl_chain *cur_tx;
struct ifnet *ifp;
ifp = &sc->sc_arpcom.ac_if;
while (sc->xl_cdata.xl_tx_head != NULL) {
cur_tx = sc->xl_cdata.xl_tx_head;
bus_dmamap_sync(sc->sc_dmat, sc->sc_listmap,
((caddr_t)cur_tx->xl_ptr - sc->sc_listkva),
sizeof(struct xl_list),
BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
if (CSR_READ_4(sc, XL_DOWNLIST_PTR))
break;
sc->xl_cdata.xl_tx_head = cur_tx->xl_next;
if (cur_tx->map->dm_nsegs != 0) {
bus_dmamap_t map = cur_tx->map;
bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(sc->sc_dmat, map);
}
if (cur_tx->xl_mbuf != NULL) {
m_freem(cur_tx->xl_mbuf);
cur_tx->xl_mbuf = NULL;
}
cur_tx->xl_next = sc->xl_cdata.xl_tx_free;
sc->xl_cdata.xl_tx_free = cur_tx;
}
if (sc->xl_cdata.xl_tx_head == NULL) {
ifq_clr_oactive(&ifp->if_snd);
ifp->if_timer = 0;
sc->xl_cdata.xl_tx_tail = NULL;
} else {
if (CSR_READ_4(sc, XL_DMACTL) & XL_DMACTL_DOWN_STALLED ||
!CSR_READ_4(sc, XL_DOWNLIST_PTR)) {
CSR_WRITE_4(sc, XL_DOWNLIST_PTR,
sc->sc_listmap->dm_segs[0].ds_addr +
((caddr_t)sc->xl_cdata.xl_tx_head->xl_ptr -
sc->sc_listkva));
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_DOWN_UNSTALL);
}
}
}
void
xl_txeof_90xB(struct xl_softc *sc)
{
struct xl_chain *cur_tx = NULL;
struct ifnet *ifp;
int idx;
ifp = &sc->sc_arpcom.ac_if;
idx = sc->xl_cdata.xl_tx_cons;
while (idx != sc->xl_cdata.xl_tx_prod) {
cur_tx = &sc->xl_cdata.xl_tx_chain[idx];
if ((cur_tx->xl_ptr->xl_status &
htole32(XL_TXSTAT_DL_COMPLETE)) == 0)
break;
if (cur_tx->xl_mbuf != NULL) {
m_freem(cur_tx->xl_mbuf);
cur_tx->xl_mbuf = NULL;
}
if (cur_tx->map->dm_nsegs != 0) {
bus_dmamap_sync(sc->sc_dmat, cur_tx->map,
0, cur_tx->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(sc->sc_dmat, cur_tx->map);
}
sc->xl_cdata.xl_tx_cnt--;
XL_INC(idx, XL_TX_LIST_CNT);
}
sc->xl_cdata.xl_tx_cons = idx;
if (cur_tx != NULL)
ifq_clr_oactive(&ifp->if_snd);
if (sc->xl_cdata.xl_tx_cnt == 0)
ifp->if_timer = 0;
}
void
xl_txeoc(struct xl_softc *sc)
{
u_int8_t txstat;
while ((txstat = CSR_READ_1(sc, XL_TX_STATUS))) {
if (txstat & XL_TXSTATUS_UNDERRUN ||
txstat & XL_TXSTATUS_JABBER ||
txstat & XL_TXSTATUS_RECLAIM) {
if (txstat != 0x90) {
printf("%s: transmission error: %x\n",
sc->sc_dev.dv_xname, txstat);
}
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_RESET);
xl_wait(sc);
if (sc->xl_type == XL_TYPE_905B) {
if (sc->xl_cdata.xl_tx_cnt) {
int i;
struct xl_chain *c;
i = sc->xl_cdata.xl_tx_cons;
c = &sc->xl_cdata.xl_tx_chain[i];
CSR_WRITE_4(sc, XL_DOWNLIST_PTR,
c->xl_phys);
CSR_WRITE_1(sc, XL_DOWN_POLL, 64);
}
} else {
if (sc->xl_cdata.xl_tx_head != NULL)
CSR_WRITE_4(sc, XL_DOWNLIST_PTR,
sc->sc_listmap->dm_segs[0].ds_addr +
((caddr_t)sc->xl_cdata.xl_tx_head->xl_ptr -
sc->sc_listkva));
}
CSR_WRITE_1(sc, XL_TX_FREETHRESH, XL_PACKET_SIZE >> 8);
if (txstat & XL_TXSTATUS_UNDERRUN &&
sc->xl_tx_thresh < XL_PACKET_SIZE) {
sc->xl_tx_thresh += XL_MIN_FRAMELEN;
#ifdef notdef
printf("%s: tx underrun, increasing tx start"
" threshold to %d\n", sc->sc_dev.dv_xname,
sc->xl_tx_thresh);
#endif
}
CSR_WRITE_2(sc, XL_COMMAND,
XL_CMD_TX_SET_START|sc->xl_tx_thresh);
if (sc->xl_type == XL_TYPE_905B) {
CSR_WRITE_2(sc, XL_COMMAND,
XL_CMD_SET_TX_RECLAIM|(XL_PACKET_SIZE >> 4));
}
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_ENABLE);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_DOWN_UNSTALL);
} else {
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_ENABLE);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_DOWN_UNSTALL);
}
CSR_WRITE_1(sc, XL_TX_STATUS, 0x01);
}
}
int
xl_intr(void *arg)
{
struct xl_softc *sc;
struct ifnet *ifp;
u_int16_t status;
int claimed = 0;
sc = arg;
ifp = &sc->sc_arpcom.ac_if;
while ((status = CSR_READ_2(sc, XL_STATUS)) & XL_INTRS && status != 0xFFFF) {
claimed = 1;
CSR_WRITE_2(sc, XL_COMMAND,
XL_CMD_INTR_ACK|(status & XL_INTRS));
if (sc->intr_ack)
(*sc->intr_ack)(sc);
if (!(ifp->if_flags & IFF_RUNNING))
return (claimed);
if (status & XL_STAT_UP_COMPLETE)
xl_rxeof(sc);
if (status & XL_STAT_DOWN_COMPLETE) {
if (sc->xl_type == XL_TYPE_905B)
xl_txeof_90xB(sc);
else
xl_txeof(sc);
}
if (status & XL_STAT_TX_COMPLETE) {
ifp->if_oerrors++;
xl_txeoc(sc);
}
if (status & XL_STAT_ADFAIL)
xl_init(sc);
if (status & XL_STAT_STATSOFLOW) {
sc->xl_stats_no_timeout = 1;
xl_stats_update(sc);
sc->xl_stats_no_timeout = 0;
}
}
if (!ifq_empty(&ifp->if_snd))
(*ifp->if_start)(ifp);
return (claimed);
}
void
xl_stats_update(void *xsc)
{
struct xl_softc *sc;
struct ifnet *ifp;
struct xl_stats xl_stats;
u_int8_t *p;
int i;
struct mii_data *mii = NULL;
bzero(&xl_stats, sizeof(struct xl_stats));
sc = xsc;
ifp = &sc->sc_arpcom.ac_if;
if (sc->xl_hasmii)
mii = &sc->sc_mii;
p = (u_int8_t *)&xl_stats;
XL_SEL_WIN(6);
for (i = 0; i < 16; i++)
*p++ = CSR_READ_1(sc, XL_W6_CARRIER_LOST + i);
ifp->if_ierrors += xl_stats.xl_rx_overrun;
ifp->if_collisions += xl_stats.xl_tx_multi_collision +
xl_stats.xl_tx_single_collision +
xl_stats.xl_tx_late_collision;
XL_SEL_WIN(4);
CSR_READ_1(sc, XL_W4_BADSSD);
if (mii != NULL && (!sc->xl_stats_no_timeout))
mii_tick(mii);
XL_SEL_WIN(7);
if (!sc->xl_stats_no_timeout)
timeout_add_sec(&sc->xl_stsup_tmo, 1);
}
int
xl_encap(struct xl_softc *sc, struct xl_chain *c, struct mbuf *m_head)
{
int error, frag, total_len;
u_int32_t status;
bus_dmamap_t map;
map = sc->sc_tx_sparemap;
reload:
error = bus_dmamap_load_mbuf(sc->sc_dmat, map,
m_head, BUS_DMA_NOWAIT);
if (error && error != EFBIG) {
m_freem(m_head);
return (1);
}
for (frag = 0, total_len = 0; frag < map->dm_nsegs; frag++) {
if (frag == XL_MAXFRAGS)
break;
total_len += map->dm_segs[frag].ds_len;
c->xl_ptr->xl_frag[frag].xl_addr =
htole32(map->dm_segs[frag].ds_addr);
c->xl_ptr->xl_frag[frag].xl_len =
htole32(map->dm_segs[frag].ds_len);
}
if (error) {
struct mbuf *m_new = NULL;
MGETHDR(m_new, M_DONTWAIT, MT_DATA);
if (m_new == NULL) {
m_freem(m_head);
return (1);
}
if (m_head->m_pkthdr.len > MHLEN) {
MCLGET(m_new, M_DONTWAIT);
if (!(m_new->m_flags & M_EXT)) {
m_freem(m_new);
m_freem(m_head);
return (1);
}
}
m_copydata(m_head, 0, m_head->m_pkthdr.len,
mtod(m_new, caddr_t));
m_new->m_pkthdr.len = m_new->m_len = m_head->m_pkthdr.len;
m_freem(m_head);
m_head = m_new;
goto reload;
}
bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
BUS_DMASYNC_PREWRITE);
if (c->map->dm_nsegs != 0) {
bus_dmamap_sync(sc->sc_dmat, c->map,
0, c->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(sc->sc_dmat, c->map);
}
c->xl_mbuf = m_head;
sc->sc_tx_sparemap = c->map;
c->map = map;
c->xl_ptr->xl_frag[frag - 1].xl_len |= htole32(XL_LAST_FRAG);
c->xl_ptr->xl_status = htole32(total_len);
c->xl_ptr->xl_next = 0;
if (sc->xl_type == XL_TYPE_905B) {
status = XL_TXSTAT_RND_DEFEAT;
#ifndef XL905B_TXCSUM_BROKEN
if (m_head->m_pkthdr.csum_flags) {
if (m_head->m_pkthdr.csum_flags & M_IPV4_CSUM_OUT)
status |= XL_TXSTAT_IPCKSUM;
if (m_head->m_pkthdr.csum_flags & M_TCP_CSUM_OUT)
status |= XL_TXSTAT_TCPCKSUM;
if (m_head->m_pkthdr.csum_flags & M_UDP_CSUM_OUT)
status |= XL_TXSTAT_UDPCKSUM;
}
#endif
c->xl_ptr->xl_status = htole32(status);
}
bus_dmamap_sync(sc->sc_dmat, sc->sc_listmap,
offsetof(struct xl_list_data, xl_tx_list[0]),
sizeof(struct xl_list) * XL_TX_LIST_CNT,
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
return (0);
}
void
xl_start(struct ifnet *ifp)
{
struct xl_softc *sc;
struct mbuf *m_head = NULL;
struct xl_chain *prev = NULL, *cur_tx = NULL, *start_tx;
struct xl_chain *prev_tx;
int error;
sc = ifp->if_softc;
if (sc->xl_cdata.xl_tx_free == NULL) {
xl_txeoc(sc);
xl_txeof(sc);
if (sc->xl_cdata.xl_tx_free == NULL) {
ifq_set_oactive(&ifp->if_snd);
return;
}
}
start_tx = sc->xl_cdata.xl_tx_free;
while (sc->xl_cdata.xl_tx_free != NULL) {
m_head = ifq_dequeue(&ifp->if_snd);
if (m_head == NULL)
break;
prev_tx = cur_tx;
cur_tx = sc->xl_cdata.xl_tx_free;
error = xl_encap(sc, cur_tx, m_head);
if (error) {
cur_tx = prev_tx;
continue;
}
sc->xl_cdata.xl_tx_free = cur_tx->xl_next;
cur_tx->xl_next = NULL;
if (prev != NULL) {
prev->xl_next = cur_tx;
prev->xl_ptr->xl_next =
sc->sc_listmap->dm_segs[0].ds_addr +
((caddr_t)cur_tx->xl_ptr - sc->sc_listkva);
}
prev = cur_tx;
#if NBPFILTER > 0
if (ifp->if_bpf)
bpf_mtap(ifp->if_bpf, cur_tx->xl_mbuf,
BPF_DIRECTION_OUT);
#endif
}
if (cur_tx == NULL)
return;
cur_tx->xl_ptr->xl_status |= htole32(XL_TXSTAT_DL_INTR);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_DOWN_STALL);
xl_wait(sc);
if (sc->xl_cdata.xl_tx_head != NULL) {
sc->xl_cdata.xl_tx_tail->xl_next = start_tx;
sc->xl_cdata.xl_tx_tail->xl_ptr->xl_next =
sc->sc_listmap->dm_segs[0].ds_addr +
((caddr_t)start_tx->xl_ptr - sc->sc_listkva);
sc->xl_cdata.xl_tx_tail->xl_ptr->xl_status &=
htole32(~XL_TXSTAT_DL_INTR);
sc->xl_cdata.xl_tx_tail = cur_tx;
} else {
sc->xl_cdata.xl_tx_head = start_tx;
sc->xl_cdata.xl_tx_tail = cur_tx;
}
if (!CSR_READ_4(sc, XL_DOWNLIST_PTR))
CSR_WRITE_4(sc, XL_DOWNLIST_PTR,
sc->sc_listmap->dm_segs[0].ds_addr +
((caddr_t)start_tx->xl_ptr - sc->sc_listkva));
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_DOWN_UNSTALL);
XL_SEL_WIN(7);
ifp->if_timer = 5;
xl_rxeof(sc);
}
void
xl_start_90xB(struct ifnet *ifp)
{
struct xl_softc *sc;
struct mbuf *m_head = NULL;
struct xl_chain *prev = NULL, *cur_tx = NULL, *start_tx;
struct xl_chain *prev_tx;
int error, idx;
sc = ifp->if_softc;
if (ifq_is_oactive(&ifp->if_snd))
return;
idx = sc->xl_cdata.xl_tx_prod;
start_tx = &sc->xl_cdata.xl_tx_chain[idx];
while (sc->xl_cdata.xl_tx_chain[idx].xl_mbuf == NULL) {
if ((XL_TX_LIST_CNT - sc->xl_cdata.xl_tx_cnt) < 3) {
ifq_set_oactive(&ifp->if_snd);
break;
}
m_head = ifq_dequeue(&ifp->if_snd);
if (m_head == NULL)
break;
prev_tx = cur_tx;
cur_tx = &sc->xl_cdata.xl_tx_chain[idx];
error = xl_encap(sc, cur_tx, m_head);
if (error) {
cur_tx = prev_tx;
continue;
}
if (prev != NULL)
prev->xl_ptr->xl_next = htole32(cur_tx->xl_phys);
prev = cur_tx;
#if NBPFILTER > 0
if (ifp->if_bpf)
bpf_mtap(ifp->if_bpf, cur_tx->xl_mbuf,
BPF_DIRECTION_OUT);
#endif
XL_INC(idx, XL_TX_LIST_CNT);
sc->xl_cdata.xl_tx_cnt++;
}
if (cur_tx == NULL)
return;
cur_tx->xl_ptr->xl_status |= htole32(XL_TXSTAT_DL_INTR);
sc->xl_cdata.xl_tx_prod = idx;
start_tx->xl_prev->xl_ptr->xl_next = htole32(start_tx->xl_phys);
ifp->if_timer = 5;
}
void
xl_init(void *xsc)
{
struct xl_softc *sc = xsc;
struct ifnet *ifp = &sc->sc_arpcom.ac_if;
int s, i;
struct mii_data *mii = NULL;
s = splnet();
xl_stop(sc);
xl_reset(sc);
if (sc->xl_hasmii)
mii = &sc->sc_mii;
if (mii == NULL) {
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_RESET);
xl_wait(sc);
}
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_RESET);
xl_wait(sc);
DELAY(10000);
XL_SEL_WIN(2);
for (i = 0; i < ETHER_ADDR_LEN; i++) {
CSR_WRITE_1(sc, XL_W2_STATION_ADDR_LO + i,
sc->sc_arpcom.ac_enaddr[i]);
}
for (i = 0; i < 3; i++)
CSR_WRITE_2(sc, XL_W2_STATION_MASK_LO + (i * 2), 0);
#ifdef notdef
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_RESET);
xl_wait(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_RESET);
xl_wait(sc);
#endif
if (xl_list_rx_init(sc) == ENOBUFS) {
printf("%s: initialization failed: no "
"memory for rx buffers\n", sc->sc_dev.dv_xname);
xl_stop(sc);
splx(s);
return;
}
if (sc->xl_type == XL_TYPE_905B)
xl_list_tx_init_90xB(sc);
else
xl_list_tx_init(sc);
CSR_WRITE_1(sc, XL_TX_FREETHRESH, XL_PACKET_SIZE >> 8);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_SET_START|sc->xl_tx_thresh);
if (sc->xl_type == XL_TYPE_905B) {
CSR_WRITE_2(sc, XL_COMMAND,
XL_CMD_SET_TX_RECLAIM|(XL_PACKET_SIZE >> 4));
}
xl_iff(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_UP_STALL);
xl_wait(sc);
CSR_WRITE_4(sc, XL_UPLIST_PTR, sc->sc_listmap->dm_segs[0].ds_addr +
offsetof(struct xl_list_data, xl_rx_list[0]));
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_UP_UNSTALL);
xl_wait(sc);
if (sc->xl_type == XL_TYPE_905B) {
CSR_WRITE_1(sc, XL_DOWN_POLL, 64);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_DOWN_STALL);
xl_wait(sc);
CSR_WRITE_4(sc, XL_DOWNLIST_PTR,
sc->sc_listmap->dm_segs[0].ds_addr +
offsetof(struct xl_list_data, xl_tx_list[0]));
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_DOWN_UNSTALL);
xl_wait(sc);
}
XL_SEL_WIN(3);
if (sc->xl_xcvr == XL_XCVR_COAX)
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_COAX_START);
else
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_COAX_STOP);
if (sc->xl_type == XL_TYPE_905B)
CSR_WRITE_2(sc, XL_W3_MAXPKTSIZE, XL_PACKET_SIZE);
else {
u_int8_t macctl;
macctl = CSR_READ_1(sc, XL_W3_MAC_CTRL);
macctl |= XL_MACCTRL_ALLOW_LARGE_PACK;
CSR_WRITE_1(sc, XL_W3_MAC_CTRL, macctl);
}
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_STATS_DISABLE);
sc->xl_stats_no_timeout = 1;
xl_stats_update(sc);
sc->xl_stats_no_timeout = 0;
XL_SEL_WIN(4);
CSR_WRITE_2(sc, XL_W4_NET_DIAG, XL_NETDIAG_UPPER_BYTES_ENABLE);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_STATS_ENABLE);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_INTR_ACK|0xFF);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_STAT_ENB|XL_INTRS);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_INTR_ENB|XL_INTRS);
if (sc->intr_ack)
(*sc->intr_ack)(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_SET_THRESH|(XL_PACKET_SIZE >>2));
CSR_WRITE_4(sc, XL_DMACTL, XL_DMACTL_UP_RX_EARLY);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_ENABLE);
xl_wait(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_ENABLE);
xl_wait(sc);
if (mii != NULL)
mii_mediachg(mii);
XL_SEL_WIN(7);
ifp->if_flags |= IFF_RUNNING;
ifq_clr_oactive(&ifp->if_snd);
splx(s);
timeout_add_sec(&sc->xl_stsup_tmo, 1);
}
int
xl_ifmedia_upd(struct ifnet *ifp)
{
struct xl_softc *sc;
struct ifmedia *ifm = NULL;
struct mii_data *mii = NULL;
sc = ifp->if_softc;
if (sc->xl_hasmii)
mii = &sc->sc_mii;
if (mii == NULL)
ifm = &sc->ifmedia;
else
ifm = &mii->mii_media;
switch(IFM_SUBTYPE(ifm->ifm_media)) {
case IFM_100_FX:
case IFM_10_FL:
case IFM_10_2:
case IFM_10_5:
xl_setmode(sc, ifm->ifm_media);
return (0);
break;
default:
break;
}
if (sc->xl_media & XL_MEDIAOPT_MII || sc->xl_media & XL_MEDIAOPT_BTX
|| sc->xl_media & XL_MEDIAOPT_BT4) {
xl_init(sc);
} else {
xl_setmode(sc, ifm->ifm_media);
}
return (0);
}
void
xl_ifmedia_sts(struct ifnet *ifp, struct ifmediareq *ifmr)
{
struct xl_softc *sc;
u_int32_t icfg;
u_int16_t status = 0;
struct mii_data *mii = NULL;
sc = ifp->if_softc;
if (sc->xl_hasmii != 0)
mii = &sc->sc_mii;
XL_SEL_WIN(4);
status = CSR_READ_2(sc, XL_W4_MEDIA_STATUS);
XL_SEL_WIN(3);
icfg = CSR_READ_4(sc, XL_W3_INTERNAL_CFG) & XL_ICFG_CONNECTOR_MASK;
icfg >>= XL_ICFG_CONNECTOR_BITS;
ifmr->ifm_active = IFM_ETHER;
ifmr->ifm_status = IFM_AVALID;
if ((status & XL_MEDIASTAT_CARRIER) == 0)
ifmr->ifm_status |= IFM_ACTIVE;
switch(icfg) {
case XL_XCVR_10BT:
ifmr->ifm_active = IFM_ETHER|IFM_10_T;
if (CSR_READ_1(sc, XL_W3_MAC_CTRL) & XL_MACCTRL_DUPLEX)
ifmr->ifm_active |= IFM_FDX;
else
ifmr->ifm_active |= IFM_HDX;
break;
case XL_XCVR_AUI:
if (sc->xl_type == XL_TYPE_905B &&
sc->xl_media == XL_MEDIAOPT_10FL) {
ifmr->ifm_active = IFM_ETHER|IFM_10_FL;
if (CSR_READ_1(sc, XL_W3_MAC_CTRL) & XL_MACCTRL_DUPLEX)
ifmr->ifm_active |= IFM_FDX;
else
ifmr->ifm_active |= IFM_HDX;
} else
ifmr->ifm_active = IFM_ETHER|IFM_10_5;
break;
case XL_XCVR_COAX:
ifmr->ifm_active = IFM_ETHER|IFM_10_2;
break;
case XL_XCVR_100BTX:
case XL_XCVR_AUTO:
case XL_XCVR_MII:
if (mii != NULL) {
mii_pollstat(mii);
ifmr->ifm_active = mii->mii_media_active;
ifmr->ifm_status = mii->mii_media_status;
}
break;
case XL_XCVR_100BFX:
ifmr->ifm_active = IFM_ETHER|IFM_100_FX;
break;
default:
printf("%s: unknown XCVR type: %d\n", sc->sc_dev.dv_xname, icfg);
break;
}
}
int
xl_ioctl(struct ifnet *ifp, u_long command, caddr_t data)
{
struct xl_softc *sc = ifp->if_softc;
struct ifreq *ifr = (struct ifreq *)data;
int s, error = 0;
struct mii_data *mii = NULL;
s = splnet();
switch(command) {
case SIOCSIFADDR:
ifp->if_flags |= IFF_UP;
if (!(ifp->if_flags & IFF_RUNNING))
xl_init(sc);
break;
case SIOCSIFFLAGS:
if (ifp->if_flags & IFF_UP) {
if (ifp->if_flags & IFF_RUNNING)
error = ENETRESET;
else
xl_init(sc);
} else {
if (ifp->if_flags & IFF_RUNNING)
xl_stop(sc);
}
break;
case SIOCGIFMEDIA:
case SIOCSIFMEDIA:
if (sc->xl_hasmii != 0)
mii = &sc->sc_mii;
if (mii == NULL)
error = ifmedia_ioctl(ifp, ifr,
&sc->ifmedia, command);
else
error = ifmedia_ioctl(ifp, ifr,
&mii->mii_media, command);
break;
case SIOCGIFRXR:
error = if_rxr_ioctl((struct if_rxrinfo *)ifr->ifr_data,
NULL, MCLBYTES, &sc->xl_cdata.xl_rx_ring);
break;
default:
error = ether_ioctl(ifp, &sc->sc_arpcom, command, data);
}
if (error == ENETRESET) {
if (ifp->if_flags & IFF_RUNNING)
xl_iff(sc);
error = 0;
}
splx(s);
return (error);
}
void
xl_watchdog(struct ifnet *ifp)
{
struct xl_softc *sc;
u_int16_t status = 0;
sc = ifp->if_softc;
ifp->if_oerrors++;
XL_SEL_WIN(4);
status = CSR_READ_2(sc, XL_W4_MEDIA_STATUS);
printf("%s: watchdog timeout\n", sc->sc_dev.dv_xname);
if (status & XL_MEDIASTAT_CARRIER)
printf("%s: no carrier - transceiver cable problem?\n",
sc->sc_dev.dv_xname);
xl_txeoc(sc);
xl_txeof(sc);
xl_rxeof(sc);
xl_init(sc);
if (!ifq_empty(&ifp->if_snd))
(*ifp->if_start)(ifp);
}
void
xl_freetxrx(struct xl_softc *sc)
{
bus_dmamap_t map;
int i;
for (i = 0; i < XL_RX_LIST_CNT; i++) {
if (sc->xl_cdata.xl_rx_chain[i].map->dm_nsegs != 0) {
map = sc->xl_cdata.xl_rx_chain[i].map;
bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
BUS_DMASYNC_POSTREAD);
bus_dmamap_unload(sc->sc_dmat, map);
}
if (sc->xl_cdata.xl_rx_chain[i].xl_mbuf != NULL) {
m_freem(sc->xl_cdata.xl_rx_chain[i].xl_mbuf);
sc->xl_cdata.xl_rx_chain[i].xl_mbuf = NULL;
}
}
bzero(&sc->xl_ldata->xl_rx_list, sizeof(sc->xl_ldata->xl_rx_list));
for (i = 0; i < XL_TX_LIST_CNT; i++) {
if (sc->xl_cdata.xl_tx_chain[i].map->dm_nsegs != 0) {
map = sc->xl_cdata.xl_tx_chain[i].map;
bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(sc->sc_dmat, map);
}
if (sc->xl_cdata.xl_tx_chain[i].xl_mbuf != NULL) {
m_freem(sc->xl_cdata.xl_tx_chain[i].xl_mbuf);
sc->xl_cdata.xl_tx_chain[i].xl_mbuf = NULL;
}
}
bzero(&sc->xl_ldata->xl_tx_list, sizeof(sc->xl_ldata->xl_tx_list));
}
void
xl_stop(struct xl_softc *sc)
{
struct ifnet *ifp;
timeout_del(&sc->xl_stsup_tmo);
ifp = &sc->sc_arpcom.ac_if;
ifp->if_flags &= ~IFF_RUNNING;
ifq_clr_oactive(&ifp->if_snd);
ifp->if_timer = 0;
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_DISABLE);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_STATS_DISABLE);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_INTR_ENB);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_DISCARD);
xl_wait(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_DISABLE);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_COAX_STOP);
DELAY(800);
#ifdef foo
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_RESET);
xl_wait(sc);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_TX_RESET);
xl_wait(sc);
#endif
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_INTR_ACK|XL_STAT_INTLATCH);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_STAT_ENB|0);
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_INTR_ENB|0);
if (sc->intr_ack)
(*sc->intr_ack)(sc);
xl_freetxrx(sc);
}
#ifndef SMALL_KERNEL
void
xl_wol_power(struct xl_softc *sc)
{
if ((sc->xl_flags & XL_FLAG_WOL) && sc->wol_power) {
CSR_WRITE_2(sc, XL_COMMAND, XL_CMD_RX_ENABLE);
sc->wol_power(sc->wol_power_arg);
}
}
#endif
void
xl_attach(struct xl_softc *sc)
{
u_int8_t enaddr[ETHER_ADDR_LEN];
u_int16_t xcvr[2];
struct ifnet *ifp = &sc->sc_arpcom.ac_if;
int i;
uint64_t media = IFM_ETHER|IFM_100_TX|IFM_FDX;
struct ifmedia *ifm;
i = splnet();
xl_reset(sc);
splx(i);
if (xl_read_eeprom(sc, (caddr_t)&enaddr, XL_EE_OEM_ADR0, 3, 1)) {
printf("\n%s: failed to read station address\n",
sc->sc_dev.dv_xname);
return;
}
memcpy(&sc->sc_arpcom.ac_enaddr, enaddr, ETHER_ADDR_LEN);
if (bus_dmamem_alloc(sc->sc_dmat, sizeof(struct xl_list_data),
PAGE_SIZE, 0, sc->sc_listseg, 1, &sc->sc_listnseg,
BUS_DMA_NOWAIT | BUS_DMA_ZERO) != 0) {
printf(": can't alloc list mem\n");
return;
}
if (bus_dmamem_map(sc->sc_dmat, sc->sc_listseg, sc->sc_listnseg,
sizeof(struct xl_list_data), &sc->sc_listkva,
BUS_DMA_NOWAIT) != 0) {
printf(": can't map list mem\n");
return;
}
if (bus_dmamap_create(sc->sc_dmat, sizeof(struct xl_list_data), 1,
sizeof(struct xl_list_data), 0, BUS_DMA_NOWAIT,
&sc->sc_listmap) != 0) {
printf(": can't alloc list map\n");
return;
}
if (bus_dmamap_load(sc->sc_dmat, sc->sc_listmap, sc->sc_listkva,
sizeof(struct xl_list_data), NULL, BUS_DMA_NOWAIT) != 0) {
printf(": can't load list map\n");
return;
}
sc->xl_ldata = (struct xl_list_data *)sc->sc_listkva;
for (i = 0; i < XL_RX_LIST_CNT; i++) {
if (bus_dmamap_create(sc->sc_dmat, MCLBYTES, 1, MCLBYTES,
0, BUS_DMA_NOWAIT,
&sc->xl_cdata.xl_rx_chain[i].map) != 0) {
printf(": can't create rx map\n");
return;
}
}
if (bus_dmamap_create(sc->sc_dmat, MCLBYTES, 1, MCLBYTES, 0,
BUS_DMA_NOWAIT, &sc->sc_rx_sparemap) != 0) {
printf(": can't create rx spare map\n");
return;
}
for (i = 0; i < XL_TX_LIST_CNT; i++) {
if (bus_dmamap_create(sc->sc_dmat, MCLBYTES,
XL_TX_LIST_CNT - 3, MCLBYTES, 0, BUS_DMA_NOWAIT,
&sc->xl_cdata.xl_tx_chain[i].map) != 0) {
printf(": can't create tx map\n");
return;
}
}
if (bus_dmamap_create(sc->sc_dmat, MCLBYTES, XL_TX_LIST_CNT - 3,
MCLBYTES, 0, BUS_DMA_NOWAIT, &sc->sc_tx_sparemap) != 0) {
printf(": can't create tx spare map\n");
return;
}
printf(", address %s\n", ether_sprintf(sc->sc_arpcom.ac_enaddr));
if (sc->xl_flags & (XL_FLAG_INVERT_LED_PWR|XL_FLAG_INVERT_MII_PWR)) {
u_int16_t n;
XL_SEL_WIN(2);
n = CSR_READ_2(sc, 12);
if (sc->xl_flags & XL_FLAG_INVERT_LED_PWR)
n |= 0x0010;
if (sc->xl_flags & XL_FLAG_INVERT_MII_PWR)
n |= 0x4000;
CSR_WRITE_2(sc, 12, n);
}
xl_read_eeprom(sc, (caddr_t)&sc->xl_caps, XL_EE_CAPS, 1, 0);
if (sc->xl_caps & XL_CAPS_NO_TXLENGTH ||
!(sc->xl_caps & XL_CAPS_LARGE_PKTS))
sc->xl_type = XL_TYPE_905B;
else
sc->xl_type = XL_TYPE_90X;
sc->xl_tx_thresh = XL_MIN_FRAMELEN;
timeout_set(&sc->xl_stsup_tmo, xl_stats_update, sc);
ifp->if_softc = sc;
ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
ifp->if_ioctl = xl_ioctl;
if (sc->xl_type == XL_TYPE_905B)
ifp->if_start = xl_start_90xB;
else
ifp->if_start = xl_start;
ifp->if_watchdog = xl_watchdog;
ifp->if_baudrate = 10000000;
ifq_init_maxlen(&ifp->if_snd, XL_TX_LIST_CNT - 1);
memcpy(ifp->if_xname, sc->sc_dev.dv_xname, IFNAMSIZ);
ifp->if_capabilities = IFCAP_VLAN_MTU;
#ifndef XL905B_TXCSUM_BROKEN
ifp->if_capabilities |= IFCAP_CSUM_IPv4|IFCAP_CSUM_TCPv4|
IFCAP_CSUM_UDPv4;
#endif
XL_SEL_WIN(3);
sc->xl_media = CSR_READ_2(sc, XL_W3_MEDIA_OPT);
xl_read_eeprom(sc, (char *)&xcvr, XL_EE_ICFG_0, 2, 0);
sc->xl_xcvr = xcvr[0] | xcvr[1] << 16;
sc->xl_xcvr &= XL_ICFG_CONNECTOR_MASK;
sc->xl_xcvr >>= XL_ICFG_CONNECTOR_BITS;
xl_mediacheck(sc);
if (sc->xl_media & XL_MEDIAOPT_MII || sc->xl_media & XL_MEDIAOPT_BTX
|| sc->xl_media & XL_MEDIAOPT_BT4) {
ifmedia_init(&sc->sc_mii.mii_media, 0,
xl_ifmedia_upd, xl_ifmedia_sts);
sc->xl_hasmii = 1;
sc->sc_mii.mii_ifp = ifp;
sc->sc_mii.mii_readreg = xl_miibus_readreg;
sc->sc_mii.mii_writereg = xl_miibus_writereg;
sc->sc_mii.mii_statchg = xl_miibus_statchg;
xl_setcfg(sc);
mii_attach((struct device *)sc, &sc->sc_mii, 0xffffffff,
MII_PHY_ANY, MII_OFFSET_ANY, 0);
if (LIST_FIRST(&sc->sc_mii.mii_phys) == NULL) {
ifmedia_add(&sc->sc_mii.mii_media, IFM_ETHER|IFM_NONE,
0, NULL);
ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER|IFM_NONE);
}
else {
ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER|IFM_AUTO);
}
ifm = &sc->sc_mii.mii_media;
}
else {
ifmedia_init(&sc->ifmedia, 0, xl_ifmedia_upd, xl_ifmedia_sts);
sc->xl_hasmii = 0;
ifm = &sc->ifmedia;
}
if (sc->xl_xcvr == XL_XCVR_AUTO)
xl_choose_xcvr(sc, 0);
if (sc->xl_media & XL_MEDIAOPT_BT) {
ifmedia_add(ifm, IFM_ETHER|IFM_10_T, 0, NULL);
ifmedia_add(ifm, IFM_ETHER|IFM_10_T|IFM_HDX, 0, NULL);
if (sc->xl_caps & XL_CAPS_FULL_DUPLEX)
ifmedia_add(ifm, IFM_ETHER|IFM_10_T|IFM_FDX, 0, NULL);
}
if (sc->xl_media & (XL_MEDIAOPT_AUI|XL_MEDIAOPT_10FL)) {
if (sc->xl_type == XL_TYPE_905B &&
sc->xl_media == XL_MEDIAOPT_10FL) {
ifmedia_add(ifm, IFM_ETHER|IFM_10_FL, 0, NULL);
ifmedia_add(ifm, IFM_ETHER|IFM_10_FL|IFM_HDX,
0, NULL);
if (sc->xl_caps & XL_CAPS_FULL_DUPLEX)
ifmedia_add(ifm,
IFM_ETHER|IFM_10_FL|IFM_FDX, 0, NULL);
} else {
ifmedia_add(ifm, IFM_ETHER|IFM_10_5, 0, NULL);
}
}
if (sc->xl_media & XL_MEDIAOPT_BNC) {
ifmedia_add(ifm, IFM_ETHER|IFM_10_2, 0, NULL);
}
if (sc->xl_media & XL_MEDIAOPT_BFX) {
ifp->if_baudrate = 100000000;
ifmedia_add(ifm, IFM_ETHER|IFM_100_FX, 0, NULL);
}
switch(sc->xl_xcvr) {
case XL_XCVR_10BT:
media = IFM_ETHER|IFM_10_T;
xl_setmode(sc, media);
break;
case XL_XCVR_AUI:
if (sc->xl_type == XL_TYPE_905B &&
sc->xl_media == XL_MEDIAOPT_10FL) {
media = IFM_ETHER|IFM_10_FL;
xl_setmode(sc, media);
} else {
media = IFM_ETHER|IFM_10_5;
xl_setmode(sc, media);
}
break;
case XL_XCVR_COAX:
media = IFM_ETHER|IFM_10_2;
xl_setmode(sc, media);
break;
case XL_XCVR_AUTO:
case XL_XCVR_100BTX:
case XL_XCVR_MII:
break;
case XL_XCVR_100BFX:
media = IFM_ETHER|IFM_100_FX;
xl_setmode(sc, media);
break;
default:
printf("%s: unknown XCVR type: %d\n", sc->sc_dev.dv_xname,
sc->xl_xcvr);
media = IFM_ETHER | IFM_10_T;
break;
}
if (sc->xl_hasmii == 0)
ifmedia_set(&sc->ifmedia, media);
if (sc->xl_flags & XL_FLAG_NO_XCVR_PWR) {
XL_SEL_WIN(0);
CSR_WRITE_2(sc, XL_W0_MFG_ID, XL_NO_XCVR_PWR_MAGICBITS);
}
#ifndef SMALL_KERNEL
if ((sc->xl_caps & XL_CAPS_PWRMGMT) != 0) {
ifp->if_capabilities |= IFCAP_WOL;
ifp->if_wol = xl_wol;
xl_wol(ifp, 0);
}
#endif
if_attach(ifp);
ether_ifattach(ifp);
}
int
xl_detach(struct xl_softc *sc)
{
struct ifnet *ifp = &sc->sc_arpcom.ac_if;
extern void xl_freetxrx(struct xl_softc *);
timeout_del(&sc->xl_stsup_tmo);
xl_freetxrx(sc);
if (sc->xl_hasmii)
mii_detach(&sc->sc_mii, MII_PHY_ANY, MII_OFFSET_ANY);
ifmedia_delete_instance(&sc->sc_mii.mii_media, IFM_INST_ANY);
ether_ifdetach(ifp);
if_detach(ifp);
return (0);
}
#ifndef SMALL_KERNEL
int
xl_wol(struct ifnet *ifp, int enable)
{
struct xl_softc *sc = ifp->if_softc;
XL_SEL_WIN(7);
if (enable) {
if (!(ifp->if_flags & IFF_RUNNING))
xl_init(sc);
CSR_WRITE_2(sc, XL_W7_BM_PME, XL_BM_PME_MAGIC);
sc->xl_flags |= XL_FLAG_WOL;
} else {
CSR_WRITE_2(sc, XL_W7_BM_PME, 0);
sc->xl_flags &= ~XL_FLAG_WOL;
}
return (0);
}
#endif
struct cfdriver xl_cd = {
NULL, "xl", DV_IFNET
};