The Design and Implementation of the FreeBSD Operating System, Second Edition
Now available: The Design and Implementation of the FreeBSD Operating System (Second Edition)


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]

FreeBSD/Linux Kernel Cross Reference
sys/dev/vnic/thunder_bgx.c

Version: -  FREEBSD  -  FREEBSD-13-STABLE  -  FREEBSD-13-0  -  FREEBSD-12-STABLE  -  FREEBSD-12-0  -  FREEBSD-11-STABLE  -  FREEBSD-11-0  -  FREEBSD-10-STABLE  -  FREEBSD-10-0  -  FREEBSD-9-STABLE  -  FREEBSD-9-0  -  FREEBSD-8-STABLE  -  FREEBSD-8-0  -  FREEBSD-7-STABLE  -  FREEBSD-7-0  -  FREEBSD-6-STABLE  -  FREEBSD-6-0  -  FREEBSD-5-STABLE  -  FREEBSD-5-0  -  FREEBSD-4-STABLE  -  FREEBSD-3-STABLE  -  FREEBSD22  -  l41  -  OPENBSD  -  linux-2.6  -  MK84  -  PLAN9  -  xnu-8792 
SearchContext: -  none  -  3  -  10 

    1 /*
    2  * Copyright (C) 2015 Cavium Inc.
    3  * All rights reserved.
    4  *
    5  * Redistribution and use in source and binary forms, with or without
    6  * modification, are permitted provided that the following conditions
    7  * are met:
    8  * 1. Redistributions of source code must retain the above copyright
    9  *    notice, this list of conditions and the following disclaimer.
   10  * 2. Redistributions in binary form must reproduce the above copyright
   11  *    notice, this list of conditions and the following disclaimer in the
   12  *    documentation and/or other materials provided with the distribution.
   13  *
   14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
   15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
   16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
   17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
   18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
   19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
   20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
   21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
   22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
   23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
   24  * SUCH DAMAGE.
   25  *
   26  * $FreeBSD$
   27  *
   28  */
   29 #include "opt_platform.h"
   30 
   31 #include <sys/cdefs.h>
   32 __FBSDID("$FreeBSD$");
   33 
   34 #include <sys/param.h>
   35 #include <sys/systm.h>
   36 #include <sys/bitset.h>
   37 #include <sys/bitstring.h>
   38 #include <sys/bus.h>
   39 #include <sys/endian.h>
   40 #include <sys/kernel.h>
   41 #include <sys/malloc.h>
   42 #include <sys/module.h>
   43 #include <sys/rman.h>
   44 #include <sys/pciio.h>
   45 #include <sys/pcpu.h>
   46 #include <sys/proc.h>
   47 #include <sys/socket.h>
   48 #include <sys/sockio.h>
   49 #include <sys/cpuset.h>
   50 #include <sys/lock.h>
   51 #include <sys/mutex.h>
   52 
   53 #include <net/ethernet.h>
   54 #include <net/if.h>
   55 #include <net/if_media.h>
   56 
   57 #include <machine/bus.h>
   58 
   59 #include <dev/pci/pcireg.h>
   60 #include <dev/pci/pcivar.h>
   61 
   62 #include "thunder_bgx.h"
   63 #include "thunder_bgx_var.h"
   64 #include "nic_reg.h"
   65 #include "nic.h"
   66 
   67 #include "lmac_if.h"
   68 
   69 #define THUNDER_BGX_DEVSTR      "ThunderX BGX Ethernet I/O Interface"
   70 
   71 MALLOC_DEFINE(M_BGX, "thunder_bgx", "ThunderX BGX dynamic memory");
   72 
   73 #define BGX_NODE_ID_MASK        0x1
   74 #define BGX_NODE_ID_SHIFT       24
   75 
   76 #define DRV_NAME        "thunder-BGX"
   77 #define DRV_VERSION     "1.0"
   78 
   79 static int bgx_init_phy(struct bgx *);
   80 
   81 static struct bgx *bgx_vnic[MAX_BGX_THUNDER];
   82 static int lmac_count __unused; /* Total no of LMACs in system */
   83 
   84 static int bgx_xaui_check_link(struct lmac *lmac);
   85 static void bgx_get_qlm_mode(struct bgx *);
   86 static void bgx_init_hw(struct bgx *);
   87 static int bgx_lmac_enable(struct bgx *, uint8_t);
   88 static void bgx_lmac_disable(struct bgx *, uint8_t);
   89 
   90 static int thunder_bgx_probe(device_t);
   91 static int thunder_bgx_attach(device_t);
   92 static int thunder_bgx_detach(device_t);
   93 
   94 static device_method_t thunder_bgx_methods[] = {
   95         /* Device interface */
   96         DEVMETHOD(device_probe,         thunder_bgx_probe),
   97         DEVMETHOD(device_attach,        thunder_bgx_attach),
   98         DEVMETHOD(device_detach,        thunder_bgx_detach),
   99 
  100         DEVMETHOD_END,
  101 };
  102 
  103 static driver_t thunder_bgx_driver = {
  104         "bgx",
  105         thunder_bgx_methods,
  106         sizeof(struct lmac),
  107 };
  108 
  109 static devclass_t thunder_bgx_devclass;
  110 
  111 DRIVER_MODULE(thunder_bgx, pci, thunder_bgx_driver, thunder_bgx_devclass, 0, 0);
  112 MODULE_VERSION(thunder_bgx, 1);
  113 MODULE_DEPEND(thunder_bgx, pci, 1, 1, 1);
  114 MODULE_DEPEND(thunder_bgx, ether, 1, 1, 1);
  115 MODULE_DEPEND(thunder_bgx, thunder_mdio, 1, 1, 1);
  116 
  117 static int
  118 thunder_bgx_probe(device_t dev)
  119 {
  120         uint16_t vendor_id;
  121         uint16_t device_id;
  122 
  123         vendor_id = pci_get_vendor(dev);
  124         device_id = pci_get_device(dev);
  125 
  126         if (vendor_id == PCI_VENDOR_ID_CAVIUM &&
  127             device_id == PCI_DEVICE_ID_THUNDER_BGX) {
  128                 device_set_desc(dev, THUNDER_BGX_DEVSTR);
  129                 return (BUS_PROBE_DEFAULT);
  130         }
  131 
  132         return (ENXIO);
  133 }
  134 
  135 static int
  136 thunder_bgx_attach(device_t dev)
  137 {
  138         struct bgx *bgx;
  139         uint8_t lmacid;
  140         int err;
  141         int rid;
  142         struct lmac *lmac;
  143 
  144         bgx = malloc(sizeof(*bgx), M_BGX, (M_WAITOK | M_ZERO));
  145         bgx->dev = dev;
  146 
  147         lmac = device_get_softc(dev);
  148         lmac->bgx = bgx;
  149         /* Enable bus mastering */
  150         pci_enable_busmaster(dev);
  151         /* Allocate resources - configuration registers */
  152         rid = PCIR_BAR(PCI_CFG_REG_BAR_NUM);
  153         bgx->reg_base = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid,
  154             RF_ACTIVE);
  155         if (bgx->reg_base == NULL) {
  156                 device_printf(dev, "Could not allocate CSR memory space\n");
  157                 err = ENXIO;
  158                 goto err_disable_device;
  159         }
  160 
  161         bgx->bgx_id = (rman_get_start(bgx->reg_base) >> BGX_NODE_ID_SHIFT) &
  162             BGX_NODE_ID_MASK;
  163         bgx->bgx_id += nic_get_node_id(bgx->reg_base) * MAX_BGX_PER_CN88XX;
  164 
  165         bgx_vnic[bgx->bgx_id] = bgx;
  166         bgx_get_qlm_mode(bgx);
  167 
  168         err = bgx_init_phy(bgx);
  169         if (err != 0)
  170                 goto err_free_res;
  171 
  172         bgx_init_hw(bgx);
  173 
  174         /* Enable all LMACs */
  175         for (lmacid = 0; lmacid < bgx->lmac_count; lmacid++) {
  176                 err = bgx_lmac_enable(bgx, lmacid);
  177                 if (err) {
  178                         device_printf(dev, "BGX%d failed to enable lmac%d\n",
  179                                 bgx->bgx_id, lmacid);
  180                         goto err_free_res;
  181                 }
  182         }
  183 
  184         return (0);
  185 
  186 err_free_res:
  187         bgx_vnic[bgx->bgx_id] = NULL;
  188         bus_release_resource(dev, SYS_RES_MEMORY,
  189             rman_get_rid(bgx->reg_base), bgx->reg_base);
  190 err_disable_device:
  191         free(bgx, M_BGX);
  192         pci_disable_busmaster(dev);
  193 
  194         return (err);
  195 }
  196 
  197 static int
  198 thunder_bgx_detach(device_t dev)
  199 {
  200         struct lmac *lmac;
  201         struct bgx *bgx;
  202         uint8_t lmacid;
  203 
  204         lmac = device_get_softc(dev);
  205         bgx = lmac->bgx;
  206         /* Disable all LMACs */
  207         for (lmacid = 0; lmacid < bgx->lmac_count; lmacid++)
  208                 bgx_lmac_disable(bgx, lmacid);
  209 
  210         bgx_vnic[bgx->bgx_id] = NULL;
  211         bus_release_resource(dev, SYS_RES_MEMORY,
  212             rman_get_rid(bgx->reg_base), bgx->reg_base);
  213         free(bgx, M_BGX);
  214         pci_disable_busmaster(dev);
  215 
  216         return (0);
  217 }
  218 
  219 /* Register read/write APIs */
  220 static uint64_t
  221 bgx_reg_read(struct bgx *bgx, uint8_t lmac, uint64_t offset)
  222 {
  223         bus_space_handle_t addr;
  224 
  225         addr = ((uint32_t)lmac << 20) + offset;
  226 
  227         return (bus_read_8(bgx->reg_base, addr));
  228 }
  229 
  230 static void
  231 bgx_reg_write(struct bgx *bgx, uint8_t lmac, uint64_t offset, uint64_t val)
  232 {
  233         bus_space_handle_t addr;
  234 
  235         addr = ((uint32_t)lmac << 20) + offset;
  236 
  237         bus_write_8(bgx->reg_base, addr, val);
  238 }
  239 
  240 static void
  241 bgx_reg_modify(struct bgx *bgx, uint8_t lmac, uint64_t offset, uint64_t val)
  242 {
  243         bus_space_handle_t addr;
  244 
  245         addr = ((uint32_t)lmac << 20) + offset;
  246 
  247         bus_write_8(bgx->reg_base, addr, val | bus_read_8(bgx->reg_base, addr));
  248 }
  249 
  250 static int
  251 bgx_poll_reg(struct bgx *bgx, uint8_t lmac, uint64_t reg, uint64_t mask,
  252     boolean_t zero)
  253 {
  254         int timeout = 10;
  255         uint64_t reg_val;
  256 
  257         while (timeout) {
  258                 reg_val = bgx_reg_read(bgx, lmac, reg);
  259                 if (zero && !(reg_val & mask))
  260                         return (0);
  261                 if (!zero && (reg_val & mask))
  262                         return (0);
  263 
  264                 DELAY(100);
  265                 timeout--;
  266         }
  267         return (ETIMEDOUT);
  268 }
  269 
  270 /* Return number of BGX present in HW */
  271 u_int
  272 bgx_get_map(int node)
  273 {
  274         int i;
  275         u_int map = 0;
  276 
  277         for (i = 0; i < MAX_BGX_PER_CN88XX; i++) {
  278                 if (bgx_vnic[(node * MAX_BGX_PER_CN88XX) + i])
  279                         map |= (1 << i);
  280         }
  281 
  282         return (map);
  283 }
  284 
  285 /* Return number of LMAC configured for this BGX */
  286 int
  287 bgx_get_lmac_count(int node, int bgx_idx)
  288 {
  289         struct bgx *bgx;
  290 
  291         bgx = bgx_vnic[(node * MAX_BGX_PER_CN88XX) + bgx_idx];
  292         if (bgx != NULL)
  293                 return (bgx->lmac_count);
  294 
  295         return (0);
  296 }
  297 
  298 /* Returns the current link status of LMAC */
  299 void
  300 bgx_get_lmac_link_state(int node, int bgx_idx, int lmacid, void *status)
  301 {
  302         struct bgx_link_status *link = (struct bgx_link_status *)status;
  303         struct bgx *bgx;
  304         struct lmac *lmac;
  305 
  306         bgx = bgx_vnic[(node * MAX_BGX_PER_CN88XX) + bgx_idx];
  307         if (bgx == NULL)
  308                 return;
  309 
  310         lmac = &bgx->lmac[lmacid];
  311         link->link_up = lmac->link_up;
  312         link->duplex = lmac->last_duplex;
  313         link->speed = lmac->last_speed;
  314 }
  315 
  316 const uint8_t
  317 *bgx_get_lmac_mac(int node, int bgx_idx, int lmacid)
  318 {
  319         struct bgx *bgx = bgx_vnic[(node * MAX_BGX_PER_CN88XX) + bgx_idx];
  320 
  321         if (bgx != NULL)
  322                 return (bgx->lmac[lmacid].mac);
  323 
  324         return (NULL);
  325 }
  326 
  327 void
  328 bgx_set_lmac_mac(int node, int bgx_idx, int lmacid, const uint8_t *mac)
  329 {
  330         struct bgx *bgx = bgx_vnic[(node * MAX_BGX_PER_CN88XX) + bgx_idx];
  331 
  332         if (bgx == NULL)
  333                 return;
  334 
  335         memcpy(bgx->lmac[lmacid].mac, mac, ETHER_ADDR_LEN);
  336 }
  337 
  338 static void
  339 bgx_sgmii_change_link_state(struct lmac *lmac)
  340 {
  341         struct bgx *bgx = lmac->bgx;
  342         uint64_t cmr_cfg;
  343         uint64_t port_cfg = 0;
  344         uint64_t misc_ctl = 0;
  345 
  346         cmr_cfg = bgx_reg_read(bgx, lmac->lmacid, BGX_CMRX_CFG);
  347         cmr_cfg &= ~CMR_EN;
  348         bgx_reg_write(bgx, lmac->lmacid, BGX_CMRX_CFG, cmr_cfg);
  349 
  350         port_cfg = bgx_reg_read(bgx, lmac->lmacid, BGX_GMP_GMI_PRTX_CFG);
  351         misc_ctl = bgx_reg_read(bgx, lmac->lmacid, BGX_GMP_PCS_MISCX_CTL);
  352 
  353         if (lmac->link_up) {
  354                 misc_ctl &= ~PCS_MISC_CTL_GMX_ENO;
  355                 port_cfg &= ~GMI_PORT_CFG_DUPLEX;
  356                 port_cfg |=  (lmac->last_duplex << 2);
  357         } else {
  358                 misc_ctl |= PCS_MISC_CTL_GMX_ENO;
  359         }
  360 
  361         switch (lmac->last_speed) {
  362         case 10:
  363                 port_cfg &= ~GMI_PORT_CFG_SPEED; /* speed 0 */
  364                 port_cfg |= GMI_PORT_CFG_SPEED_MSB;  /* speed_msb 1 */
  365                 port_cfg &= ~GMI_PORT_CFG_SLOT_TIME; /* slottime 0 */
  366                 misc_ctl &= ~PCS_MISC_CTL_SAMP_PT_MASK;
  367                 misc_ctl |= 50; /* samp_pt */
  368                 bgx_reg_write(bgx, lmac->lmacid, BGX_GMP_GMI_TXX_SLOT, 64);
  369                 bgx_reg_write(bgx, lmac->lmacid, BGX_GMP_GMI_TXX_BURST, 0);
  370                 break;
  371         case 100:
  372                 port_cfg &= ~GMI_PORT_CFG_SPEED; /* speed 0 */
  373                 port_cfg &= ~GMI_PORT_CFG_SPEED_MSB; /* speed_msb 0 */
  374                 port_cfg &= ~GMI_PORT_CFG_SLOT_TIME; /* slottime 0 */
  375                 misc_ctl &= ~PCS_MISC_CTL_SAMP_PT_MASK;
  376                 misc_ctl |= 5; /* samp_pt */
  377                 bgx_reg_write(bgx, lmac->lmacid, BGX_GMP_GMI_TXX_SLOT, 64);
  378                 bgx_reg_write(bgx, lmac->lmacid, BGX_GMP_GMI_TXX_BURST, 0);
  379                 break;
  380         case 1000:
  381                 port_cfg |= GMI_PORT_CFG_SPEED; /* speed 1 */
  382                 port_cfg &= ~GMI_PORT_CFG_SPEED_MSB; /* speed_msb 0 */
  383                 port_cfg |= GMI_PORT_CFG_SLOT_TIME; /* slottime 1 */
  384                 misc_ctl &= ~PCS_MISC_CTL_SAMP_PT_MASK;
  385                 misc_ctl |= 1; /* samp_pt */
  386                 bgx_reg_write(bgx, lmac->lmacid, BGX_GMP_GMI_TXX_SLOT, 512);
  387                 if (lmac->last_duplex)
  388                         bgx_reg_write(bgx, lmac->lmacid,
  389                                       BGX_GMP_GMI_TXX_BURST, 0);
  390                 else
  391                         bgx_reg_write(bgx, lmac->lmacid,
  392                                       BGX_GMP_GMI_TXX_BURST, 8192);
  393                 break;
  394         default:
  395                 break;
  396         }
  397         bgx_reg_write(bgx, lmac->lmacid, BGX_GMP_PCS_MISCX_CTL, misc_ctl);
  398         bgx_reg_write(bgx, lmac->lmacid, BGX_GMP_GMI_PRTX_CFG, port_cfg);
  399 
  400         port_cfg = bgx_reg_read(bgx, lmac->lmacid, BGX_GMP_GMI_PRTX_CFG);
  401 
  402         /* renable lmac */
  403         cmr_cfg |= CMR_EN;
  404         bgx_reg_write(bgx, lmac->lmacid, BGX_CMRX_CFG, cmr_cfg);
  405 }
  406 
  407 static void
  408 bgx_lmac_handler(void *arg)
  409 {
  410         struct lmac *lmac;
  411         int link, duplex, speed;
  412         int link_changed = 0;
  413         int err;
  414 
  415         lmac = (struct lmac *)arg;
  416 
  417         err = LMAC_MEDIA_STATUS(lmac->phy_if_dev, lmac->lmacid,
  418             &link, &duplex, &speed);
  419         if (err != 0)
  420                 goto out;
  421 
  422         if (!link && lmac->last_link)
  423                 link_changed = -1;
  424 
  425         if (link &&
  426             (lmac->last_duplex != duplex ||
  427              lmac->last_link != link ||
  428              lmac->last_speed != speed)) {
  429                         link_changed = 1;
  430         }
  431 
  432         lmac->last_link = link;
  433         lmac->last_speed = speed;
  434         lmac->last_duplex = duplex;
  435 
  436         if (!link_changed)
  437                 goto out;
  438 
  439         if (link_changed > 0)
  440                 lmac->link_up = true;
  441         else
  442                 lmac->link_up = false;
  443 
  444         if (lmac->is_sgmii)
  445                 bgx_sgmii_change_link_state(lmac);
  446         else
  447                 bgx_xaui_check_link(lmac);
  448 
  449 out:
  450         callout_reset(&lmac->check_link, hz * 2, bgx_lmac_handler, lmac);
  451 }
  452 
  453 uint64_t
  454 bgx_get_rx_stats(int node, int bgx_idx, int lmac, int idx)
  455 {
  456         struct bgx *bgx;
  457 
  458         bgx = bgx_vnic[(node * MAX_BGX_PER_CN88XX) + bgx_idx];
  459         if (bgx == NULL)
  460                 return (0);
  461 
  462         if (idx > 8)
  463                 lmac = (0);
  464         return (bgx_reg_read(bgx, lmac, BGX_CMRX_RX_STAT0 + (idx * 8)));
  465 }
  466 
  467 uint64_t
  468 bgx_get_tx_stats(int node, int bgx_idx, int lmac, int idx)
  469 {
  470         struct bgx *bgx;
  471 
  472         bgx = bgx_vnic[(node * MAX_BGX_PER_CN88XX) + bgx_idx];
  473         if (bgx == NULL)
  474                 return (0);
  475 
  476         return (bgx_reg_read(bgx, lmac, BGX_CMRX_TX_STAT0 + (idx * 8)));
  477 }
  478 
  479 static void
  480 bgx_flush_dmac_addrs(struct bgx *bgx, int lmac)
  481 {
  482         uint64_t offset;
  483 
  484         while (bgx->lmac[lmac].dmac > 0) {
  485                 offset = ((bgx->lmac[lmac].dmac - 1) * sizeof(uint64_t)) +
  486                     (lmac * MAX_DMAC_PER_LMAC * sizeof(uint64_t));
  487                 bgx_reg_write(bgx, 0, BGX_CMR_RX_DMACX_CAM + offset, 0);
  488                 bgx->lmac[lmac].dmac--;
  489         }
  490 }
  491 
  492 void
  493 bgx_add_dmac_addr(uint64_t dmac, int node, int bgx_idx, int lmac)
  494 {
  495         uint64_t offset;
  496         struct bgx *bgx;
  497 
  498 #ifdef BGX_IN_PROMISCUOUS_MODE
  499         return;
  500 #endif
  501 
  502         bgx_idx += node * MAX_BGX_PER_CN88XX;
  503         bgx = bgx_vnic[bgx_idx];
  504 
  505         if (bgx == NULL) {
  506                 printf("BGX%d not yet initialized, ignoring DMAC addition\n",
  507                     bgx_idx);
  508                 return;
  509         }
  510 
  511         dmac = dmac | (1UL << 48) | ((uint64_t)lmac << 49); /* Enable DMAC */
  512         if (bgx->lmac[lmac].dmac == MAX_DMAC_PER_LMAC) {
  513                 device_printf(bgx->dev,
  514                     "Max DMAC filters for LMAC%d reached, ignoring\n",
  515                     lmac);
  516                 return;
  517         }
  518 
  519         if (bgx->lmac[lmac].dmac == MAX_DMAC_PER_LMAC_TNS_BYPASS_MODE)
  520                 bgx->lmac[lmac].dmac = 1;
  521 
  522         offset = (bgx->lmac[lmac].dmac * sizeof(uint64_t)) +
  523             (lmac * MAX_DMAC_PER_LMAC * sizeof(uint64_t));
  524         bgx_reg_write(bgx, 0, BGX_CMR_RX_DMACX_CAM + offset, dmac);
  525         bgx->lmac[lmac].dmac++;
  526 
  527         bgx_reg_write(bgx, lmac, BGX_CMRX_RX_DMAC_CTL,
  528             (CAM_ACCEPT << 3) | (MCAST_MODE_CAM_FILTER << 1) |
  529             (BCAST_ACCEPT << 0));
  530 }
  531 
  532 /* Configure BGX LMAC in internal loopback mode */
  533 void
  534 bgx_lmac_internal_loopback(int node, int bgx_idx,
  535     int lmac_idx, boolean_t enable)
  536 {
  537         struct bgx *bgx;
  538         struct lmac *lmac;
  539         uint64_t cfg;
  540 
  541         bgx = bgx_vnic[(node * MAX_BGX_PER_CN88XX) + bgx_idx];
  542         if (bgx == NULL)
  543                 return;
  544 
  545         lmac = &bgx->lmac[lmac_idx];
  546         if (lmac->is_sgmii) {
  547                 cfg = bgx_reg_read(bgx, lmac_idx, BGX_GMP_PCS_MRX_CTL);
  548                 if (enable)
  549                         cfg |= PCS_MRX_CTL_LOOPBACK1;
  550                 else
  551                         cfg &= ~PCS_MRX_CTL_LOOPBACK1;
  552                 bgx_reg_write(bgx, lmac_idx, BGX_GMP_PCS_MRX_CTL, cfg);
  553         } else {
  554                 cfg = bgx_reg_read(bgx, lmac_idx, BGX_SPUX_CONTROL1);
  555                 if (enable)
  556                         cfg |= SPU_CTL_LOOPBACK;
  557                 else
  558                         cfg &= ~SPU_CTL_LOOPBACK;
  559                 bgx_reg_write(bgx, lmac_idx, BGX_SPUX_CONTROL1, cfg);
  560         }
  561 }
  562 
  563 static int
  564 bgx_lmac_sgmii_init(struct bgx *bgx, int lmacid)
  565 {
  566         uint64_t cfg;
  567 
  568         bgx_reg_modify(bgx, lmacid, BGX_GMP_GMI_TXX_THRESH, 0x30);
  569         /* max packet size */
  570         bgx_reg_modify(bgx, lmacid, BGX_GMP_GMI_RXX_JABBER, MAX_FRAME_SIZE);
  571 
  572         /* Disable frame alignment if using preamble */
  573         cfg = bgx_reg_read(bgx, lmacid, BGX_GMP_GMI_TXX_APPEND);
  574         if (cfg & 1)
  575                 bgx_reg_write(bgx, lmacid, BGX_GMP_GMI_TXX_SGMII_CTL, 0);
  576 
  577         /* Enable lmac */
  578         bgx_reg_modify(bgx, lmacid, BGX_CMRX_CFG, CMR_EN);
  579 
  580         /* PCS reset */
  581         bgx_reg_modify(bgx, lmacid, BGX_GMP_PCS_MRX_CTL, PCS_MRX_CTL_RESET);
  582         if (bgx_poll_reg(bgx, lmacid, BGX_GMP_PCS_MRX_CTL,
  583             PCS_MRX_CTL_RESET, TRUE) != 0) {
  584                 device_printf(bgx->dev, "BGX PCS reset not completed\n");
  585                 return (ENXIO);
  586         }
  587 
  588         /* power down, reset autoneg, autoneg enable */
  589         cfg = bgx_reg_read(bgx, lmacid, BGX_GMP_PCS_MRX_CTL);
  590         cfg &= ~PCS_MRX_CTL_PWR_DN;
  591         cfg |= (PCS_MRX_CTL_RST_AN | PCS_MRX_CTL_AN_EN);
  592         bgx_reg_write(bgx, lmacid, BGX_GMP_PCS_MRX_CTL, cfg);
  593 
  594         if (bgx_poll_reg(bgx, lmacid, BGX_GMP_PCS_MRX_STATUS,
  595             PCS_MRX_STATUS_AN_CPT, FALSE) != 0) {
  596                 device_printf(bgx->dev, "BGX AN_CPT not completed\n");
  597                 return (ENXIO);
  598         }
  599 
  600         return (0);
  601 }
  602 
  603 static int
  604 bgx_lmac_xaui_init(struct bgx *bgx, int lmacid, int lmac_type)
  605 {
  606         uint64_t cfg;
  607 
  608         /* Reset SPU */
  609         bgx_reg_modify(bgx, lmacid, BGX_SPUX_CONTROL1, SPU_CTL_RESET);
  610         if (bgx_poll_reg(bgx, lmacid, BGX_SPUX_CONTROL1,
  611             SPU_CTL_RESET, TRUE) != 0) {
  612                 device_printf(bgx->dev, "BGX SPU reset not completed\n");
  613                 return (ENXIO);
  614         }
  615 
  616         /* Disable LMAC */
  617         cfg = bgx_reg_read(bgx, lmacid, BGX_CMRX_CFG);
  618         cfg &= ~CMR_EN;
  619         bgx_reg_write(bgx, lmacid, BGX_CMRX_CFG, cfg);
  620 
  621         bgx_reg_modify(bgx, lmacid, BGX_SPUX_CONTROL1, SPU_CTL_LOW_POWER);
  622         /* Set interleaved running disparity for RXAUI */
  623         if (bgx->lmac_type != BGX_MODE_RXAUI) {
  624                 bgx_reg_modify(bgx, lmacid,
  625                     BGX_SPUX_MISC_CONTROL, SPU_MISC_CTL_RX_DIS);
  626         } else {
  627                 bgx_reg_modify(bgx, lmacid, BGX_SPUX_MISC_CONTROL,
  628                     SPU_MISC_CTL_RX_DIS | SPU_MISC_CTL_INTLV_RDISP);
  629         }
  630 
  631         /* clear all interrupts */
  632         cfg = bgx_reg_read(bgx, lmacid, BGX_SMUX_RX_INT);
  633         bgx_reg_write(bgx, lmacid, BGX_SMUX_RX_INT, cfg);
  634         cfg = bgx_reg_read(bgx, lmacid, BGX_SMUX_TX_INT);
  635         bgx_reg_write(bgx, lmacid, BGX_SMUX_TX_INT, cfg);
  636         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_INT);
  637         bgx_reg_write(bgx, lmacid, BGX_SPUX_INT, cfg);
  638 
  639         if (bgx->use_training) {
  640                 bgx_reg_write(bgx, lmacid, BGX_SPUX_BR_PMD_LP_CUP, 0x00);
  641                 bgx_reg_write(bgx, lmacid, BGX_SPUX_BR_PMD_LD_CUP, 0x00);
  642                 bgx_reg_write(bgx, lmacid, BGX_SPUX_BR_PMD_LD_REP, 0x00);
  643                 /* training enable */
  644                 bgx_reg_modify(bgx, lmacid, BGX_SPUX_BR_PMD_CRTL,
  645                     SPU_PMD_CRTL_TRAIN_EN);
  646         }
  647 
  648         /* Append FCS to each packet */
  649         bgx_reg_modify(bgx, lmacid, BGX_SMUX_TX_APPEND, SMU_TX_APPEND_FCS_D);
  650 
  651         /* Disable forward error correction */
  652         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_FEC_CONTROL);
  653         cfg &= ~SPU_FEC_CTL_FEC_EN;
  654         bgx_reg_write(bgx, lmacid, BGX_SPUX_FEC_CONTROL, cfg);
  655 
  656         /* Disable autoneg */
  657         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_AN_CONTROL);
  658         cfg = cfg & ~(SPU_AN_CTL_AN_EN | SPU_AN_CTL_XNP_EN);
  659         bgx_reg_write(bgx, lmacid, BGX_SPUX_AN_CONTROL, cfg);
  660 
  661         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_AN_ADV);
  662         if (bgx->lmac_type == BGX_MODE_10G_KR)
  663                 cfg |= (1 << 23);
  664         else if (bgx->lmac_type == BGX_MODE_40G_KR)
  665                 cfg |= (1 << 24);
  666         else
  667                 cfg &= ~((1 << 23) | (1 << 24));
  668         cfg = cfg & (~((1UL << 25) | (1UL << 22) | (1UL << 12)));
  669         bgx_reg_write(bgx, lmacid, BGX_SPUX_AN_ADV, cfg);
  670 
  671         cfg = bgx_reg_read(bgx, 0, BGX_SPU_DBG_CONTROL);
  672         cfg &= ~SPU_DBG_CTL_AN_ARB_LINK_CHK_EN;
  673         bgx_reg_write(bgx, 0, BGX_SPU_DBG_CONTROL, cfg);
  674 
  675         /* Enable lmac */
  676         bgx_reg_modify(bgx, lmacid, BGX_CMRX_CFG, CMR_EN);
  677 
  678         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_CONTROL1);
  679         cfg &= ~SPU_CTL_LOW_POWER;
  680         bgx_reg_write(bgx, lmacid, BGX_SPUX_CONTROL1, cfg);
  681 
  682         cfg = bgx_reg_read(bgx, lmacid, BGX_SMUX_TX_CTL);
  683         cfg &= ~SMU_TX_CTL_UNI_EN;
  684         cfg |= SMU_TX_CTL_DIC_EN;
  685         bgx_reg_write(bgx, lmacid, BGX_SMUX_TX_CTL, cfg);
  686 
  687         /* take lmac_count into account */
  688         bgx_reg_modify(bgx, lmacid, BGX_SMUX_TX_THRESH, (0x100 - 1));
  689         /* max packet size */
  690         bgx_reg_modify(bgx, lmacid, BGX_SMUX_RX_JABBER, MAX_FRAME_SIZE);
  691 
  692         return (0);
  693 }
  694 
  695 static int
  696 bgx_xaui_check_link(struct lmac *lmac)
  697 {
  698         struct bgx *bgx = lmac->bgx;
  699         int lmacid = lmac->lmacid;
  700         int lmac_type = bgx->lmac_type;
  701         uint64_t cfg;
  702 
  703         bgx_reg_modify(bgx, lmacid, BGX_SPUX_MISC_CONTROL, SPU_MISC_CTL_RX_DIS);
  704         if (bgx->use_training) {
  705                 cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_INT);
  706                 if ((cfg & (1UL << 13)) == 0) {
  707                         cfg = (1UL << 13) | (1UL << 14);
  708                         bgx_reg_write(bgx, lmacid, BGX_SPUX_INT, cfg);
  709                         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_BR_PMD_CRTL);
  710                         cfg |= (1UL << 0);
  711                         bgx_reg_write(bgx, lmacid, BGX_SPUX_BR_PMD_CRTL, cfg);
  712                         return (ENXIO);
  713                 }
  714         }
  715 
  716         /* wait for PCS to come out of reset */
  717         if (bgx_poll_reg(bgx, lmacid, BGX_SPUX_CONTROL1,
  718             SPU_CTL_RESET, TRUE) != 0) {
  719                 device_printf(bgx->dev, "BGX SPU reset not completed\n");
  720                 return (ENXIO);
  721         }
  722 
  723         if ((lmac_type == BGX_MODE_10G_KR) || (lmac_type == BGX_MODE_XFI) ||
  724             (lmac_type == BGX_MODE_40G_KR) || (lmac_type == BGX_MODE_XLAUI)) {
  725                 if (bgx_poll_reg(bgx, lmacid, BGX_SPUX_BR_STATUS1,
  726                     SPU_BR_STATUS_BLK_LOCK, FALSE)) {
  727                         device_printf(bgx->dev,
  728                             "SPU_BR_STATUS_BLK_LOCK not completed\n");
  729                         return (ENXIO);
  730                 }
  731         } else {
  732                 if (bgx_poll_reg(bgx, lmacid, BGX_SPUX_BX_STATUS,
  733                     SPU_BX_STATUS_RX_ALIGN, FALSE) != 0) {
  734                         device_printf(bgx->dev,
  735                             "SPU_BX_STATUS_RX_ALIGN not completed\n");
  736                         return (ENXIO);
  737                 }
  738         }
  739 
  740         /* Clear rcvflt bit (latching high) and read it back */
  741         bgx_reg_modify(bgx, lmacid, BGX_SPUX_STATUS2, SPU_STATUS2_RCVFLT);
  742         if (bgx_reg_read(bgx, lmacid, BGX_SPUX_STATUS2) & SPU_STATUS2_RCVFLT) {
  743                 device_printf(bgx->dev, "Receive fault, retry training\n");
  744                 if (bgx->use_training) {
  745                         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_INT);
  746                         if ((cfg & (1UL << 13)) == 0) {
  747                                 cfg = (1UL << 13) | (1UL << 14);
  748                                 bgx_reg_write(bgx, lmacid, BGX_SPUX_INT, cfg);
  749                                 cfg = bgx_reg_read(bgx, lmacid,
  750                                     BGX_SPUX_BR_PMD_CRTL);
  751                                 cfg |= (1UL << 0);
  752                                 bgx_reg_write(bgx, lmacid,
  753                                     BGX_SPUX_BR_PMD_CRTL, cfg);
  754                                 return (ENXIO);
  755                         }
  756                 }
  757                 return (ENXIO);
  758         }
  759 
  760         /* Wait for MAC RX to be ready */
  761         if (bgx_poll_reg(bgx, lmacid, BGX_SMUX_RX_CTL,
  762             SMU_RX_CTL_STATUS, TRUE) != 0) {
  763                 device_printf(bgx->dev, "SMU RX link not okay\n");
  764                 return (ENXIO);
  765         }
  766 
  767         /* Wait for BGX RX to be idle */
  768         if (bgx_poll_reg(bgx, lmacid, BGX_SMUX_CTL,
  769             SMU_CTL_RX_IDLE, FALSE) != 0) {
  770                 device_printf(bgx->dev, "SMU RX not idle\n");
  771                 return (ENXIO);
  772         }
  773 
  774         /* Wait for BGX TX to be idle */
  775         if (bgx_poll_reg(bgx, lmacid, BGX_SMUX_CTL,
  776             SMU_CTL_TX_IDLE, FALSE) != 0) {
  777                 device_printf(bgx->dev, "SMU TX not idle\n");
  778                 return (ENXIO);
  779         }
  780 
  781         if ((bgx_reg_read(bgx, lmacid, BGX_SPUX_STATUS2) &
  782             SPU_STATUS2_RCVFLT) != 0) {
  783                 device_printf(bgx->dev, "Receive fault\n");
  784                 return (ENXIO);
  785         }
  786 
  787         /* Receive link is latching low. Force it high and verify it */
  788         bgx_reg_modify(bgx, lmacid, BGX_SPUX_STATUS1, SPU_STATUS1_RCV_LNK);
  789         if (bgx_poll_reg(bgx, lmacid, BGX_SPUX_STATUS1,
  790             SPU_STATUS1_RCV_LNK, FALSE) != 0) {
  791                 device_printf(bgx->dev, "SPU receive link down\n");
  792                 return (ENXIO);
  793         }
  794 
  795         cfg = bgx_reg_read(bgx, lmacid, BGX_SPUX_MISC_CONTROL);
  796         cfg &= ~SPU_MISC_CTL_RX_DIS;
  797         bgx_reg_write(bgx, lmacid, BGX_SPUX_MISC_CONTROL, cfg);
  798         return (0);
  799 }
  800 
  801 static void
  802 bgx_poll_for_link(void *arg)
  803 {
  804         struct lmac *lmac;
  805         uint64_t link;
  806 
  807         lmac = (struct lmac *)arg;
  808 
  809         /* Receive link is latching low. Force it high and verify it */
  810         bgx_reg_modify(lmac->bgx, lmac->lmacid,
  811                        BGX_SPUX_STATUS1, SPU_STATUS1_RCV_LNK);
  812         bgx_poll_reg(lmac->bgx, lmac->lmacid, BGX_SPUX_STATUS1,
  813                      SPU_STATUS1_RCV_LNK, false);
  814 
  815         link = bgx_reg_read(lmac->bgx, lmac->lmacid, BGX_SPUX_STATUS1);
  816         if (link & SPU_STATUS1_RCV_LNK) {
  817                 lmac->link_up = 1;
  818                 if (lmac->bgx->lmac_type == BGX_MODE_XLAUI)
  819                         lmac->last_speed = 40000;
  820                 else
  821                         lmac->last_speed = 10000;
  822                 lmac->last_duplex = 1;
  823         } else {
  824                 lmac->link_up = 0;
  825         }
  826 
  827         if (lmac->last_link != lmac->link_up) {
  828                 lmac->last_link = lmac->link_up;
  829                 if (lmac->link_up)
  830                         bgx_xaui_check_link(lmac);
  831         }
  832 
  833         callout_reset(&lmac->check_link, hz * 2, bgx_poll_for_link, lmac);
  834 }
  835 
  836 static int
  837 bgx_lmac_enable(struct bgx *bgx, uint8_t lmacid)
  838 {
  839         uint64_t __unused dmac_bcast = (1UL << 48) - 1;
  840         struct lmac *lmac;
  841         uint64_t cfg;
  842 
  843         lmac = &bgx->lmac[lmacid];
  844         lmac->bgx = bgx;
  845 
  846         if (bgx->lmac_type == BGX_MODE_SGMII) {
  847                 lmac->is_sgmii = 1;
  848                 if (bgx_lmac_sgmii_init(bgx, lmacid) != 0)
  849                         return -1;
  850         } else {
  851                 lmac->is_sgmii = 0;
  852                 if (bgx_lmac_xaui_init(bgx, lmacid, bgx->lmac_type))
  853                         return -1;
  854         }
  855 
  856         if (lmac->is_sgmii) {
  857                 cfg = bgx_reg_read(bgx, lmacid, BGX_GMP_GMI_TXX_APPEND);
  858                 cfg |= ((1UL << 2) | (1UL << 1)); /* FCS and PAD */
  859                 bgx_reg_modify(bgx, lmacid, BGX_GMP_GMI_TXX_APPEND, cfg);
  860                 bgx_reg_write(bgx, lmacid, BGX_GMP_GMI_TXX_MIN_PKT, 60 - 1);
  861         } else {
  862                 cfg = bgx_reg_read(bgx, lmacid, BGX_SMUX_TX_APPEND);
  863                 cfg |= ((1UL << 2) | (1UL << 1)); /* FCS and PAD */
  864                 bgx_reg_modify(bgx, lmacid, BGX_SMUX_TX_APPEND, cfg);
  865                 bgx_reg_write(bgx, lmacid, BGX_SMUX_TX_MIN_PKT, 60 + 4);
  866         }
  867 
  868         /* Enable lmac */
  869         bgx_reg_modify(bgx, lmacid, BGX_CMRX_CFG,
  870                        CMR_EN | CMR_PKT_RX_EN | CMR_PKT_TX_EN);
  871 
  872         /* Restore default cfg, incase low level firmware changed it */
  873         bgx_reg_write(bgx, lmacid, BGX_CMRX_RX_DMAC_CTL, 0x03);
  874 
  875         /* Add broadcast MAC into all LMAC's DMAC filters */
  876         bgx_add_dmac_addr(dmac_bcast, 0, bgx->bgx_id, lmacid);
  877 
  878         if ((bgx->lmac_type != BGX_MODE_XFI) &&
  879             (bgx->lmac_type != BGX_MODE_XAUI) &&
  880             (bgx->lmac_type != BGX_MODE_XLAUI) &&
  881             (bgx->lmac_type != BGX_MODE_40G_KR) &&
  882             (bgx->lmac_type != BGX_MODE_10G_KR)) {
  883                 if (lmac->phy_if_dev == NULL) {
  884                         device_printf(bgx->dev,
  885                             "LMAC%d missing interface to PHY\n", lmacid);
  886                         return (ENXIO);
  887                 }
  888 
  889                 if (LMAC_PHY_CONNECT(lmac->phy_if_dev, lmac->phyaddr,
  890                     lmacid) != 0) {
  891                         device_printf(bgx->dev,
  892                             "LMAC%d could not connect to PHY\n", lmacid);
  893                         return (ENXIO);
  894                 }
  895                 mtx_init(&lmac->check_link_mtx, "BGX link poll", NULL, MTX_DEF);
  896                 callout_init_mtx(&lmac->check_link, &lmac->check_link_mtx, 0);
  897                 mtx_lock(&lmac->check_link_mtx);
  898                 bgx_lmac_handler(lmac);
  899                 mtx_unlock(&lmac->check_link_mtx);
  900         } else {
  901                 mtx_init(&lmac->check_link_mtx, "BGX link poll", NULL, MTX_DEF);
  902                 callout_init_mtx(&lmac->check_link, &lmac->check_link_mtx, 0);
  903                 mtx_lock(&lmac->check_link_mtx);
  904                 bgx_poll_for_link(lmac);
  905                 mtx_unlock(&lmac->check_link_mtx);
  906         }
  907 
  908         return (0);
  909 }
  910 
  911 static void
  912 bgx_lmac_disable(struct bgx *bgx, uint8_t lmacid)
  913 {
  914         struct lmac *lmac;
  915         uint64_t cmrx_cfg;
  916 
  917         lmac = &bgx->lmac[lmacid];
  918 
  919         /* Stop callout */
  920         callout_drain(&lmac->check_link);
  921         mtx_destroy(&lmac->check_link_mtx);
  922 
  923         cmrx_cfg = bgx_reg_read(bgx, lmacid, BGX_CMRX_CFG);
  924         cmrx_cfg &= ~(1 << 15);
  925         bgx_reg_write(bgx, lmacid, BGX_CMRX_CFG, cmrx_cfg);
  926         bgx_flush_dmac_addrs(bgx, lmacid);
  927 
  928         if ((bgx->lmac_type != BGX_MODE_XFI) &&
  929             (bgx->lmac_type != BGX_MODE_XLAUI) &&
  930             (bgx->lmac_type != BGX_MODE_40G_KR) &&
  931             (bgx->lmac_type != BGX_MODE_10G_KR)) {
  932                 if (lmac->phy_if_dev == NULL) {
  933                         device_printf(bgx->dev,
  934                             "LMAC%d missing interface to PHY\n", lmacid);
  935                         return;
  936                 }
  937                 if (LMAC_PHY_DISCONNECT(lmac->phy_if_dev, lmac->phyaddr,
  938                     lmacid) != 0) {
  939                         device_printf(bgx->dev,
  940                             "LMAC%d could not disconnect PHY\n", lmacid);
  941                         return;
  942                 }
  943                 lmac->phy_if_dev = NULL;
  944         }
  945 }
  946 
  947 static void
  948 bgx_set_num_ports(struct bgx *bgx)
  949 {
  950         uint64_t lmac_count;
  951 
  952         switch (bgx->qlm_mode) {
  953         case QLM_MODE_SGMII:
  954                 bgx->lmac_count = 4;
  955                 bgx->lmac_type = BGX_MODE_SGMII;
  956                 bgx->lane_to_sds = 0;
  957                 break;
  958         case QLM_MODE_XAUI_1X4:
  959                 bgx->lmac_count = 1;
  960                 bgx->lmac_type = BGX_MODE_XAUI;
  961                 bgx->lane_to_sds = 0xE4;
  962                         break;
  963         case QLM_MODE_RXAUI_2X2:
  964                 bgx->lmac_count = 2;
  965                 bgx->lmac_type = BGX_MODE_RXAUI;
  966                 bgx->lane_to_sds = 0xE4;
  967                         break;
  968         case QLM_MODE_XFI_4X1:
  969                 bgx->lmac_count = 4;
  970                 bgx->lmac_type = BGX_MODE_XFI;
  971                 bgx->lane_to_sds = 0;
  972                 break;
  973         case QLM_MODE_XLAUI_1X4:
  974                 bgx->lmac_count = 1;
  975                 bgx->lmac_type = BGX_MODE_XLAUI;
  976                 bgx->lane_to_sds = 0xE4;
  977                 break;
  978         case QLM_MODE_10G_KR_4X1:
  979                 bgx->lmac_count = 4;
  980                 bgx->lmac_type = BGX_MODE_10G_KR;
  981                 bgx->lane_to_sds = 0;
  982                 bgx->use_training = 1;
  983                 break;
  984         case QLM_MODE_40G_KR4_1X4:
  985                 bgx->lmac_count = 1;
  986                 bgx->lmac_type = BGX_MODE_40G_KR;
  987                 bgx->lane_to_sds = 0xE4;
  988                 bgx->use_training = 1;
  989                 break;
  990         default:
  991                 bgx->lmac_count = 0;
  992                 break;
  993         }
  994 
  995         /*
  996          * Check if low level firmware has programmed LMAC count
  997          * based on board type, if yes consider that otherwise
  998          * the default static values
  999          */
 1000         lmac_count = bgx_reg_read(bgx, 0, BGX_CMR_RX_LMACS) & 0x7;
 1001         if (lmac_count != 4)
 1002                 bgx->lmac_count = lmac_count;
 1003 }
 1004 
 1005 static void
 1006 bgx_init_hw(struct bgx *bgx)
 1007 {
 1008         int i;
 1009 
 1010         bgx_set_num_ports(bgx);
 1011 
 1012         bgx_reg_modify(bgx, 0, BGX_CMR_GLOBAL_CFG, CMR_GLOBAL_CFG_FCS_STRIP);
 1013         if (bgx_reg_read(bgx, 0, BGX_CMR_BIST_STATUS))
 1014                 device_printf(bgx->dev, "BGX%d BIST failed\n", bgx->bgx_id);
 1015 
 1016         /* Set lmac type and lane2serdes mapping */
 1017         for (i = 0; i < bgx->lmac_count; i++) {
 1018                 if (bgx->lmac_type == BGX_MODE_RXAUI) {
 1019                         if (i)
 1020                                 bgx->lane_to_sds = 0x0e;
 1021                         else
 1022                                 bgx->lane_to_sds = 0x04;
 1023                         bgx_reg_write(bgx, i, BGX_CMRX_CFG,
 1024                             (bgx->lmac_type << 8) | bgx->lane_to_sds);
 1025                         continue;
 1026                 }
 1027                 bgx_reg_write(bgx, i, BGX_CMRX_CFG,
 1028                     (bgx->lmac_type << 8) | (bgx->lane_to_sds + i));
 1029                 bgx->lmac[i].lmacid_bd = lmac_count;
 1030                 lmac_count++;
 1031         }
 1032 
 1033         bgx_reg_write(bgx, 0, BGX_CMR_TX_LMACS, bgx->lmac_count);
 1034         bgx_reg_write(bgx, 0, BGX_CMR_RX_LMACS, bgx->lmac_count);
 1035 
 1036         /* Set the backpressure AND mask */
 1037         for (i = 0; i < bgx->lmac_count; i++) {
 1038                 bgx_reg_modify(bgx, 0, BGX_CMR_CHAN_MSK_AND,
 1039                     ((1UL << MAX_BGX_CHANS_PER_LMAC) - 1) <<
 1040                     (i * MAX_BGX_CHANS_PER_LMAC));
 1041         }
 1042 
 1043         /* Disable all MAC filtering */
 1044         for (i = 0; i < RX_DMAC_COUNT; i++)
 1045                 bgx_reg_write(bgx, 0, BGX_CMR_RX_DMACX_CAM + (i * 8), 0x00);
 1046 
 1047         /* Disable MAC steering (NCSI traffic) */
 1048         for (i = 0; i < RX_TRAFFIC_STEER_RULE_COUNT; i++)
 1049                 bgx_reg_write(bgx, 0, BGX_CMR_RX_STREERING + (i * 8), 0x00);
 1050 }
 1051 
 1052 static void
 1053 bgx_get_qlm_mode(struct bgx *bgx)
 1054 {
 1055         device_t dev = bgx->dev;;
 1056         int lmac_type;
 1057         int train_en;
 1058 
 1059         /* Read LMAC0 type to figure out QLM mode
 1060          * This is configured by low level firmware
 1061          */
 1062         lmac_type = bgx_reg_read(bgx, 0, BGX_CMRX_CFG);
 1063         lmac_type = (lmac_type >> 8) & 0x07;
 1064 
 1065         train_en = bgx_reg_read(bgx, 0, BGX_SPUX_BR_PMD_CRTL) &
 1066             SPU_PMD_CRTL_TRAIN_EN;
 1067 
 1068         switch (lmac_type) {
 1069         case BGX_MODE_SGMII:
 1070                 bgx->qlm_mode = QLM_MODE_SGMII;
 1071                 if (bootverbose) {
 1072                         device_printf(dev, "BGX%d QLM mode: SGMII\n",
 1073                             bgx->bgx_id);
 1074                 }
 1075                 break;
 1076         case BGX_MODE_XAUI:
 1077                 bgx->qlm_mode = QLM_MODE_XAUI_1X4;
 1078                 if (bootverbose) {
 1079                         device_printf(dev, "BGX%d QLM mode: XAUI\n",
 1080                             bgx->bgx_id);
 1081                 }
 1082                 break;
 1083         case BGX_MODE_RXAUI:
 1084                 bgx->qlm_mode = QLM_MODE_RXAUI_2X2;
 1085                 if (bootverbose) {
 1086                         device_printf(dev, "BGX%d QLM mode: RXAUI\n",
 1087                             bgx->bgx_id);
 1088                 }
 1089                 break;
 1090         case BGX_MODE_XFI:
 1091                 if (!train_en) {
 1092                         bgx->qlm_mode = QLM_MODE_XFI_4X1;
 1093                         if (bootverbose) {
 1094                                 device_printf(dev, "BGX%d QLM mode: XFI\n",
 1095                                     bgx->bgx_id);
 1096                         }
 1097                 } else {
 1098                         bgx->qlm_mode = QLM_MODE_10G_KR_4X1;
 1099                         if (bootverbose) {
 1100                                 device_printf(dev, "BGX%d QLM mode: 10G_KR\n",
 1101                                     bgx->bgx_id);
 1102                         }
 1103                 }
 1104                 break;
 1105         case BGX_MODE_XLAUI:
 1106                 if (!train_en) {
 1107                         bgx->qlm_mode = QLM_MODE_XLAUI_1X4;
 1108                         if (bootverbose) {
 1109                                 device_printf(dev, "BGX%d QLM mode: XLAUI\n",
 1110                                     bgx->bgx_id);
 1111                         }
 1112                 } else {
 1113                         bgx->qlm_mode = QLM_MODE_40G_KR4_1X4;
 1114                         if (bootverbose) {
 1115                                 device_printf(dev, "BGX%d QLM mode: 40G_KR4\n",
 1116                                     bgx->bgx_id);
 1117                         }
 1118                 }
 1119                 break;
 1120         default:
 1121                 bgx->qlm_mode = QLM_MODE_SGMII;
 1122                 if (bootverbose) {
 1123                         device_printf(dev, "BGX%d QLM default mode: SGMII\n",
 1124                             bgx->bgx_id);
 1125                 }
 1126         }
 1127 }
 1128 
 1129 static int
 1130 bgx_init_phy(struct bgx *bgx)
 1131 {
 1132         int err;
 1133 
 1134         /* By default we fail */
 1135         err = ENXIO;
 1136 #ifdef FDT
 1137         err = bgx_fdt_init_phy(bgx);
 1138 #endif
 1139 #ifdef ACPI
 1140         if (err != 0) {
 1141                 /* ARM64TODO: Add ACPI function here */
 1142         }
 1143 #endif
 1144         return (err);
 1145 }

Cache object: 0370e3677a00778f2e2691e1f8675620


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]


This page is part of the FreeBSD/Linux Linux Kernel Cross-Reference, and was automatically generated using a modified version of the LXR engine.