The Design and Implementation of the FreeBSD Operating System, Second Edition
Now available: The Design and Implementation of the FreeBSD Operating System (Second Edition)


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]

FreeBSD/Linux Kernel Cross Reference
sys/contrib/dev/iwlwifi/pcie/tx.c

Version: -  FREEBSD  -  FREEBSD-13-STABLE  -  FREEBSD-13-0  -  FREEBSD-12-STABLE  -  FREEBSD-12-0  -  FREEBSD-11-STABLE  -  FREEBSD-11-0  -  FREEBSD-10-STABLE  -  FREEBSD-10-0  -  FREEBSD-9-STABLE  -  FREEBSD-9-0  -  FREEBSD-8-STABLE  -  FREEBSD-8-0  -  FREEBSD-7-STABLE  -  FREEBSD-7-0  -  FREEBSD-6-STABLE  -  FREEBSD-6-0  -  FREEBSD-5-STABLE  -  FREEBSD-5-0  -  FREEBSD-4-STABLE  -  FREEBSD-3-STABLE  -  FREEBSD22  -  l41  -  OPENBSD  -  linux-2.6  -  MK84  -  PLAN9  -  xnu-8792 
SearchContext: -  none  -  3  -  10 

    1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
    2 /*
    3  * Copyright (C) 2003-2014, 2018-2021 Intel Corporation
    4  * Copyright (C) 2013-2015 Intel Mobile Communications GmbH
    5  * Copyright (C) 2016-2017 Intel Deutschland GmbH
    6  */
    7 #include <linux/etherdevice.h>
    8 #include <linux/ieee80211.h>
    9 #include <linux/slab.h>
   10 #include <linux/sched.h>
   11 #ifdef CONFIG_INET
   12 #include <net/ip6_checksum.h>
   13 #include <net/tso.h>
   14 #endif
   15 #if defined(__FreeBSD__)
   16 #include <net/mac80211.h>
   17 #endif
   18 
   19 #include "iwl-debug.h"
   20 #include "iwl-csr.h"
   21 #include "iwl-prph.h"
   22 #include "iwl-io.h"
   23 #include "iwl-scd.h"
   24 #include "iwl-op-mode.h"
   25 #include "internal.h"
   26 #include "fw/api/tx.h"
   27 
   28 /*************** DMA-QUEUE-GENERAL-FUNCTIONS  *****
   29  * DMA services
   30  *
   31  * Theory of operation
   32  *
   33  * A Tx or Rx queue resides in host DRAM, and is comprised of a circular buffer
   34  * of buffer descriptors, each of which points to one or more data buffers for
   35  * the device to read from or fill.  Driver and device exchange status of each
   36  * queue via "read" and "write" pointers.  Driver keeps minimum of 2 empty
   37  * entries in each circular buffer, to protect against confusing empty and full
   38  * queue states.
   39  *
   40  * The device reads or writes the data in the queues via the device's several
   41  * DMA/FIFO channels.  Each queue is mapped to a single DMA channel.
   42  *
   43  * For Tx queue, there are low mark and high mark limits. If, after queuing
   44  * the packet for Tx, free space become < low mark, Tx queue stopped. When
   45  * reclaiming packets (on 'tx done IRQ), if free space become > high mark,
   46  * Tx queue resumed.
   47  *
   48  ***************************************************/
   49 
   50 
   51 int iwl_pcie_alloc_dma_ptr(struct iwl_trans *trans,
   52                            struct iwl_dma_ptr *ptr, size_t size)
   53 {
   54         if (WARN_ON(ptr->addr))
   55                 return -EINVAL;
   56 
   57         ptr->addr = dma_alloc_coherent(trans->dev, size,
   58                                        &ptr->dma, GFP_KERNEL);
   59         if (!ptr->addr)
   60                 return -ENOMEM;
   61         ptr->size = size;
   62         return 0;
   63 }
   64 
   65 void iwl_pcie_free_dma_ptr(struct iwl_trans *trans, struct iwl_dma_ptr *ptr)
   66 {
   67         if (unlikely(!ptr->addr))
   68                 return;
   69 
   70         dma_free_coherent(trans->dev, ptr->size, ptr->addr, ptr->dma);
   71         memset(ptr, 0, sizeof(*ptr));
   72 }
   73 
   74 /*
   75  * iwl_pcie_txq_inc_wr_ptr - Send new write index to hardware
   76  */
   77 static void iwl_pcie_txq_inc_wr_ptr(struct iwl_trans *trans,
   78                                     struct iwl_txq *txq)
   79 {
   80         u32 reg = 0;
   81         int txq_id = txq->id;
   82 
   83         lockdep_assert_held(&txq->lock);
   84 
   85         /*
   86          * explicitly wake up the NIC if:
   87          * 1. shadow registers aren't enabled
   88          * 2. NIC is woken up for CMD regardless of shadow outside this function
   89          * 3. there is a chance that the NIC is asleep
   90          */
   91         if (!trans->trans_cfg->base_params->shadow_reg_enable &&
   92             txq_id != trans->txqs.cmd.q_id &&
   93             test_bit(STATUS_TPOWER_PMI, &trans->status)) {
   94                 /*
   95                  * wake up nic if it's powered down ...
   96                  * uCode will wake up, and interrupt us again, so next
   97                  * time we'll skip this part.
   98                  */
   99                 reg = iwl_read32(trans, CSR_UCODE_DRV_GP1);
  100 
  101                 if (reg & CSR_UCODE_DRV_GP1_BIT_MAC_SLEEP) {
  102                         IWL_DEBUG_INFO(trans, "Tx queue %d requesting wakeup, GP1 = 0x%x\n",
  103                                        txq_id, reg);
  104                         iwl_set_bit(trans, CSR_GP_CNTRL,
  105                                     CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ);
  106                         txq->need_update = true;
  107                         return;
  108                 }
  109         }
  110 
  111         /*
  112          * if not in power-save mode, uCode will never sleep when we're
  113          * trying to tx (during RFKILL, we're not trying to tx).
  114          */
  115         IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr);
  116         if (!txq->block)
  117                 iwl_write32(trans, HBUS_TARG_WRPTR,
  118                             txq->write_ptr | (txq_id << 8));
  119 }
  120 
  121 void iwl_pcie_txq_check_wrptrs(struct iwl_trans *trans)
  122 {
  123         int i;
  124 
  125         for (i = 0; i < trans->trans_cfg->base_params->num_of_queues; i++) {
  126                 struct iwl_txq *txq = trans->txqs.txq[i];
  127 
  128                 if (!test_bit(i, trans->txqs.queue_used))
  129                         continue;
  130 
  131                 spin_lock_bh(&txq->lock);
  132                 if (txq->need_update) {
  133                         iwl_pcie_txq_inc_wr_ptr(trans, txq);
  134                         txq->need_update = false;
  135                 }
  136                 spin_unlock_bh(&txq->lock);
  137         }
  138 }
  139 
  140 static inline void iwl_pcie_tfd_set_tb(struct iwl_trans *trans, void *tfd,
  141                                        u8 idx, dma_addr_t addr, u16 len)
  142 {
  143         struct iwl_tfd *tfd_fh = (void *)tfd;
  144         struct iwl_tfd_tb *tb = &tfd_fh->tbs[idx];
  145 
  146         u16 hi_n_len = len << 4;
  147 
  148         put_unaligned_le32(addr, &tb->lo);
  149         hi_n_len |= iwl_get_dma_hi_addr(addr);
  150 
  151         tb->hi_n_len = cpu_to_le16(hi_n_len);
  152 
  153         tfd_fh->num_tbs = idx + 1;
  154 }
  155 
  156 static int iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq,
  157                                   dma_addr_t addr, u16 len, bool reset)
  158 {
  159         void *tfd;
  160         u32 num_tbs;
  161 
  162         tfd = (u8 *)txq->tfds + trans->txqs.tfd.size * txq->write_ptr;
  163 
  164         if (reset)
  165                 memset(tfd, 0, trans->txqs.tfd.size);
  166 
  167         num_tbs = iwl_txq_gen1_tfd_get_num_tbs(trans, tfd);
  168 
  169         /* Each TFD can point to a maximum max_tbs Tx buffers */
  170         if (num_tbs >= trans->txqs.tfd.max_tbs) {
  171                 IWL_ERR(trans, "Error can not send more than %d chunks\n",
  172                         trans->txqs.tfd.max_tbs);
  173                 return -EINVAL;
  174         }
  175 
  176         if (WARN(addr & ~IWL_TX_DMA_MASK,
  177                  "Unaligned address = %llx\n", (unsigned long long)addr))
  178                 return -EINVAL;
  179 
  180         iwl_pcie_tfd_set_tb(trans, tfd, num_tbs, addr, len);
  181 
  182         return num_tbs;
  183 }
  184 
  185 static void iwl_pcie_clear_cmd_in_flight(struct iwl_trans *trans)
  186 {
  187         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  188 
  189         if (!trans->trans_cfg->base_params->apmg_wake_up_wa)
  190                 return;
  191 
  192         spin_lock(&trans_pcie->reg_lock);
  193 
  194         if (WARN_ON(!trans_pcie->cmd_hold_nic_awake)) {
  195                 spin_unlock(&trans_pcie->reg_lock);
  196                 return;
  197         }
  198 
  199         trans_pcie->cmd_hold_nic_awake = false;
  200         __iwl_trans_pcie_clear_bit(trans, CSR_GP_CNTRL,
  201                                    CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ);
  202         spin_unlock(&trans_pcie->reg_lock);
  203 }
  204 
  205 /*
  206  * iwl_pcie_txq_unmap -  Unmap any remaining DMA mappings and free skb's
  207  */
  208 static void iwl_pcie_txq_unmap(struct iwl_trans *trans, int txq_id)
  209 {
  210         struct iwl_txq *txq = trans->txqs.txq[txq_id];
  211 
  212         if (!txq) {
  213                 IWL_ERR(trans, "Trying to free a queue that wasn't allocated?\n");
  214                 return;
  215         }
  216 
  217         spin_lock_bh(&txq->lock);
  218         while (txq->write_ptr != txq->read_ptr) {
  219                 IWL_DEBUG_TX_REPLY(trans, "Q %d Free %d\n",
  220                                    txq_id, txq->read_ptr);
  221 
  222                 if (txq_id != trans->txqs.cmd.q_id) {
  223                         struct sk_buff *skb = txq->entries[txq->read_ptr].skb;
  224 
  225                         if (WARN_ON_ONCE(!skb))
  226                                 continue;
  227 
  228                         iwl_txq_free_tso_page(trans, skb);
  229                 }
  230                 iwl_txq_free_tfd(trans, txq);
  231                 txq->read_ptr = iwl_txq_inc_wrap(trans, txq->read_ptr);
  232 
  233                 if (txq->read_ptr == txq->write_ptr &&
  234                     txq_id == trans->txqs.cmd.q_id)
  235                         iwl_pcie_clear_cmd_in_flight(trans);
  236         }
  237 
  238         while (!skb_queue_empty(&txq->overflow_q)) {
  239                 struct sk_buff *skb = __skb_dequeue(&txq->overflow_q);
  240 
  241                 iwl_op_mode_free_skb(trans->op_mode, skb);
  242         }
  243 
  244         spin_unlock_bh(&txq->lock);
  245 
  246         /* just in case - this queue may have been stopped */
  247         iwl_wake_queue(trans, txq);
  248 }
  249 
  250 /*
  251  * iwl_pcie_txq_free - Deallocate DMA queue.
  252  * @txq: Transmit queue to deallocate.
  253  *
  254  * Empty queue by removing and destroying all BD's.
  255  * Free all buffers.
  256  * 0-fill, but do not free "txq" descriptor structure.
  257  */
  258 static void iwl_pcie_txq_free(struct iwl_trans *trans, int txq_id)
  259 {
  260         struct iwl_txq *txq = trans->txqs.txq[txq_id];
  261         struct device *dev = trans->dev;
  262         int i;
  263 
  264         if (WARN_ON(!txq))
  265                 return;
  266 
  267         iwl_pcie_txq_unmap(trans, txq_id);
  268 
  269         /* De-alloc array of command/tx buffers */
  270         if (txq_id == trans->txqs.cmd.q_id)
  271                 for (i = 0; i < txq->n_window; i++) {
  272                         kfree_sensitive(txq->entries[i].cmd);
  273                         kfree_sensitive(txq->entries[i].free_buf);
  274                 }
  275 
  276         /* De-alloc circular buffer of TFDs */
  277         if (txq->tfds) {
  278                 dma_free_coherent(dev,
  279                                   trans->txqs.tfd.size *
  280                                   trans->trans_cfg->base_params->max_tfd_queue_size,
  281                                   txq->tfds, txq->dma_addr);
  282                 txq->dma_addr = 0;
  283                 txq->tfds = NULL;
  284 
  285                 dma_free_coherent(dev,
  286                                   sizeof(*txq->first_tb_bufs) * txq->n_window,
  287                                   txq->first_tb_bufs, txq->first_tb_dma);
  288         }
  289 
  290         kfree(txq->entries);
  291         txq->entries = NULL;
  292 
  293         del_timer_sync(&txq->stuck_timer);
  294 
  295         /* 0-fill queue descriptor structure */
  296         memset(txq, 0, sizeof(*txq));
  297 }
  298 
  299 void iwl_pcie_tx_start(struct iwl_trans *trans, u32 scd_base_addr)
  300 {
  301         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  302         int nq = trans->trans_cfg->base_params->num_of_queues;
  303         int chan;
  304         u32 reg_val;
  305         int clear_dwords = (SCD_TRANS_TBL_OFFSET_QUEUE(nq) -
  306                                 SCD_CONTEXT_MEM_LOWER_BOUND) / sizeof(u32);
  307 
  308         /* make sure all queue are not stopped/used */
  309         memset(trans->txqs.queue_stopped, 0,
  310                sizeof(trans->txqs.queue_stopped));
  311         memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used));
  312 
  313         trans_pcie->scd_base_addr =
  314                 iwl_read_prph(trans, SCD_SRAM_BASE_ADDR);
  315 
  316         WARN_ON(scd_base_addr != 0 &&
  317                 scd_base_addr != trans_pcie->scd_base_addr);
  318 
  319         /* reset context data, TX status and translation data */
  320         iwl_trans_write_mem(trans, trans_pcie->scd_base_addr +
  321                                    SCD_CONTEXT_MEM_LOWER_BOUND,
  322                             NULL, clear_dwords);
  323 
  324         iwl_write_prph(trans, SCD_DRAM_BASE_ADDR,
  325                        trans->txqs.scd_bc_tbls.dma >> 10);
  326 
  327         /* The chain extension of the SCD doesn't work well. This feature is
  328          * enabled by default by the HW, so we need to disable it manually.
  329          */
  330         if (trans->trans_cfg->base_params->scd_chain_ext_wa)
  331                 iwl_write_prph(trans, SCD_CHAINEXT_EN, 0);
  332 
  333         iwl_trans_ac_txq_enable(trans, trans->txqs.cmd.q_id,
  334                                 trans->txqs.cmd.fifo,
  335                                 trans->txqs.cmd.wdg_timeout);
  336 
  337         /* Activate all Tx DMA/FIFO channels */
  338         iwl_scd_activate_fifos(trans);
  339 
  340         /* Enable DMA channel */
  341         for (chan = 0; chan < FH_TCSR_CHNL_NUM; chan++)
  342                 iwl_write_direct32(trans, FH_TCSR_CHNL_TX_CONFIG_REG(chan),
  343                                    FH_TCSR_TX_CONFIG_REG_VAL_DMA_CHNL_ENABLE |
  344                                    FH_TCSR_TX_CONFIG_REG_VAL_DMA_CREDIT_ENABLE);
  345 
  346         /* Update FH chicken bits */
  347         reg_val = iwl_read_direct32(trans, FH_TX_CHICKEN_BITS_REG);
  348         iwl_write_direct32(trans, FH_TX_CHICKEN_BITS_REG,
  349                            reg_val | FH_TX_CHICKEN_BITS_SCD_AUTO_RETRY_EN);
  350 
  351         /* Enable L1-Active */
  352         if (trans->trans_cfg->device_family < IWL_DEVICE_FAMILY_8000)
  353                 iwl_clear_bits_prph(trans, APMG_PCIDEV_STT_REG,
  354                                     APMG_PCIDEV_STT_VAL_L1_ACT_DIS);
  355 }
  356 
  357 void iwl_trans_pcie_tx_reset(struct iwl_trans *trans)
  358 {
  359         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  360         int txq_id;
  361 
  362         /*
  363          * we should never get here in gen2 trans mode return early to avoid
  364          * having invalid accesses
  365          */
  366         if (WARN_ON_ONCE(trans->trans_cfg->gen2))
  367                 return;
  368 
  369         for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues;
  370              txq_id++) {
  371                 struct iwl_txq *txq = trans->txqs.txq[txq_id];
  372                 if (trans->trans_cfg->use_tfh)
  373                         iwl_write_direct64(trans,
  374                                            FH_MEM_CBBC_QUEUE(trans, txq_id),
  375                                            txq->dma_addr);
  376                 else
  377                         iwl_write_direct32(trans,
  378                                            FH_MEM_CBBC_QUEUE(trans, txq_id),
  379                                            txq->dma_addr >> 8);
  380                 iwl_pcie_txq_unmap(trans, txq_id);
  381                 txq->read_ptr = 0;
  382                 txq->write_ptr = 0;
  383         }
  384 
  385         /* Tell NIC where to find the "keep warm" buffer */
  386         iwl_write_direct32(trans, FH_KW_MEM_ADDR_REG,
  387                            trans_pcie->kw.dma >> 4);
  388 
  389         /*
  390          * Send 0 as the scd_base_addr since the device may have be reset
  391          * while we were in WoWLAN in which case SCD_SRAM_BASE_ADDR will
  392          * contain garbage.
  393          */
  394         iwl_pcie_tx_start(trans, 0);
  395 }
  396 
  397 static void iwl_pcie_tx_stop_fh(struct iwl_trans *trans)
  398 {
  399         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  400         int ch, ret;
  401         u32 mask = 0;
  402 
  403         spin_lock_bh(&trans_pcie->irq_lock);
  404 
  405         if (!iwl_trans_grab_nic_access(trans))
  406                 goto out;
  407 
  408         /* Stop each Tx DMA channel */
  409         for (ch = 0; ch < FH_TCSR_CHNL_NUM; ch++) {
  410                 iwl_write32(trans, FH_TCSR_CHNL_TX_CONFIG_REG(ch), 0x0);
  411                 mask |= FH_TSSR_TX_STATUS_REG_MSK_CHNL_IDLE(ch);
  412         }
  413 
  414         /* Wait for DMA channels to be idle */
  415         ret = iwl_poll_bit(trans, FH_TSSR_TX_STATUS_REG, mask, mask, 5000);
  416         if (ret < 0)
  417                 IWL_ERR(trans,
  418                         "Failing on timeout while stopping DMA channel %d [0x%08x]\n",
  419                         ch, iwl_read32(trans, FH_TSSR_TX_STATUS_REG));
  420 
  421         iwl_trans_release_nic_access(trans);
  422 
  423 out:
  424         spin_unlock_bh(&trans_pcie->irq_lock);
  425 }
  426 
  427 /*
  428  * iwl_pcie_tx_stop - Stop all Tx DMA channels
  429  */
  430 int iwl_pcie_tx_stop(struct iwl_trans *trans)
  431 {
  432         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  433         int txq_id;
  434 
  435         /* Turn off all Tx DMA fifos */
  436         iwl_scd_deactivate_fifos(trans);
  437 
  438         /* Turn off all Tx DMA channels */
  439         iwl_pcie_tx_stop_fh(trans);
  440 
  441         /*
  442          * This function can be called before the op_mode disabled the
  443          * queues. This happens when we have an rfkill interrupt.
  444          * Since we stop Tx altogether - mark the queues as stopped.
  445          */
  446         memset(trans->txqs.queue_stopped, 0,
  447                sizeof(trans->txqs.queue_stopped));
  448         memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used));
  449 
  450         /* This can happen: start_hw, stop_device */
  451         if (!trans_pcie->txq_memory)
  452                 return 0;
  453 
  454         /* Unmap DMA from host system and free skb's */
  455         for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues;
  456              txq_id++)
  457                 iwl_pcie_txq_unmap(trans, txq_id);
  458 
  459         return 0;
  460 }
  461 
  462 /*
  463  * iwl_trans_tx_free - Free TXQ Context
  464  *
  465  * Destroy all TX DMA queues and structures
  466  */
  467 void iwl_pcie_tx_free(struct iwl_trans *trans)
  468 {
  469         int txq_id;
  470         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  471 
  472         memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used));
  473 
  474         /* Tx queues */
  475         if (trans_pcie->txq_memory) {
  476                 for (txq_id = 0;
  477                      txq_id < trans->trans_cfg->base_params->num_of_queues;
  478                      txq_id++) {
  479                         iwl_pcie_txq_free(trans, txq_id);
  480                         trans->txqs.txq[txq_id] = NULL;
  481                 }
  482         }
  483 
  484         kfree(trans_pcie->txq_memory);
  485         trans_pcie->txq_memory = NULL;
  486 
  487         iwl_pcie_free_dma_ptr(trans, &trans_pcie->kw);
  488 
  489         iwl_pcie_free_dma_ptr(trans, &trans->txqs.scd_bc_tbls);
  490 }
  491 
  492 /*
  493  * iwl_pcie_tx_alloc - allocate TX context
  494  * Allocate all Tx DMA structures and initialize them
  495  */
  496 static int iwl_pcie_tx_alloc(struct iwl_trans *trans)
  497 {
  498         int ret;
  499         int txq_id, slots_num;
  500         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  501         u16 bc_tbls_size = trans->trans_cfg->base_params->num_of_queues;
  502 
  503         if (WARN_ON(trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_AX210))
  504                 return -EINVAL;
  505 
  506         bc_tbls_size *= sizeof(struct iwlagn_scd_bc_tbl);
  507 
  508         /*It is not allowed to alloc twice, so warn when this happens.
  509          * We cannot rely on the previous allocation, so free and fail */
  510         if (WARN_ON(trans_pcie->txq_memory)) {
  511                 ret = -EINVAL;
  512                 goto error;
  513         }
  514 
  515         ret = iwl_pcie_alloc_dma_ptr(trans, &trans->txqs.scd_bc_tbls,
  516                                      bc_tbls_size);
  517         if (ret) {
  518                 IWL_ERR(trans, "Scheduler BC Table allocation failed\n");
  519                 goto error;
  520         }
  521 
  522         /* Alloc keep-warm buffer */
  523         ret = iwl_pcie_alloc_dma_ptr(trans, &trans_pcie->kw, IWL_KW_SIZE);
  524         if (ret) {
  525                 IWL_ERR(trans, "Keep Warm allocation failed\n");
  526                 goto error;
  527         }
  528 
  529         trans_pcie->txq_memory =
  530                 kcalloc(trans->trans_cfg->base_params->num_of_queues,
  531                         sizeof(struct iwl_txq), GFP_KERNEL);
  532         if (!trans_pcie->txq_memory) {
  533                 IWL_ERR(trans, "Not enough memory for txq\n");
  534                 ret = -ENOMEM;
  535                 goto error;
  536         }
  537 
  538         /* Alloc and init all Tx queues, including the command queue (#4/#9) */
  539         for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues;
  540              txq_id++) {
  541                 bool cmd_queue = (txq_id == trans->txqs.cmd.q_id);
  542 
  543                 if (cmd_queue)
  544                         slots_num = max_t(u32, IWL_CMD_QUEUE_SIZE,
  545                                           trans->cfg->min_txq_size);
  546                 else
  547                         slots_num = max_t(u32, IWL_DEFAULT_QUEUE_SIZE,
  548                                           trans->cfg->min_ba_txq_size);
  549                 trans->txqs.txq[txq_id] = &trans_pcie->txq_memory[txq_id];
  550                 ret = iwl_txq_alloc(trans, trans->txqs.txq[txq_id], slots_num,
  551                                     cmd_queue);
  552                 if (ret) {
  553                         IWL_ERR(trans, "Tx %d queue alloc failed\n", txq_id);
  554                         goto error;
  555                 }
  556                 trans->txqs.txq[txq_id]->id = txq_id;
  557         }
  558 
  559         return 0;
  560 
  561 error:
  562         iwl_pcie_tx_free(trans);
  563 
  564         return ret;
  565 }
  566 
  567 int iwl_pcie_tx_init(struct iwl_trans *trans)
  568 {
  569         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  570         int ret;
  571         int txq_id, slots_num;
  572         bool alloc = false;
  573 
  574         if (!trans_pcie->txq_memory) {
  575                 ret = iwl_pcie_tx_alloc(trans);
  576                 if (ret)
  577                         goto error;
  578                 alloc = true;
  579         }
  580 
  581         spin_lock_bh(&trans_pcie->irq_lock);
  582 
  583         /* Turn off all Tx DMA fifos */
  584         iwl_scd_deactivate_fifos(trans);
  585 
  586         /* Tell NIC where to find the "keep warm" buffer */
  587         iwl_write_direct32(trans, FH_KW_MEM_ADDR_REG,
  588                            trans_pcie->kw.dma >> 4);
  589 
  590         spin_unlock_bh(&trans_pcie->irq_lock);
  591 
  592         /* Alloc and init all Tx queues, including the command queue (#4/#9) */
  593         for (txq_id = 0; txq_id < trans->trans_cfg->base_params->num_of_queues;
  594              txq_id++) {
  595                 bool cmd_queue = (txq_id == trans->txqs.cmd.q_id);
  596 
  597                 if (cmd_queue)
  598                         slots_num = max_t(u32, IWL_CMD_QUEUE_SIZE,
  599                                           trans->cfg->min_txq_size);
  600                 else
  601                         slots_num = max_t(u32, IWL_DEFAULT_QUEUE_SIZE,
  602                                           trans->cfg->min_ba_txq_size);
  603                 ret = iwl_txq_init(trans, trans->txqs.txq[txq_id], slots_num,
  604                                    cmd_queue);
  605                 if (ret) {
  606                         IWL_ERR(trans, "Tx %d queue init failed\n", txq_id);
  607                         goto error;
  608                 }
  609 
  610                 /*
  611                  * Tell nic where to find circular buffer of TFDs for a
  612                  * given Tx queue, and enable the DMA channel used for that
  613                  * queue.
  614                  * Circular buffer (TFD queue in DRAM) physical base address
  615                  */
  616                 iwl_write_direct32(trans, FH_MEM_CBBC_QUEUE(trans, txq_id),
  617                                    trans->txqs.txq[txq_id]->dma_addr >> 8);
  618         }
  619 
  620         iwl_set_bits_prph(trans, SCD_GP_CTRL, SCD_GP_CTRL_AUTO_ACTIVE_MODE);
  621         if (trans->trans_cfg->base_params->num_of_queues > 20)
  622                 iwl_set_bits_prph(trans, SCD_GP_CTRL,
  623                                   SCD_GP_CTRL_ENABLE_31_QUEUES);
  624 
  625         return 0;
  626 error:
  627         /*Upon error, free only if we allocated something */
  628         if (alloc)
  629                 iwl_pcie_tx_free(trans);
  630         return ret;
  631 }
  632 
  633 static int iwl_pcie_set_cmd_in_flight(struct iwl_trans *trans,
  634                                       const struct iwl_host_cmd *cmd)
  635 {
  636         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  637 
  638         /* Make sure the NIC is still alive in the bus */
  639         if (test_bit(STATUS_TRANS_DEAD, &trans->status))
  640                 return -ENODEV;
  641 
  642         if (!trans->trans_cfg->base_params->apmg_wake_up_wa)
  643                 return 0;
  644 
  645         /*
  646          * wake up the NIC to make sure that the firmware will see the host
  647          * command - we will let the NIC sleep once all the host commands
  648          * returned. This needs to be done only on NICs that have
  649          * apmg_wake_up_wa set (see above.)
  650          */
  651         if (!_iwl_trans_pcie_grab_nic_access(trans))
  652                 return -EIO;
  653 
  654         /*
  655          * In iwl_trans_grab_nic_access(), we've acquired the reg_lock.
  656          * There, we also returned immediately if cmd_hold_nic_awake is
  657          * already true, so it's OK to unconditionally set it to true.
  658          */
  659         trans_pcie->cmd_hold_nic_awake = true;
  660         spin_unlock(&trans_pcie->reg_lock);
  661 
  662         return 0;
  663 }
  664 
  665 /*
  666  * iwl_pcie_cmdq_reclaim - Reclaim TX command queue entries already Tx'd
  667  *
  668  * When FW advances 'R' index, all entries between old and new 'R' index
  669  * need to be reclaimed. As result, some free space forms.  If there is
  670  * enough free space (> low mark), wake the stack that feeds us.
  671  */
  672 static void iwl_pcie_cmdq_reclaim(struct iwl_trans *trans, int txq_id, int idx)
  673 {
  674         struct iwl_txq *txq = trans->txqs.txq[txq_id];
  675         int nfreed = 0;
  676         u16 r;
  677 
  678         lockdep_assert_held(&txq->lock);
  679 
  680         idx = iwl_txq_get_cmd_index(txq, idx);
  681         r = iwl_txq_get_cmd_index(txq, txq->read_ptr);
  682 
  683         if (idx >= trans->trans_cfg->base_params->max_tfd_queue_size ||
  684             (!iwl_txq_used(txq, idx))) {
  685                 WARN_ONCE(test_bit(txq_id, trans->txqs.queue_used),
  686                           "%s: Read index for DMA queue txq id (%d), index %d is out of range [0-%d] %d %d.\n",
  687                           __func__, txq_id, idx,
  688                           trans->trans_cfg->base_params->max_tfd_queue_size,
  689                           txq->write_ptr, txq->read_ptr);
  690                 return;
  691         }
  692 
  693         for (idx = iwl_txq_inc_wrap(trans, idx); r != idx;
  694              r = iwl_txq_inc_wrap(trans, r)) {
  695                 txq->read_ptr = iwl_txq_inc_wrap(trans, txq->read_ptr);
  696 
  697                 if (nfreed++ > 0) {
  698                         IWL_ERR(trans, "HCMD skipped: index (%d) %d %d\n",
  699                                 idx, txq->write_ptr, r);
  700                         iwl_force_nmi(trans);
  701                 }
  702         }
  703 
  704         if (txq->read_ptr == txq->write_ptr)
  705                 iwl_pcie_clear_cmd_in_flight(trans);
  706 
  707         iwl_txq_progress(txq);
  708 }
  709 
  710 static int iwl_pcie_txq_set_ratid_map(struct iwl_trans *trans, u16 ra_tid,
  711                                  u16 txq_id)
  712 {
  713         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  714         u32 tbl_dw_addr;
  715         u32 tbl_dw;
  716         u16 scd_q2ratid;
  717 
  718         scd_q2ratid = ra_tid & SCD_QUEUE_RA_TID_MAP_RATID_MSK;
  719 
  720         tbl_dw_addr = trans_pcie->scd_base_addr +
  721                         SCD_TRANS_TBL_OFFSET_QUEUE(txq_id);
  722 
  723         tbl_dw = iwl_trans_read_mem32(trans, tbl_dw_addr);
  724 
  725         if (txq_id & 0x1)
  726                 tbl_dw = (scd_q2ratid << 16) | (tbl_dw & 0x0000FFFF);
  727         else
  728                 tbl_dw = scd_q2ratid | (tbl_dw & 0xFFFF0000);
  729 
  730         iwl_trans_write_mem32(trans, tbl_dw_addr, tbl_dw);
  731 
  732         return 0;
  733 }
  734 
  735 /* Receiver address (actually, Rx station's index into station table),
  736  * combined with Traffic ID (QOS priority), in format used by Tx Scheduler */
  737 #define BUILD_RAxTID(sta_id, tid)       (((sta_id) << 4) + (tid))
  738 
  739 bool iwl_trans_pcie_txq_enable(struct iwl_trans *trans, int txq_id, u16 ssn,
  740                                const struct iwl_trans_txq_scd_cfg *cfg,
  741                                unsigned int wdg_timeout)
  742 {
  743         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  744         struct iwl_txq *txq = trans->txqs.txq[txq_id];
  745         int fifo = -1;
  746         bool scd_bug = false;
  747 
  748         if (test_and_set_bit(txq_id, trans->txqs.queue_used))
  749                 WARN_ONCE(1, "queue %d already used - expect issues", txq_id);
  750 
  751         txq->wd_timeout = msecs_to_jiffies(wdg_timeout);
  752 
  753         if (cfg) {
  754                 fifo = cfg->fifo;
  755 
  756                 /* Disable the scheduler prior configuring the cmd queue */
  757                 if (txq_id == trans->txqs.cmd.q_id &&
  758                     trans_pcie->scd_set_active)
  759                         iwl_scd_enable_set_active(trans, 0);
  760 
  761                 /* Stop this Tx queue before configuring it */
  762                 iwl_scd_txq_set_inactive(trans, txq_id);
  763 
  764                 /* Set this queue as a chain-building queue unless it is CMD */
  765                 if (txq_id != trans->txqs.cmd.q_id)
  766                         iwl_scd_txq_set_chain(trans, txq_id);
  767 
  768                 if (cfg->aggregate) {
  769                         u16 ra_tid = BUILD_RAxTID(cfg->sta_id, cfg->tid);
  770 
  771                         /* Map receiver-address / traffic-ID to this queue */
  772                         iwl_pcie_txq_set_ratid_map(trans, ra_tid, txq_id);
  773 
  774                         /* enable aggregations for the queue */
  775                         iwl_scd_txq_enable_agg(trans, txq_id);
  776                         txq->ampdu = true;
  777                 } else {
  778                         /*
  779                          * disable aggregations for the queue, this will also
  780                          * make the ra_tid mapping configuration irrelevant
  781                          * since it is now a non-AGG queue.
  782                          */
  783                         iwl_scd_txq_disable_agg(trans, txq_id);
  784 
  785                         ssn = txq->read_ptr;
  786                 }
  787         } else {
  788                 /*
  789                  * If we need to move the SCD write pointer by steps of
  790                  * 0x40, 0x80 or 0xc0, it gets stuck. Avoids this and let
  791                  * the op_mode know by returning true later.
  792                  * Do this only in case cfg is NULL since this trick can
  793                  * be done only if we have DQA enabled which is true for mvm
  794                  * only. And mvm never sets a cfg pointer.
  795                  * This is really ugly, but this is the easiest way out for
  796                  * this sad hardware issue.
  797                  * This bug has been fixed on devices 9000 and up.
  798                  */
  799                 scd_bug = !trans->trans_cfg->mq_rx_supported &&
  800                         !((ssn - txq->write_ptr) & 0x3f) &&
  801                         (ssn != txq->write_ptr);
  802                 if (scd_bug)
  803                         ssn++;
  804         }
  805 
  806         /* Place first TFD at index corresponding to start sequence number.
  807          * Assumes that ssn_idx is valid (!= 0xFFF) */
  808         txq->read_ptr = (ssn & 0xff);
  809         txq->write_ptr = (ssn & 0xff);
  810         iwl_write_direct32(trans, HBUS_TARG_WRPTR,
  811                            (ssn & 0xff) | (txq_id << 8));
  812 
  813         if (cfg) {
  814                 u8 frame_limit = cfg->frame_limit;
  815 
  816                 iwl_write_prph(trans, SCD_QUEUE_RDPTR(txq_id), ssn);
  817 
  818                 /* Set up Tx window size and frame limit for this queue */
  819                 iwl_trans_write_mem32(trans, trans_pcie->scd_base_addr +
  820                                 SCD_CONTEXT_QUEUE_OFFSET(txq_id), 0);
  821                 iwl_trans_write_mem32(trans,
  822                         trans_pcie->scd_base_addr +
  823                         SCD_CONTEXT_QUEUE_OFFSET(txq_id) + sizeof(u32),
  824                         SCD_QUEUE_CTX_REG2_VAL(WIN_SIZE, frame_limit) |
  825                         SCD_QUEUE_CTX_REG2_VAL(FRAME_LIMIT, frame_limit));
  826 
  827                 /* Set up status area in SRAM, map to Tx DMA/FIFO, activate */
  828                 iwl_write_prph(trans, SCD_QUEUE_STATUS_BITS(txq_id),
  829                                (1 << SCD_QUEUE_STTS_REG_POS_ACTIVE) |
  830                                (cfg->fifo << SCD_QUEUE_STTS_REG_POS_TXF) |
  831                                (1 << SCD_QUEUE_STTS_REG_POS_WSL) |
  832                                SCD_QUEUE_STTS_REG_MSK);
  833 
  834                 /* enable the scheduler for this queue (only) */
  835                 if (txq_id == trans->txqs.cmd.q_id &&
  836                     trans_pcie->scd_set_active)
  837                         iwl_scd_enable_set_active(trans, BIT(txq_id));
  838 
  839                 IWL_DEBUG_TX_QUEUES(trans,
  840                                     "Activate queue %d on FIFO %d WrPtr: %d\n",
  841                                     txq_id, fifo, ssn & 0xff);
  842         } else {
  843                 IWL_DEBUG_TX_QUEUES(trans,
  844                                     "Activate queue %d WrPtr: %d\n",
  845                                     txq_id, ssn & 0xff);
  846         }
  847 
  848         return scd_bug;
  849 }
  850 
  851 void iwl_trans_pcie_txq_set_shared_mode(struct iwl_trans *trans, u32 txq_id,
  852                                         bool shared_mode)
  853 {
  854         struct iwl_txq *txq = trans->txqs.txq[txq_id];
  855 
  856         txq->ampdu = !shared_mode;
  857 }
  858 
  859 void iwl_trans_pcie_txq_disable(struct iwl_trans *trans, int txq_id,
  860                                 bool configure_scd)
  861 {
  862         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
  863         u32 stts_addr = trans_pcie->scd_base_addr +
  864                         SCD_TX_STTS_QUEUE_OFFSET(txq_id);
  865         static const u32 zero_val[4] = {};
  866 
  867         trans->txqs.txq[txq_id]->frozen_expiry_remainder = 0;
  868         trans->txqs.txq[txq_id]->frozen = false;
  869 
  870         /*
  871          * Upon HW Rfkill - we stop the device, and then stop the queues
  872          * in the op_mode. Just for the sake of the simplicity of the op_mode,
  873          * allow the op_mode to call txq_disable after it already called
  874          * stop_device.
  875          */
  876         if (!test_and_clear_bit(txq_id, trans->txqs.queue_used)) {
  877                 WARN_ONCE(test_bit(STATUS_DEVICE_ENABLED, &trans->status),
  878                           "queue %d not used", txq_id);
  879                 return;
  880         }
  881 
  882         if (configure_scd) {
  883                 iwl_scd_txq_set_inactive(trans, txq_id);
  884 
  885                 iwl_trans_write_mem(trans, stts_addr, (const void *)zero_val,
  886                                     ARRAY_SIZE(zero_val));
  887         }
  888 
  889         iwl_pcie_txq_unmap(trans, txq_id);
  890         trans->txqs.txq[txq_id]->ampdu = false;
  891 
  892         IWL_DEBUG_TX_QUEUES(trans, "Deactivate queue %d\n", txq_id);
  893 }
  894 
  895 /*************** HOST COMMAND QUEUE FUNCTIONS   *****/
  896 
  897 /*
  898  * iwl_pcie_enqueue_hcmd - enqueue a uCode command
  899  * @priv: device private data point
  900  * @cmd: a pointer to the ucode command structure
  901  *
  902  * The function returns < 0 values to indicate the operation
  903  * failed. On success, it returns the index (>= 0) of command in the
  904  * command queue.
  905  */
  906 int iwl_pcie_enqueue_hcmd(struct iwl_trans *trans,
  907                           struct iwl_host_cmd *cmd)
  908 {
  909         struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id];
  910         struct iwl_device_cmd *out_cmd;
  911         struct iwl_cmd_meta *out_meta;
  912         void *dup_buf = NULL;
  913         dma_addr_t phys_addr;
  914         int idx;
  915         u16 copy_size, cmd_size, tb0_size;
  916         bool had_nocopy = false;
  917         u8 group_id = iwl_cmd_groupid(cmd->id);
  918         int i, ret;
  919         u32 cmd_pos;
  920         const u8 *cmddata[IWL_MAX_CMD_TBS_PER_TFD];
  921         u16 cmdlen[IWL_MAX_CMD_TBS_PER_TFD];
  922         unsigned long flags;
  923 
  924         if (WARN(!trans->wide_cmd_header &&
  925                  group_id > IWL_ALWAYS_LONG_GROUP,
  926                  "unsupported wide command %#x\n", cmd->id))
  927                 return -EINVAL;
  928 
  929         if (group_id != 0) {
  930                 copy_size = sizeof(struct iwl_cmd_header_wide);
  931                 cmd_size = sizeof(struct iwl_cmd_header_wide);
  932         } else {
  933                 copy_size = sizeof(struct iwl_cmd_header);
  934                 cmd_size = sizeof(struct iwl_cmd_header);
  935         }
  936 
  937         /* need one for the header if the first is NOCOPY */
  938         BUILD_BUG_ON(IWL_MAX_CMD_TBS_PER_TFD > IWL_NUM_OF_TBS - 1);
  939 
  940         for (i = 0; i < IWL_MAX_CMD_TBS_PER_TFD; i++) {
  941                 cmddata[i] = cmd->data[i];
  942                 cmdlen[i] = cmd->len[i];
  943 
  944                 if (!cmd->len[i])
  945                         continue;
  946 
  947                 /* need at least IWL_FIRST_TB_SIZE copied */
  948                 if (copy_size < IWL_FIRST_TB_SIZE) {
  949                         int copy = IWL_FIRST_TB_SIZE - copy_size;
  950 
  951                         if (copy > cmdlen[i])
  952                                 copy = cmdlen[i];
  953                         cmdlen[i] -= copy;
  954                         cmddata[i] += copy;
  955                         copy_size += copy;
  956                 }
  957 
  958                 if (cmd->dataflags[i] & IWL_HCMD_DFL_NOCOPY) {
  959                         had_nocopy = true;
  960                         if (WARN_ON(cmd->dataflags[i] & IWL_HCMD_DFL_DUP)) {
  961                                 idx = -EINVAL;
  962                                 goto free_dup_buf;
  963                         }
  964                 } else if (cmd->dataflags[i] & IWL_HCMD_DFL_DUP) {
  965                         /*
  966                          * This is also a chunk that isn't copied
  967                          * to the static buffer so set had_nocopy.
  968                          */
  969                         had_nocopy = true;
  970 
  971                         /* only allowed once */
  972                         if (WARN_ON(dup_buf)) {
  973                                 idx = -EINVAL;
  974                                 goto free_dup_buf;
  975                         }
  976 
  977                         dup_buf = kmemdup(cmddata[i], cmdlen[i],
  978                                           GFP_ATOMIC);
  979                         if (!dup_buf)
  980                                 return -ENOMEM;
  981                 } else {
  982                         /* NOCOPY must not be followed by normal! */
  983                         if (WARN_ON(had_nocopy)) {
  984                                 idx = -EINVAL;
  985                                 goto free_dup_buf;
  986                         }
  987                         copy_size += cmdlen[i];
  988                 }
  989                 cmd_size += cmd->len[i];
  990         }
  991 
  992         /*
  993          * If any of the command structures end up being larger than
  994          * the TFD_MAX_PAYLOAD_SIZE and they aren't dynamically
  995          * allocated into separate TFDs, then we will need to
  996          * increase the size of the buffers.
  997          */
  998         if (WARN(copy_size > TFD_MAX_PAYLOAD_SIZE,
  999                  "Command %s (%#x) is too large (%d bytes)\n",
 1000                  iwl_get_cmd_string(trans, cmd->id),
 1001                  cmd->id, copy_size)) {
 1002                 idx = -EINVAL;
 1003                 goto free_dup_buf;
 1004         }
 1005 
 1006         spin_lock_irqsave(&txq->lock, flags);
 1007 
 1008         if (iwl_txq_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) {
 1009                 spin_unlock_irqrestore(&txq->lock, flags);
 1010 
 1011                 IWL_ERR(trans, "No space in command queue\n");
 1012                 iwl_op_mode_cmd_queue_full(trans->op_mode);
 1013                 idx = -ENOSPC;
 1014                 goto free_dup_buf;
 1015         }
 1016 
 1017         idx = iwl_txq_get_cmd_index(txq, txq->write_ptr);
 1018         out_cmd = txq->entries[idx].cmd;
 1019         out_meta = &txq->entries[idx].meta;
 1020 
 1021         memset(out_meta, 0, sizeof(*out_meta)); /* re-initialize to NULL */
 1022         if (cmd->flags & CMD_WANT_SKB)
 1023                 out_meta->source = cmd;
 1024 
 1025         /* set up the header */
 1026         if (group_id != 0) {
 1027                 out_cmd->hdr_wide.cmd = iwl_cmd_opcode(cmd->id);
 1028                 out_cmd->hdr_wide.group_id = group_id;
 1029                 out_cmd->hdr_wide.version = iwl_cmd_version(cmd->id);
 1030                 out_cmd->hdr_wide.length =
 1031                         cpu_to_le16(cmd_size -
 1032                                     sizeof(struct iwl_cmd_header_wide));
 1033                 out_cmd->hdr_wide.reserved = 0;
 1034                 out_cmd->hdr_wide.sequence =
 1035                         cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) |
 1036                                                  INDEX_TO_SEQ(txq->write_ptr));
 1037 
 1038                 cmd_pos = sizeof(struct iwl_cmd_header_wide);
 1039                 copy_size = sizeof(struct iwl_cmd_header_wide);
 1040         } else {
 1041                 out_cmd->hdr.cmd = iwl_cmd_opcode(cmd->id);
 1042                 out_cmd->hdr.sequence =
 1043                         cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) |
 1044                                                  INDEX_TO_SEQ(txq->write_ptr));
 1045                 out_cmd->hdr.group_id = 0;
 1046 
 1047                 cmd_pos = sizeof(struct iwl_cmd_header);
 1048                 copy_size = sizeof(struct iwl_cmd_header);
 1049         }
 1050 
 1051         /* and copy the data that needs to be copied */
 1052         for (i = 0; i < IWL_MAX_CMD_TBS_PER_TFD; i++) {
 1053                 int copy;
 1054 
 1055                 if (!cmd->len[i])
 1056                         continue;
 1057 
 1058                 /* copy everything if not nocopy/dup */
 1059                 if (!(cmd->dataflags[i] & (IWL_HCMD_DFL_NOCOPY |
 1060                                            IWL_HCMD_DFL_DUP))) {
 1061                         copy = cmd->len[i];
 1062 
 1063                         memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy);
 1064                         cmd_pos += copy;
 1065                         copy_size += copy;
 1066                         continue;
 1067                 }
 1068 
 1069                 /*
 1070                  * Otherwise we need at least IWL_FIRST_TB_SIZE copied
 1071                  * in total (for bi-directional DMA), but copy up to what
 1072                  * we can fit into the payload for debug dump purposes.
 1073                  */
 1074                 copy = min_t(int, TFD_MAX_PAYLOAD_SIZE - cmd_pos, cmd->len[i]);
 1075 
 1076                 memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy);
 1077                 cmd_pos += copy;
 1078 
 1079                 /* However, treat copy_size the proper way, we need it below */
 1080                 if (copy_size < IWL_FIRST_TB_SIZE) {
 1081                         copy = IWL_FIRST_TB_SIZE - copy_size;
 1082 
 1083                         if (copy > cmd->len[i])
 1084                                 copy = cmd->len[i];
 1085                         copy_size += copy;
 1086                 }
 1087         }
 1088 
 1089         IWL_DEBUG_HC(trans,
 1090                      "Sending command %s (%.2x.%.2x), seq: 0x%04X, %d bytes at %d[%d]:%d\n",
 1091                      iwl_get_cmd_string(trans, cmd->id),
 1092                      group_id, out_cmd->hdr.cmd,
 1093                      le16_to_cpu(out_cmd->hdr.sequence),
 1094                      cmd_size, txq->write_ptr, idx, trans->txqs.cmd.q_id);
 1095 
 1096         /* start the TFD with the minimum copy bytes */
 1097         tb0_size = min_t(int, copy_size, IWL_FIRST_TB_SIZE);
 1098         memcpy(&txq->first_tb_bufs[idx], &out_cmd->hdr, tb0_size);
 1099         iwl_pcie_txq_build_tfd(trans, txq,
 1100                                iwl_txq_get_first_tb_dma(txq, idx),
 1101                                tb0_size, true);
 1102 
 1103         /* map first command fragment, if any remains */
 1104         if (copy_size > tb0_size) {
 1105                 phys_addr = dma_map_single(trans->dev,
 1106                                            ((u8 *)&out_cmd->hdr) + tb0_size,
 1107                                            copy_size - tb0_size,
 1108                                            DMA_TO_DEVICE);
 1109                 if (dma_mapping_error(trans->dev, phys_addr)) {
 1110                         iwl_txq_gen1_tfd_unmap(trans, out_meta, txq,
 1111                                                txq->write_ptr);
 1112                         idx = -ENOMEM;
 1113                         goto out;
 1114                 }
 1115 
 1116                 iwl_pcie_txq_build_tfd(trans, txq, phys_addr,
 1117                                        copy_size - tb0_size, false);
 1118         }
 1119 
 1120         /* map the remaining (adjusted) nocopy/dup fragments */
 1121         for (i = 0; i < IWL_MAX_CMD_TBS_PER_TFD; i++) {
 1122                 void *data = (void *)(uintptr_t)cmddata[i];
 1123 
 1124                 if (!cmdlen[i])
 1125                         continue;
 1126                 if (!(cmd->dataflags[i] & (IWL_HCMD_DFL_NOCOPY |
 1127                                            IWL_HCMD_DFL_DUP)))
 1128                         continue;
 1129                 if (cmd->dataflags[i] & IWL_HCMD_DFL_DUP)
 1130                         data = dup_buf;
 1131                 phys_addr = dma_map_single(trans->dev, data,
 1132                                            cmdlen[i], DMA_TO_DEVICE);
 1133                 if (dma_mapping_error(trans->dev, phys_addr)) {
 1134                         iwl_txq_gen1_tfd_unmap(trans, out_meta, txq,
 1135                                                txq->write_ptr);
 1136                         idx = -ENOMEM;
 1137                         goto out;
 1138                 }
 1139 
 1140                 iwl_pcie_txq_build_tfd(trans, txq, phys_addr, cmdlen[i], false);
 1141         }
 1142 
 1143         BUILD_BUG_ON(IWL_TFH_NUM_TBS > sizeof(out_meta->tbs) * BITS_PER_BYTE);
 1144         out_meta->flags = cmd->flags;
 1145         if (WARN_ON_ONCE(txq->entries[idx].free_buf))
 1146                 kfree_sensitive(txq->entries[idx].free_buf);
 1147         txq->entries[idx].free_buf = dup_buf;
 1148 
 1149         trace_iwlwifi_dev_hcmd(trans->dev, cmd, cmd_size, &out_cmd->hdr_wide);
 1150 
 1151         /* start timer if queue currently empty */
 1152         if (txq->read_ptr == txq->write_ptr && txq->wd_timeout)
 1153                 mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout);
 1154 
 1155         ret = iwl_pcie_set_cmd_in_flight(trans, cmd);
 1156         if (ret < 0) {
 1157                 idx = ret;
 1158                 goto out;
 1159         }
 1160 
 1161         /* Increment and update queue's write index */
 1162         txq->write_ptr = iwl_txq_inc_wrap(trans, txq->write_ptr);
 1163         iwl_pcie_txq_inc_wr_ptr(trans, txq);
 1164 
 1165  out:
 1166         spin_unlock_irqrestore(&txq->lock, flags);
 1167  free_dup_buf:
 1168         if (idx < 0)
 1169                 kfree(dup_buf);
 1170         return idx;
 1171 }
 1172 
 1173 /*
 1174  * iwl_pcie_hcmd_complete - Pull unused buffers off the queue and reclaim them
 1175  * @rxb: Rx buffer to reclaim
 1176  */
 1177 void iwl_pcie_hcmd_complete(struct iwl_trans *trans,
 1178                             struct iwl_rx_cmd_buffer *rxb)
 1179 {
 1180         struct iwl_rx_packet *pkt = rxb_addr(rxb);
 1181         u16 sequence = le16_to_cpu(pkt->hdr.sequence);
 1182         u8 group_id;
 1183         u32 cmd_id;
 1184         int txq_id = SEQ_TO_QUEUE(sequence);
 1185         int index = SEQ_TO_INDEX(sequence);
 1186         int cmd_index;
 1187         struct iwl_device_cmd *cmd;
 1188         struct iwl_cmd_meta *meta;
 1189         struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
 1190         struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id];
 1191 
 1192         /* If a Tx command is being handled and it isn't in the actual
 1193          * command queue then there a command routing bug has been introduced
 1194          * in the queue management code. */
 1195         if (WARN(txq_id != trans->txqs.cmd.q_id,
 1196                  "wrong command queue %d (should be %d), sequence 0x%X readp=%d writep=%d\n",
 1197                  txq_id, trans->txqs.cmd.q_id, sequence, txq->read_ptr,
 1198                  txq->write_ptr)) {
 1199                 iwl_print_hex_error(trans, pkt, 32);
 1200                 return;
 1201         }
 1202 
 1203         spin_lock_bh(&txq->lock);
 1204 
 1205         cmd_index = iwl_txq_get_cmd_index(txq, index);
 1206         cmd = txq->entries[cmd_index].cmd;
 1207         meta = &txq->entries[cmd_index].meta;
 1208         group_id = cmd->hdr.group_id;
 1209         cmd_id = WIDE_ID(group_id, cmd->hdr.cmd);
 1210 
 1211         iwl_txq_gen1_tfd_unmap(trans, meta, txq, index);
 1212 
 1213         /* Input error checking is done when commands are added to queue. */
 1214         if (meta->flags & CMD_WANT_SKB) {
 1215                 struct page *p = rxb_steal_page(rxb);
 1216 
 1217                 meta->source->resp_pkt = pkt;
 1218 #if defined(__linux__)
 1219                 meta->source->_rx_page_addr = (unsigned long)page_address(p);
 1220 #elif defined(__FreeBSD__)
 1221                 meta->source->_page = p;
 1222 #endif
 1223                 meta->source->_rx_page_order = trans_pcie->rx_page_order;
 1224         }
 1225 
 1226         if (meta->flags & CMD_WANT_ASYNC_CALLBACK)
 1227                 iwl_op_mode_async_cb(trans->op_mode, cmd);
 1228 
 1229         iwl_pcie_cmdq_reclaim(trans, txq_id, index);
 1230 
 1231         if (!(meta->flags & CMD_ASYNC)) {
 1232                 if (!test_bit(STATUS_SYNC_HCMD_ACTIVE, &trans->status)) {
 1233                         IWL_WARN(trans,
 1234                                  "HCMD_ACTIVE already clear for command %s\n",
 1235                                  iwl_get_cmd_string(trans, cmd_id));
 1236                 }
 1237                 clear_bit(STATUS_SYNC_HCMD_ACTIVE, &trans->status);
 1238                 IWL_DEBUG_INFO(trans, "Clearing HCMD_ACTIVE for command %s\n",
 1239                                iwl_get_cmd_string(trans, cmd_id));
 1240                 wake_up(&trans->wait_command_queue);
 1241         }
 1242 
 1243         meta->flags = 0;
 1244 
 1245         spin_unlock_bh(&txq->lock);
 1246 }
 1247 
 1248 static int iwl_fill_data_tbs(struct iwl_trans *trans, struct sk_buff *skb,
 1249                              struct iwl_txq *txq, u8 hdr_len,
 1250                              struct iwl_cmd_meta *out_meta)
 1251 {
 1252         u16 head_tb_len;
 1253         int i;
 1254 
 1255         /*
 1256          * Set up TFD's third entry to point directly to remainder
 1257          * of skb's head, if any
 1258          */
 1259         head_tb_len = skb_headlen(skb) - hdr_len;
 1260 
 1261         if (head_tb_len > 0) {
 1262                 dma_addr_t tb_phys = dma_map_single(trans->dev,
 1263                                                     skb->data + hdr_len,
 1264                                                     head_tb_len, DMA_TO_DEVICE);
 1265                 if (unlikely(dma_mapping_error(trans->dev, tb_phys)))
 1266                         return -EINVAL;
 1267                 trace_iwlwifi_dev_tx_tb(trans->dev, skb, skb->data + hdr_len,
 1268                                         tb_phys, head_tb_len);
 1269                 iwl_pcie_txq_build_tfd(trans, txq, tb_phys, head_tb_len, false);
 1270         }
 1271 
 1272         /* set up the remaining entries to point to the data */
 1273         for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
 1274                 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
 1275                 dma_addr_t tb_phys;
 1276                 int tb_idx;
 1277 
 1278                 if (!skb_frag_size(frag))
 1279                         continue;
 1280 
 1281                 tb_phys = skb_frag_dma_map(trans->dev, frag, 0,
 1282                                            skb_frag_size(frag), DMA_TO_DEVICE);
 1283 
 1284                 if (unlikely(dma_mapping_error(trans->dev, tb_phys)))
 1285                         return -EINVAL;
 1286                 trace_iwlwifi_dev_tx_tb(trans->dev, skb, skb_frag_address(frag),
 1287                                         tb_phys, skb_frag_size(frag));
 1288                 tb_idx = iwl_pcie_txq_build_tfd(trans, txq, tb_phys,
 1289                                                 skb_frag_size(frag), false);
 1290                 if (tb_idx < 0)
 1291                         return tb_idx;
 1292 
 1293                 out_meta->tbs |= BIT(tb_idx);
 1294         }
 1295 
 1296         return 0;
 1297 }
 1298 
 1299 #ifdef CONFIG_INET
 1300 static int iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb,
 1301                                    struct iwl_txq *txq, u8 hdr_len,
 1302                                    struct iwl_cmd_meta *out_meta,
 1303                                    struct iwl_device_tx_cmd *dev_cmd,
 1304                                    u16 tb1_len)
 1305 {
 1306         struct iwl_tx_cmd *tx_cmd = (void *)dev_cmd->payload;
 1307         struct ieee80211_hdr *hdr = (void *)skb->data;
 1308         unsigned int snap_ip_tcp_hdrlen, ip_hdrlen, total_len, hdr_room;
 1309         unsigned int mss = skb_shinfo(skb)->gso_size;
 1310         u16 length, iv_len, amsdu_pad;
 1311         u8 *start_hdr;
 1312         struct iwl_tso_hdr_page *hdr_page;
 1313         struct tso_t tso;
 1314 
 1315         /* if the packet is protected, then it must be CCMP or GCMP */
 1316         BUILD_BUG_ON(IEEE80211_CCMP_HDR_LEN != IEEE80211_GCMP_HDR_LEN);
 1317         iv_len = ieee80211_has_protected(hdr->frame_control) ?
 1318                 IEEE80211_CCMP_HDR_LEN : 0;
 1319 
 1320         trace_iwlwifi_dev_tx(trans->dev, skb,
 1321                              iwl_txq_get_tfd(trans, txq, txq->write_ptr),
 1322                              trans->txqs.tfd.size,
 1323                              &dev_cmd->hdr, IWL_FIRST_TB_SIZE + tb1_len, 0);
 1324 
 1325         ip_hdrlen = skb_transport_header(skb) - skb_network_header(skb);
 1326         snap_ip_tcp_hdrlen = 8 + ip_hdrlen + tcp_hdrlen(skb);
 1327         total_len = skb->len - snap_ip_tcp_hdrlen - hdr_len - iv_len;
 1328         amsdu_pad = 0;
 1329 
 1330         /* total amount of header we may need for this A-MSDU */
 1331         hdr_room = DIV_ROUND_UP(total_len, mss) *
 1332                 (3 + snap_ip_tcp_hdrlen + sizeof(struct ethhdr)) + iv_len;
 1333 
 1334         /* Our device supports 9 segments at most, it will fit in 1 page */
 1335         hdr_page = get_page_hdr(trans, hdr_room, skb);
 1336         if (!hdr_page)
 1337                 return -ENOMEM;
 1338 
 1339         start_hdr = hdr_page->pos;
 1340         memcpy(hdr_page->pos, skb->data + hdr_len, iv_len);
 1341         hdr_page->pos += iv_len;
 1342 
 1343         /*
 1344          * Pull the ieee80211 header + IV to be able to use TSO core,
 1345          * we will restore it for the tx_status flow.
 1346          */
 1347         skb_pull(skb, hdr_len + iv_len);
 1348 
 1349         /*
 1350          * Remove the length of all the headers that we don't actually
 1351          * have in the MPDU by themselves, but that we duplicate into
 1352          * all the different MSDUs inside the A-MSDU.
 1353          */
 1354         le16_add_cpu(&tx_cmd->len, -snap_ip_tcp_hdrlen);
 1355 
 1356         tso_start(skb, &tso);
 1357 
 1358         while (total_len) {
 1359                 /* this is the data left for this subframe */
 1360                 unsigned int data_left =
 1361                         min_t(unsigned int, mss, total_len);
 1362                 unsigned int hdr_tb_len;
 1363                 dma_addr_t hdr_tb_phys;
 1364                 u8 *subf_hdrs_start = hdr_page->pos;
 1365 
 1366                 total_len -= data_left;
 1367 
 1368                 memset(hdr_page->pos, 0, amsdu_pad);
 1369                 hdr_page->pos += amsdu_pad;
 1370                 amsdu_pad = (4 - (sizeof(struct ethhdr) + snap_ip_tcp_hdrlen +
 1371                                   data_left)) & 0x3;
 1372                 ether_addr_copy(hdr_page->pos, ieee80211_get_DA(hdr));
 1373                 hdr_page->pos += ETH_ALEN;
 1374                 ether_addr_copy(hdr_page->pos, ieee80211_get_SA(hdr));
 1375                 hdr_page->pos += ETH_ALEN;
 1376 
 1377                 length = snap_ip_tcp_hdrlen + data_left;
 1378                 *((__be16 *)hdr_page->pos) = cpu_to_be16(length);
 1379                 hdr_page->pos += sizeof(length);
 1380 
 1381                 /*
 1382                  * This will copy the SNAP as well which will be considered
 1383                  * as MAC header.
 1384                  */
 1385                 tso_build_hdr(skb, hdr_page->pos, &tso, data_left, !total_len);
 1386 
 1387                 hdr_page->pos += snap_ip_tcp_hdrlen;
 1388 
 1389                 hdr_tb_len = hdr_page->pos - start_hdr;
 1390                 hdr_tb_phys = dma_map_single(trans->dev, start_hdr,
 1391                                              hdr_tb_len, DMA_TO_DEVICE);
 1392                 if (unlikely(dma_mapping_error(trans->dev, hdr_tb_phys)))
 1393                         return -EINVAL;
 1394                 iwl_pcie_txq_build_tfd(trans, txq, hdr_tb_phys,
 1395                                        hdr_tb_len, false);
 1396                 trace_iwlwifi_dev_tx_tb(trans->dev, skb, start_hdr,
 1397                                         hdr_tb_phys, hdr_tb_len);
 1398                 /* add this subframe's headers' length to the tx_cmd */
 1399                 le16_add_cpu(&tx_cmd->len, hdr_page->pos - subf_hdrs_start);
 1400 
 1401                 /* prepare the start_hdr for the next subframe */
 1402                 start_hdr = hdr_page->pos;
 1403 
 1404                 /* put the payload */
 1405                 while (data_left) {
 1406                         unsigned int size = min_t(unsigned int, tso.size,
 1407                                                   data_left);
 1408                         dma_addr_t tb_phys;
 1409 
 1410                         tb_phys = dma_map_single(trans->dev, tso.data,
 1411                                                  size, DMA_TO_DEVICE);
 1412                         if (unlikely(dma_mapping_error(trans->dev, tb_phys)))
 1413                                 return -EINVAL;
 1414 
 1415                         iwl_pcie_txq_build_tfd(trans, txq, tb_phys,
 1416                                                size, false);
 1417                         trace_iwlwifi_dev_tx_tb(trans->dev, skb, tso.data,
 1418                                                 tb_phys, size);
 1419 
 1420                         data_left -= size;
 1421                         tso_build_data(skb, &tso, size);
 1422                 }
 1423         }
 1424 
 1425         /* re -add the WiFi header and IV */
 1426         skb_push(skb, hdr_len + iv_len);
 1427 
 1428         return 0;
 1429 }
 1430 #else /* CONFIG_INET */
 1431 static int iwl_fill_data_tbs_amsdu(struct iwl_trans *trans, struct sk_buff *skb,
 1432                                    struct iwl_txq *txq, u8 hdr_len,
 1433                                    struct iwl_cmd_meta *out_meta,
 1434                                    struct iwl_device_tx_cmd *dev_cmd,
 1435                                    u16 tb1_len)
 1436 {
 1437         /* No A-MSDU without CONFIG_INET */
 1438         WARN_ON(1);
 1439 
 1440         return -1;
 1441 }
 1442 #endif /* CONFIG_INET */
 1443 
 1444 int iwl_trans_pcie_tx(struct iwl_trans *trans, struct sk_buff *skb,
 1445                       struct iwl_device_tx_cmd *dev_cmd, int txq_id)
 1446 {
 1447         struct ieee80211_hdr *hdr;
 1448         struct iwl_tx_cmd *tx_cmd = (struct iwl_tx_cmd *)dev_cmd->payload;
 1449         struct iwl_cmd_meta *out_meta;
 1450         struct iwl_txq *txq;
 1451         dma_addr_t tb0_phys, tb1_phys, scratch_phys;
 1452         void *tb1_addr;
 1453         void *tfd;
 1454         u16 len, tb1_len;
 1455         bool wait_write_ptr;
 1456         __le16 fc;
 1457         u8 hdr_len;
 1458         u16 wifi_seq;
 1459         bool amsdu;
 1460 
 1461         txq = trans->txqs.txq[txq_id];
 1462 
 1463         if (WARN_ONCE(!test_bit(txq_id, trans->txqs.queue_used),
 1464                       "TX on unused queue %d\n", txq_id))
 1465                 return -EINVAL;
 1466 
 1467         if (skb_is_nonlinear(skb) &&
 1468             skb_shinfo(skb)->nr_frags > IWL_TRANS_MAX_FRAGS(trans) &&
 1469             __skb_linearize(skb))
 1470                 return -ENOMEM;
 1471 
 1472         /* mac80211 always puts the full header into the SKB's head,
 1473          * so there's no need to check if it's readable there
 1474          */
 1475         hdr = (struct ieee80211_hdr *)skb->data;
 1476         fc = hdr->frame_control;
 1477         hdr_len = ieee80211_hdrlen(fc);
 1478 
 1479         spin_lock(&txq->lock);
 1480 
 1481         if (iwl_txq_space(trans, txq) < txq->high_mark) {
 1482                 iwl_txq_stop(trans, txq);
 1483 
 1484                 /* don't put the packet on the ring, if there is no room */
 1485                 if (unlikely(iwl_txq_space(trans, txq) < 3)) {
 1486                         struct iwl_device_tx_cmd **dev_cmd_ptr;
 1487 
 1488                         dev_cmd_ptr = (void *)((u8 *)skb->cb +
 1489                                                trans->txqs.dev_cmd_offs);
 1490 
 1491                         *dev_cmd_ptr = dev_cmd;
 1492                         __skb_queue_tail(&txq->overflow_q, skb);
 1493 
 1494                         spin_unlock(&txq->lock);
 1495                         return 0;
 1496                 }
 1497         }
 1498 
 1499         /* In AGG mode, the index in the ring must correspond to the WiFi
 1500          * sequence number. This is a HW requirements to help the SCD to parse
 1501          * the BA.
 1502          * Check here that the packets are in the right place on the ring.
 1503          */
 1504         wifi_seq = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl));
 1505         WARN_ONCE(txq->ampdu &&
 1506                   (wifi_seq & 0xff) != txq->write_ptr,
 1507                   "Q: %d WiFi Seq %d tfdNum %d",
 1508                   txq_id, wifi_seq, txq->write_ptr);
 1509 
 1510         /* Set up driver data for this TFD */
 1511         txq->entries[txq->write_ptr].skb = skb;
 1512         txq->entries[txq->write_ptr].cmd = dev_cmd;
 1513 
 1514         dev_cmd->hdr.sequence =
 1515                 cpu_to_le16((u16)(QUEUE_TO_SEQ(txq_id) |
 1516                             INDEX_TO_SEQ(txq->write_ptr)));
 1517 
 1518         tb0_phys = iwl_txq_get_first_tb_dma(txq, txq->write_ptr);
 1519         scratch_phys = tb0_phys + sizeof(struct iwl_cmd_header) +
 1520                        offsetof(struct iwl_tx_cmd, scratch);
 1521 
 1522         tx_cmd->dram_lsb_ptr = cpu_to_le32(scratch_phys);
 1523         tx_cmd->dram_msb_ptr = iwl_get_dma_hi_addr(scratch_phys);
 1524 
 1525         /* Set up first empty entry in queue's array of Tx/cmd buffers */
 1526         out_meta = &txq->entries[txq->write_ptr].meta;
 1527         out_meta->flags = 0;
 1528 
 1529         /*
 1530          * The second TB (tb1) points to the remainder of the TX command
 1531          * and the 802.11 header - dword aligned size
 1532          * (This calculation modifies the TX command, so do it before the
 1533          * setup of the first TB)
 1534          */
 1535         len = sizeof(struct iwl_tx_cmd) + sizeof(struct iwl_cmd_header) +
 1536               hdr_len - IWL_FIRST_TB_SIZE;
 1537         /* do not align A-MSDU to dword as the subframe header aligns it */
 1538         amsdu = ieee80211_is_data_qos(fc) &&
 1539                 (*ieee80211_get_qos_ctl(hdr) &
 1540                  IEEE80211_QOS_CTL_A_MSDU_PRESENT);
 1541         if (!amsdu) {
 1542                 tb1_len = ALIGN(len, 4);
 1543                 /* Tell NIC about any 2-byte padding after MAC header */
 1544                 if (tb1_len != len)
 1545                         tx_cmd->tx_flags |= cpu_to_le32(TX_CMD_FLG_MH_PAD);
 1546         } else {
 1547                 tb1_len = len;
 1548         }
 1549 
 1550         /*
 1551          * The first TB points to bi-directional DMA data, we'll
 1552          * memcpy the data into it later.
 1553          */
 1554         iwl_pcie_txq_build_tfd(trans, txq, tb0_phys,
 1555                                IWL_FIRST_TB_SIZE, true);
 1556 
 1557         /* there must be data left over for TB1 or this code must be changed */
 1558         BUILD_BUG_ON(sizeof(struct iwl_tx_cmd) < IWL_FIRST_TB_SIZE);
 1559 
 1560         /* map the data for TB1 */
 1561         tb1_addr = ((u8 *)&dev_cmd->hdr) + IWL_FIRST_TB_SIZE;
 1562         tb1_phys = dma_map_single(trans->dev, tb1_addr, tb1_len, DMA_TO_DEVICE);
 1563         if (unlikely(dma_mapping_error(trans->dev, tb1_phys)))
 1564                 goto out_err;
 1565         iwl_pcie_txq_build_tfd(trans, txq, tb1_phys, tb1_len, false);
 1566 
 1567         trace_iwlwifi_dev_tx(trans->dev, skb,
 1568                              iwl_txq_get_tfd(trans, txq, txq->write_ptr),
 1569                              trans->txqs.tfd.size,
 1570                              &dev_cmd->hdr, IWL_FIRST_TB_SIZE + tb1_len,
 1571                              hdr_len);
 1572 
 1573         /*
 1574          * If gso_size wasn't set, don't give the frame "amsdu treatment"
 1575          * (adding subframes, etc.).
 1576          * This can happen in some testing flows when the amsdu was already
 1577          * pre-built, and we just need to send the resulting skb.
 1578          */
 1579         if (amsdu && skb_shinfo(skb)->gso_size) {
 1580                 if (unlikely(iwl_fill_data_tbs_amsdu(trans, skb, txq, hdr_len,
 1581                                                      out_meta, dev_cmd,
 1582                                                      tb1_len)))
 1583                         goto out_err;
 1584         } else {
 1585                 struct sk_buff *frag;
 1586 
 1587                 if (unlikely(iwl_fill_data_tbs(trans, skb, txq, hdr_len,
 1588                                                out_meta)))
 1589                         goto out_err;
 1590 
 1591                 skb_walk_frags(skb, frag) {
 1592                         if (unlikely(iwl_fill_data_tbs(trans, frag, txq, 0,
 1593                                                        out_meta)))
 1594                                 goto out_err;
 1595                 }
 1596         }
 1597 
 1598         /* building the A-MSDU might have changed this data, so memcpy it now */
 1599         memcpy(&txq->first_tb_bufs[txq->write_ptr], dev_cmd, IWL_FIRST_TB_SIZE);
 1600 
 1601         tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr);
 1602         /* Set up entry for this TFD in Tx byte-count array */
 1603         iwl_txq_gen1_update_byte_cnt_tbl(trans, txq, le16_to_cpu(tx_cmd->len),
 1604                                          iwl_txq_gen1_tfd_get_num_tbs(trans,
 1605                                                                       tfd));
 1606 
 1607         wait_write_ptr = ieee80211_has_morefrags(fc);
 1608 
 1609         /* start timer if queue currently empty */
 1610         if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) {
 1611                 /*
 1612                  * If the TXQ is active, then set the timer, if not,
 1613                  * set the timer in remainder so that the timer will
 1614                  * be armed with the right value when the station will
 1615                  * wake up.
 1616                  */
 1617                 if (!txq->frozen)
 1618                         mod_timer(&txq->stuck_timer,
 1619                                   jiffies + txq->wd_timeout);
 1620                 else
 1621                         txq->frozen_expiry_remainder = txq->wd_timeout;
 1622         }
 1623 
 1624         /* Tell device the write index *just past* this latest filled TFD */
 1625         txq->write_ptr = iwl_txq_inc_wrap(trans, txq->write_ptr);
 1626         if (!wait_write_ptr)
 1627                 iwl_pcie_txq_inc_wr_ptr(trans, txq);
 1628 
 1629         /*
 1630          * At this point the frame is "transmitted" successfully
 1631          * and we will get a TX status notification eventually.
 1632          */
 1633         spin_unlock(&txq->lock);
 1634         return 0;
 1635 out_err:
 1636         iwl_txq_gen1_tfd_unmap(trans, out_meta, txq, txq->write_ptr);
 1637         spin_unlock(&txq->lock);
 1638         return -1;
 1639 }

Cache object: b5e9786e804863c9926941599c569e4c


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]


This page is part of the FreeBSD/Linux Linux Kernel Cross-Reference, and was automatically generated using a modified version of the LXR engine.