The Design and Implementation of the FreeBSD Operating System, Second Edition
Now available: The Design and Implementation of the FreeBSD Operating System (Second Edition)


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]

FreeBSD/Linux Kernel Cross Reference
sys/arm/broadcom/bcm2835/bcm2835_dma.c

Version: -  FREEBSD  -  FREEBSD-13-STABLE  -  FREEBSD-13-0  -  FREEBSD-12-STABLE  -  FREEBSD-12-0  -  FREEBSD-11-STABLE  -  FREEBSD-11-0  -  FREEBSD-10-STABLE  -  FREEBSD-10-0  -  FREEBSD-9-STABLE  -  FREEBSD-9-0  -  FREEBSD-8-STABLE  -  FREEBSD-8-0  -  FREEBSD-7-STABLE  -  FREEBSD-7-0  -  FREEBSD-6-STABLE  -  FREEBSD-6-0  -  FREEBSD-5-STABLE  -  FREEBSD-5-0  -  FREEBSD-4-STABLE  -  FREEBSD-3-STABLE  -  FREEBSD22  -  l41  -  OPENBSD  -  linux-2.6  -  MK84  -  PLAN9  -  xnu-8792 
SearchContext: -  none  -  3  -  10 

    1 /*-
    2  * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
    3  *
    4  * Copyright (c) 2013 Daisuke Aoyama <aoyama@peach.ne.jp>
    5  * Copyright (c) 2013 Oleksandr Tymoshenko <gonzo@bluezbox.com>
    6  *
    7  * Redistribution and use in source and binary forms, with or without
    8  * modification, are permitted provided that the following conditions
    9  * are met:
   10  * 1. Redistributions of source code must retain the above copyright
   11  *    notice, this list of conditions and the following disclaimer.
   12  * 2. Redistributions in binary form must reproduce the above copyright
   13  *    notice, this list of conditions and the following disclaimer in the
   14  *    documentation and/or other materials provided with the distribution.
   15  *
   16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
   17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
   18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
   19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
   20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
   21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
   22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
   23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
   24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
   25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
   26  * SUCH DAMAGE.
   27  *
   28  */
   29 
   30 #include <sys/cdefs.h>
   31 __FBSDID("$FreeBSD$");
   32 
   33 #include <sys/param.h>
   34 #include <sys/systm.h>
   35 #include <sys/bus.h>
   36 #include <sys/kernel.h>
   37 #include <sys/lock.h>
   38 #include <sys/malloc.h>
   39 #include <sys/module.h>
   40 #include <sys/mutex.h>
   41 #include <sys/queue.h>
   42 #include <sys/resource.h>
   43 #include <sys/rman.h>
   44 
   45 #include <dev/ofw/openfirm.h>
   46 #include <dev/ofw/ofw_bus.h>
   47 #include <dev/ofw/ofw_bus_subr.h>
   48 
   49 #include <vm/vm.h>
   50 #include <vm/pmap.h>
   51 #include <machine/bus.h>
   52 
   53 #include "bcm2835_dma.h"
   54 #include "bcm2835_vcbus.h"
   55 
   56 #define MAX_REG                 9
   57 
   58 /* private flags */
   59 #define BCM_DMA_CH_USED         0x00000001
   60 #define BCM_DMA_CH_FREE         0x40000000
   61 #define BCM_DMA_CH_UNMAP        0x80000000
   62 
   63 /* Register Map (4.2.1.2) */
   64 #define BCM_DMA_CS(n)           (0x100*(n) + 0x00)
   65 #define         CS_ACTIVE               (1 <<  0)
   66 #define         CS_END                  (1 <<  1)
   67 #define         CS_INT                  (1 <<  2)
   68 #define         CS_DREQ                 (1 <<  3)
   69 #define         CS_ISPAUSED             (1 <<  4)
   70 #define         CS_ISHELD               (1 <<  5)
   71 #define         CS_ISWAIT               (1 <<  6)
   72 #define         CS_ERR                  (1 <<  8)
   73 #define         CS_WAITWRT              (1 << 28)
   74 #define         CS_DISDBG               (1 << 29)
   75 #define         CS_ABORT                (1 << 30)
   76 #define         CS_RESET                (1U << 31)
   77 #define BCM_DMA_CBADDR(n)       (0x100*(n) + 0x04)
   78 #define BCM_DMA_INFO(n)         (0x100*(n) + 0x08)
   79 #define         INFO_INT_EN             (1 << 0)
   80 #define         INFO_TDMODE             (1 << 1)
   81 #define         INFO_WAIT_RESP          (1 << 3)
   82 #define         INFO_D_INC              (1 << 4)
   83 #define         INFO_D_WIDTH            (1 << 5)
   84 #define         INFO_D_DREQ             (1 << 6)
   85 #define         INFO_S_INC              (1 << 8)
   86 #define         INFO_S_WIDTH            (1 << 9)
   87 #define         INFO_S_DREQ             (1 << 10)
   88 #define         INFO_WAITS_SHIFT        (21)
   89 #define         INFO_PERMAP_SHIFT       (16)
   90 #define         INFO_PERMAP_MASK        (0x1f << INFO_PERMAP_SHIFT)
   91 
   92 #define BCM_DMA_SRC(n)          (0x100*(n) + 0x0C)
   93 #define BCM_DMA_DST(n)          (0x100*(n) + 0x10)
   94 #define BCM_DMA_LEN(n)          (0x100*(n) + 0x14)
   95 #define BCM_DMA_STRIDE(n)       (0x100*(n) + 0x18)
   96 #define BCM_DMA_CBNEXT(n)       (0x100*(n) + 0x1C)
   97 #define BCM_DMA_DEBUG(n)        (0x100*(n) + 0x20)
   98 #define         DEBUG_ERROR_MASK        (7)
   99 
  100 #define BCM_DMA_INT_STATUS      0xfe0
  101 #define BCM_DMA_ENABLE          0xff0
  102 
  103 /* relative offset from BCM_VC_DMA0_BASE (p.39) */
  104 #define BCM_DMA_CH(n)           (0x100*(n))
  105 
  106 /* channels used by GPU */
  107 #define BCM_DMA_CH_BULK         0
  108 #define BCM_DMA_CH_FAST1        2
  109 #define BCM_DMA_CH_FAST2        3
  110 
  111 #define BCM_DMA_CH_GPU_MASK     ((1 << BCM_DMA_CH_BULK) |       \
  112                                  (1 << BCM_DMA_CH_FAST1) |      \
  113                                  (1 << BCM_DMA_CH_FAST2))
  114 
  115 /* DMA Control Block - 256bit aligned (p.40) */
  116 struct bcm_dma_cb {
  117         uint32_t info;          /* Transfer Information */
  118         uint32_t src;           /* Source Address */
  119         uint32_t dst;           /* Destination Address */
  120         uint32_t len;           /* Transfer Length */
  121         uint32_t stride;        /* 2D Mode Stride */
  122         uint32_t next;          /* Next Control Block Address */
  123         uint32_t rsvd1;         /* Reserved */
  124         uint32_t rsvd2;         /* Reserved */
  125 };
  126 
  127 #ifdef DEBUG
  128 static void bcm_dma_cb_dump(struct bcm_dma_cb *cb);
  129 static void bcm_dma_reg_dump(int ch);
  130 #endif
  131 
  132 /* DMA channel private info */
  133 struct bcm_dma_ch {
  134         int                     ch;
  135         uint32_t                flags;
  136         struct bcm_dma_cb *     cb;
  137         uint32_t                vc_cb;
  138         bus_dmamap_t            dma_map;
  139         void                    (*intr_func)(int, void *);
  140         void *                  intr_arg;
  141 };
  142 
  143 struct bcm_dma_softc {
  144         device_t                sc_dev;
  145         struct mtx              sc_mtx;
  146         struct resource *       sc_mem;
  147         struct resource *       sc_irq[BCM_DMA_CH_MAX];
  148         void *                  sc_intrhand[BCM_DMA_CH_MAX];
  149         struct bcm_dma_ch       sc_dma_ch[BCM_DMA_CH_MAX];
  150         bus_dma_tag_t           sc_dma_tag;
  151 };
  152 
  153 static struct bcm_dma_softc *bcm_dma_sc = NULL;
  154 static uint32_t bcm_dma_channel_mask;
  155 
  156 static struct ofw_compat_data compat_data[] = {
  157         {"broadcom,bcm2835-dma",        1},
  158         {"brcm,bcm2835-dma",            1},
  159         {NULL,                          0}
  160 };
  161 
  162 static void
  163 bcm_dmamap_cb(void *arg, bus_dma_segment_t *segs,
  164         int nseg, int err)
  165 {
  166         bus_addr_t *addr;
  167 
  168         if (err)
  169                 return;
  170 
  171         addr = (bus_addr_t*)arg;
  172         *addr = ARMC_TO_VCBUS(segs[0].ds_addr);
  173 }
  174 
  175 static void
  176 bcm_dma_reset(device_t dev, int ch)
  177 {
  178         struct bcm_dma_softc *sc = device_get_softc(dev);
  179         struct bcm_dma_cb *cb;
  180         uint32_t cs;
  181         int count;
  182 
  183         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  184                 return;
  185 
  186         cs = bus_read_4(sc->sc_mem, BCM_DMA_CS(ch));
  187 
  188         if (cs & CS_ACTIVE) {
  189                 /* pause current task */
  190                 bus_write_4(sc->sc_mem, BCM_DMA_CS(ch), 0);
  191 
  192                 count = 1000;
  193                 do {
  194                         cs = bus_read_4(sc->sc_mem, BCM_DMA_CS(ch));
  195                 } while (!(cs & CS_ISPAUSED) && (count-- > 0));
  196 
  197                 if (!(cs & CS_ISPAUSED)) {
  198                         device_printf(dev,
  199                             "Can't abort DMA transfer at channel %d\n", ch);
  200                 }
  201 
  202                 bus_write_4(sc->sc_mem, BCM_DMA_CBNEXT(ch), 0);
  203 
  204                 /* Complete everything, clear interrupt */
  205                 bus_write_4(sc->sc_mem, BCM_DMA_CS(ch),
  206                     CS_ABORT | CS_INT | CS_END| CS_ACTIVE);
  207         }
  208 
  209         /* clear control blocks */
  210         bus_write_4(sc->sc_mem, BCM_DMA_CBADDR(ch), 0);
  211         bus_write_4(sc->sc_mem, BCM_DMA_CBNEXT(ch), 0);
  212 
  213         /* Reset control block */
  214         cb = sc->sc_dma_ch[ch].cb;
  215         bzero(cb, sizeof(*cb));
  216         cb->info = INFO_WAIT_RESP;
  217 }
  218 
  219 static int
  220 bcm_dma_init(device_t dev)
  221 {
  222         struct bcm_dma_softc *sc = device_get_softc(dev);
  223         uint32_t reg;
  224         struct bcm_dma_ch *ch;
  225         void *cb_virt;
  226         vm_paddr_t cb_phys;
  227         int err;
  228         int i;
  229 
  230         /*
  231          * Only channels set in bcm_dma_channel_mask can be controlled by us.
  232          * The others are out of our control as well as the corresponding bits
  233          * in both BCM_DMA_ENABLE and BCM_DMA_INT_STATUS global registers. As
  234          * these registers are RW ones, there is no safe way how to write only
  235          * the bits which can be controlled by us.
  236          *
  237          * Fortunately, after reset, all channels are enabled in BCM_DMA_ENABLE
  238          * register and all statuses are cleared in BCM_DMA_INT_STATUS one.
  239          * Not touching these registers is a trade off between correct
  240          * initialization which does not count on anything and not messing up
  241          * something we have no control over.
  242          */
  243         reg = bus_read_4(sc->sc_mem, BCM_DMA_ENABLE);
  244         if ((reg & bcm_dma_channel_mask) != bcm_dma_channel_mask)
  245                 device_printf(dev, "channels are not enabled\n");
  246         reg = bus_read_4(sc->sc_mem, BCM_DMA_INT_STATUS);
  247         if ((reg & bcm_dma_channel_mask) != 0)
  248                 device_printf(dev, "statuses are not cleared\n");
  249 
  250         /*
  251          * Allocate DMA chunks control blocks based on p.40 of the peripheral
  252          * spec - control block should be 32-bit aligned.  The DMA controller
  253          * has a full 32-bit register dedicated to this address, so we do not
  254          * need to bother with the per-SoC peripheral restrictions.
  255          */
  256         err = bus_dma_tag_create(bus_get_dma_tag(dev),
  257             1, 0, BUS_SPACE_MAXADDR_32BIT,
  258             BUS_SPACE_MAXADDR, NULL, NULL,
  259             sizeof(struct bcm_dma_cb), 1,
  260             sizeof(struct bcm_dma_cb),
  261             BUS_DMA_ALLOCNOW, NULL, NULL,
  262             &sc->sc_dma_tag);
  263 
  264         if (err) {
  265                 device_printf(dev, "failed allocate DMA tag\n");
  266                 return (err);
  267         }
  268 
  269         /* setup initial settings */
  270         for (i = 0; i < BCM_DMA_CH_MAX; i++) {
  271                 ch = &sc->sc_dma_ch[i];
  272 
  273                 bzero(ch, sizeof(struct bcm_dma_ch));
  274                 ch->ch = i;
  275                 ch->flags = BCM_DMA_CH_UNMAP;
  276 
  277                 if ((bcm_dma_channel_mask & (1 << i)) == 0)
  278                         continue;
  279 
  280                 err = bus_dmamem_alloc(sc->sc_dma_tag, &cb_virt,
  281                     BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO,
  282                     &ch->dma_map);
  283                 if (err) {
  284                         device_printf(dev, "cannot allocate DMA memory\n");
  285                         break;
  286                 }
  287 
  288                 /* 
  289                  * Least alignment for busdma-allocated stuff is cache 
  290                  * line size, so just make sure nothing stupid happened
  291                  * and we got properly aligned address
  292                  */
  293                 if ((uintptr_t)cb_virt & 0x1f) {
  294                         device_printf(dev,
  295                             "DMA address is not 32-bytes aligned: %p\n",
  296                             (void*)cb_virt);
  297                         break;
  298                 }
  299 
  300                 err = bus_dmamap_load(sc->sc_dma_tag, ch->dma_map, cb_virt,
  301                     sizeof(struct bcm_dma_cb), bcm_dmamap_cb, &cb_phys,
  302                     BUS_DMA_WAITOK);
  303                 if (err) {
  304                         device_printf(dev, "cannot load DMA memory\n");
  305                         break;
  306                 }
  307 
  308                 ch->cb = cb_virt;
  309                 ch->vc_cb = cb_phys;
  310                 ch->flags = BCM_DMA_CH_FREE;
  311                 ch->cb->info = INFO_WAIT_RESP;
  312 
  313                 /* reset DMA engine */
  314                 bus_write_4(sc->sc_mem, BCM_DMA_CS(i), CS_RESET);
  315         }
  316 
  317         return (0);
  318 }
  319 
  320 /*
  321  * Allocate DMA channel for further use, returns channel # or
  322  *     BCM_DMA_CH_INVALID
  323  */
  324 int
  325 bcm_dma_allocate(int req_ch)
  326 {
  327         struct bcm_dma_softc *sc = bcm_dma_sc;
  328         int ch = BCM_DMA_CH_INVALID;
  329         int i;
  330 
  331         if (req_ch >= BCM_DMA_CH_MAX)
  332                 return (BCM_DMA_CH_INVALID);
  333 
  334         /* Auto(req_ch < 0) or CH specified */
  335         mtx_lock(&sc->sc_mtx);
  336 
  337         if (req_ch < 0) {
  338                 for (i = 0; i < BCM_DMA_CH_MAX; i++) {
  339                         if (sc->sc_dma_ch[i].flags & BCM_DMA_CH_FREE) {
  340                                 ch = i;
  341                                 sc->sc_dma_ch[ch].flags &= ~BCM_DMA_CH_FREE;
  342                                 sc->sc_dma_ch[ch].flags |= BCM_DMA_CH_USED;
  343                                 break;
  344                         }
  345                 }
  346         }
  347         else {
  348                 if (sc->sc_dma_ch[req_ch].flags & BCM_DMA_CH_FREE) {
  349                         ch = req_ch;
  350                         sc->sc_dma_ch[ch].flags &= ~BCM_DMA_CH_FREE;
  351                         sc->sc_dma_ch[ch].flags |= BCM_DMA_CH_USED;
  352                 }
  353         }
  354 
  355         mtx_unlock(&sc->sc_mtx);
  356         return (ch);
  357 }
  358 
  359 /*
  360  * Frees allocated channel. Returns 0 on success, -1 otherwise
  361  */
  362 int
  363 bcm_dma_free(int ch)
  364 {
  365         struct bcm_dma_softc *sc = bcm_dma_sc;
  366 
  367         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  368                 return (-1);
  369 
  370         mtx_lock(&sc->sc_mtx);
  371         if (sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED) {
  372                 sc->sc_dma_ch[ch].flags |= BCM_DMA_CH_FREE;
  373                 sc->sc_dma_ch[ch].flags &= ~BCM_DMA_CH_USED;
  374                 sc->sc_dma_ch[ch].intr_func = NULL;
  375                 sc->sc_dma_ch[ch].intr_arg = NULL;
  376 
  377                 /* reset DMA engine */
  378                 bcm_dma_reset(sc->sc_dev, ch);
  379         }
  380 
  381         mtx_unlock(&sc->sc_mtx);
  382         return (0);
  383 }
  384 
  385 /*
  386  * Assign handler function for channel interrupt
  387  * Returns 0 on success, -1 otherwise
  388  */
  389 int
  390 bcm_dma_setup_intr(int ch, void (*func)(int, void *), void *arg)
  391 {
  392         struct bcm_dma_softc *sc = bcm_dma_sc;
  393         struct bcm_dma_cb *cb;
  394 
  395         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  396                 return (-1);
  397 
  398         if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED))
  399                 return (-1);
  400 
  401         sc->sc_dma_ch[ch].intr_func = func;
  402         sc->sc_dma_ch[ch].intr_arg = arg;
  403         cb = sc->sc_dma_ch[ch].cb;
  404         cb->info |= INFO_INT_EN;
  405 
  406         return (0);
  407 }
  408 
  409 /*
  410  * Setup DMA source parameters
  411  *     ch - channel number
  412  *     dreq - hardware DREQ # or BCM_DMA_DREQ_NONE if
  413  *         source is physical memory
  414  *     inc_addr - BCM_DMA_INC_ADDR if source address
  415  *         should be increased after each access or 
  416  *         BCM_DMA_SAME_ADDR if address should remain 
  417  *         the same
  418  *     width - size of read operation, BCM_DMA_32BIT
  419  *         for 32bit bursts, BCM_DMA_128BIT for 128 bits
  420  *        
  421  * Returns 0 on success, -1 otherwise
  422  */
  423 int
  424 bcm_dma_setup_src(int ch, int dreq, int inc_addr, int width)
  425 {
  426         struct bcm_dma_softc *sc = bcm_dma_sc;
  427         uint32_t info;
  428 
  429         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  430                 return (-1);
  431 
  432         if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED))
  433                 return (-1);
  434 
  435         info = sc->sc_dma_ch[ch].cb->info;
  436         info &= ~INFO_PERMAP_MASK;
  437         info |= (dreq << INFO_PERMAP_SHIFT) & INFO_PERMAP_MASK;
  438 
  439         if (dreq)
  440                 info |= INFO_S_DREQ;
  441         else
  442                 info &= ~INFO_S_DREQ;
  443 
  444         if (width == BCM_DMA_128BIT)
  445                 info |= INFO_S_WIDTH;
  446         else
  447                 info &= ~INFO_S_WIDTH;
  448 
  449         if (inc_addr == BCM_DMA_INC_ADDR)
  450                 info |= INFO_S_INC;
  451         else
  452                 info &= ~INFO_S_INC;
  453 
  454         sc->sc_dma_ch[ch].cb->info = info;
  455 
  456         return (0);
  457 }
  458 
  459 /*
  460  * Setup DMA destination parameters
  461  *     ch - channel number
  462  *     dreq - hardware DREQ # or BCM_DMA_DREQ_NONE if
  463  *         destination is physical memory
  464  *     inc_addr - BCM_DMA_INC_ADDR if source address
  465  *         should be increased after each access or 
  466  *         BCM_DMA_SAME_ADDR if address should remain 
  467  *         the same
  468  *     width - size of write operation, BCM_DMA_32BIT
  469  *         for 32bit bursts, BCM_DMA_128BIT for 128 bits
  470  *        
  471  * Returns 0 on success, -1 otherwise
  472  */
  473 int
  474 bcm_dma_setup_dst(int ch, int dreq, int inc_addr, int width)
  475 {
  476         struct bcm_dma_softc *sc = bcm_dma_sc;
  477         uint32_t info;
  478 
  479         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  480                 return (-1);
  481 
  482         if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED))
  483                 return (-1);
  484 
  485         info = sc->sc_dma_ch[ch].cb->info;
  486         info &= ~INFO_PERMAP_MASK;
  487         info |= (dreq << INFO_PERMAP_SHIFT) & INFO_PERMAP_MASK;
  488 
  489         if (dreq)
  490                 info |= INFO_D_DREQ;
  491         else
  492                 info &= ~INFO_D_DREQ;
  493 
  494         if (width == BCM_DMA_128BIT)
  495                 info |= INFO_D_WIDTH;
  496         else
  497                 info &= ~INFO_D_WIDTH;
  498 
  499         if (inc_addr == BCM_DMA_INC_ADDR)
  500                 info |= INFO_D_INC;
  501         else
  502                 info &= ~INFO_D_INC;
  503 
  504         sc->sc_dma_ch[ch].cb->info = info;
  505 
  506         return (0);
  507 }
  508 
  509 #ifdef DEBUG
  510 void
  511 bcm_dma_cb_dump(struct bcm_dma_cb *cb)
  512 {
  513 
  514         printf("DMA CB ");
  515         printf("INFO: %8.8x ", cb->info);
  516         printf("SRC: %8.8x ", cb->src);
  517         printf("DST: %8.8x ", cb->dst);
  518         printf("LEN: %8.8x ", cb->len);
  519         printf("\n");
  520         printf("STRIDE: %8.8x ", cb->stride);
  521         printf("NEXT: %8.8x ", cb->next);
  522         printf("RSVD1: %8.8x ", cb->rsvd1);
  523         printf("RSVD2: %8.8x ", cb->rsvd2);
  524         printf("\n");
  525 }
  526 
  527 void
  528 bcm_dma_reg_dump(int ch)
  529 {
  530         struct bcm_dma_softc *sc = bcm_dma_sc;
  531         int i;
  532         uint32_t reg;
  533 
  534         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  535                 return;
  536 
  537         printf("DMA%d: ", ch);
  538         for (i = 0; i < MAX_REG; i++) {
  539                 reg = bus_read_4(sc->sc_mem, BCM_DMA_CH(ch) + i*4);
  540                 printf("%8.8x ", reg);
  541         }
  542         printf("\n");
  543 }
  544 #endif
  545 
  546 /*
  547  * Start DMA transaction
  548  *     ch - channel number
  549  *     src, dst - source and destination address in
  550  *         ARM physical memory address space. 
  551  *     len - amount of bytes to be transferred
  552  *        
  553  * Returns 0 on success, -1 otherwise
  554  */
  555 int
  556 bcm_dma_start(int ch, vm_paddr_t src, vm_paddr_t dst, int len)
  557 {
  558         struct bcm_dma_softc *sc = bcm_dma_sc;
  559         struct bcm_dma_cb *cb;
  560 
  561         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  562                 return (-1);
  563 
  564         if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED))
  565                 return (-1);
  566 
  567         cb = sc->sc_dma_ch[ch].cb;
  568         cb->src = ARMC_TO_VCBUS(src);
  569         cb->dst = ARMC_TO_VCBUS(dst);
  570 
  571         cb->len = len;
  572 
  573         bus_dmamap_sync(sc->sc_dma_tag,
  574             sc->sc_dma_ch[ch].dma_map, BUS_DMASYNC_PREWRITE);
  575 
  576         bus_write_4(sc->sc_mem, BCM_DMA_CBADDR(ch),
  577             sc->sc_dma_ch[ch].vc_cb);
  578         bus_write_4(sc->sc_mem, BCM_DMA_CS(ch), CS_ACTIVE);
  579 
  580 #ifdef DEBUG
  581         bcm_dma_cb_dump(sc->sc_dma_ch[ch].cb);
  582         bcm_dma_reg_dump(ch);
  583 #endif
  584 
  585         return (0);
  586 }
  587 
  588 /*
  589  * Get length requested for DMA transaction
  590  *     ch - channel number
  591  *        
  592  * Returns size of transaction, 0 if channel is invalid
  593  */
  594 uint32_t
  595 bcm_dma_length(int ch)
  596 {
  597         struct bcm_dma_softc *sc = bcm_dma_sc;
  598         struct bcm_dma_cb *cb;
  599 
  600         if (ch < 0 || ch >= BCM_DMA_CH_MAX)
  601                 return (0);
  602 
  603         if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED))
  604                 return (0);
  605 
  606         cb = sc->sc_dma_ch[ch].cb;
  607 
  608         return (cb->len);
  609 }
  610 
  611 static void
  612 bcm_dma_intr(void *arg)
  613 {
  614         struct bcm_dma_softc *sc = bcm_dma_sc;
  615         struct bcm_dma_ch *ch = (struct bcm_dma_ch *)arg;
  616         uint32_t cs, debug;
  617 
  618         /* my interrupt? */
  619         cs = bus_read_4(sc->sc_mem, BCM_DMA_CS(ch->ch));
  620 
  621         /*
  622          * Is it an active channel?  Our diagnostics could be better here, but
  623          * it's not necessarily an easy task to resolve a rid/resource to an
  624          * actual irq number.  We'd want to do this to set a flag indicating
  625          * whether the irq is shared or not, so we know to complain.
  626          */
  627         if (!(ch->flags & BCM_DMA_CH_USED))
  628                 return;
  629 
  630         /* Again, we can't complain here.  The same logic applies. */
  631         if (!(cs & (CS_INT | CS_ERR)))
  632                 return;
  633 
  634         if (cs & CS_ERR) {
  635                 debug = bus_read_4(sc->sc_mem, BCM_DMA_DEBUG(ch->ch));
  636                 device_printf(sc->sc_dev, "DMA error %d on CH%d\n",
  637                         debug & DEBUG_ERROR_MASK, ch->ch);
  638                 bus_write_4(sc->sc_mem, BCM_DMA_DEBUG(ch->ch), 
  639                     debug & DEBUG_ERROR_MASK);
  640                 bcm_dma_reset(sc->sc_dev, ch->ch);
  641         }
  642 
  643         if (cs & CS_INT) {
  644                 /* acknowledge interrupt */
  645                 bus_write_4(sc->sc_mem, BCM_DMA_CS(ch->ch), 
  646                     CS_INT | CS_END);
  647 
  648                 /* Prepare for possible access to len field */
  649                 bus_dmamap_sync(sc->sc_dma_tag, ch->dma_map,
  650                     BUS_DMASYNC_POSTWRITE);
  651 
  652                 /* save callback function and argument */
  653                 if (ch->intr_func)
  654                         ch->intr_func(ch->ch, ch->intr_arg);
  655         }
  656 }
  657 
  658 static int
  659 bcm_dma_probe(device_t dev)
  660 {
  661 
  662         if (!ofw_bus_status_okay(dev))
  663                 return (ENXIO);
  664 
  665         if (ofw_bus_search_compatible(dev, compat_data)->ocd_data == 0)
  666                 return (ENXIO);
  667 
  668         device_set_desc(dev, "BCM2835 DMA Controller");
  669         return (BUS_PROBE_DEFAULT);
  670 }
  671 
  672 static int
  673 bcm_dma_attach(device_t dev)
  674 {
  675         struct bcm_dma_softc *sc = device_get_softc(dev);
  676         phandle_t node;
  677         int rid, err = 0;
  678         int i;
  679 
  680         sc->sc_dev = dev;
  681 
  682         if (bcm_dma_sc)
  683                 return (ENXIO);
  684 
  685         for (i = 0; i < BCM_DMA_CH_MAX; i++) {
  686                 sc->sc_irq[i] = NULL;
  687                 sc->sc_intrhand[i] = NULL;
  688         }
  689 
  690         /* Get DMA channel mask. */
  691         node = ofw_bus_get_node(sc->sc_dev);
  692         if (OF_getencprop(node, "brcm,dma-channel-mask", &bcm_dma_channel_mask,
  693             sizeof(bcm_dma_channel_mask)) == -1 &&
  694             OF_getencprop(node, "broadcom,channels", &bcm_dma_channel_mask,
  695             sizeof(bcm_dma_channel_mask)) == -1) {
  696                 device_printf(dev, "could not get channel mask property\n");
  697                 return (ENXIO);
  698         }
  699 
  700         /* Mask out channels used by GPU. */
  701         bcm_dma_channel_mask &= ~BCM_DMA_CH_GPU_MASK;
  702 
  703         /* DMA0 - DMA14 */
  704         rid = 0;
  705         sc->sc_mem = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid, RF_ACTIVE);
  706         if (sc->sc_mem == NULL) {
  707                 device_printf(dev, "could not allocate memory resource\n");
  708                 return (ENXIO);
  709         }
  710 
  711         /* IRQ DMA0 - DMA11 XXX NOT USE DMA12(spurious?) */
  712         for (rid = 0; rid < BCM_DMA_CH_MAX; rid++) {
  713                 if ((bcm_dma_channel_mask & (1 << rid)) == 0)
  714                         continue;
  715 
  716                 sc->sc_irq[rid] = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid,
  717                     RF_ACTIVE | RF_SHAREABLE);
  718                 if (sc->sc_irq[rid] == NULL) {
  719                         device_printf(dev, "cannot allocate interrupt\n");
  720                         err = ENXIO;
  721                         goto fail;
  722                 }
  723                 if (bus_setup_intr(dev, sc->sc_irq[rid], INTR_TYPE_MISC | INTR_MPSAFE,
  724                                    NULL, bcm_dma_intr, &sc->sc_dma_ch[rid],
  725                                    &sc->sc_intrhand[rid])) {
  726                         device_printf(dev, "cannot setup interrupt handler\n");
  727                         err = ENXIO;
  728                         goto fail;
  729                 }
  730         }
  731 
  732         mtx_init(&sc->sc_mtx, "bcmdma", "bcmdma", MTX_DEF);
  733         bcm_dma_sc = sc;
  734 
  735         err = bcm_dma_init(dev);
  736         if (err)
  737                 goto fail;
  738 
  739         return (err);
  740 
  741 fail:
  742         if (sc->sc_mem)
  743                 bus_release_resource(dev, SYS_RES_MEMORY, 0, sc->sc_mem);
  744 
  745         for (i = 0; i < BCM_DMA_CH_MAX; i++) {
  746                 if (sc->sc_intrhand[i])
  747                         bus_teardown_intr(dev, sc->sc_irq[i], sc->sc_intrhand[i]);
  748                 if (sc->sc_irq[i])
  749                         bus_release_resource(dev, SYS_RES_IRQ, 0, sc->sc_irq[i]);
  750         }
  751 
  752         return (err);
  753 }
  754 
  755 static device_method_t bcm_dma_methods[] = {
  756         DEVMETHOD(device_probe,         bcm_dma_probe),
  757         DEVMETHOD(device_attach,        bcm_dma_attach),
  758         { 0, 0 }
  759 };
  760 
  761 static driver_t bcm_dma_driver = {
  762         "bcm_dma",
  763         bcm_dma_methods,
  764         sizeof(struct bcm_dma_softc),
  765 };
  766 
  767 DRIVER_MODULE(bcm_dma, simplebus, bcm_dma_driver, 0, 0);
  768 MODULE_VERSION(bcm_dma, 1);

Cache object: 46feefc9ed27e37ed19c55d7a3488dc9


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]


This page is part of the FreeBSD/Linux Linux Kernel Cross-Reference, and was automatically generated using a modified version of the LXR engine.