The Design and Implementation of the FreeBSD Operating System, Second Edition
Now available: The Design and Implementation of the FreeBSD Operating System (Second Edition)


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]

FreeBSD/Linux Kernel Cross Reference
sys/contrib/openzfs/include/os/freebsd/spl/sys/atomic.h

Version: -  FREEBSD  -  FREEBSD-13-STABLE  -  FREEBSD-13-0  -  FREEBSD-12-STABLE  -  FREEBSD-12-0  -  FREEBSD-11-STABLE  -  FREEBSD-11-0  -  FREEBSD-10-STABLE  -  FREEBSD-10-0  -  FREEBSD-9-STABLE  -  FREEBSD-9-0  -  FREEBSD-8-STABLE  -  FREEBSD-8-0  -  FREEBSD-7-STABLE  -  FREEBSD-7-0  -  FREEBSD-6-STABLE  -  FREEBSD-6-0  -  FREEBSD-5-STABLE  -  FREEBSD-5-0  -  FREEBSD-4-STABLE  -  FREEBSD-3-STABLE  -  FREEBSD22  -  l41  -  OPENBSD  -  linux-2.6  -  MK84  -  PLAN9  -  xnu-8792 
SearchContext: -  none  -  3  -  10 

    1 /*
    2  * Copyright (c) 2007 Pawel Jakub Dawidek <pjd@FreeBSD.org>
    3  * All rights reserved.
    4  *
    5  * Redistribution and use in source and binary forms, with or without
    6  * modification, are permitted provided that the following conditions
    7  * are met:
    8  * 1. Redistributions of source code must retain the above copyright
    9  *    notice, this list of conditions and the following disclaimer.
   10  * 2. Redistributions in binary form must reproduce the above copyright
   11  *    notice, this list of conditions and the following disclaimer in the
   12  *    documentation and/or other materials provided with the distribution.
   13  *
   14  * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
   15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
   16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
   17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE LIABLE
   18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
   19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
   20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
   21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
   22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
   23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
   24  * SUCH DAMAGE.
   25  *
   26  * $FreeBSD$
   27  */
   28 
   29 #ifndef _OPENSOLARIS_SYS_ATOMIC_H_
   30 #define _OPENSOLARIS_SYS_ATOMIC_H_
   31 
   32 #ifndef _STANDALONE
   33 
   34 #include <sys/types.h>
   35 #include <machine/atomic.h>
   36 
   37 #define atomic_sub_64   atomic_subtract_64
   38 
   39 #if defined(__i386__) && (defined(_KERNEL) || defined(KLD_MODULE))
   40 #define I386_HAVE_ATOMIC64
   41 #endif
   42 
   43 #if defined(__i386__) || defined(__amd64__) || defined(__arm__)
   44 /* No spurious failures from fcmpset. */
   45 #define STRONG_FCMPSET
   46 #endif
   47 
   48 #if !defined(__LP64__) && !defined(__mips_n32) && \
   49         !defined(ARM_HAVE_ATOMIC64) && !defined(I386_HAVE_ATOMIC64) && \
   50         !defined(HAS_EMULATED_ATOMIC64)
   51 extern void atomic_add_64(volatile uint64_t *target, int64_t delta);
   52 extern void atomic_dec_64(volatile uint64_t *target);
   53 extern uint64_t atomic_swap_64(volatile uint64_t *a, uint64_t value);
   54 extern uint64_t atomic_load_64(volatile uint64_t *a);
   55 extern uint64_t atomic_add_64_nv(volatile uint64_t *target, int64_t delta);
   56 extern uint64_t atomic_cas_64(volatile uint64_t *target, uint64_t cmp,
   57     uint64_t newval);
   58 #endif
   59 
   60 #define membar_consumer()               atomic_thread_fence_acq()
   61 #define membar_producer()               atomic_thread_fence_rel()
   62 #define membar_sync()                   atomic_thread_fence_seq_cst()
   63 
   64 static __inline uint32_t
   65 atomic_add_32_nv(volatile uint32_t *target, int32_t delta)
   66 {
   67         return (atomic_fetchadd_32(target, delta) + delta);
   68 }
   69 
   70 static __inline uint_t
   71 atomic_add_int_nv(volatile uint_t *target, int delta)
   72 {
   73         return (atomic_add_32_nv(target, delta));
   74 }
   75 
   76 static __inline void
   77 atomic_inc_32(volatile uint32_t *target)
   78 {
   79         atomic_add_32(target, 1);
   80 }
   81 
   82 static __inline uint32_t
   83 atomic_inc_32_nv(volatile uint32_t *target)
   84 {
   85         return (atomic_add_32_nv(target, 1));
   86 }
   87 
   88 static __inline void
   89 atomic_dec_32(volatile uint32_t *target)
   90 {
   91         atomic_subtract_32(target, 1);
   92 }
   93 
   94 static __inline uint32_t
   95 atomic_dec_32_nv(volatile uint32_t *target)
   96 {
   97         return (atomic_add_32_nv(target, -1));
   98 }
   99 
  100 #ifndef __sparc64__
  101 static inline uint32_t
  102 atomic_cas_32(volatile uint32_t *target, uint32_t cmp, uint32_t newval)
  103 {
  104 #ifdef STRONG_FCMPSET
  105         (void) atomic_fcmpset_32(target, &cmp, newval);
  106 #else
  107         uint32_t expected = cmp;
  108 
  109         do {
  110                 if (atomic_fcmpset_32(target, &cmp, newval))
  111                         break;
  112         } while (cmp == expected);
  113 #endif
  114         return (cmp);
  115 }
  116 #endif
  117 
  118 #if defined(__LP64__) || defined(__mips_n32) || \
  119         defined(ARM_HAVE_ATOMIC64) || defined(I386_HAVE_ATOMIC64) || \
  120         defined(HAS_EMULATED_ATOMIC64)
  121 static __inline void
  122 atomic_dec_64(volatile uint64_t *target)
  123 {
  124         atomic_subtract_64(target, 1);
  125 }
  126 
  127 static inline uint64_t
  128 atomic_add_64_nv(volatile uint64_t *target, int64_t delta)
  129 {
  130         return (atomic_fetchadd_64(target, delta) + delta);
  131 }
  132 
  133 #ifndef __sparc64__
  134 static inline uint64_t
  135 atomic_cas_64(volatile uint64_t *target, uint64_t cmp, uint64_t newval)
  136 {
  137 #ifdef STRONG_FCMPSET
  138         (void) atomic_fcmpset_64(target, &cmp, newval);
  139 #else
  140         uint64_t expected = cmp;
  141 
  142         do {
  143                 if (atomic_fcmpset_64(target, &cmp, newval))
  144                         break;
  145         } while (cmp == expected);
  146 #endif
  147         return (cmp);
  148 }
  149 #endif
  150 #endif
  151 
  152 static __inline void
  153 atomic_inc_64(volatile uint64_t *target)
  154 {
  155         atomic_add_64(target, 1);
  156 }
  157 
  158 static __inline uint64_t
  159 atomic_inc_64_nv(volatile uint64_t *target)
  160 {
  161         return (atomic_add_64_nv(target, 1));
  162 }
  163 
  164 static __inline uint64_t
  165 atomic_dec_64_nv(volatile uint64_t *target)
  166 {
  167         return (atomic_add_64_nv(target, -1));
  168 }
  169 
  170 #if !defined(COMPAT_32BIT) && defined(__LP64__)
  171 static __inline void *
  172 atomic_cas_ptr(volatile void *target, void *cmp,  void *newval)
  173 {
  174         return ((void *)atomic_cas_64((volatile uint64_t *)target,
  175             (uint64_t)cmp, (uint64_t)newval));
  176 }
  177 #else
  178 static __inline void *
  179 atomic_cas_ptr(volatile void *target, void *cmp,  void *newval)
  180 {
  181         return ((void *)atomic_cas_32((volatile uint32_t *)target,
  182             (uint32_t)cmp, (uint32_t)newval));
  183 }
  184 #endif  /* !defined(COMPAT_32BIT) && defined(__LP64__) */
  185 
  186 #else /* _STANDALONE */
  187 /*
  188  * sometimes atomic_add_64 is defined, sometimes not, but the
  189  * following is always right for the boot loader.
  190  */
  191 #undef atomic_add_64
  192 #define atomic_add_64(ptr, val) *(ptr) += val
  193 #undef atomic_sub_64
  194 #define atomic_sub_64(ptr, val) *(ptr) -= val
  195 #endif /* !_STANDALONE */
  196 
  197 #endif  /* !_OPENSOLARIS_SYS_ATOMIC_H_ */

Cache object: c6e1cb697951a4533538236b1a38e53a


[ source navigation ] [ diff markup ] [ identifier search ] [ freetext search ] [ file search ] [ list types ] [ track identifier ]


This page is part of the FreeBSD/Linux Linux Kernel Cross-Reference, and was automatically generated using a modified version of the LXR engine.