1 /*
2 * Copyright (C) 2007-2010 Lawrence Livermore National Security, LLC.
3 * Copyright (C) 2007 The Regents of the University of California.
4 * Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER).
5 * Written by Brian Behlendorf <behlendorf1@llnl.gov>.
6 * UCRL-CODE-235197
7 *
8 * This file is part of the SPL, Solaris Porting Layer.
9 *
10 * The SPL is free software; you can redistribute it and/or modify it
11 * under the terms of the GNU General Public License as published by the
12 * Free Software Foundation; either version 2 of the License, or (at your
13 * option) any later version.
14 *
15 * The SPL is distributed in the hope that it will be useful, but WITHOUT
16 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
17 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
18 * for more details.
19 *
20 * You should have received a copy of the GNU General Public License along
21 * with the SPL. If not, see <http://www.gnu.org/licenses/>.
22 */
23
24 #ifndef _SPL_BYTEORDER_H
25 #define _SPL_BYTEORDER_H
26
27 #include <asm/byteorder.h>
28
29 #if defined(__BIG_ENDIAN) && !defined(_ZFS_BIG_ENDIAN)
30 #define _ZFS_BIG_ENDIAN
31 #endif
32
33 #if defined(__LITTLE_ENDIAN) && !defined(_ZFS_LITTLE_ENDIAN)
34 #define _ZFS_LITTLE_ENDIAN
35 #endif
36
37 #include <sys/isa_defs.h>
38
39 #ifdef __COVERITY__
40 /*
41 * Coverity's taint warnings from byteswapping are false positives for us.
42 * Suppress them by hiding byteswapping from Coverity.
43 */
44
45 #define BSWAP_8(x) ((x) & 0xff)
46 #define BSWAP_16(x) ((x) & 0xffff)
47 #define BSWAP_32(x) ((x) & 0xffffffff)
48 #define BSWAP_64(x) (x)
49
50 #else /* __COVERITY__ */
51
52 #define BSWAP_8(x) ((x) & 0xff)
53 #define BSWAP_16(x) ((BSWAP_8(x) << 8) | BSWAP_8((x) >> 8))
54 #define BSWAP_32(x) ((BSWAP_16(x) << 16) | BSWAP_16((x) >> 16))
55 #define BSWAP_64(x) ((BSWAP_32(x) << 32) | BSWAP_32((x) >> 32))
56
57 #endif /* __COVERITY__ */
58
59 #define LE_16(x) cpu_to_le16(x)
60 #define LE_32(x) cpu_to_le32(x)
61 #define LE_64(x) cpu_to_le64(x)
62 #define BE_16(x) cpu_to_be16(x)
63 #define BE_32(x) cpu_to_be32(x)
64 #define BE_64(x) cpu_to_be64(x)
65
66 #define BE_IN8(xa) \
67 *((uint8_t *)(xa))
68
69 #define BE_IN16(xa) \
70 (((uint16_t)BE_IN8(xa) << 8) | BE_IN8((uint8_t *)(xa)+1))
71
72 #define BE_IN32(xa) \
73 (((uint32_t)BE_IN16(xa) << 16) | BE_IN16((uint8_t *)(xa)+2))
74
75 #ifdef _ZFS_BIG_ENDIAN
76 static __inline__ uint64_t
77 htonll(uint64_t n)
78 {
79 return (n);
80 }
81
82 static __inline__ uint64_t
83 ntohll(uint64_t n)
84 {
85 return (n);
86 }
87 #else
88 static __inline__ uint64_t
89 htonll(uint64_t n)
90 {
91 return ((((uint64_t)htonl(n)) << 32) + htonl(n >> 32));
92 }
93
94 static __inline__ uint64_t
95 ntohll(uint64_t n)
96 {
97 return ((((uint64_t)ntohl(n)) << 32) + ntohl(n >> 32));
98 }
99 #endif
100
101 #endif /* SPL_BYTEORDER_H */
Cache object: 5191251430d40bb20ee96328c3f9be1c
|