1 /* $NetBSD: cpu.h,v 1.123 2022/06/25 12:41:56 jmcneill Exp $ */
2
3 /*
4 * Copyright (c) 1994-1996 Mark Brinicombe.
5 * Copyright (c) 1994 Brini.
6 * All rights reserved.
7 *
8 * This code is derived from software written for Brini by Mark Brinicombe
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions and the following disclaimer.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by Brini.
21 * 4. The name of the company nor the name of the author may be used to
22 * endorse or promote products derived from this software without specific
23 * prior written permission.
24 *
25 * THIS SOFTWARE IS PROVIDED BY BRINI ``AS IS'' AND ANY EXPRESS OR IMPLIED
26 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
27 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
28 * IN NO EVENT SHALL BRINI OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,
29 * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
30 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
31 * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
32 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
33 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
34 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
35 * SUCH DAMAGE.
36 *
37 * RiscBSD kernel project
38 *
39 * cpu.h
40 *
41 * CPU specific symbols
42 *
43 * Created : 18/09/94
44 *
45 * Based on kate/katelib/arm6.h
46 */
47
48 #ifndef _ARM_CPU_H_
49 #define _ARM_CPU_H_
50
51 #ifdef _KERNEL
52 #ifndef _LOCORE
53
54 typedef unsigned long mpidr_t;
55
56 #ifdef MULTIPROCESSOR
57 extern u_int arm_cpu_max;
58 extern mpidr_t cpu_mpidr[];
59
60 void cpu_init_secondary_processor(int);
61 void cpu_boot_secondary_processors(void);
62 void cpu_mpstart(void);
63 bool cpu_hatched_p(u_int);
64
65 void cpu_clr_mbox(int);
66 void cpu_set_hatched(int);
67
68 #endif
69
70 struct proc;
71
72 void cpu_proc_fork(struct proc *, struct proc *);
73
74 #endif /* !_LOCORE */
75 #endif /* _KERNEL */
76
77 #ifdef __arm__
78
79 /*
80 * User-visible definitions
81 */
82
83 /* CTL_MACHDEP definitions. */
84 #define CPU_DEBUG 1 /* int: misc kernel debug control */
85 #define CPU_BOOTED_DEVICE 2 /* string: device we booted from */
86 #define CPU_BOOTED_KERNEL 3 /* string: kernel we booted */
87 #define CPU_CONSDEV 4 /* struct: dev_t of our console */
88 #define CPU_POWERSAVE 5 /* int: use CPU powersave mode */
89
90 #if defined(_KERNEL) || defined(_KMEMUSER)
91
92 /*
93 * Kernel-only definitions
94 */
95
96 #if !defined(_MODULE) && defined(_KERNEL_OPT)
97 #include "opt_gprof.h"
98 #include "opt_multiprocessor.h"
99 #include "opt_cpuoptions.h"
100 #include "opt_lockdebug.h"
101 #include "opt_cputypes.h"
102 #endif /* !_MODULE && _KERNEL_OPT */
103
104 #ifndef _LOCORE
105 #if defined(TPIDRPRW_IS_CURLWP) || defined(TPIDRPRW_IS_CURCPU)
106 #include <arm/armreg.h>
107 #endif /* TPIDRPRW_IS_CURLWP || TPIDRPRW_IS_CURCPU */
108
109 /* 1 == use cpu_sleep(), 0 == don't */
110 extern int cpu_do_powersave;
111 extern int cpu_fpu_present;
112
113 /* All the CLKF_* macros take a struct clockframe * as an argument. */
114
115 /*
116 * CLKF_USERMODE: Return TRUE/FALSE (1/0) depending on whether the
117 * frame came from USR mode or not.
118 */
119 #define CLKF_USERMODE(cf) (((cf)->cf_tf.tf_spsr & PSR_MODE) == PSR_USR32_MODE)
120
121 /*
122 * CLKF_INTR: True if we took the interrupt from inside another
123 * interrupt handler.
124 */
125 #if !defined(__ARM_EABI__)
126 /* Hack to treat FPE time as interrupt time so we can measure it */
127 #define CLKF_INTR(cf) \
128 ((curcpu()->ci_intr_depth > 1) || \
129 ((cf)->cf_tf.tf_spsr & PSR_MODE) == PSR_UND32_MODE)
130 #else
131 #define CLKF_INTR(cf) ((void)(cf), curcpu()->ci_intr_depth > 1)
132 #endif
133
134 /*
135 * CLKF_PC: Extract the program counter from a clockframe
136 */
137 #define CLKF_PC(frame) (frame->cf_tf.tf_pc)
138
139 /*
140 * LWP_PC: Find out the program counter for the given lwp.
141 */
142 #define LWP_PC(l) (lwp_trapframe(l)->tf_pc)
143
144 /*
145 * Per-CPU information. For now we assume one CPU.
146 */
147 #ifdef _KERNEL
148 static inline int curcpl(void);
149 static inline void set_curcpl(int);
150 static inline void cpu_dosoftints(void);
151 #endif
152
153 #ifdef _KMEMUSER
154 #include <sys/intr.h>
155 #endif
156 #include <sys/atomic.h>
157 #include <sys/cpu_data.h>
158 #include <sys/device_if.h>
159 #include <sys/evcnt.h>
160
161 #include <machine/param.h>
162
163 /*
164 * Cache info variables.
165 */
166 #define CACHE_TYPE_VIVT 0
167 #define CACHE_TYPE_xxPT 1
168 #define CACHE_TYPE_VIPT 1
169 #define CACHE_TYPE_PIxx 2
170 #define CACHE_TYPE_PIPT 3
171
172 /* PRIMARY CACHE VARIABLES */
173 struct arm_cache_info {
174 u_int icache_size;
175 u_int icache_line_size;
176 u_int icache_ways;
177 u_int icache_way_size;
178 u_int icache_sets;
179
180 u_int dcache_size;
181 u_int dcache_line_size;
182 u_int dcache_ways;
183 u_int dcache_way_size;
184 u_int dcache_sets;
185
186 uint8_t cache_type;
187 bool cache_unified;
188 uint8_t icache_type;
189 uint8_t dcache_type;
190 };
191
192 struct cpu_info {
193 struct cpu_data ci_data; /* MI per-cpu data */
194 device_t ci_dev; /* Device corresponding to this CPU */
195 cpuid_t ci_cpuid;
196 uint32_t ci_arm_cpuid; /* aggregate CPU id */
197 uint32_t ci_arm_cputype; /* CPU type */
198 uint32_t ci_arm_cpurev; /* CPU revision */
199 uint32_t ci_ctrl; /* The CPU control register */
200
201 /*
202 * the following are in their own cache line, as they are stored to
203 * regularly by remote CPUs; when they were mixed with other fields
204 * we observed frequent cache misses.
205 */
206 int ci_want_resched __aligned(COHERENCY_UNIT);
207 /* resched() was called */
208 lwp_t * ci_curlwp __aligned(COHERENCY_UNIT);
209 /* current lwp */
210 lwp_t * ci_onproc; /* current user LWP / kthread */
211
212 /*
213 * largely CPU-private.
214 */
215 lwp_t * ci_softlwps[SOFTINT_COUNT] __aligned(COHERENCY_UNIT);
216
217 struct cpu_softc *
218 ci_softc; /* platform softc */
219
220 int ci_cpl; /* current processor level (spl) */
221 volatile int ci_hwpl; /* current hardware priority */
222 int ci_kfpu_spl;
223
224 volatile u_int ci_intr_depth; /* */
225 volatile u_int ci_softints;
226 volatile uint32_t ci_blocked_pics;
227 volatile uint32_t ci_pending_pics;
228 volatile uint32_t ci_pending_ipls;
229
230 lwp_t * ci_lastlwp; /* last lwp */
231
232 struct evcnt ci_arm700bugcount;
233 int32_t ci_mtx_count;
234 int ci_mtx_oldspl;
235 register_t ci_undefsave[3];
236 uint32_t ci_vfp_id;
237 uint64_t ci_lastintr;
238
239 struct pmap_tlb_info *
240 ci_tlb_info;
241 struct pmap * ci_pmap_lastuser;
242 struct pmap * ci_pmap_cur;
243 tlb_asid_t ci_pmap_asid_cur;
244
245 struct trapframe *
246 ci_ddb_regs;
247
248 struct evcnt ci_abt_evs[16];
249 struct evcnt ci_und_ev;
250 struct evcnt ci_und_cp15_ev;
251 struct evcnt ci_vfp_evs[3];
252
253 uint32_t ci_midr;
254 uint32_t ci_actlr;
255 uint32_t ci_revidr;
256 uint32_t ci_mpidr;
257 uint32_t ci_mvfr[2];
258
259 uint32_t ci_capacity_dmips_mhz;
260
261 struct arm_cache_info
262 ci_cacheinfo;
263
264 #if defined(GPROF) && defined(MULTIPROCESSOR)
265 struct gmonparam *ci_gmon; /* MI per-cpu GPROF */
266 #endif
267 };
268
269 extern struct cpu_info cpu_info_store[];
270
271 struct lwp *arm_curlwp(void);
272 struct cpu_info *arm_curcpu(void);
273
274 #ifdef _KERNEL
275 #if defined(_MODULE)
276
277 #define curlwp arm_curlwp()
278 #define curcpu() arm_curcpu()
279
280 #elif defined(TPIDRPRW_IS_CURLWP)
281 static inline struct lwp *
282 _curlwp(void)
283 {
284 return (struct lwp *) armreg_tpidrprw_read();
285 }
286
287 static inline void
288 _curlwp_set(struct lwp *l)
289 {
290 armreg_tpidrprw_write((uintptr_t)l);
291 }
292
293 // Also in <sys/lwp.h> but also here if this was included before <sys/lwp.h>
294 static inline struct cpu_info *lwp_getcpu(struct lwp *);
295
296 #define curlwp _curlwp()
297 // curcpu() expands into two instructions: a mrc and a ldr
298 #define curcpu() lwp_getcpu(_curlwp())
299 #elif defined(TPIDRPRW_IS_CURCPU)
300 #ifdef __HAVE_PREEMPTION
301 #error __HAVE_PREEMPTION requires TPIDRPRW_IS_CURLWP
302 #endif
303 static inline struct cpu_info *
304 curcpu(void)
305 {
306 return (struct cpu_info *) armreg_tpidrprw_read();
307 }
308 #elif !defined(MULTIPROCESSOR)
309 #define curcpu() (&cpu_info_store[0])
310 #elif !defined(__HAVE_PREEMPTION)
311 #error MULTIPROCESSOR && !__HAVE_PREEMPTION requires TPIDRPRW_IS_CURCPU or TPIDRPRW_IS_CURLWP
312 #else
313 #error MULTIPROCESSOR && __HAVE_PREEMPTION requires TPIDRPRW_IS_CURLWP
314 #endif /* !TPIDRPRW_IS_CURCPU && !TPIDRPRW_IS_CURLWP */
315
316 #ifndef curlwp
317 #define curlwp (curcpu()->ci_curlwp)
318 #endif
319 #define curpcb ((struct pcb *)lwp_getpcb(curlwp))
320
321 #define CPU_INFO_ITERATOR int
322 #if defined(_MODULE) || defined(MULTIPROCESSOR)
323 extern struct cpu_info *cpu_info[];
324 #define cpu_number() (curcpu()->ci_index)
325 #define CPU_IS_PRIMARY(ci) ((ci)->ci_index == 0)
326 #define CPU_INFO_FOREACH(cii, ci) \
327 cii = 0, ci = cpu_info[0]; cii < (ncpu ? ncpu : 1) && (ci = cpu_info[cii]) != NULL; cii++
328 #else
329 #define cpu_number() 0
330
331 #define CPU_IS_PRIMARY(ci) true
332 #define CPU_INFO_FOREACH(cii, ci) \
333 cii = 0, __USE(cii), ci = curcpu(); ci != NULL; ci = NULL
334 #endif
335
336 #define LWP0_CPU_INFO (&cpu_info_store[0])
337
338 static inline int
339 curcpl(void)
340 {
341 return curcpu()->ci_cpl;
342 }
343
344 static inline void
345 set_curcpl(int pri)
346 {
347 curcpu()->ci_cpl = pri;
348 }
349
350 static inline void
351 cpu_dosoftints(void)
352 {
353 #ifdef __HAVE_FAST_SOFTINTS
354 void dosoftints(void);
355 #ifndef __HAVE_PIC_FAST_SOFTINTS
356 struct cpu_info * const ci = curcpu();
357 if (ci->ci_intr_depth == 0 && (ci->ci_softints >> ci->ci_cpl) > 0)
358 dosoftints();
359 #endif
360 #endif
361 }
362
363 /*
364 * Scheduling glue
365 */
366 void cpu_signotify(struct lwp *);
367 #define setsoftast(ci) (cpu_signotify((ci)->ci_onproc))
368
369 /*
370 * Give a profiling tick to the current process when the user profiling
371 * buffer pages are invalid. On the i386, request an ast to send us
372 * through trap(), marking the proc as needing a profiling tick.
373 */
374 #define cpu_need_proftick(l) ((l)->l_pflag |= LP_OWEUPC, \
375 setsoftast(lwp_getcpu(l)))
376
377 /*
378 * We've already preallocated the stack for the idlelwps for additional CPUs.
379 * This hook allows to return them.
380 */
381 vaddr_t cpu_uarea_alloc_idlelwp(struct cpu_info *);
382
383 #ifdef _ARM_ARCH_6
384 int cpu_maxproc_hook(int);
385 #endif
386
387 #endif /* _KERNEL */
388
389 #endif /* !_LOCORE */
390
391 #endif /* _KERNEL || _KMEMUSER */
392
393 #elif defined(__aarch64__)
394
395 #include <aarch64/cpu.h>
396
397 #endif /* __arm__/__aarch64__ */
398
399 #endif /* !_ARM_CPU_H_ */
Cache object: 60a6c46c00a25d7d5c1cdfeec6714a91
|