1 /* $NetBSD: kern_kthread.c,v 1.47 2022/09/13 09:37:49 riastradh Exp $ */
2
3 /*-
4 * Copyright (c) 1998, 1999, 2007, 2009, 2019 The NetBSD Foundation, Inc.
5 * All rights reserved.
6 *
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility,
9 * NASA Ames Research Center, and by Andrew Doran.
10 *
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
13 * are met:
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in the
18 * documentation and/or other materials provided with the distribution.
19 *
20 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
21 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
22 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
23 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
24 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30 * POSSIBILITY OF SUCH DAMAGE.
31 */
32
33 #include <sys/cdefs.h>
34 __KERNEL_RCSID(0, "$NetBSD: kern_kthread.c,v 1.47 2022/09/13 09:37:49 riastradh Exp $");
35
36 #include <sys/param.h>
37 #include <sys/cpu.h>
38 #include <sys/systm.h>
39 #include <sys/kernel.h>
40 #include <sys/kthread.h>
41 #include <sys/mutex.h>
42 #include <sys/sched.h>
43 #include <sys/kmem.h>
44 #include <sys/msan.h>
45
46 #include <uvm/uvm_extern.h>
47
48 static lwp_t * kthread_jtarget;
49 static kmutex_t kthread_lock;
50 static kcondvar_t kthread_cv;
51
52 void
53 kthread_sysinit(void)
54 {
55
56 mutex_init(&kthread_lock, MUTEX_DEFAULT, IPL_NONE);
57 cv_init(&kthread_cv, "kthrwait");
58 kthread_jtarget = NULL;
59 }
60
61 /*
62 * kthread_create: create a kernel thread, that is, system-only LWP.
63 */
64 int
65 kthread_create(pri_t pri, int flag, struct cpu_info *ci,
66 void (*func)(void *), void *arg, lwp_t **lp, const char *fmt, ...)
67 {
68 lwp_t *l;
69 vaddr_t uaddr;
70 int error, lc;
71 va_list ap;
72
73 KASSERT((flag & KTHREAD_INTR) == 0 || (flag & KTHREAD_MPSAFE) != 0);
74
75 uaddr = uvm_uarea_system_alloc(
76 (flag & (KTHREAD_INTR|KTHREAD_IDLE)) == KTHREAD_IDLE ? ci : NULL);
77 if (uaddr == 0) {
78 return ENOMEM;
79 }
80 kmsan_orig((void *)uaddr, USPACE, KMSAN_TYPE_POOL, __RET_ADDR);
81 if ((flag & KTHREAD_TS) != 0) {
82 lc = SCHED_OTHER;
83 } else {
84 lc = SCHED_RR;
85 }
86
87 error = lwp_create(&lwp0, &proc0, uaddr, LWP_DETACHED, NULL,
88 0, func, arg, &l, lc, &lwp0.l_sigmask, &lwp0.l_sigstk);
89 if (error) {
90 uvm_uarea_system_free(uaddr);
91 return error;
92 }
93 if (fmt != NULL) {
94 l->l_name = kmem_alloc(MAXCOMLEN, KM_SLEEP);
95 va_start(ap, fmt);
96 vsnprintf(l->l_name, MAXCOMLEN, fmt, ap);
97 va_end(ap);
98 }
99
100 /*
101 * Set parameters.
102 */
103 if (pri == PRI_NONE) {
104 if ((flag & KTHREAD_TS) != 0) {
105 /* Maximum user priority level. */
106 pri = MAXPRI_USER;
107 } else {
108 /* Minimum kernel priority level. */
109 pri = PRI_KTHREAD;
110 }
111 }
112 mutex_enter(proc0.p_lock);
113 lwp_lock(l);
114 lwp_changepri(l, pri);
115 if (ci != NULL) {
116 if (ci != l->l_cpu) {
117 lwp_unlock_to(l, ci->ci_schedstate.spc_lwplock);
118 lwp_lock(l);
119 }
120 l->l_pflag |= LP_BOUND;
121 l->l_cpu = ci;
122 }
123
124 if ((flag & KTHREAD_MUSTJOIN) != 0) {
125 KASSERT(lp != NULL);
126 l->l_pflag |= LP_MUSTJOIN;
127 }
128 if ((flag & KTHREAD_INTR) != 0) {
129 l->l_pflag |= LP_INTR;
130 }
131 if ((flag & KTHREAD_MPSAFE) == 0) {
132 l->l_pflag &= ~LP_MPSAFE;
133 }
134
135 /*
136 * Set the new LWP running, unless the caller has requested
137 * otherwise.
138 */
139 KASSERT(l->l_stat == LSIDL);
140 if ((flag & KTHREAD_IDLE) == 0) {
141 setrunnable(l);
142 /* LWP now unlocked */
143 } else {
144 lwp_unlock(l);
145 }
146 mutex_exit(proc0.p_lock);
147
148 /* All done! */
149 if (lp != NULL) {
150 *lp = l;
151 }
152 return 0;
153 }
154
155 /*
156 * Cause a kernel thread to exit. Assumes the exiting thread is the
157 * current context.
158 */
159 void
160 kthread_exit(int ecode)
161 {
162 const char *name;
163 lwp_t *l = curlwp;
164
165 /* We can't do much with the exit code, so just report it. */
166 if (ecode != 0) {
167 if ((name = l->l_name) == NULL)
168 name = "unnamed";
169 printf("WARNING: kthread `%s' (%d) exits with status %d\n",
170 name, l->l_lid, ecode);
171 }
172
173 /* Barrier for joining. */
174 if (l->l_pflag & LP_MUSTJOIN) {
175 mutex_enter(&kthread_lock);
176 while (kthread_jtarget != l) {
177 cv_wait(&kthread_cv, &kthread_lock);
178 }
179 kthread_jtarget = NULL;
180 cv_broadcast(&kthread_cv);
181 mutex_exit(&kthread_lock);
182 }
183
184 /* If the kernel lock is held, we need to drop it now. */
185 if ((l->l_pflag & LP_MPSAFE) == 0) {
186 KERNEL_UNLOCK_LAST(l);
187 }
188
189 /* And exit.. */
190 lwp_exit(l);
191 panic("kthread_exit");
192 }
193
194 /*
195 * Wait for a kthread to exit, as pthread_join().
196 */
197 int
198 kthread_join(lwp_t *l)
199 {
200
201 KASSERT((l->l_flag & LW_SYSTEM) != 0);
202 KASSERT((l->l_pflag & LP_MUSTJOIN) != 0);
203
204 /*
205 * - Wait if some other thread has occupied the target.
206 * - Specify our kthread as a target and notify it.
207 * - Wait for the target kthread to notify us.
208 */
209 mutex_enter(&kthread_lock);
210 while (kthread_jtarget) {
211 cv_wait(&kthread_cv, &kthread_lock);
212 }
213 kthread_jtarget = l;
214 cv_broadcast(&kthread_cv);
215 while (kthread_jtarget == l) {
216 cv_wait(&kthread_cv, &kthread_lock);
217 }
218 mutex_exit(&kthread_lock);
219
220 return 0;
221 }
222
223 /*
224 * kthread_fpu_enter()
225 *
226 * Allow the current lwp, which must be a kthread, to use the FPU.
227 * Return a cookie that must be passed to kthread_fpu_exit when
228 * done. Must be used only in thread context. Recursive -- you
229 * can call kthread_fpu_enter several times in a row as long as
230 * you pass the cookies in reverse order to kthread_fpu_exit.
231 */
232 int
233 kthread_fpu_enter(void)
234 {
235 struct lwp *l = curlwp;
236 int s;
237
238 KASSERTMSG(!cpu_intr_p(),
239 "%s is not allowed in interrupt context", __func__);
240 KASSERTMSG(!cpu_softintr_p(),
241 "%s is not allowed in interrupt context", __func__);
242
243 /*
244 * Remember whether this thread already had FPU access, and
245 * mark this thread as having FPU access.
246 */
247 lwp_lock(l);
248 KASSERTMSG(l->l_flag & LW_SYSTEM,
249 "%s is allowed only in kthreads", __func__);
250 s = l->l_flag & LW_SYSTEM_FPU;
251 l->l_flag |= LW_SYSTEM_FPU;
252 lwp_unlock(l);
253
254 /* Take MD steps to enable the FPU if necessary. */
255 if (s == 0)
256 kthread_fpu_enter_md();
257
258 return s;
259 }
260
261 /*
262 * kthread_fpu_exit(s)
263 *
264 * Restore the current lwp's FPU access to what it was before the
265 * matching call to kthread_fpu_enter() that returned s. Must be
266 * used only in thread context.
267 */
268 void
269 kthread_fpu_exit(int s)
270 {
271 struct lwp *l = curlwp;
272
273 KASSERT(s == (s & LW_SYSTEM_FPU));
274 KASSERTMSG(!cpu_intr_p(),
275 "%s is not allowed in interrupt context", __func__);
276 KASSERTMSG(!cpu_softintr_p(),
277 "%s is not allowed in interrupt context", __func__);
278
279 lwp_lock(l);
280 KASSERTMSG(l->l_flag & LW_SYSTEM,
281 "%s is allowed only in kthreads", __func__);
282 KASSERT(l->l_flag & LW_SYSTEM_FPU);
283 l->l_flag ^= s ^ LW_SYSTEM_FPU;
284 lwp_unlock(l);
285
286 /* Take MD steps to zero and disable the FPU if necessary. */
287 if (s == 0)
288 kthread_fpu_exit_md();
289 }
Cache object: a5d9c7d889ff6cae694f4687cf4ac8e4
|