1 /*-
2 * Copyright (c) 2005 Pawel Jakub Dawidek <pjd@FreeBSD.org>
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
13 *
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24 * SUCH DAMAGE.
25 */
26
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
29
30 #include <sys/param.h>
31 #include <sys/systm.h>
32 #include <sys/kernel.h>
33 #include <sys/module.h>
34 #include <sys/lock.h>
35 #include <sys/mutex.h>
36 #include <sys/bio.h>
37 #include <sys/sysctl.h>
38 #include <sys/malloc.h>
39 #include <vm/uma.h>
40 #include <geom/geom.h>
41 #include <geom/shsec/g_shsec.h>
42
43
44 static MALLOC_DEFINE(M_SHSEC, "shsec data", "GEOM_SHSEC Data");
45
46 static uma_zone_t g_shsec_zone;
47
48 static int g_shsec_destroy(struct g_shsec_softc *sc, boolean_t force);
49 static int g_shsec_destroy_geom(struct gctl_req *req, struct g_class *mp,
50 struct g_geom *gp);
51
52 static g_taste_t g_shsec_taste;
53 static g_ctl_req_t g_shsec_config;
54 static g_dumpconf_t g_shsec_dumpconf;
55 static g_init_t g_shsec_init;
56 static g_fini_t g_shsec_fini;
57
58 struct g_class g_shsec_class = {
59 .name = G_SHSEC_CLASS_NAME,
60 .version = G_VERSION,
61 .ctlreq = g_shsec_config,
62 .taste = g_shsec_taste,
63 .destroy_geom = g_shsec_destroy_geom,
64 .init = g_shsec_init,
65 .fini = g_shsec_fini
66 };
67
68 SYSCTL_DECL(_kern_geom);
69 SYSCTL_NODE(_kern_geom, OID_AUTO, shsec, CTLFLAG_RW, 0, "GEOM_SHSEC stuff");
70 static u_int g_shsec_debug = 0;
71 TUNABLE_INT("kern.geom.shsec.debug", &g_shsec_debug);
72 SYSCTL_UINT(_kern_geom_shsec, OID_AUTO, debug, CTLFLAG_RW, &g_shsec_debug, 0,
73 "Debug level");
74 static u_int g_shsec_maxmem = MAXPHYS * 100;
75 TUNABLE_INT("kern.geom.shsec.maxmem", &g_shsec_maxmem);
76 SYSCTL_UINT(_kern_geom_shsec, OID_AUTO, maxmem, CTLFLAG_RD, &g_shsec_maxmem,
77 0, "Maximum memory that can be allocated for I/O (in bytes)");
78 static u_int g_shsec_alloc_failed = 0;
79 SYSCTL_UINT(_kern_geom_shsec, OID_AUTO, alloc_failed, CTLFLAG_RD,
80 &g_shsec_alloc_failed, 0, "How many times I/O allocation failed");
81
82 /*
83 * Greatest Common Divisor.
84 */
85 static u_int
86 gcd(u_int a, u_int b)
87 {
88 u_int c;
89
90 while (b != 0) {
91 c = a;
92 a = b;
93 b = (c % b);
94 }
95 return (a);
96 }
97
98 /*
99 * Least Common Multiple.
100 */
101 static u_int
102 lcm(u_int a, u_int b)
103 {
104
105 return ((a * b) / gcd(a, b));
106 }
107
108 static void
109 g_shsec_init(struct g_class *mp __unused)
110 {
111
112 g_shsec_zone = uma_zcreate("g_shsec_zone", MAXPHYS, NULL, NULL, NULL,
113 NULL, 0, 0);
114 g_shsec_maxmem -= g_shsec_maxmem % MAXPHYS;
115 uma_zone_set_max(g_shsec_zone, g_shsec_maxmem / MAXPHYS);
116 }
117
118 static void
119 g_shsec_fini(struct g_class *mp __unused)
120 {
121
122 uma_zdestroy(g_shsec_zone);
123 }
124
125 /*
126 * Return the number of valid disks.
127 */
128 static u_int
129 g_shsec_nvalid(struct g_shsec_softc *sc)
130 {
131 u_int i, no;
132
133 no = 0;
134 for (i = 0; i < sc->sc_ndisks; i++) {
135 if (sc->sc_disks[i] != NULL)
136 no++;
137 }
138
139 return (no);
140 }
141
142 static void
143 g_shsec_remove_disk(struct g_consumer *cp)
144 {
145 struct g_shsec_softc *sc;
146 u_int no;
147
148 KASSERT(cp != NULL, ("Non-valid disk in %s.", __func__));
149 sc = (struct g_shsec_softc *)cp->private;
150 KASSERT(sc != NULL, ("NULL sc in %s.", __func__));
151 no = cp->index;
152
153 G_SHSEC_DEBUG(0, "Disk %s removed from %s.", cp->provider->name,
154 sc->sc_name);
155
156 sc->sc_disks[no] = NULL;
157 if (sc->sc_provider != NULL) {
158 g_orphan_provider(sc->sc_provider, ENXIO);
159 sc->sc_provider = NULL;
160 G_SHSEC_DEBUG(0, "Device %s removed.", sc->sc_name);
161 }
162
163 if (cp->acr > 0 || cp->acw > 0 || cp->ace > 0)
164 g_access(cp, -cp->acr, -cp->acw, -cp->ace);
165 g_detach(cp);
166 g_destroy_consumer(cp);
167 }
168
169 static void
170 g_shsec_orphan(struct g_consumer *cp)
171 {
172 struct g_shsec_softc *sc;
173 struct g_geom *gp;
174
175 g_topology_assert();
176 gp = cp->geom;
177 sc = gp->softc;
178 if (sc == NULL)
179 return;
180
181 g_shsec_remove_disk(cp);
182 /* If there are no valid disks anymore, remove device. */
183 if (g_shsec_nvalid(sc) == 0)
184 g_shsec_destroy(sc, 1);
185 }
186
187 static int
188 g_shsec_access(struct g_provider *pp, int dr, int dw, int de)
189 {
190 struct g_consumer *cp1, *cp2;
191 struct g_shsec_softc *sc;
192 struct g_geom *gp;
193 int error;
194
195 gp = pp->geom;
196 sc = gp->softc;
197
198 if (sc == NULL) {
199 /*
200 * It looks like geom is being withered.
201 * In that case we allow only negative requests.
202 */
203 KASSERT(dr <= 0 && dw <= 0 && de <= 0,
204 ("Positive access request (device=%s).", pp->name));
205 if ((pp->acr + dr) == 0 && (pp->acw + dw) == 0 &&
206 (pp->ace + de) == 0) {
207 G_SHSEC_DEBUG(0, "Device %s definitely destroyed.",
208 gp->name);
209 }
210 return (0);
211 }
212
213 /* On first open, grab an extra "exclusive" bit */
214 if (pp->acr == 0 && pp->acw == 0 && pp->ace == 0)
215 de++;
216 /* ... and let go of it on last close */
217 if ((pp->acr + dr) == 0 && (pp->acw + dw) == 0 && (pp->ace + de) == 0)
218 de--;
219
220 error = ENXIO;
221 LIST_FOREACH(cp1, &gp->consumer, consumer) {
222 error = g_access(cp1, dr, dw, de);
223 if (error == 0)
224 continue;
225 /*
226 * If we fail here, backout all previous changes.
227 */
228 LIST_FOREACH(cp2, &gp->consumer, consumer) {
229 if (cp1 == cp2)
230 return (error);
231 g_access(cp2, -dr, -dw, -de);
232 }
233 /* NOTREACHED */
234 }
235
236 return (error);
237 }
238
239 static void
240 g_shsec_xor1(uint32_t *src, uint32_t *dst, ssize_t len)
241 {
242
243 for (; len > 0; len -= sizeof(uint32_t), dst++)
244 *dst = *dst ^ *src++;
245 KASSERT(len == 0, ("len != 0 (len=%zd)", len));
246 }
247
248 static void
249 g_shsec_done(struct bio *bp)
250 {
251 struct g_shsec_softc *sc;
252 struct bio *pbp;
253
254 pbp = bp->bio_parent;
255 sc = pbp->bio_to->geom->softc;
256 if (bp->bio_error == 0)
257 G_SHSEC_LOGREQ(2, bp, "Request done.");
258 else {
259 G_SHSEC_LOGREQ(0, bp, "Request failed (error=%d).",
260 bp->bio_error);
261 if (pbp->bio_error == 0)
262 pbp->bio_error = bp->bio_error;
263 }
264 if (pbp->bio_cmd == BIO_READ) {
265 if ((pbp->bio_pflags & G_SHSEC_BFLAG_FIRST) != 0) {
266 bcopy(bp->bio_data, pbp->bio_data, pbp->bio_length);
267 pbp->bio_pflags = 0;
268 } else {
269 g_shsec_xor1((uint32_t *)bp->bio_data,
270 (uint32_t *)pbp->bio_data,
271 (ssize_t)pbp->bio_length);
272 }
273 }
274 bzero(bp->bio_data, bp->bio_length);
275 uma_zfree(g_shsec_zone, bp->bio_data);
276 g_destroy_bio(bp);
277 pbp->bio_inbed++;
278 if (pbp->bio_children == pbp->bio_inbed) {
279 pbp->bio_completed = pbp->bio_length;
280 g_io_deliver(pbp, pbp->bio_error);
281 }
282 }
283
284 static void
285 g_shsec_xor2(uint32_t *rand, uint32_t *dst, ssize_t len)
286 {
287
288 for (; len > 0; len -= sizeof(uint32_t), dst++) {
289 *rand = arc4random();
290 *dst = *dst ^ *rand++;
291 }
292 KASSERT(len == 0, ("len != 0 (len=%zd)", len));
293 }
294
295 static void
296 g_shsec_start(struct bio *bp)
297 {
298 TAILQ_HEAD(, bio) queue = TAILQ_HEAD_INITIALIZER(queue);
299 struct g_shsec_softc *sc;
300 struct bio *cbp;
301 uint32_t *dst;
302 ssize_t len;
303 u_int no;
304 int error;
305
306 sc = bp->bio_to->geom->softc;
307 /*
308 * If sc == NULL, provider's error should be set and g_shsec_start()
309 * should not be called at all.
310 */
311 KASSERT(sc != NULL,
312 ("Provider's error should be set (error=%d)(device=%s).",
313 bp->bio_to->error, bp->bio_to->name));
314
315 G_SHSEC_LOGREQ(2, bp, "Request received.");
316
317 switch (bp->bio_cmd) {
318 case BIO_READ:
319 case BIO_WRITE:
320 /*
321 * Only those requests are supported.
322 */
323 break;
324 case BIO_DELETE:
325 case BIO_GETATTR:
326 /* To which provider it should be delivered? */
327 default:
328 g_io_deliver(bp, EOPNOTSUPP);
329 return;
330 }
331
332 /*
333 * Allocate all bios first and calculate XOR.
334 */
335 dst = NULL;
336 len = bp->bio_length;
337 if (bp->bio_cmd == BIO_READ)
338 bp->bio_pflags = G_SHSEC_BFLAG_FIRST;
339 for (no = 0; no < sc->sc_ndisks; no++) {
340 cbp = g_clone_bio(bp);
341 if (cbp == NULL) {
342 error = ENOMEM;
343 goto failure;
344 }
345 TAILQ_INSERT_TAIL(&queue, cbp, bio_queue);
346
347 /*
348 * Fill in the component buf structure.
349 */
350 cbp->bio_done = g_shsec_done;
351 cbp->bio_data = uma_zalloc(g_shsec_zone, M_NOWAIT);
352 if (cbp->bio_data == NULL) {
353 g_shsec_alloc_failed++;
354 error = ENOMEM;
355 goto failure;
356 }
357 cbp->bio_caller2 = sc->sc_disks[no];
358 if (bp->bio_cmd == BIO_WRITE) {
359 if (no == 0) {
360 dst = (uint32_t *)cbp->bio_data;
361 bcopy(bp->bio_data, dst, len);
362 } else {
363 g_shsec_xor2((uint32_t *)cbp->bio_data, dst,
364 len);
365 }
366 }
367 }
368 /*
369 * Fire off all allocated requests!
370 */
371 while ((cbp = TAILQ_FIRST(&queue)) != NULL) {
372 struct g_consumer *cp;
373
374 TAILQ_REMOVE(&queue, cbp, bio_queue);
375 cp = cbp->bio_caller2;
376 cbp->bio_caller2 = NULL;
377 cbp->bio_to = cp->provider;
378 G_SHSEC_LOGREQ(2, cbp, "Sending request.");
379 g_io_request(cbp, cp);
380 }
381 return;
382 failure:
383 while ((cbp = TAILQ_FIRST(&queue)) != NULL) {
384 TAILQ_REMOVE(&queue, cbp, bio_queue);
385 bp->bio_children--;
386 if (cbp->bio_data != NULL) {
387 bzero(cbp->bio_data, cbp->bio_length);
388 uma_zfree(g_shsec_zone, cbp->bio_data);
389 }
390 g_destroy_bio(cbp);
391 }
392 if (bp->bio_error == 0)
393 bp->bio_error = error;
394 g_io_deliver(bp, bp->bio_error);
395 }
396
397 static void
398 g_shsec_check_and_run(struct g_shsec_softc *sc)
399 {
400 off_t mediasize, ms;
401 u_int no, sectorsize = 0;
402
403 if (g_shsec_nvalid(sc) != sc->sc_ndisks)
404 return;
405
406 sc->sc_provider = g_new_providerf(sc->sc_geom, "shsec/%s", sc->sc_name);
407 /*
408 * Find the smallest disk.
409 */
410 mediasize = sc->sc_disks[0]->provider->mediasize;
411 mediasize -= sc->sc_disks[0]->provider->sectorsize;
412 sectorsize = sc->sc_disks[0]->provider->sectorsize;
413 for (no = 1; no < sc->sc_ndisks; no++) {
414 ms = sc->sc_disks[no]->provider->mediasize;
415 ms -= sc->sc_disks[no]->provider->sectorsize;
416 if (ms < mediasize)
417 mediasize = ms;
418 sectorsize = lcm(sectorsize,
419 sc->sc_disks[no]->provider->sectorsize);
420 }
421 sc->sc_provider->sectorsize = sectorsize;
422 sc->sc_provider->mediasize = mediasize;
423 g_error_provider(sc->sc_provider, 0);
424
425 G_SHSEC_DEBUG(0, "Device %s activated.", sc->sc_name);
426 }
427
428 static int
429 g_shsec_read_metadata(struct g_consumer *cp, struct g_shsec_metadata *md)
430 {
431 struct g_provider *pp;
432 u_char *buf;
433 int error;
434
435 g_topology_assert();
436
437 error = g_access(cp, 1, 0, 0);
438 if (error != 0)
439 return (error);
440 pp = cp->provider;
441 g_topology_unlock();
442 buf = g_read_data(cp, pp->mediasize - pp->sectorsize, pp->sectorsize,
443 &error);
444 g_topology_lock();
445 g_access(cp, -1, 0, 0);
446 if (buf == NULL)
447 return (error);
448
449 /* Decode metadata. */
450 shsec_metadata_decode(buf, md);
451 g_free(buf);
452
453 return (0);
454 }
455
456 /*
457 * Add disk to given device.
458 */
459 static int
460 g_shsec_add_disk(struct g_shsec_softc *sc, struct g_provider *pp, u_int no)
461 {
462 struct g_consumer *cp, *fcp;
463 struct g_geom *gp;
464 struct g_shsec_metadata md;
465 int error;
466
467 /* Metadata corrupted? */
468 if (no >= sc->sc_ndisks)
469 return (EINVAL);
470
471 /* Check if disk is not already attached. */
472 if (sc->sc_disks[no] != NULL)
473 return (EEXIST);
474
475 gp = sc->sc_geom;
476 fcp = LIST_FIRST(&gp->consumer);
477
478 cp = g_new_consumer(gp);
479 error = g_attach(cp, pp);
480 if (error != 0) {
481 g_destroy_consumer(cp);
482 return (error);
483 }
484
485 if (fcp != NULL && (fcp->acr > 0 || fcp->acw > 0 || fcp->ace > 0)) {
486 error = g_access(cp, fcp->acr, fcp->acw, fcp->ace);
487 if (error != 0) {
488 g_detach(cp);
489 g_destroy_consumer(cp);
490 return (error);
491 }
492 }
493
494 /* Reread metadata. */
495 error = g_shsec_read_metadata(cp, &md);
496 if (error != 0)
497 goto fail;
498
499 if (strcmp(md.md_magic, G_SHSEC_MAGIC) != 0 ||
500 strcmp(md.md_name, sc->sc_name) != 0 || md.md_id != sc->sc_id) {
501 G_SHSEC_DEBUG(0, "Metadata on %s changed.", pp->name);
502 goto fail;
503 }
504
505 cp->private = sc;
506 cp->index = no;
507 sc->sc_disks[no] = cp;
508
509 G_SHSEC_DEBUG(0, "Disk %s attached to %s.", pp->name, sc->sc_name);
510
511 g_shsec_check_and_run(sc);
512
513 return (0);
514 fail:
515 if (fcp != NULL && (fcp->acr > 0 || fcp->acw > 0 || fcp->ace > 0))
516 g_access(cp, -fcp->acr, -fcp->acw, -fcp->ace);
517 g_detach(cp);
518 g_destroy_consumer(cp);
519 return (error);
520 }
521
522 static struct g_geom *
523 g_shsec_create(struct g_class *mp, const struct g_shsec_metadata *md)
524 {
525 struct g_shsec_softc *sc;
526 struct g_geom *gp;
527 u_int no;
528
529 G_SHSEC_DEBUG(1, "Creating device %s (id=%u).", md->md_name, md->md_id);
530
531 /* Two disks is minimum. */
532 if (md->md_all < 2) {
533 G_SHSEC_DEBUG(0, "Too few disks defined for %s.", md->md_name);
534 return (NULL);
535 }
536
537 /* Check for duplicate unit */
538 LIST_FOREACH(gp, &mp->geom, geom) {
539 sc = gp->softc;
540 if (sc != NULL && strcmp(sc->sc_name, md->md_name) == 0) {
541 G_SHSEC_DEBUG(0, "Device %s already configured.",
542 sc->sc_name);
543 return (NULL);
544 }
545 }
546 gp = g_new_geomf(mp, "%s", md->md_name);
547 gp->softc = NULL; /* for a moment */
548
549 sc = malloc(sizeof(*sc), M_SHSEC, M_WAITOK | M_ZERO);
550 gp->start = g_shsec_start;
551 gp->spoiled = g_shsec_orphan;
552 gp->orphan = g_shsec_orphan;
553 gp->access = g_shsec_access;
554 gp->dumpconf = g_shsec_dumpconf;
555
556 sc->sc_id = md->md_id;
557 sc->sc_ndisks = md->md_all;
558 sc->sc_disks = malloc(sizeof(struct g_consumer *) * sc->sc_ndisks,
559 M_SHSEC, M_WAITOK | M_ZERO);
560 for (no = 0; no < sc->sc_ndisks; no++)
561 sc->sc_disks[no] = NULL;
562
563 gp->softc = sc;
564 sc->sc_geom = gp;
565 sc->sc_provider = NULL;
566
567 G_SHSEC_DEBUG(0, "Device %s created (id=%u).", sc->sc_name, sc->sc_id);
568
569 return (gp);
570 }
571
572 static int
573 g_shsec_destroy(struct g_shsec_softc *sc, boolean_t force)
574 {
575 struct g_provider *pp;
576 struct g_geom *gp;
577 u_int no;
578
579 g_topology_assert();
580
581 if (sc == NULL)
582 return (ENXIO);
583
584 pp = sc->sc_provider;
585 if (pp != NULL && (pp->acr != 0 || pp->acw != 0 || pp->ace != 0)) {
586 if (force) {
587 G_SHSEC_DEBUG(0, "Device %s is still open, so it "
588 "can't be definitely removed.", pp->name);
589 } else {
590 G_SHSEC_DEBUG(1,
591 "Device %s is still open (r%dw%de%d).", pp->name,
592 pp->acr, pp->acw, pp->ace);
593 return (EBUSY);
594 }
595 }
596
597 for (no = 0; no < sc->sc_ndisks; no++) {
598 if (sc->sc_disks[no] != NULL)
599 g_shsec_remove_disk(sc->sc_disks[no]);
600 }
601
602 gp = sc->sc_geom;
603 gp->softc = NULL;
604 KASSERT(sc->sc_provider == NULL, ("Provider still exists? (device=%s)",
605 gp->name));
606 free(sc->sc_disks, M_SHSEC);
607 free(sc, M_SHSEC);
608
609 pp = LIST_FIRST(&gp->provider);
610 if (pp == NULL || (pp->acr == 0 && pp->acw == 0 && pp->ace == 0))
611 G_SHSEC_DEBUG(0, "Device %s destroyed.", gp->name);
612
613 g_wither_geom(gp, ENXIO);
614
615 return (0);
616 }
617
618 static int
619 g_shsec_destroy_geom(struct gctl_req *req __unused, struct g_class *mp __unused,
620 struct g_geom *gp)
621 {
622 struct g_shsec_softc *sc;
623
624 sc = gp->softc;
625 return (g_shsec_destroy(sc, 0));
626 }
627
628 static struct g_geom *
629 g_shsec_taste(struct g_class *mp, struct g_provider *pp, int flags __unused)
630 {
631 struct g_shsec_metadata md;
632 struct g_shsec_softc *sc;
633 struct g_consumer *cp;
634 struct g_geom *gp;
635 int error;
636
637 g_trace(G_T_TOPOLOGY, "%s(%s, %s)", __func__, mp->name, pp->name);
638 g_topology_assert();
639
640 G_SHSEC_DEBUG(3, "Tasting %s.", pp->name);
641
642 gp = g_new_geomf(mp, "shsec:taste");
643 gp->start = g_shsec_start;
644 gp->access = g_shsec_access;
645 gp->orphan = g_shsec_orphan;
646 cp = g_new_consumer(gp);
647 g_attach(cp, pp);
648 error = g_shsec_read_metadata(cp, &md);
649 g_detach(cp);
650 g_destroy_consumer(cp);
651 g_destroy_geom(gp);
652 if (error != 0)
653 return (NULL);
654 gp = NULL;
655
656 if (strcmp(md.md_magic, G_SHSEC_MAGIC) != 0)
657 return (NULL);
658 if (md.md_version > G_SHSEC_VERSION) {
659 G_SHSEC_DEBUG(0, "Kernel module is too old to handle %s.\n",
660 pp->name);
661 return (NULL);
662 }
663 /*
664 * Backward compatibility:
665 */
666 /* There was no md_provsize field in earlier versions of metadata. */
667 if (md.md_version < 1)
668 md.md_provsize = pp->mediasize;
669
670 if (md.md_provider[0] != '\0' && strcmp(md.md_provider, pp->name) != 0)
671 return (NULL);
672 if (md.md_provsize != pp->mediasize)
673 return (NULL);
674
675 /*
676 * Let's check if device already exists.
677 */
678 sc = NULL;
679 LIST_FOREACH(gp, &mp->geom, geom) {
680 sc = gp->softc;
681 if (sc == NULL)
682 continue;
683 if (strcmp(md.md_name, sc->sc_name) != 0)
684 continue;
685 if (md.md_id != sc->sc_id)
686 continue;
687 break;
688 }
689 if (gp != NULL) {
690 G_SHSEC_DEBUG(1, "Adding disk %s to %s.", pp->name, gp->name);
691 error = g_shsec_add_disk(sc, pp, md.md_no);
692 if (error != 0) {
693 G_SHSEC_DEBUG(0, "Cannot add disk %s to %s (error=%d).",
694 pp->name, gp->name, error);
695 return (NULL);
696 }
697 } else {
698 gp = g_shsec_create(mp, &md);
699 if (gp == NULL) {
700 G_SHSEC_DEBUG(0, "Cannot create device %s.", md.md_name);
701 return (NULL);
702 }
703 sc = gp->softc;
704 G_SHSEC_DEBUG(1, "Adding disk %s to %s.", pp->name, gp->name);
705 error = g_shsec_add_disk(sc, pp, md.md_no);
706 if (error != 0) {
707 G_SHSEC_DEBUG(0, "Cannot add disk %s to %s (error=%d).",
708 pp->name, gp->name, error);
709 g_shsec_destroy(sc, 1);
710 return (NULL);
711 }
712 }
713 return (gp);
714 }
715
716 static struct g_shsec_softc *
717 g_shsec_find_device(struct g_class *mp, const char *name)
718 {
719 struct g_shsec_softc *sc;
720 struct g_geom *gp;
721
722 LIST_FOREACH(gp, &mp->geom, geom) {
723 sc = gp->softc;
724 if (sc == NULL)
725 continue;
726 if (strcmp(sc->sc_name, name) == 0)
727 return (sc);
728 }
729 return (NULL);
730 }
731
732 static void
733 g_shsec_ctl_destroy(struct gctl_req *req, struct g_class *mp)
734 {
735 struct g_shsec_softc *sc;
736 int *force, *nargs, error;
737 const char *name;
738 char param[16];
739 u_int i;
740
741 g_topology_assert();
742
743 nargs = gctl_get_paraml(req, "nargs", sizeof(*nargs));
744 if (nargs == NULL) {
745 gctl_error(req, "No '%s' argument.", "nargs");
746 return;
747 }
748 if (*nargs <= 0) {
749 gctl_error(req, "Missing device(s).");
750 return;
751 }
752 force = gctl_get_paraml(req, "force", sizeof(*force));
753 if (force == NULL) {
754 gctl_error(req, "No '%s' argument.", "force");
755 return;
756 }
757
758 for (i = 0; i < (u_int)*nargs; i++) {
759 snprintf(param, sizeof(param), "arg%u", i);
760 name = gctl_get_asciiparam(req, param);
761 if (name == NULL) {
762 gctl_error(req, "No 'arg%u' argument.", i);
763 return;
764 }
765 sc = g_shsec_find_device(mp, name);
766 if (sc == NULL) {
767 gctl_error(req, "No such device: %s.", name);
768 return;
769 }
770 error = g_shsec_destroy(sc, *force);
771 if (error != 0) {
772 gctl_error(req, "Cannot destroy device %s (error=%d).",
773 sc->sc_name, error);
774 return;
775 }
776 }
777 }
778
779 static void
780 g_shsec_config(struct gctl_req *req, struct g_class *mp, const char *verb)
781 {
782 uint32_t *version;
783
784 g_topology_assert();
785
786 version = gctl_get_paraml(req, "version", sizeof(*version));
787 if (version == NULL) {
788 gctl_error(req, "No '%s' argument.", "version");
789 return;
790 }
791 if (*version != G_SHSEC_VERSION) {
792 gctl_error(req, "Userland and kernel parts are out of sync.");
793 return;
794 }
795
796 if (strcmp(verb, "stop") == 0) {
797 g_shsec_ctl_destroy(req, mp);
798 return;
799 }
800
801 gctl_error(req, "Unknown verb.");
802 }
803
804 static void
805 g_shsec_dumpconf(struct sbuf *sb, const char *indent, struct g_geom *gp,
806 struct g_consumer *cp, struct g_provider *pp)
807 {
808 struct g_shsec_softc *sc;
809
810 sc = gp->softc;
811 if (sc == NULL)
812 return;
813 if (pp != NULL) {
814 /* Nothing here. */
815 } else if (cp != NULL) {
816 sbuf_printf(sb, "%s<Number>%u</Number>\n", indent,
817 (u_int)cp->index);
818 } else {
819 sbuf_printf(sb, "%s<ID>%u</ID>\n", indent, (u_int)sc->sc_id);
820 sbuf_printf(sb, "%s<Status>Total=%u, Online=%u</Status>\n",
821 indent, sc->sc_ndisks, g_shsec_nvalid(sc));
822 sbuf_printf(sb, "%s<State>", indent);
823 if (sc->sc_provider != NULL && sc->sc_provider->error == 0)
824 sbuf_printf(sb, "UP");
825 else
826 sbuf_printf(sb, "DOWN");
827 sbuf_printf(sb, "</State>\n");
828 }
829 }
830
831 DECLARE_GEOM_CLASS(g_shsec_class, g_shsec);
Cache object: fadac977d584555a69098ffa408568d2
|