FreeBSD/Linux Kernel Cross Reference
sys/dev/buslogic/bt.c
1 /*-
2 * Generic driver for the BusLogic MultiMaster SCSI host adapters
3 * Product specific probe and attach routines can be found in:
4 * sys/dev/buslogic/bt_isa.c BT-54X, BT-445 cards
5 * sys/dev/buslogic/bt_mca.c BT-64X, SDC3211B, SDC3211F
6 * sys/dev/buslogic/bt_pci.c BT-946, BT-948, BT-956, BT-958 cards
7 *
8 * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
9 *
10 * Copyright (c) 1998, 1999 Justin T. Gibbs.
11 * All rights reserved.
12 *
13 * Redistribution and use in source and binary forms, with or without
14 * modification, are permitted provided that the following conditions
15 * are met:
16 * 1. Redistributions of source code must retain the above copyright
17 * notice, this list of conditions, and the following disclaimer,
18 * without modification, immediately at the beginning of the file.
19 * 2. The name of the author may not be used to endorse or promote products
20 * derived from this software without specific prior written permission.
21 *
22 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
26 * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 * SUCH DAMAGE.
33 */
34
35 #include <sys/cdefs.h>
36 __FBSDID("$FreeBSD: releng/12.0/sys/dev/buslogic/bt.c 328523 2018-01-29 00:14:46Z imp $");
37
38 /*
39 * Special thanks to Leonard N. Zubkoff for writing such a complete and
40 * well documented Mylex/BusLogic MultiMaster driver for Linux. Support
41 * in this driver for the wide range of MultiMaster controllers and
42 * firmware revisions, with their otherwise undocumented quirks, would not
43 * have been possible without his efforts.
44 */
45
46 #include <sys/param.h>
47 #include <sys/conf.h>
48 #include <sys/systm.h>
49 #include <sys/malloc.h>
50 #include <sys/kernel.h>
51 #include <sys/lock.h>
52 #include <sys/module.h>
53 #include <sys/mutex.h>
54 #include <sys/sysctl.h>
55 #include <sys/bus.h>
56
57 #include <machine/bus.h>
58 #include <sys/rman.h>
59
60 #include <cam/cam.h>
61 #include <cam/cam_ccb.h>
62 #include <cam/cam_sim.h>
63 #include <cam/cam_xpt_sim.h>
64 #include <cam/cam_debug.h>
65
66 #include <cam/scsi/scsi_message.h>
67
68 #include <vm/vm.h>
69 #include <vm/pmap.h>
70
71 #include <dev/buslogic/btreg.h>
72
73 /* MailBox Management functions */
74 static __inline void btnextinbox(struct bt_softc *bt);
75 static __inline void btnextoutbox(struct bt_softc *bt);
76
77 static __inline void
78 btnextinbox(struct bt_softc *bt)
79 {
80 if (bt->cur_inbox == bt->last_inbox)
81 bt->cur_inbox = bt->in_boxes;
82 else
83 bt->cur_inbox++;
84 }
85
86 static __inline void
87 btnextoutbox(struct bt_softc *bt)
88 {
89 if (bt->cur_outbox == bt->last_outbox)
90 bt->cur_outbox = bt->out_boxes;
91 else
92 bt->cur_outbox++;
93 }
94
95 /* CCB Mangement functions */
96 static __inline u_int32_t btccbvtop(struct bt_softc *bt,
97 struct bt_ccb *bccb);
98 static __inline struct bt_ccb* btccbptov(struct bt_softc *bt,
99 u_int32_t ccb_addr);
100 static __inline u_int32_t btsensepaddr(struct bt_softc *bt,
101 struct bt_ccb *bccb);
102 static __inline struct scsi_sense_data* btsensevaddr(struct bt_softc *bt,
103 struct bt_ccb *bccb);
104
105 static __inline u_int32_t
106 btccbvtop(struct bt_softc *bt, struct bt_ccb *bccb)
107 {
108 return (bt->bt_ccb_physbase
109 + (u_int32_t)((caddr_t)bccb - (caddr_t)bt->bt_ccb_array));
110 }
111
112 static __inline struct bt_ccb *
113 btccbptov(struct bt_softc *bt, u_int32_t ccb_addr)
114 {
115 return (bt->bt_ccb_array +
116 ((struct bt_ccb*)(uintptr_t)ccb_addr - (struct bt_ccb*)(uintptr_t)bt->bt_ccb_physbase));
117 }
118
119 static __inline u_int32_t
120 btsensepaddr(struct bt_softc *bt, struct bt_ccb *bccb)
121 {
122 u_int index;
123
124 index = (u_int)(bccb - bt->bt_ccb_array);
125 return (bt->sense_buffers_physbase
126 + (index * sizeof(struct scsi_sense_data)));
127 }
128
129 static __inline struct scsi_sense_data *
130 btsensevaddr(struct bt_softc *bt, struct bt_ccb *bccb)
131 {
132 u_int index;
133
134 index = (u_int)(bccb - bt->bt_ccb_array);
135 return (bt->sense_buffers + index);
136 }
137
138 static __inline struct bt_ccb* btgetccb(struct bt_softc *bt);
139 static __inline void btfreeccb(struct bt_softc *bt,
140 struct bt_ccb *bccb);
141 static void btallocccbs(struct bt_softc *bt);
142 static bus_dmamap_callback_t btexecuteccb;
143 static void btdone(struct bt_softc *bt, struct bt_ccb *bccb,
144 bt_mbi_comp_code_t comp_code);
145 static void bt_intr_locked(struct bt_softc *bt);
146
147 /* Host adapter command functions */
148 static int btreset(struct bt_softc* bt, int hard_reset);
149
150 /* Initialization functions */
151 static int btinitmboxes(struct bt_softc *bt);
152 static bus_dmamap_callback_t btmapmboxes;
153 static bus_dmamap_callback_t btmapccbs;
154 static bus_dmamap_callback_t btmapsgs;
155
156 /* Transfer Negotiation Functions */
157 static void btfetchtransinfo(struct bt_softc *bt,
158 struct ccb_trans_settings *cts);
159
160 /* CAM SIM entry points */
161 #define ccb_bccb_ptr spriv_ptr0
162 #define ccb_bt_ptr spriv_ptr1
163 static void btaction(struct cam_sim *sim, union ccb *ccb);
164 static void btpoll(struct cam_sim *sim);
165
166 /* Our timeout handler */
167 static void bttimeout(void *arg);
168
169 /*
170 * XXX
171 * Do our own re-probe protection until a configuration
172 * manager can do it for us. This ensures that we don't
173 * reprobe a card already found by the PCI probes.
174 */
175 struct bt_isa_port bt_isa_ports[] =
176 {
177 { 0x130, 0, 4 },
178 { 0x134, 0, 5 },
179 { 0x230, 0, 2 },
180 { 0x234, 0, 3 },
181 { 0x330, 0, 0 },
182 { 0x334, 0, 1 }
183 };
184
185 /*
186 * I/O ports listed in the order enumerated by the
187 * card for certain op codes.
188 */
189 u_int16_t bt_board_ports[] =
190 {
191 0x330,
192 0x334,
193 0x230,
194 0x234,
195 0x130,
196 0x134
197 };
198
199 /* Exported functions */
200 void
201 bt_init_softc(device_t dev, struct resource *port,
202 struct resource *irq, struct resource *drq)
203 {
204 struct bt_softc *bt = device_get_softc(dev);
205
206 SLIST_INIT(&bt->free_bt_ccbs);
207 LIST_INIT(&bt->pending_ccbs);
208 SLIST_INIT(&bt->sg_maps);
209 bt->dev = dev;
210 bt->port = port;
211 bt->irq = irq;
212 bt->drq = drq;
213 mtx_init(&bt->lock, "bt", NULL, MTX_DEF);
214 }
215
216 void
217 bt_free_softc(device_t dev)
218 {
219 struct bt_softc *bt = device_get_softc(dev);
220
221 switch (bt->init_level) {
222 default:
223 case 11:
224 bus_dmamap_unload(bt->sense_dmat, bt->sense_dmamap);
225 case 10:
226 bus_dmamem_free(bt->sense_dmat, bt->sense_buffers,
227 bt->sense_dmamap);
228 case 9:
229 bus_dma_tag_destroy(bt->sense_dmat);
230 case 8:
231 {
232 struct sg_map_node *sg_map;
233
234 while ((sg_map = SLIST_FIRST(&bt->sg_maps))!= NULL) {
235 SLIST_REMOVE_HEAD(&bt->sg_maps, links);
236 bus_dmamap_unload(bt->sg_dmat,
237 sg_map->sg_dmamap);
238 bus_dmamem_free(bt->sg_dmat, sg_map->sg_vaddr,
239 sg_map->sg_dmamap);
240 free(sg_map, M_DEVBUF);
241 }
242 bus_dma_tag_destroy(bt->sg_dmat);
243 }
244 case 7:
245 bus_dmamap_unload(bt->ccb_dmat, bt->ccb_dmamap);
246 /* FALLTHROUGH */
247 case 6:
248 bus_dmamem_free(bt->ccb_dmat, bt->bt_ccb_array,
249 bt->ccb_dmamap);
250 /* FALLTHROUGH */
251 case 5:
252 bus_dma_tag_destroy(bt->ccb_dmat);
253 /* FALLTHROUGH */
254 case 4:
255 bus_dmamap_unload(bt->mailbox_dmat, bt->mailbox_dmamap);
256 /* FALLTHROUGH */
257 case 3:
258 bus_dmamem_free(bt->mailbox_dmat, bt->in_boxes,
259 bt->mailbox_dmamap);
260 /* FALLTHROUGH */
261 case 2:
262 bus_dma_tag_destroy(bt->buffer_dmat);
263 /* FALLTHROUGH */
264 case 1:
265 bus_dma_tag_destroy(bt->mailbox_dmat);
266 /* FALLTHROUGH */
267 case 0:
268 break;
269 }
270 mtx_destroy(&bt->lock);
271 }
272
273 int
274 bt_port_probe(device_t dev, struct bt_probe_info *info)
275 {
276 struct bt_softc *bt = device_get_softc(dev);
277 config_data_t config_data;
278 int error;
279
280 /* See if there is really a card present */
281 if (bt_probe(dev) || bt_fetch_adapter_info(dev))
282 return(1);
283
284 /*
285 * Determine our IRQ, and DMA settings and
286 * export them to the configuration system.
287 */
288 mtx_lock(&bt->lock);
289 error = bt_cmd(bt, BOP_INQUIRE_CONFIG, NULL, /*parmlen*/0,
290 (u_int8_t*)&config_data, sizeof(config_data),
291 DEFAULT_CMD_TIMEOUT);
292 mtx_unlock(&bt->lock);
293 if (error != 0) {
294 printf("bt_port_probe: Could not determine IRQ or DMA "
295 "settings for adapter.\n");
296 return (1);
297 }
298
299 if (bt->model[0] == '5') {
300 /* DMA settings only make sense for ISA cards */
301 switch (config_data.dma_chan) {
302 case DMA_CHAN_5:
303 info->drq = 5;
304 break;
305 case DMA_CHAN_6:
306 info->drq = 6;
307 break;
308 case DMA_CHAN_7:
309 info->drq = 7;
310 break;
311 default:
312 printf("bt_port_probe: Invalid DMA setting "
313 "detected for adapter.\n");
314 return (1);
315 }
316 } else {
317 info->drq = -1;
318 }
319 switch (config_data.irq) {
320 case IRQ_9:
321 case IRQ_10:
322 case IRQ_11:
323 case IRQ_12:
324 case IRQ_14:
325 case IRQ_15:
326 info->irq = ffs(config_data.irq) + 8;
327 break;
328 default:
329 printf("bt_port_probe: Invalid IRQ setting %x"
330 "detected for adapter.\n", config_data.irq);
331 return (1);
332 }
333 return (0);
334 }
335
336 /*
337 * Probe the adapter and verify that the card is a BusLogic.
338 */
339 int
340 bt_probe(device_t dev)
341 {
342 struct bt_softc *bt = device_get_softc(dev);
343 esetup_info_data_t esetup_info;
344 u_int status;
345 u_int intstat;
346 u_int geometry;
347 int error;
348 u_int8_t param;
349
350 /*
351 * See if the three I/O ports look reasonable.
352 * Touch the minimal number of registers in the
353 * failure case.
354 */
355 status = bt_inb(bt, STATUS_REG);
356 if ((status == 0)
357 || (status & (DIAG_ACTIVE|CMD_REG_BUSY|
358 STATUS_REG_RSVD|CMD_INVALID)) != 0) {
359 if (bootverbose)
360 device_printf(dev, "Failed Status Reg Test - %x\n",
361 status);
362 return (ENXIO);
363 }
364
365 intstat = bt_inb(bt, INTSTAT_REG);
366 if ((intstat & INTSTAT_REG_RSVD) != 0) {
367 device_printf(dev, "Failed Intstat Reg Test\n");
368 return (ENXIO);
369 }
370
371 geometry = bt_inb(bt, GEOMETRY_REG);
372 if (geometry == 0xFF) {
373 if (bootverbose)
374 device_printf(dev, "Failed Geometry Reg Test\n");
375 return (ENXIO);
376 }
377
378 /*
379 * Looking good so far. Final test is to reset the
380 * adapter and attempt to fetch the extended setup
381 * information. This should filter out all 1542 cards.
382 */
383 mtx_lock(&bt->lock);
384 if ((error = btreset(bt, /*hard_reset*/TRUE)) != 0) {
385 mtx_unlock(&bt->lock);
386 if (bootverbose)
387 device_printf(dev, "Failed Reset\n");
388 return (ENXIO);
389 }
390
391 param = sizeof(esetup_info);
392 error = bt_cmd(bt, BOP_INQUIRE_ESETUP_INFO, ¶m, /*parmlen*/1,
393 (u_int8_t*)&esetup_info, sizeof(esetup_info),
394 DEFAULT_CMD_TIMEOUT);
395 mtx_unlock(&bt->lock);
396 if (error != 0) {
397 return (ENXIO);
398 }
399
400 return (0);
401 }
402
403 /*
404 * Pull the boards setup information and record it in our softc.
405 */
406 int
407 bt_fetch_adapter_info(device_t dev)
408 {
409 struct bt_softc *bt = device_get_softc(dev);
410 board_id_data_t board_id;
411 esetup_info_data_t esetup_info;
412 config_data_t config_data;
413 int error;
414 u_int8_t length_param;
415
416 /* First record the firmware version */
417 mtx_lock(&bt->lock);
418 error = bt_cmd(bt, BOP_INQUIRE_BOARD_ID, NULL, /*parmlen*/0,
419 (u_int8_t*)&board_id, sizeof(board_id),
420 DEFAULT_CMD_TIMEOUT);
421 if (error != 0) {
422 mtx_unlock(&bt->lock);
423 device_printf(dev, "bt_fetch_adapter_info - Failed Get Board Info\n");
424 return (error);
425 }
426 bt->firmware_ver[0] = board_id.firmware_rev_major;
427 bt->firmware_ver[1] = '.';
428 bt->firmware_ver[2] = board_id.firmware_rev_minor;
429 bt->firmware_ver[3] = '\0';
430
431 /*
432 * Depending on the firmware major and minor version,
433 * we may be able to fetch additional minor version info.
434 */
435 if (bt->firmware_ver[0] > '') {
436
437 error = bt_cmd(bt, BOP_INQUIRE_FW_VER_3DIG, NULL, /*parmlen*/0,
438 (u_int8_t*)&bt->firmware_ver[3], 1,
439 DEFAULT_CMD_TIMEOUT);
440 if (error != 0) {
441 mtx_unlock(&bt->lock);
442 device_printf(dev,
443 "bt_fetch_adapter_info - Failed Get "
444 "Firmware 3rd Digit\n");
445 return (error);
446 }
447 if (bt->firmware_ver[3] == ' ')
448 bt->firmware_ver[3] = '\0';
449 bt->firmware_ver[4] = '\0';
450 }
451
452 if (strcmp(bt->firmware_ver, "3.3") >= 0) {
453
454 error = bt_cmd(bt, BOP_INQUIRE_FW_VER_4DIG, NULL, /*parmlen*/0,
455 (u_int8_t*)&bt->firmware_ver[4], 1,
456 DEFAULT_CMD_TIMEOUT);
457 if (error != 0) {
458 mtx_unlock(&bt->lock);
459 device_printf(dev,
460 "bt_fetch_adapter_info - Failed Get "
461 "Firmware 4th Digit\n");
462 return (error);
463 }
464 if (bt->firmware_ver[4] == ' ')
465 bt->firmware_ver[4] = '\0';
466 bt->firmware_ver[5] = '\0';
467 }
468
469 /*
470 * Some boards do not handle the "recently documented"
471 * Inquire Board Model Number command correctly or do not give
472 * exact information. Use the Firmware and Extended Setup
473 * information in these cases to come up with the right answer.
474 * The major firmware revision number indicates:
475 *
476 * 5.xx BusLogic "W" Series Host Adapters:
477 * BT-948/958/958D
478 * 4.xx BusLogic "C" Series Host Adapters:
479 * BT-946C/956C/956CD/747C/757C/757CD/445C/545C/540CF
480 * 3.xx BusLogic "S" Series Host Adapters:
481 * BT-747S/747D/757S/757D/445S/545S/542D
482 * BT-542B/742A (revision H)
483 * 2.xx BusLogic "A" Series Host Adapters:
484 * BT-542B/742A (revision G and below)
485 */
486 length_param = sizeof(esetup_info);
487 error = bt_cmd(bt, BOP_INQUIRE_ESETUP_INFO, &length_param, /*parmlen*/1,
488 (u_int8_t*)&esetup_info, sizeof(esetup_info),
489 DEFAULT_CMD_TIMEOUT);
490 if (error != 0) {
491 mtx_unlock(&bt->lock);
492 return (error);
493 }
494
495 bt->bios_addr = esetup_info.bios_addr << 12;
496
497 bt->mailbox_addrlimit = BUS_SPACE_MAXADDR;
498 if (esetup_info.bus_type == 'A'
499 && bt->firmware_ver[0] == '2') {
500 snprintf(bt->model, sizeof(bt->model), "542B");
501 } else {
502 ha_model_data_t model_data;
503 int i;
504
505 length_param = sizeof(model_data);
506 error = bt_cmd(bt, BOP_INQUIRE_MODEL, &length_param, 1,
507 (u_int8_t*)&model_data, sizeof(model_data),
508 DEFAULT_CMD_TIMEOUT);
509 if (error != 0) {
510 mtx_unlock(&bt->lock);
511 device_printf(dev,
512 "bt_fetch_adapter_info - Failed Inquire "
513 "Model Number\n");
514 return (error);
515 }
516 for (i = 0; i < sizeof(model_data.ascii_model); i++) {
517 bt->model[i] = model_data.ascii_model[i];
518 if (bt->model[i] == ' ')
519 break;
520 }
521 bt->model[i] = '\0';
522 }
523
524 bt->level_trigger_ints = esetup_info.level_trigger_ints ? 1 : 0;
525
526 /* SG element limits */
527 bt->max_sg = esetup_info.max_sg;
528
529 /* Set feature flags */
530 bt->wide_bus = esetup_info.wide_bus;
531 bt->diff_bus = esetup_info.diff_bus;
532 bt->ultra_scsi = esetup_info.ultra_scsi;
533
534 if ((bt->firmware_ver[0] == '5')
535 || (bt->firmware_ver[0] == '4' && bt->wide_bus))
536 bt->extended_lun = TRUE;
537
538 bt->strict_rr = (strcmp(bt->firmware_ver, "3.31") >= 0);
539
540 bt->extended_trans =
541 ((bt_inb(bt, GEOMETRY_REG) & EXTENDED_TRANSLATION) != 0);
542
543 /*
544 * Determine max CCB count and whether tagged queuing is
545 * available based on controller type. Tagged queuing
546 * only works on 'W' series adapters, 'C' series adapters
547 * with firmware of rev 4.42 and higher, and 'S' series
548 * adapters with firmware of rev 3.35 and higher. The
549 * maximum CCB counts are as follows:
550 *
551 * 192 BT-948/958/958D
552 * 100 BT-946C/956C/956CD/747C/757C/757CD/445C
553 * 50 BT-545C/540CF
554 * 30 BT-747S/747D/757S/757D/445S/545S/542D/542B/742A
555 */
556 if (bt->firmware_ver[0] == '5') {
557 bt->max_ccbs = 192;
558 bt->tag_capable = TRUE;
559 } else if (bt->firmware_ver[0] == '4') {
560 if (bt->model[0] == '5')
561 bt->max_ccbs = 50;
562 else
563 bt->max_ccbs = 100;
564 bt->tag_capable = (strcmp(bt->firmware_ver, "4.22") >= 0);
565 } else {
566 bt->max_ccbs = 30;
567 if (bt->firmware_ver[0] == '3'
568 && (strcmp(bt->firmware_ver, "3.35") >= 0))
569 bt->tag_capable = TRUE;
570 else
571 bt->tag_capable = FALSE;
572 }
573
574 if (bt->tag_capable != FALSE)
575 bt->tags_permitted = ALL_TARGETS;
576
577 /* Determine Sync/Wide/Disc settings */
578 if (bt->firmware_ver[0] >= '4') {
579 auto_scsi_data_t auto_scsi_data;
580 fetch_lram_params_t fetch_lram_params;
581 int error;
582
583 /*
584 * These settings are stored in the
585 * AutoSCSI data in LRAM of 'W' and 'C'
586 * adapters.
587 */
588 fetch_lram_params.offset = AUTO_SCSI_BYTE_OFFSET;
589 fetch_lram_params.response_len = sizeof(auto_scsi_data);
590 error = bt_cmd(bt, BOP_FETCH_LRAM,
591 (u_int8_t*)&fetch_lram_params,
592 sizeof(fetch_lram_params),
593 (u_int8_t*)&auto_scsi_data,
594 sizeof(auto_scsi_data), DEFAULT_CMD_TIMEOUT);
595
596 if (error != 0) {
597 mtx_unlock(&bt->lock);
598 device_printf(dev,
599 "bt_fetch_adapter_info - Failed "
600 "Get Auto SCSI Info\n");
601 return (error);
602 }
603
604 bt->disc_permitted = auto_scsi_data.low_disc_permitted
605 | (auto_scsi_data.high_disc_permitted << 8);
606 bt->sync_permitted = auto_scsi_data.low_sync_permitted
607 | (auto_scsi_data.high_sync_permitted << 8);
608 bt->fast_permitted = auto_scsi_data.low_fast_permitted
609 | (auto_scsi_data.high_fast_permitted << 8);
610 bt->ultra_permitted = auto_scsi_data.low_ultra_permitted
611 | (auto_scsi_data.high_ultra_permitted << 8);
612 bt->wide_permitted = auto_scsi_data.low_wide_permitted
613 | (auto_scsi_data.high_wide_permitted << 8);
614
615 if (bt->ultra_scsi == FALSE)
616 bt->ultra_permitted = 0;
617
618 if (bt->wide_bus == FALSE)
619 bt->wide_permitted = 0;
620 } else {
621 /*
622 * 'S' and 'A' series have this information in the setup
623 * information structure.
624 */
625 setup_data_t setup_info;
626
627 length_param = sizeof(setup_info);
628 error = bt_cmd(bt, BOP_INQUIRE_SETUP_INFO, &length_param,
629 /*paramlen*/1, (u_int8_t*)&setup_info,
630 sizeof(setup_info), DEFAULT_CMD_TIMEOUT);
631
632 if (error != 0) {
633 mtx_unlock(&bt->lock);
634 device_printf(dev,
635 "bt_fetch_adapter_info - Failed "
636 "Get Setup Info\n");
637 return (error);
638 }
639
640 if (setup_info.initiate_sync != 0) {
641 bt->sync_permitted = ALL_TARGETS;
642
643 if (bt->model[0] == '7') {
644 if (esetup_info.sync_neg10MB != 0)
645 bt->fast_permitted = ALL_TARGETS;
646 if (strcmp(bt->model, "757") == 0)
647 bt->wide_permitted = ALL_TARGETS;
648 }
649 }
650 bt->disc_permitted = ALL_TARGETS;
651 }
652
653 /* We need as many mailboxes as we can have ccbs */
654 bt->num_boxes = bt->max_ccbs;
655
656 /* Determine our SCSI ID */
657
658 error = bt_cmd(bt, BOP_INQUIRE_CONFIG, NULL, /*parmlen*/0,
659 (u_int8_t*)&config_data, sizeof(config_data),
660 DEFAULT_CMD_TIMEOUT);
661 mtx_unlock(&bt->lock);
662 if (error != 0) {
663 device_printf(dev,
664 "bt_fetch_adapter_info - Failed Get Config\n");
665 return (error);
666 }
667 bt->scsi_id = config_data.scsi_id;
668
669 return (0);
670 }
671
672 /*
673 * Start the board, ready for normal operation
674 */
675 int
676 bt_init(device_t dev)
677 {
678 struct bt_softc *bt = device_get_softc(dev);
679
680 /* Announce the Adapter */
681 device_printf(dev, "BT-%s FW Rev. %s ", bt->model, bt->firmware_ver);
682
683 if (bt->ultra_scsi != 0)
684 printf("Ultra ");
685
686 if (bt->wide_bus != 0)
687 printf("Wide ");
688 else
689 printf("Narrow ");
690
691 if (bt->diff_bus != 0)
692 printf("Diff ");
693
694 printf("SCSI Host Adapter, SCSI ID %d, %d CCBs\n", bt->scsi_id,
695 bt->max_ccbs);
696
697 /*
698 * Create our DMA tags. These tags define the kinds of device
699 * accessible memory allocations and memory mappings we will
700 * need to perform during normal operation.
701 *
702 * Unless we need to further restrict the allocation, we rely
703 * on the restrictions of the parent dmat, hence the common
704 * use of MAXADDR and MAXSIZE.
705 */
706
707 /* DMA tag for mapping buffers into device visible space. */
708 if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
709 /* alignment */ 1,
710 /* boundary */ 0,
711 /* lowaddr */ BUS_SPACE_MAXADDR,
712 /* highaddr */ BUS_SPACE_MAXADDR,
713 /* filter */ NULL,
714 /* filterarg */ NULL,
715 /* maxsize */ DFLTPHYS,
716 /* nsegments */ BT_NSEG,
717 /* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
718 /* flags */ BUS_DMA_ALLOCNOW,
719 /* lockfunc */ busdma_lock_mutex,
720 /* lockarg */ &bt->lock,
721 &bt->buffer_dmat) != 0) {
722 goto error_exit;
723 }
724
725 bt->init_level++;
726 /* DMA tag for our mailboxes */
727 if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
728 /* alignment */ 1,
729 /* boundary */ 0,
730 /* lowaddr */ bt->mailbox_addrlimit,
731 /* highaddr */ BUS_SPACE_MAXADDR,
732 /* filter */ NULL,
733 /* filterarg */ NULL,
734 /* maxsize */ bt->num_boxes *
735 (sizeof(bt_mbox_in_t) +
736 sizeof(bt_mbox_out_t)),
737 /* nsegments */ 1,
738 /* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
739 /* flags */ 0,
740 /* lockfunc */ NULL,
741 /* lockarg */ NULL,
742 &bt->mailbox_dmat) != 0) {
743 goto error_exit;
744 }
745
746 bt->init_level++;
747
748 /* Allocation for our mailboxes */
749 if (bus_dmamem_alloc(bt->mailbox_dmat, (void **)&bt->out_boxes,
750 BUS_DMA_NOWAIT, &bt->mailbox_dmamap) != 0) {
751 goto error_exit;
752 }
753
754 bt->init_level++;
755
756 /* And permanently map them */
757 bus_dmamap_load(bt->mailbox_dmat, bt->mailbox_dmamap,
758 bt->out_boxes,
759 bt->num_boxes * (sizeof(bt_mbox_in_t)
760 + sizeof(bt_mbox_out_t)),
761 btmapmboxes, bt, /*flags*/0);
762
763 bt->init_level++;
764
765 bt->in_boxes = (bt_mbox_in_t *)&bt->out_boxes[bt->num_boxes];
766
767 mtx_lock(&bt->lock);
768 btinitmboxes(bt);
769 mtx_unlock(&bt->lock);
770
771 /* DMA tag for our ccb structures */
772 if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
773 /* alignment */ 1,
774 /* boundary */ 0,
775 /* lowaddr */ BUS_SPACE_MAXADDR,
776 /* highaddr */ BUS_SPACE_MAXADDR,
777 /* filter */ NULL,
778 /* filterarg */ NULL,
779 /* maxsize */ bt->max_ccbs *
780 sizeof(struct bt_ccb),
781 /* nsegments */ 1,
782 /* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
783 /* flags */ 0,
784 /* lockfunc */ NULL,
785 /* lockarg */ NULL,
786 &bt->ccb_dmat) != 0) {
787 goto error_exit;
788 }
789
790 bt->init_level++;
791
792 /* Allocation for our ccbs */
793 if (bus_dmamem_alloc(bt->ccb_dmat, (void **)&bt->bt_ccb_array,
794 BUS_DMA_NOWAIT, &bt->ccb_dmamap) != 0) {
795 goto error_exit;
796 }
797
798 bt->init_level++;
799
800 /* And permanently map them */
801 bus_dmamap_load(bt->ccb_dmat, bt->ccb_dmamap,
802 bt->bt_ccb_array,
803 bt->max_ccbs * sizeof(struct bt_ccb),
804 btmapccbs, bt, /*flags*/0);
805
806 bt->init_level++;
807
808 /* DMA tag for our S/G structures. We allocate in page sized chunks */
809 if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
810 /* alignment */ 1,
811 /* boundary */ 0,
812 /* lowaddr */ BUS_SPACE_MAXADDR,
813 /* highaddr */ BUS_SPACE_MAXADDR,
814 /* filter */ NULL,
815 /* filterarg */ NULL,
816 /* maxsize */ PAGE_SIZE,
817 /* nsegments */ 1,
818 /* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
819 /* flags */ 0,
820 /* lockfunc */ NULL,
821 /* lockarg */ NULL,
822 &bt->sg_dmat) != 0) {
823 goto error_exit;
824 }
825
826 bt->init_level++;
827
828 /* Perform initial CCB allocation */
829 bzero(bt->bt_ccb_array, bt->max_ccbs * sizeof(struct bt_ccb));
830 btallocccbs(bt);
831
832 if (bt->num_ccbs == 0) {
833 device_printf(dev,
834 "bt_init - Unable to allocate initial ccbs\n");
835 goto error_exit;
836 }
837
838 /*
839 * Note that we are going and return (to attach)
840 */
841 return 0;
842
843 error_exit:
844
845 return (ENXIO);
846 }
847
848 int
849 bt_attach(device_t dev)
850 {
851 struct bt_softc *bt = device_get_softc(dev);
852 int tagged_dev_openings;
853 struct cam_devq *devq;
854 int error;
855
856 /*
857 * We reserve 1 ccb for error recovery, so don't
858 * tell the XPT about it.
859 */
860 if (bt->tag_capable != 0)
861 tagged_dev_openings = bt->max_ccbs - 1;
862 else
863 tagged_dev_openings = 0;
864
865 /*
866 * Create the device queue for our SIM.
867 */
868 devq = cam_simq_alloc(bt->max_ccbs - 1);
869 if (devq == NULL)
870 return (ENOMEM);
871
872 /*
873 * Construct our SIM entry
874 */
875 bt->sim = cam_sim_alloc(btaction, btpoll, "bt", bt,
876 device_get_unit(bt->dev), &bt->lock, 2, tagged_dev_openings, devq);
877 if (bt->sim == NULL) {
878 cam_simq_free(devq);
879 return (ENOMEM);
880 }
881
882 mtx_lock(&bt->lock);
883 if (xpt_bus_register(bt->sim, dev, 0) != CAM_SUCCESS) {
884 cam_sim_free(bt->sim, /*free_devq*/TRUE);
885 mtx_unlock(&bt->lock);
886 return (ENXIO);
887 }
888
889 if (xpt_create_path(&bt->path, /*periph*/NULL,
890 cam_sim_path(bt->sim), CAM_TARGET_WILDCARD,
891 CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
892 xpt_bus_deregister(cam_sim_path(bt->sim));
893 cam_sim_free(bt->sim, /*free_devq*/TRUE);
894 mtx_unlock(&bt->lock);
895 return (ENXIO);
896 }
897 mtx_unlock(&bt->lock);
898
899 /*
900 * Setup interrupt.
901 */
902 error = bus_setup_intr(dev, bt->irq, INTR_TYPE_CAM | INTR_ENTROPY |
903 INTR_MPSAFE, NULL, bt_intr, bt, &bt->ih);
904 if (error) {
905 device_printf(dev, "bus_setup_intr() failed: %d\n", error);
906 return (error);
907 }
908 gone_in_dev(dev, 12, "bt(4) driver");
909
910 return (0);
911 }
912
913 int
914 bt_check_probed_iop(u_int ioport)
915 {
916 u_int i;
917
918 for (i = 0; i < BT_NUM_ISAPORTS; i++) {
919 if (bt_isa_ports[i].addr == ioport) {
920 if (bt_isa_ports[i].probed != 0)
921 return (1);
922 else {
923 return (0);
924 }
925 }
926 }
927 return (1);
928 }
929
930 void
931 bt_mark_probed_bio(isa_compat_io_t port)
932 {
933 if (port < BIO_DISABLED)
934 bt_mark_probed_iop(bt_board_ports[port]);
935 }
936
937 void
938 bt_mark_probed_iop(u_int ioport)
939 {
940 u_int i;
941
942 for (i = 0; i < BT_NUM_ISAPORTS; i++) {
943 if (ioport == bt_isa_ports[i].addr) {
944 bt_isa_ports[i].probed = 1;
945 break;
946 }
947 }
948 }
949
950 void
951 bt_find_probe_range(int ioport, int *port_index, int *max_port_index)
952 {
953 if (ioport > 0) {
954 int i;
955
956 for (i = 0;i < BT_NUM_ISAPORTS; i++)
957 if (ioport <= bt_isa_ports[i].addr)
958 break;
959 if ((i >= BT_NUM_ISAPORTS)
960 || (ioport != bt_isa_ports[i].addr)) {
961 printf(
962 "bt_find_probe_range: Invalid baseport of 0x%x specified.\n"
963 "bt_find_probe_range: Nearest valid baseport is 0x%x.\n"
964 "bt_find_probe_range: Failing probe.\n",
965 ioport,
966 (i < BT_NUM_ISAPORTS)
967 ? bt_isa_ports[i].addr
968 : bt_isa_ports[BT_NUM_ISAPORTS - 1].addr);
969 *port_index = *max_port_index = -1;
970 return;
971 }
972 *port_index = *max_port_index = bt_isa_ports[i].bio;
973 } else {
974 *port_index = 0;
975 *max_port_index = BT_NUM_ISAPORTS - 1;
976 }
977 }
978
979 int
980 bt_iop_from_bio(isa_compat_io_t bio_index)
981 {
982 if (bio_index < BT_NUM_ISAPORTS)
983 return (bt_board_ports[bio_index]);
984 return (-1);
985 }
986
987
988 static void
989 btallocccbs(struct bt_softc *bt)
990 {
991 struct bt_ccb *next_ccb;
992 struct sg_map_node *sg_map;
993 bus_addr_t physaddr;
994 bt_sg_t *segs;
995 int newcount;
996 int i;
997
998 if (bt->num_ccbs >= bt->max_ccbs)
999 /* Can't allocate any more */
1000 return;
1001
1002 next_ccb = &bt->bt_ccb_array[bt->num_ccbs];
1003
1004 sg_map = malloc(sizeof(*sg_map), M_DEVBUF, M_NOWAIT);
1005
1006 if (sg_map == NULL)
1007 goto error_exit;
1008
1009 /* Allocate S/G space for the next batch of CCBS */
1010 if (bus_dmamem_alloc(bt->sg_dmat, (void **)&sg_map->sg_vaddr,
1011 BUS_DMA_NOWAIT, &sg_map->sg_dmamap) != 0) {
1012 free(sg_map, M_DEVBUF);
1013 goto error_exit;
1014 }
1015
1016 SLIST_INSERT_HEAD(&bt->sg_maps, sg_map, links);
1017
1018 bus_dmamap_load(bt->sg_dmat, sg_map->sg_dmamap, sg_map->sg_vaddr,
1019 PAGE_SIZE, btmapsgs, bt, /*flags*/0);
1020
1021 segs = sg_map->sg_vaddr;
1022 physaddr = sg_map->sg_physaddr;
1023
1024 newcount = (PAGE_SIZE / (BT_NSEG * sizeof(bt_sg_t)));
1025 for (i = 0; bt->num_ccbs < bt->max_ccbs && i < newcount; i++) {
1026 int error;
1027
1028 next_ccb->sg_list = segs;
1029 next_ccb->sg_list_phys = physaddr;
1030 next_ccb->flags = BCCB_FREE;
1031 callout_init_mtx(&next_ccb->timer, &bt->lock, 0);
1032 error = bus_dmamap_create(bt->buffer_dmat, /*flags*/0,
1033 &next_ccb->dmamap);
1034 if (error != 0)
1035 break;
1036 SLIST_INSERT_HEAD(&bt->free_bt_ccbs, next_ccb, links);
1037 segs += BT_NSEG;
1038 physaddr += (BT_NSEG * sizeof(bt_sg_t));
1039 next_ccb++;
1040 bt->num_ccbs++;
1041 }
1042
1043 /* Reserve a CCB for error recovery */
1044 if (bt->recovery_bccb == NULL) {
1045 bt->recovery_bccb = SLIST_FIRST(&bt->free_bt_ccbs);
1046 SLIST_REMOVE_HEAD(&bt->free_bt_ccbs, links);
1047 }
1048
1049 if (SLIST_FIRST(&bt->free_bt_ccbs) != NULL)
1050 return;
1051
1052 error_exit:
1053 device_printf(bt->dev, "Can't malloc BCCBs\n");
1054 }
1055
1056 static __inline void
1057 btfreeccb(struct bt_softc *bt, struct bt_ccb *bccb)
1058 {
1059
1060 if (!dumping)
1061 mtx_assert(&bt->lock, MA_OWNED);
1062 if ((bccb->flags & BCCB_ACTIVE) != 0)
1063 LIST_REMOVE(&bccb->ccb->ccb_h, sim_links.le);
1064 if (bt->resource_shortage != 0
1065 && (bccb->ccb->ccb_h.status & CAM_RELEASE_SIMQ) == 0) {
1066 bccb->ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
1067 bt->resource_shortage = FALSE;
1068 }
1069 bccb->flags = BCCB_FREE;
1070 SLIST_INSERT_HEAD(&bt->free_bt_ccbs, bccb, links);
1071 bt->active_ccbs--;
1072 }
1073
1074 static __inline struct bt_ccb*
1075 btgetccb(struct bt_softc *bt)
1076 {
1077 struct bt_ccb* bccb;
1078
1079 if (!dumping)
1080 mtx_assert(&bt->lock, MA_OWNED);
1081 if ((bccb = SLIST_FIRST(&bt->free_bt_ccbs)) != NULL) {
1082 SLIST_REMOVE_HEAD(&bt->free_bt_ccbs, links);
1083 bt->active_ccbs++;
1084 } else {
1085 btallocccbs(bt);
1086 bccb = SLIST_FIRST(&bt->free_bt_ccbs);
1087 if (bccb != NULL) {
1088 SLIST_REMOVE_HEAD(&bt->free_bt_ccbs, links);
1089 bt->active_ccbs++;
1090 }
1091 }
1092
1093 return (bccb);
1094 }
1095
1096 static void
1097 btaction(struct cam_sim *sim, union ccb *ccb)
1098 {
1099 struct bt_softc *bt;
1100
1101 CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("btaction\n"));
1102
1103 bt = (struct bt_softc *)cam_sim_softc(sim);
1104 mtx_assert(&bt->lock, MA_OWNED);
1105
1106 switch (ccb->ccb_h.func_code) {
1107 /* Common cases first */
1108 case XPT_SCSI_IO: /* Execute the requested I/O operation */
1109 case XPT_RESET_DEV: /* Bus Device Reset the specified SCSI device */
1110 {
1111 struct bt_ccb *bccb;
1112 struct bt_hccb *hccb;
1113
1114 /*
1115 * get a bccb to use.
1116 */
1117 if ((bccb = btgetccb(bt)) == NULL) {
1118
1119 bt->resource_shortage = TRUE;
1120 xpt_freeze_simq(bt->sim, /*count*/1);
1121 ccb->ccb_h.status = CAM_REQUEUE_REQ;
1122 xpt_done(ccb);
1123 return;
1124 }
1125
1126 hccb = &bccb->hccb;
1127
1128 /*
1129 * So we can find the BCCB when an abort is requested
1130 */
1131 bccb->ccb = ccb;
1132 ccb->ccb_h.ccb_bccb_ptr = bccb;
1133 ccb->ccb_h.ccb_bt_ptr = bt;
1134
1135 /*
1136 * Put all the arguments for the xfer in the bccb
1137 */
1138 hccb->target_id = ccb->ccb_h.target_id;
1139 hccb->target_lun = ccb->ccb_h.target_lun;
1140 hccb->btstat = 0;
1141 hccb->sdstat = 0;
1142
1143 if (ccb->ccb_h.func_code == XPT_SCSI_IO) {
1144 struct ccb_scsiio *csio;
1145 struct ccb_hdr *ccbh;
1146 int error;
1147
1148 csio = &ccb->csio;
1149 ccbh = &csio->ccb_h;
1150 hccb->opcode = INITIATOR_CCB_WRESID;
1151 hccb->datain = (ccb->ccb_h.flags & CAM_DIR_IN) ? 1 : 0;
1152 hccb->dataout =(ccb->ccb_h.flags & CAM_DIR_OUT) ? 1 : 0;
1153 hccb->cmd_len = csio->cdb_len;
1154 if (hccb->cmd_len > sizeof(hccb->scsi_cdb)) {
1155 ccb->ccb_h.status = CAM_REQ_INVALID;
1156 btfreeccb(bt, bccb);
1157 xpt_done(ccb);
1158 return;
1159 }
1160 hccb->sense_len = csio->sense_len;
1161 if ((ccbh->flags & CAM_TAG_ACTION_VALID) != 0
1162 && ccb->csio.tag_action != CAM_TAG_ACTION_NONE) {
1163 hccb->tag_enable = TRUE;
1164 hccb->tag_type = (ccb->csio.tag_action & 0x3);
1165 } else {
1166 hccb->tag_enable = FALSE;
1167 hccb->tag_type = 0;
1168 }
1169 if ((ccbh->flags & CAM_CDB_POINTER) != 0) {
1170 if ((ccbh->flags & CAM_CDB_PHYS) == 0) {
1171 bcopy(csio->cdb_io.cdb_ptr,
1172 hccb->scsi_cdb, hccb->cmd_len);
1173 } else {
1174 /* I guess I could map it in... */
1175 ccbh->status = CAM_REQ_INVALID;
1176 btfreeccb(bt, bccb);
1177 xpt_done(ccb);
1178 return;
1179 }
1180 } else {
1181 bcopy(csio->cdb_io.cdb_bytes,
1182 hccb->scsi_cdb, hccb->cmd_len);
1183 }
1184 /* If need be, bounce our sense buffer */
1185 if (bt->sense_buffers != NULL) {
1186 hccb->sense_addr = btsensepaddr(bt, bccb);
1187 } else {
1188 hccb->sense_addr = vtophys(&csio->sense_data);
1189 }
1190 /*
1191 * If we have any data to send with this command,
1192 * map it into bus space.
1193 */
1194 error = bus_dmamap_load_ccb(
1195 bt->buffer_dmat,
1196 bccb->dmamap,
1197 ccb,
1198 btexecuteccb,
1199 bccb,
1200 /*flags*/0);
1201 if (error == EINPROGRESS) {
1202 /*
1203 * So as to maintain ordering, freeze the
1204 * controller queue until our mapping is
1205 * returned.
1206 */
1207 xpt_freeze_simq(bt->sim, 1);
1208 csio->ccb_h.status |= CAM_RELEASE_SIMQ;
1209 }
1210 } else {
1211 hccb->opcode = INITIATOR_BUS_DEV_RESET;
1212 /* No data transfer */
1213 hccb->datain = TRUE;
1214 hccb->dataout = TRUE;
1215 hccb->cmd_len = 0;
1216 hccb->sense_len = 0;
1217 hccb->tag_enable = FALSE;
1218 hccb->tag_type = 0;
1219 btexecuteccb(bccb, NULL, 0, 0);
1220 }
1221 break;
1222 }
1223 case XPT_ABORT: /* Abort the specified CCB */
1224 /* XXX Implement */
1225 ccb->ccb_h.status = CAM_REQ_INVALID;
1226 xpt_done(ccb);
1227 break;
1228 case XPT_SET_TRAN_SETTINGS:
1229 {
1230 /* XXX Implement */
1231 ccb->ccb_h.status = CAM_PROVIDE_FAIL;
1232 xpt_done(ccb);
1233 break;
1234 }
1235 case XPT_GET_TRAN_SETTINGS:
1236 /* Get default/user set transfer settings for the target */
1237 {
1238 struct ccb_trans_settings *cts;
1239 u_int target_mask;
1240
1241 cts = &ccb->cts;
1242 target_mask = 0x01 << ccb->ccb_h.target_id;
1243 if (cts->type == CTS_TYPE_CURRENT_SETTINGS) {
1244 struct ccb_trans_settings_scsi *scsi =
1245 &cts->proto_specific.scsi;
1246 struct ccb_trans_settings_spi *spi =
1247 &cts->xport_specific.spi;
1248 cts->protocol = PROTO_SCSI;
1249 cts->protocol_version = SCSI_REV_2;
1250 cts->transport = XPORT_SPI;
1251 cts->transport_version = 2;
1252
1253 scsi->flags &= ~CTS_SCSI_FLAGS_TAG_ENB;
1254 spi->flags &= ~CTS_SPI_FLAGS_DISC_ENB;
1255
1256 if ((bt->disc_permitted & target_mask) != 0)
1257 spi->flags |= CTS_SPI_FLAGS_DISC_ENB;
1258 if ((bt->tags_permitted & target_mask) != 0)
1259 scsi->flags |= CTS_SCSI_FLAGS_TAG_ENB;
1260
1261 if ((bt->ultra_permitted & target_mask) != 0)
1262 spi->sync_period = 12;
1263 else if ((bt->fast_permitted & target_mask) != 0)
1264 spi->sync_period = 25;
1265 else if ((bt->sync_permitted & target_mask) != 0)
1266 spi->sync_period = 50;
1267 else
1268 spi->sync_period = 0;
1269
1270 if (spi->sync_period != 0)
1271 spi->sync_offset = 15;
1272
1273 spi->valid |= CTS_SPI_VALID_SYNC_RATE;
1274 spi->valid |= CTS_SPI_VALID_SYNC_OFFSET;
1275
1276 spi->valid |= CTS_SPI_VALID_BUS_WIDTH;
1277 if ((bt->wide_permitted & target_mask) != 0)
1278 spi->bus_width = MSG_EXT_WDTR_BUS_16_BIT;
1279 else
1280 spi->bus_width = MSG_EXT_WDTR_BUS_8_BIT;
1281
1282 if (cts->ccb_h.target_lun != CAM_LUN_WILDCARD) {
1283 scsi->valid = CTS_SCSI_VALID_TQ;
1284 spi->valid |= CTS_SPI_VALID_DISC;
1285 } else
1286 scsi->valid = 0;
1287 } else {
1288 btfetchtransinfo(bt, cts);
1289 }
1290
1291 ccb->ccb_h.status = CAM_REQ_CMP;
1292 xpt_done(ccb);
1293 break;
1294 }
1295 case XPT_CALC_GEOMETRY:
1296 {
1297 struct ccb_calc_geometry *ccg;
1298 u_int32_t size_mb;
1299 u_int32_t secs_per_cylinder;
1300
1301 ccg = &ccb->ccg;
1302 size_mb = ccg->volume_size
1303 / ((1024L * 1024L) / ccg->block_size);
1304
1305 if (size_mb >= 1024 && (bt->extended_trans != 0)) {
1306 if (size_mb >= 2048) {
1307 ccg->heads = 255;
1308 ccg->secs_per_track = 63;
1309 } else {
1310 ccg->heads = 128;
1311 ccg->secs_per_track = 32;
1312 }
1313 } else {
1314 ccg->heads = 64;
1315 ccg->secs_per_track = 32;
1316 }
1317 secs_per_cylinder = ccg->heads * ccg->secs_per_track;
1318 ccg->cylinders = ccg->volume_size / secs_per_cylinder;
1319 ccb->ccb_h.status = CAM_REQ_CMP;
1320 xpt_done(ccb);
1321 break;
1322 }
1323 case XPT_RESET_BUS: /* Reset the specified SCSI bus */
1324 {
1325 btreset(bt, /*hardreset*/TRUE);
1326 ccb->ccb_h.status = CAM_REQ_CMP;
1327 xpt_done(ccb);
1328 break;
1329 }
1330 case XPT_TERM_IO: /* Terminate the I/O process */
1331 /* XXX Implement */
1332 ccb->ccb_h.status = CAM_REQ_INVALID;
1333 xpt_done(ccb);
1334 break;
1335 case XPT_PATH_INQ: /* Path routing inquiry */
1336 {
1337 struct ccb_pathinq *cpi = &ccb->cpi;
1338
1339 cpi->version_num = 1; /* XXX??? */
1340 cpi->hba_inquiry = PI_SDTR_ABLE;
1341 if (bt->tag_capable != 0)
1342 cpi->hba_inquiry |= PI_TAG_ABLE;
1343 if (bt->wide_bus != 0)
1344 cpi->hba_inquiry |= PI_WIDE_16;
1345 cpi->target_sprt = 0;
1346 cpi->hba_misc = 0;
1347 cpi->hba_eng_cnt = 0;
1348 cpi->max_target = bt->wide_bus ? 15 : 7;
1349 cpi->max_lun = 7;
1350 cpi->initiator_id = bt->scsi_id;
1351 cpi->bus_id = cam_sim_bus(sim);
1352 cpi->base_transfer_speed = 3300;
1353 strlcpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
1354 strlcpy(cpi->hba_vid, "BusLogic", HBA_IDLEN);
1355 strlcpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
1356 cpi->unit_number = cam_sim_unit(sim);
1357 cpi->ccb_h.status = CAM_REQ_CMP;
1358 cpi->transport = XPORT_SPI;
1359 cpi->transport_version = 2;
1360 cpi->protocol = PROTO_SCSI;
1361 cpi->protocol_version = SCSI_REV_2;
1362 xpt_done(ccb);
1363 break;
1364 }
1365 default:
1366 ccb->ccb_h.status = CAM_REQ_INVALID;
1367 xpt_done(ccb);
1368 break;
1369 }
1370 }
1371
1372 static void
1373 btexecuteccb(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
1374 {
1375 struct bt_ccb *bccb;
1376 union ccb *ccb;
1377 struct bt_softc *bt;
1378
1379 bccb = (struct bt_ccb *)arg;
1380 ccb = bccb->ccb;
1381 bt = (struct bt_softc *)ccb->ccb_h.ccb_bt_ptr;
1382
1383 if (error != 0) {
1384 if (error != EFBIG)
1385 device_printf(bt->dev,
1386 "Unexepected error 0x%x returned from "
1387 "bus_dmamap_load\n", error);
1388 if (ccb->ccb_h.status == CAM_REQ_INPROG) {
1389 xpt_freeze_devq(ccb->ccb_h.path, /*count*/1);
1390 ccb->ccb_h.status = CAM_REQ_TOO_BIG|CAM_DEV_QFRZN;
1391 }
1392 btfreeccb(bt, bccb);
1393 xpt_done(ccb);
1394 return;
1395 }
1396
1397 if (nseg != 0) {
1398 bt_sg_t *sg;
1399 bus_dma_segment_t *end_seg;
1400 bus_dmasync_op_t op;
1401
1402 end_seg = dm_segs + nseg;
1403
1404 /* Copy the segments into our SG list */
1405 sg = bccb->sg_list;
1406 while (dm_segs < end_seg) {
1407 sg->len = dm_segs->ds_len;
1408 sg->addr = dm_segs->ds_addr;
1409 sg++;
1410 dm_segs++;
1411 }
1412
1413 if (nseg > 1) {
1414 bccb->hccb.opcode = INITIATOR_SG_CCB_WRESID;
1415 bccb->hccb.data_len = sizeof(bt_sg_t) * nseg;
1416 bccb->hccb.data_addr = bccb->sg_list_phys;
1417 } else {
1418 bccb->hccb.data_len = bccb->sg_list->len;
1419 bccb->hccb.data_addr = bccb->sg_list->addr;
1420 }
1421
1422 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
1423 op = BUS_DMASYNC_PREREAD;
1424 else
1425 op = BUS_DMASYNC_PREWRITE;
1426
1427 bus_dmamap_sync(bt->buffer_dmat, bccb->dmamap, op);
1428
1429 } else {
1430 bccb->hccb.opcode = INITIATOR_CCB;
1431 bccb->hccb.data_len = 0;
1432 bccb->hccb.data_addr = 0;
1433 }
1434
1435 /*
1436 * Last time we need to check if this CCB needs to
1437 * be aborted.
1438 */
1439 if (ccb->ccb_h.status != CAM_REQ_INPROG) {
1440 if (nseg != 0)
1441 bus_dmamap_unload(bt->buffer_dmat, bccb->dmamap);
1442 btfreeccb(bt, bccb);
1443 xpt_done(ccb);
1444 return;
1445 }
1446
1447 bccb->flags = BCCB_ACTIVE;
1448 ccb->ccb_h.status |= CAM_SIM_QUEUED;
1449 LIST_INSERT_HEAD(&bt->pending_ccbs, &ccb->ccb_h, sim_links.le);
1450
1451 callout_reset_sbt(&bccb->timer, SBT_1MS * ccb->ccb_h.timeout, 0,
1452 bttimeout, bccb, 0);
1453
1454 /* Tell the adapter about this command */
1455 bt->cur_outbox->ccb_addr = btccbvtop(bt, bccb);
1456 if (bt->cur_outbox->action_code != BMBO_FREE) {
1457 /*
1458 * We should never encounter a busy mailbox.
1459 * If we do, warn the user, and treat it as
1460 * a resource shortage. If the controller is
1461 * hung, one of the pending transactions will
1462 * timeout causing us to start recovery operations.
1463 */
1464 device_printf(bt->dev,
1465 "Encountered busy mailbox with %d out of %d "
1466 "commands active!!!\n", bt->active_ccbs,
1467 bt->max_ccbs);
1468 callout_stop(&bccb->timer);
1469 if (nseg != 0)
1470 bus_dmamap_unload(bt->buffer_dmat, bccb->dmamap);
1471 btfreeccb(bt, bccb);
1472 bt->resource_shortage = TRUE;
1473 xpt_freeze_simq(bt->sim, /*count*/1);
1474 ccb->ccb_h.status = CAM_REQUEUE_REQ;
1475 xpt_done(ccb);
1476 return;
1477 }
1478 bt->cur_outbox->action_code = BMBO_START;
1479 bt_outb(bt, COMMAND_REG, BOP_START_MBOX);
1480 btnextoutbox(bt);
1481 }
1482
1483 void
1484 bt_intr(void *arg)
1485 {
1486 struct bt_softc *bt;
1487
1488 bt = arg;
1489 mtx_lock(&bt->lock);
1490 bt_intr_locked(bt);
1491 mtx_unlock(&bt->lock);
1492 }
1493
1494 void
1495 bt_intr_locked(struct bt_softc *bt)
1496 {
1497 u_int intstat;
1498
1499 while (((intstat = bt_inb(bt, INTSTAT_REG)) & INTR_PENDING) != 0) {
1500
1501 if ((intstat & CMD_COMPLETE) != 0) {
1502 bt->latched_status = bt_inb(bt, STATUS_REG);
1503 bt->command_cmp = TRUE;
1504 }
1505
1506 bt_outb(bt, CONTROL_REG, RESET_INTR);
1507
1508 if ((intstat & IMB_LOADED) != 0) {
1509 while (bt->cur_inbox->comp_code != BMBI_FREE) {
1510 btdone(bt,
1511 btccbptov(bt, bt->cur_inbox->ccb_addr),
1512 bt->cur_inbox->comp_code);
1513 bt->cur_inbox->comp_code = BMBI_FREE;
1514 btnextinbox(bt);
1515 }
1516 }
1517
1518 if ((intstat & SCSI_BUS_RESET) != 0) {
1519 btreset(bt, /*hardreset*/FALSE);
1520 }
1521 }
1522 }
1523
1524 static void
1525 btdone(struct bt_softc *bt, struct bt_ccb *bccb, bt_mbi_comp_code_t comp_code)
1526 {
1527 union ccb *ccb;
1528 struct ccb_scsiio *csio;
1529
1530 ccb = bccb->ccb;
1531 csio = &bccb->ccb->csio;
1532
1533 if ((bccb->flags & BCCB_ACTIVE) == 0) {
1534 device_printf(bt->dev,
1535 "btdone - Attempt to free non-active BCCB %p\n",
1536 (void *)bccb);
1537 return;
1538 }
1539
1540 if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) {
1541 bus_dmasync_op_t op;
1542
1543 if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
1544 op = BUS_DMASYNC_POSTREAD;
1545 else
1546 op = BUS_DMASYNC_POSTWRITE;
1547 bus_dmamap_sync(bt->buffer_dmat, bccb->dmamap, op);
1548 bus_dmamap_unload(bt->buffer_dmat, bccb->dmamap);
1549 }
1550
1551 if (bccb == bt->recovery_bccb) {
1552 /*
1553 * The recovery BCCB does not have a CCB associated
1554 * with it, so short circuit the normal error handling.
1555 * We now traverse our list of pending CCBs and process
1556 * any that were terminated by the recovery CCBs action.
1557 * We also reinstate timeouts for all remaining, pending,
1558 * CCBs.
1559 */
1560 struct cam_path *path;
1561 struct ccb_hdr *ccb_h;
1562 cam_status error;
1563
1564 /* Notify all clients that a BDR occurred */
1565 error = xpt_create_path(&path, /*periph*/NULL,
1566 cam_sim_path(bt->sim),
1567 bccb->hccb.target_id,
1568 CAM_LUN_WILDCARD);
1569
1570 if (error == CAM_REQ_CMP) {
1571 xpt_async(AC_SENT_BDR, path, NULL);
1572 xpt_free_path(path);
1573 }
1574
1575 ccb_h = LIST_FIRST(&bt->pending_ccbs);
1576 while (ccb_h != NULL) {
1577 struct bt_ccb *pending_bccb;
1578
1579 pending_bccb = (struct bt_ccb *)ccb_h->ccb_bccb_ptr;
1580 if (pending_bccb->hccb.target_id
1581 == bccb->hccb.target_id) {
1582 pending_bccb->hccb.btstat = BTSTAT_HA_BDR;
1583 ccb_h = LIST_NEXT(ccb_h, sim_links.le);
1584 btdone(bt, pending_bccb, BMBI_ERROR);
1585 } else {
1586 callout_reset_sbt(&pending_bccb->timer,
1587 SBT_1MS * ccb_h->timeout, 0, bttimeout,
1588 pending_bccb, 0);
1589 ccb_h = LIST_NEXT(ccb_h, sim_links.le);
1590 }
1591 }
1592 device_printf(bt->dev, "No longer in timeout\n");
1593 return;
1594 }
1595
1596 callout_stop(&bccb->timer);
1597
1598 switch (comp_code) {
1599 case BMBI_FREE:
1600 device_printf(bt->dev,
1601 "btdone - CCB completed with free status!\n");
1602 break;
1603 case BMBI_NOT_FOUND:
1604 device_printf(bt->dev,
1605 "btdone - CCB Abort failed to find CCB\n");
1606 break;
1607 case BMBI_ABORT:
1608 case BMBI_ERROR:
1609 if (bootverbose) {
1610 printf("bt: ccb %p - error %x occurred. "
1611 "btstat = %x, sdstat = %x\n",
1612 (void *)bccb, comp_code, bccb->hccb.btstat,
1613 bccb->hccb.sdstat);
1614 }
1615 /* An error occurred */
1616 switch(bccb->hccb.btstat) {
1617 case BTSTAT_DATARUN_ERROR:
1618 if (bccb->hccb.data_len == 0) {
1619 /*
1620 * At least firmware 4.22, does this
1621 * for a QUEUE FULL condition.
1622 */
1623 bccb->hccb.sdstat = SCSI_STATUS_QUEUE_FULL;
1624 } else if (bccb->hccb.data_len < 0) {
1625 csio->ccb_h.status = CAM_DATA_RUN_ERR;
1626 break;
1627 }
1628 /* FALLTHROUGH */
1629 case BTSTAT_NOERROR:
1630 case BTSTAT_LINKED_CMD_COMPLETE:
1631 case BTSTAT_LINKED_CMD_FLAG_COMPLETE:
1632 case BTSTAT_DATAUNDERUN_ERROR:
1633
1634 csio->scsi_status = bccb->hccb.sdstat;
1635 csio->ccb_h.status |= CAM_SCSI_STATUS_ERROR;
1636 switch(csio->scsi_status) {
1637 case SCSI_STATUS_CHECK_COND:
1638 case SCSI_STATUS_CMD_TERMINATED:
1639 csio->ccb_h.status |= CAM_AUTOSNS_VALID;
1640 /* Bounce sense back if necessary */
1641 if (bt->sense_buffers != NULL) {
1642 csio->sense_data =
1643 *btsensevaddr(bt, bccb);
1644 }
1645 break;
1646 default:
1647 break;
1648 case SCSI_STATUS_OK:
1649 csio->ccb_h.status = CAM_REQ_CMP;
1650 break;
1651 }
1652 csio->resid = bccb->hccb.data_len;
1653 break;
1654 case BTSTAT_SELTIMEOUT:
1655 csio->ccb_h.status = CAM_SEL_TIMEOUT;
1656 break;
1657 case BTSTAT_UNEXPECTED_BUSFREE:
1658 csio->ccb_h.status = CAM_UNEXP_BUSFREE;
1659 break;
1660 case BTSTAT_INVALID_PHASE:
1661 csio->ccb_h.status = CAM_SEQUENCE_FAIL;
1662 break;
1663 case BTSTAT_INVALID_ACTION_CODE:
1664 panic("%s: Inavlid Action code", bt_name(bt));
1665 break;
1666 case BTSTAT_INVALID_OPCODE:
1667 panic("%s: Inavlid CCB Opcode code", bt_name(bt));
1668 break;
1669 case BTSTAT_LINKED_CCB_LUN_MISMATCH:
1670 /* We don't even support linked commands... */
1671 panic("%s: Linked CCB Lun Mismatch", bt_name(bt));
1672 break;
1673 case BTSTAT_INVALID_CCB_OR_SG_PARAM:
1674 panic("%s: Invalid CCB or SG list", bt_name(bt));
1675 break;
1676 case BTSTAT_AUTOSENSE_FAILED:
1677 csio->ccb_h.status = CAM_AUTOSENSE_FAIL;
1678 break;
1679 case BTSTAT_TAGGED_MSG_REJECTED:
1680 {
1681 struct ccb_trans_settings neg;
1682 struct ccb_trans_settings_scsi *scsi =
1683 &neg.proto_specific.scsi;
1684
1685 neg.protocol = PROTO_SCSI;
1686 neg.protocol_version = SCSI_REV_2;
1687 neg.transport = XPORT_SPI;
1688 neg.transport_version = 2;
1689 scsi->valid = CTS_SCSI_VALID_TQ;
1690 scsi->flags = 0;
1691 xpt_print_path(csio->ccb_h.path);
1692 printf("refuses tagged commands. Performing "
1693 "non-tagged I/O\n");
1694 xpt_setup_ccb(&neg.ccb_h, csio->ccb_h.path,
1695 /*priority*/1);
1696 xpt_async(AC_TRANSFER_NEG, csio->ccb_h.path, &neg);
1697 bt->tags_permitted &= ~(0x01 << csio->ccb_h.target_id);
1698 csio->ccb_h.status = CAM_MSG_REJECT_REC;
1699 break;
1700 }
1701 case BTSTAT_UNSUPPORTED_MSG_RECEIVED:
1702 /*
1703 * XXX You would think that this is
1704 * a recoverable error... Hmmm.
1705 */
1706 csio->ccb_h.status = CAM_REQ_CMP_ERR;
1707 break;
1708 case BTSTAT_HA_SOFTWARE_ERROR:
1709 case BTSTAT_HA_WATCHDOG_ERROR:
1710 case BTSTAT_HARDWARE_FAILURE:
1711 /* Hardware reset ??? Can we recover ??? */
1712 csio->ccb_h.status = CAM_NO_HBA;
1713 break;
1714 case BTSTAT_TARGET_IGNORED_ATN:
1715 case BTSTAT_OTHER_SCSI_BUS_RESET:
1716 case BTSTAT_HA_SCSI_BUS_RESET:
1717 if ((csio->ccb_h.status & CAM_STATUS_MASK)
1718 != CAM_CMD_TIMEOUT)
1719 csio->ccb_h.status = CAM_SCSI_BUS_RESET;
1720 break;
1721 case BTSTAT_HA_BDR:
1722 if ((bccb->flags & BCCB_DEVICE_RESET) == 0)
1723 csio->ccb_h.status = CAM_BDR_SENT;
1724 else
1725 csio->ccb_h.status = CAM_CMD_TIMEOUT;
1726 break;
1727 case BTSTAT_INVALID_RECONNECT:
1728 case BTSTAT_ABORT_QUEUE_GENERATED:
1729 csio->ccb_h.status = CAM_REQ_TERMIO;
1730 break;
1731 case BTSTAT_SCSI_PERROR_DETECTED:
1732 csio->ccb_h.status = CAM_UNCOR_PARITY;
1733 break;
1734 }
1735 if (csio->ccb_h.status != CAM_REQ_CMP) {
1736 xpt_freeze_devq(csio->ccb_h.path, /*count*/1);
1737 csio->ccb_h.status |= CAM_DEV_QFRZN;
1738 }
1739 if ((bccb->flags & BCCB_RELEASE_SIMQ) != 0)
1740 ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
1741 btfreeccb(bt, bccb);
1742 xpt_done(ccb);
1743 break;
1744 case BMBI_OK:
1745 /* All completed without incident */
1746 ccb->ccb_h.status |= CAM_REQ_CMP;
1747 if ((bccb->flags & BCCB_RELEASE_SIMQ) != 0)
1748 ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
1749 btfreeccb(bt, bccb);
1750 xpt_done(ccb);
1751 break;
1752 }
1753 }
1754
1755 static int
1756 btreset(struct bt_softc* bt, int hard_reset)
1757 {
1758 struct ccb_hdr *ccb_h;
1759 u_int status;
1760 u_int timeout;
1761 u_int8_t reset_type;
1762
1763 if (hard_reset != 0)
1764 reset_type = HARD_RESET;
1765 else
1766 reset_type = SOFT_RESET;
1767 bt_outb(bt, CONTROL_REG, reset_type);
1768
1769 /* Wait 5sec. for Diagnostic start */
1770 timeout = 5 * 10000;
1771 while (--timeout) {
1772 status = bt_inb(bt, STATUS_REG);
1773 if ((status & DIAG_ACTIVE) != 0)
1774 break;
1775 DELAY(100);
1776 }
1777 if (timeout == 0) {
1778 if (bootverbose)
1779 device_printf(bt->dev,
1780 "btreset - Diagnostic Active failed to "
1781 "assert. status = 0x%x\n", status);
1782 return (ETIMEDOUT);
1783 }
1784
1785 /* Wait 10sec. for Diagnostic end */
1786 timeout = 10 * 10000;
1787 while (--timeout) {
1788 status = bt_inb(bt, STATUS_REG);
1789 if ((status & DIAG_ACTIVE) == 0)
1790 break;
1791 DELAY(100);
1792 }
1793 if (timeout == 0) {
1794 panic("%s: btreset - Diagnostic Active failed to drop. "
1795 "status = 0x%x\n", bt_name(bt), status);
1796 return (ETIMEDOUT);
1797 }
1798
1799 /* Wait for the host adapter to become ready or report a failure */
1800 timeout = 10000;
1801 while (--timeout) {
1802 status = bt_inb(bt, STATUS_REG);
1803 if ((status & (DIAG_FAIL|HA_READY|DATAIN_REG_READY)) != 0)
1804 break;
1805 DELAY(100);
1806 }
1807 if (timeout == 0) {
1808 device_printf(bt->dev,
1809 "btreset - Host adapter failed to come ready. "
1810 "status = 0x%x\n", status);
1811 return (ETIMEDOUT);
1812 }
1813
1814 /* If the diagnostics failed, tell the user */
1815 if ((status & DIAG_FAIL) != 0
1816 || (status & HA_READY) == 0) {
1817 device_printf(bt->dev,
1818 "btreset - Adapter failed diagnostics\n");
1819
1820 if ((status & DATAIN_REG_READY) != 0)
1821 device_printf(bt->dev,
1822 "btreset - Host Adapter Error code = 0x%x\n",
1823 bt_inb(bt, DATAIN_REG));
1824 return (ENXIO);
1825 }
1826
1827 /* If we've allocated mailboxes, initialize them */
1828 if (bt->init_level > 4)
1829 btinitmboxes(bt);
1830
1831 /* If we've attached to the XPT, tell it about the event */
1832 if (bt->path != NULL)
1833 xpt_async(AC_BUS_RESET, bt->path, NULL);
1834
1835 /*
1836 * Perform completion processing for all outstanding CCBs.
1837 */
1838 while ((ccb_h = LIST_FIRST(&bt->pending_ccbs)) != NULL) {
1839 struct bt_ccb *pending_bccb;
1840
1841 pending_bccb = (struct bt_ccb *)ccb_h->ccb_bccb_ptr;
1842 pending_bccb->hccb.btstat = BTSTAT_HA_SCSI_BUS_RESET;
1843 btdone(bt, pending_bccb, BMBI_ERROR);
1844 }
1845
1846 return (0);
1847 }
1848
1849 /*
1850 * Send a command to the adapter.
1851 */
1852 int
1853 bt_cmd(struct bt_softc *bt, bt_op_t opcode, u_int8_t *params, u_int param_len,
1854 u_int8_t *reply_data, u_int reply_len, u_int cmd_timeout)
1855 {
1856 u_int timeout;
1857 u_int status;
1858 u_int saved_status;
1859 u_int intstat;
1860 u_int reply_buf_size;
1861 int cmd_complete;
1862 int error;
1863
1864 /* No data returned to start */
1865 reply_buf_size = reply_len;
1866 reply_len = 0;
1867 intstat = 0;
1868 cmd_complete = 0;
1869 saved_status = 0;
1870 error = 0;
1871
1872 bt->command_cmp = 0;
1873 /*
1874 * Wait up to 10 sec. for the adapter to become
1875 * ready to accept commands.
1876 */
1877 timeout = 100000;
1878 while (--timeout) {
1879 status = bt_inb(bt, STATUS_REG);
1880 if ((status & HA_READY) != 0
1881 && (status & CMD_REG_BUSY) == 0)
1882 break;
1883 /*
1884 * Throw away any pending data which may be
1885 * left over from earlier commands that we
1886 * timedout on.
1887 */
1888 if ((status & DATAIN_REG_READY) != 0)
1889 (void)bt_inb(bt, DATAIN_REG);
1890 DELAY(100);
1891 }
1892 if (timeout == 0) {
1893 device_printf(bt->dev,
1894 "bt_cmd: Timeout waiting for adapter ready, "
1895 "status = 0x%x\n", status);
1896 return (ETIMEDOUT);
1897 }
1898
1899 /*
1900 * Send the opcode followed by any necessary parameter bytes.
1901 */
1902 bt_outb(bt, COMMAND_REG, opcode);
1903
1904 /*
1905 * Wait for up to 1sec for each byte of the
1906 * parameter list sent to be sent.
1907 */
1908 timeout = 10000;
1909 while (param_len && --timeout) {
1910 DELAY(100);
1911 status = bt_inb(bt, STATUS_REG);
1912 intstat = bt_inb(bt, INTSTAT_REG);
1913
1914 if ((intstat & (INTR_PENDING|CMD_COMPLETE))
1915 == (INTR_PENDING|CMD_COMPLETE)) {
1916 saved_status = status;
1917 cmd_complete = 1;
1918 break;
1919 }
1920 if (bt->command_cmp != 0) {
1921 saved_status = bt->latched_status;
1922 cmd_complete = 1;
1923 break;
1924 }
1925 if ((status & DATAIN_REG_READY) != 0)
1926 break;
1927 if ((status & CMD_REG_BUSY) == 0) {
1928 bt_outb(bt, COMMAND_REG, *params++);
1929 param_len--;
1930 timeout = 10000;
1931 }
1932 }
1933 if (timeout == 0) {
1934 device_printf(bt->dev, "bt_cmd: Timeout sending parameters, "
1935 "status = 0x%x\n", status);
1936 cmd_complete = 1;
1937 saved_status = status;
1938 error = ETIMEDOUT;
1939 }
1940
1941 /*
1942 * Wait for the command to complete.
1943 */
1944 while (cmd_complete == 0 && --cmd_timeout) {
1945
1946 status = bt_inb(bt, STATUS_REG);
1947 intstat = bt_inb(bt, INTSTAT_REG);
1948 /*
1949 * It may be that this command was issued with
1950 * controller interrupts disabled. We'll never
1951 * get to our command if an incoming mailbox
1952 * interrupt is pending, so take care of completed
1953 * mailbox commands by calling our interrupt handler.
1954 */
1955 if ((intstat & (INTR_PENDING|IMB_LOADED))
1956 == (INTR_PENDING|IMB_LOADED))
1957 bt_intr_locked(bt);
1958
1959 if (bt->command_cmp != 0) {
1960 /*
1961 * Our interrupt handler saw CMD_COMPLETE
1962 * status before we did.
1963 */
1964 cmd_complete = 1;
1965 saved_status = bt->latched_status;
1966 } else if ((intstat & (INTR_PENDING|CMD_COMPLETE))
1967 == (INTR_PENDING|CMD_COMPLETE)) {
1968 /*
1969 * Our poll (in case interrupts are blocked)
1970 * saw the CMD_COMPLETE interrupt.
1971 */
1972 cmd_complete = 1;
1973 saved_status = status;
1974 } else if (opcode == BOP_MODIFY_IO_ADDR
1975 && (status & CMD_REG_BUSY) == 0) {
1976 /*
1977 * The BOP_MODIFY_IO_ADDR does not issue a CMD_COMPLETE,
1978 * but it should update the status register. So, we
1979 * consider this command complete when the CMD_REG_BUSY
1980 * status clears.
1981 */
1982 saved_status = status;
1983 cmd_complete = 1;
1984 } else if ((status & DATAIN_REG_READY) != 0) {
1985 u_int8_t data;
1986
1987 data = bt_inb(bt, DATAIN_REG);
1988 if (reply_len < reply_buf_size) {
1989 *reply_data++ = data;
1990 } else {
1991 device_printf(bt->dev,
1992 "bt_cmd - Discarded reply data byte "
1993 "for opcode 0x%x\n", opcode);
1994 }
1995 /*
1996 * Reset timeout to ensure at least a second
1997 * between response bytes.
1998 */
1999 cmd_timeout = MAX(cmd_timeout, 10000);
2000 reply_len++;
2001
2002 } else if ((opcode == BOP_FETCH_LRAM)
2003 && (status & HA_READY) != 0) {
2004 saved_status = status;
2005 cmd_complete = 1;
2006 }
2007 DELAY(100);
2008 }
2009 if (cmd_timeout == 0) {
2010 device_printf(bt->dev,
2011 "bt_cmd: Timeout waiting for command (%x) "
2012 "to complete.\n", opcode);
2013 device_printf(bt->dev, "status = 0x%x, intstat = 0x%x, "
2014 "rlen %d\n", status, intstat, reply_len);
2015 error = (ETIMEDOUT);
2016 }
2017
2018 /*
2019 * Clear any pending interrupts.
2020 */
2021 bt_intr_locked(bt);
2022
2023 if (error != 0)
2024 return (error);
2025
2026 /*
2027 * If the command was rejected by the controller, tell the caller.
2028 */
2029 if ((saved_status & CMD_INVALID) != 0) {
2030 /*
2031 * Some early adapters may not recover properly from
2032 * an invalid command. If it appears that the controller
2033 * has wedged (i.e. status was not cleared by our interrupt
2034 * reset above), perform a soft reset.
2035 */
2036 if (bootverbose)
2037 device_printf(bt->dev, "Invalid Command 0x%x\n",
2038 opcode);
2039 DELAY(1000);
2040 status = bt_inb(bt, STATUS_REG);
2041 if ((status & (CMD_INVALID|STATUS_REG_RSVD|DATAIN_REG_READY|
2042 CMD_REG_BUSY|DIAG_FAIL|DIAG_ACTIVE)) != 0
2043 || (status & (HA_READY|INIT_REQUIRED))
2044 != (HA_READY|INIT_REQUIRED)) {
2045 btreset(bt, /*hard_reset*/FALSE);
2046 }
2047 return (EINVAL);
2048 }
2049
2050 if (param_len > 0) {
2051 /* The controller did not accept the full argument list */
2052 return (E2BIG);
2053 }
2054
2055 if (reply_len != reply_buf_size) {
2056 /* Too much or too little data received */
2057 return (EMSGSIZE);
2058 }
2059
2060 /* We were successful */
2061 return (0);
2062 }
2063
2064 static int
2065 btinitmboxes(struct bt_softc *bt) {
2066 init_32b_mbox_params_t init_mbox;
2067 int error;
2068
2069 bzero(bt->in_boxes, sizeof(bt_mbox_in_t) * bt->num_boxes);
2070 bzero(bt->out_boxes, sizeof(bt_mbox_out_t) * bt->num_boxes);
2071 bt->cur_inbox = bt->in_boxes;
2072 bt->last_inbox = bt->in_boxes + bt->num_boxes - 1;
2073 bt->cur_outbox = bt->out_boxes;
2074 bt->last_outbox = bt->out_boxes + bt->num_boxes - 1;
2075
2076 /* Tell the adapter about them */
2077 init_mbox.num_boxes = bt->num_boxes;
2078 init_mbox.base_addr[0] = bt->mailbox_physbase & 0xFF;
2079 init_mbox.base_addr[1] = (bt->mailbox_physbase >> 8) & 0xFF;
2080 init_mbox.base_addr[2] = (bt->mailbox_physbase >> 16) & 0xFF;
2081 init_mbox.base_addr[3] = (bt->mailbox_physbase >> 24) & 0xFF;
2082 error = bt_cmd(bt, BOP_INITIALIZE_32BMBOX, (u_int8_t *)&init_mbox,
2083 /*parmlen*/sizeof(init_mbox), /*reply_buf*/NULL,
2084 /*reply_len*/0, DEFAULT_CMD_TIMEOUT);
2085
2086 if (error != 0)
2087 printf("btinitmboxes: Initialization command failed\n");
2088 else if (bt->strict_rr != 0) {
2089 /*
2090 * If the controller supports
2091 * strict round robin mode,
2092 * enable it
2093 */
2094 u_int8_t param;
2095
2096 param = 0;
2097 error = bt_cmd(bt, BOP_ENABLE_STRICT_RR, ¶m, 1,
2098 /*reply_buf*/NULL, /*reply_len*/0,
2099 DEFAULT_CMD_TIMEOUT);
2100
2101 if (error != 0) {
2102 printf("btinitmboxes: Unable to enable strict RR\n");
2103 error = 0;
2104 } else if (bootverbose) {
2105 device_printf(bt->dev,
2106 "Using Strict Round Robin Mailbox Mode\n");
2107 }
2108 }
2109
2110 return (error);
2111 }
2112
2113 /*
2114 * Update the XPT's idea of the negotiated transfer
2115 * parameters for a particular target.
2116 */
2117 static void
2118 btfetchtransinfo(struct bt_softc *bt, struct ccb_trans_settings *cts)
2119 {
2120 setup_data_t setup_info;
2121 u_int target;
2122 u_int targ_offset;
2123 u_int targ_mask;
2124 u_int sync_period;
2125 u_int sync_offset;
2126 u_int bus_width;
2127 int error;
2128 u_int8_t param;
2129 targ_syncinfo_t sync_info;
2130 struct ccb_trans_settings_scsi *scsi =
2131 &cts->proto_specific.scsi;
2132 struct ccb_trans_settings_spi *spi =
2133 &cts->xport_specific.spi;
2134
2135 spi->valid = 0;
2136 scsi->valid = 0;
2137
2138 target = cts->ccb_h.target_id;
2139 targ_offset = (target & 0x7);
2140 targ_mask = (0x01 << targ_offset);
2141
2142 /*
2143 * Inquire Setup Information. This command retreives the
2144 * Wide negotiation status for recent adapters as well as
2145 * the sync info for older models.
2146 */
2147 param = sizeof(setup_info);
2148 error = bt_cmd(bt, BOP_INQUIRE_SETUP_INFO, ¶m, /*paramlen*/1,
2149 (u_int8_t*)&setup_info, sizeof(setup_info),
2150 DEFAULT_CMD_TIMEOUT);
2151
2152 if (error != 0) {
2153 device_printf(bt->dev,
2154 "btfetchtransinfo - Inquire Setup Info Failed %x\n",
2155 error);
2156 return;
2157 }
2158
2159 sync_info = (target < 8) ? setup_info.low_syncinfo[targ_offset]
2160 : setup_info.high_syncinfo[targ_offset];
2161
2162 if (sync_info.sync == 0)
2163 sync_offset = 0;
2164 else
2165 sync_offset = sync_info.offset;
2166
2167
2168 bus_width = MSG_EXT_WDTR_BUS_8_BIT;
2169 if (strcmp(bt->firmware_ver, "5.06L") >= 0) {
2170 u_int wide_active;
2171
2172 wide_active =
2173 (target < 8) ? (setup_info.low_wide_active & targ_mask)
2174 : (setup_info.high_wide_active & targ_mask);
2175
2176 if (wide_active)
2177 bus_width = MSG_EXT_WDTR_BUS_16_BIT;
2178 } else if ((bt->wide_permitted & targ_mask) != 0) {
2179 struct ccb_getdev cgd;
2180
2181 /*
2182 * Prior to rev 5.06L, wide status isn't provided,
2183 * so we "guess" that wide transfers are in effect
2184 * if the user settings allow for wide and the inquiry
2185 * data for the device indicates that it can handle
2186 * wide transfers.
2187 */
2188 xpt_setup_ccb(&cgd.ccb_h, cts->ccb_h.path, /*priority*/1);
2189 cgd.ccb_h.func_code = XPT_GDEV_TYPE;
2190 xpt_action((union ccb *)&cgd);
2191 if ((cgd.ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP
2192 && (cgd.inq_data.flags & SID_WBus16) != 0)
2193 bus_width = MSG_EXT_WDTR_BUS_16_BIT;
2194 }
2195
2196 if (bt->firmware_ver[0] >= '3') {
2197 /*
2198 * For adapters that can do fast or ultra speeds,
2199 * use the more exact Target Sync Information command.
2200 */
2201 target_sync_info_data_t sync_info;
2202
2203 param = sizeof(sync_info);
2204 error = bt_cmd(bt, BOP_TARG_SYNC_INFO, ¶m, /*paramlen*/1,
2205 (u_int8_t*)&sync_info, sizeof(sync_info),
2206 DEFAULT_CMD_TIMEOUT);
2207
2208 if (error != 0) {
2209 device_printf(bt->dev,
2210 "btfetchtransinfo - Inquire Sync "
2211 "Info Failed 0x%x\n", error);
2212 return;
2213 }
2214 sync_period = sync_info.sync_rate[target] * 100;
2215 } else {
2216 sync_period = 2000 + (500 * sync_info.period);
2217 }
2218
2219 cts->protocol = PROTO_SCSI;
2220 cts->protocol_version = SCSI_REV_2;
2221 cts->transport = XPORT_SPI;
2222 cts->transport_version = 2;
2223
2224 spi->sync_period = sync_period;
2225 spi->valid |= CTS_SPI_VALID_SYNC_RATE;
2226 spi->sync_offset = sync_offset;
2227 spi->valid |= CTS_SPI_VALID_SYNC_OFFSET;
2228
2229 spi->valid |= CTS_SPI_VALID_BUS_WIDTH;
2230 spi->bus_width = bus_width;
2231
2232 if (cts->ccb_h.target_lun != CAM_LUN_WILDCARD) {
2233 scsi->valid = CTS_SCSI_VALID_TQ;
2234 spi->valid |= CTS_SPI_VALID_DISC;
2235 } else
2236 scsi->valid = 0;
2237
2238 xpt_async(AC_TRANSFER_NEG, cts->ccb_h.path, cts);
2239 }
2240
2241 static void
2242 btmapmboxes(void *arg, bus_dma_segment_t *segs, int nseg, int error)
2243 {
2244 struct bt_softc* bt;
2245
2246 bt = (struct bt_softc*)arg;
2247 bt->mailbox_physbase = segs->ds_addr;
2248 }
2249
2250 static void
2251 btmapccbs(void *arg, bus_dma_segment_t *segs, int nseg, int error)
2252 {
2253 struct bt_softc* bt;
2254
2255 bt = (struct bt_softc*)arg;
2256 bt->bt_ccb_physbase = segs->ds_addr;
2257 }
2258
2259 static void
2260 btmapsgs(void *arg, bus_dma_segment_t *segs, int nseg, int error)
2261 {
2262
2263 struct bt_softc* bt;
2264
2265 bt = (struct bt_softc*)arg;
2266 SLIST_FIRST(&bt->sg_maps)->sg_physaddr = segs->ds_addr;
2267 }
2268
2269 static void
2270 btpoll(struct cam_sim *sim)
2271 {
2272 bt_intr_locked(cam_sim_softc(sim));
2273 }
2274
2275 void
2276 bttimeout(void *arg)
2277 {
2278 struct bt_ccb *bccb;
2279 union ccb *ccb;
2280 struct bt_softc *bt;
2281
2282 bccb = (struct bt_ccb *)arg;
2283 ccb = bccb->ccb;
2284 bt = (struct bt_softc *)ccb->ccb_h.ccb_bt_ptr;
2285 mtx_assert(&bt->lock, MA_OWNED);
2286 xpt_print_path(ccb->ccb_h.path);
2287 printf("CCB %p - timed out\n", (void *)bccb);
2288
2289 if ((bccb->flags & BCCB_ACTIVE) == 0) {
2290 xpt_print_path(ccb->ccb_h.path);
2291 printf("CCB %p - timed out CCB already completed\n",
2292 (void *)bccb);
2293 return;
2294 }
2295
2296 /*
2297 * In order to simplify the recovery process, we ask the XPT
2298 * layer to halt the queue of new transactions and we traverse
2299 * the list of pending CCBs and remove their timeouts. This
2300 * means that the driver attempts to clear only one error
2301 * condition at a time. In general, timeouts that occur
2302 * close together are related anyway, so there is no benefit
2303 * in attempting to handle errors in parallel. Timeouts will
2304 * be reinstated when the recovery process ends.
2305 */
2306 if ((bccb->flags & BCCB_DEVICE_RESET) == 0) {
2307 struct ccb_hdr *ccb_h;
2308
2309 if ((bccb->flags & BCCB_RELEASE_SIMQ) == 0) {
2310 xpt_freeze_simq(bt->sim, /*count*/1);
2311 bccb->flags |= BCCB_RELEASE_SIMQ;
2312 }
2313
2314 ccb_h = LIST_FIRST(&bt->pending_ccbs);
2315 while (ccb_h != NULL) {
2316 struct bt_ccb *pending_bccb;
2317
2318 pending_bccb = (struct bt_ccb *)ccb_h->ccb_bccb_ptr;
2319 callout_stop(&pending_bccb->timer);
2320 ccb_h = LIST_NEXT(ccb_h, sim_links.le);
2321 }
2322 }
2323
2324 if ((bccb->flags & BCCB_DEVICE_RESET) != 0
2325 || bt->cur_outbox->action_code != BMBO_FREE
2326 || ((bccb->hccb.tag_enable == TRUE)
2327 && (bt->firmware_ver[0] < '5'))) {
2328 /*
2329 * Try a full host adapter/SCSI bus reset.
2330 * We do this only if we have already attempted
2331 * to clear the condition with a BDR, or we cannot
2332 * attempt a BDR for lack of mailbox resources
2333 * or because of faulty firmware. It turns out
2334 * that firmware versions prior to 5.xx treat BDRs
2335 * as untagged commands that cannot be sent until
2336 * all outstanding tagged commands have been processed.
2337 * This makes it somewhat difficult to use a BDR to
2338 * clear up a problem with an uncompleted tagged command.
2339 */
2340 ccb->ccb_h.status = CAM_CMD_TIMEOUT;
2341 btreset(bt, /*hardreset*/TRUE);
2342 device_printf(bt->dev, "No longer in timeout\n");
2343 } else {
2344 /*
2345 * Send a Bus Device Reset message:
2346 * The target that is holding up the bus may not
2347 * be the same as the one that triggered this timeout
2348 * (different commands have different timeout lengths),
2349 * but we have no way of determining this from our
2350 * timeout handler. Our strategy here is to queue a
2351 * BDR message to the target of the timed out command.
2352 * If this fails, we'll get another timeout 2 seconds
2353 * later which will attempt a bus reset.
2354 */
2355 bccb->flags |= BCCB_DEVICE_RESET;
2356 callout_reset(&bccb->timer, 2 * hz, bttimeout, bccb);
2357
2358 bt->recovery_bccb->hccb.opcode = INITIATOR_BUS_DEV_RESET;
2359
2360 /* No Data Transfer */
2361 bt->recovery_bccb->hccb.datain = TRUE;
2362 bt->recovery_bccb->hccb.dataout = TRUE;
2363 bt->recovery_bccb->hccb.btstat = 0;
2364 bt->recovery_bccb->hccb.sdstat = 0;
2365 bt->recovery_bccb->hccb.target_id = ccb->ccb_h.target_id;
2366
2367 /* Tell the adapter about this command */
2368 bt->cur_outbox->ccb_addr = btccbvtop(bt, bt->recovery_bccb);
2369 bt->cur_outbox->action_code = BMBO_START;
2370 bt_outb(bt, COMMAND_REG, BOP_START_MBOX);
2371 btnextoutbox(bt);
2372 }
2373 }
2374
2375 MODULE_VERSION(bt, 1);
2376 MODULE_DEPEND(bt, cam, 1, 1, 1);
Cache object: d68c4738166564e159ddbf766a6aacbc
|