1 /* $NetBSD: aic79xx_osm.c,v 1.36 2021/08/07 16:19:12 thorpej Exp $ */
2
3 /*
4 * Bus independent NetBSD shim for the aic7xxx based adaptec SCSI controllers
5 *
6 * Copyright (c) 1994-2002 Justin T. Gibbs.
7 * Copyright (c) 2001-2002 Adaptec Inc.
8 * All rights reserved.
9 *
10 * Redistribution and use in source and binary forms, with or without
11 * modification, are permitted provided that the following conditions
12 * are met:
13 * 1. Redistributions of source code must retain the above copyright
14 * notice, this list of conditions, and the following disclaimer,
15 * without modification.
16 * 2. The name of the author may not be used to endorse or promote products
17 * derived from this software without specific prior written permission.
18 *
19 * Alternatively, this software may be distributed under the terms of the
20 * GNU Public License ("GPL").
21 *
22 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
26 * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 * SUCH DAMAGE.
33 *
34 * //depot/aic7xxx/freebsd/dev/aic7xxx/aic79xx_osm.c#26 $
35 *
36 * $FreeBSD: src/sys/dev/aic7xxx/aic79xx_osm.c,v 1.11 2003/05/04 00:20:07 gibbs Exp $
37 */
38 /*
39 * Ported from FreeBSD by Pascal Renauld, Network Storage Solutions, Inc.
40 * - April 2003
41 */
42
43 #include <sys/cdefs.h>
44 __KERNEL_RCSID(0, "$NetBSD: aic79xx_osm.c,v 1.36 2021/08/07 16:19:12 thorpej Exp $");
45
46 #include <dev/ic/aic79xx_osm.h>
47 #include <dev/ic/aic79xx_inline.h>
48
49 #ifndef AHD_TMODE_ENABLE
50 #define AHD_TMODE_ENABLE 0
51 #endif
52
53 static int ahd_ioctl(struct scsipi_channel *channel, u_long cmd,
54 void *addr, int flag, struct proc *p);
55 static void ahd_action(struct scsipi_channel *chan,
56 scsipi_adapter_req_t req, void *arg);
57 static void ahd_execute_scb(void *arg, bus_dma_segment_t *dm_segs,
58 int nsegments);
59 static int ahd_poll(struct ahd_softc *ahd, int wait);
60 static void ahd_setup_data(struct ahd_softc *ahd, struct scsipi_xfer *xs,
61 struct scb *scb);
62
63 #if NOT_YET
64 static void ahd_set_recoveryscb(struct ahd_softc *ahd, struct scb *scb);
65 #endif
66
67 static bool ahd_pmf_suspend(device_t, const pmf_qual_t *);
68 static bool ahd_pmf_resume(device_t, const pmf_qual_t *);
69 static bool ahd_pmf_shutdown(device_t, int);
70
71 /*
72 * Attach all the sub-devices we can find
73 */
74 int
75 ahd_attach(struct ahd_softc *ahd)
76 {
77 int s;
78 char ahd_info[256];
79
80 ahd_controller_info(ahd, ahd_info, sizeof(ahd_info));
81 aprint_normal("%s: %s\n", ahd_name(ahd), ahd_info);
82
83 ahd_lock(ahd, &s);
84
85 ahd->sc_adapter.adapt_dev = ahd->sc_dev;
86 ahd->sc_adapter.adapt_nchannels = 1;
87
88 ahd->sc_adapter.adapt_openings = ahd->scb_data.numscbs - 1;
89 ahd->sc_adapter.adapt_max_periph = 32;
90
91 ahd->sc_adapter.adapt_ioctl = ahd_ioctl;
92 ahd->sc_adapter.adapt_minphys = ahd_minphys;
93 ahd->sc_adapter.adapt_request = ahd_action;
94
95 ahd->sc_channel.chan_adapter = &ahd->sc_adapter;
96 ahd->sc_channel.chan_bustype = &scsi_bustype;
97 ahd->sc_channel.chan_channel = 0;
98 ahd->sc_channel.chan_ntargets = AHD_NUM_TARGETS;
99 ahd->sc_channel.chan_nluns = 8 /*AHD_NUM_LUNS*/;
100 ahd->sc_channel.chan_id = ahd->our_id;
101 ahd->sc_channel.chan_flags |= SCSIPI_CHAN_CANGROW;
102
103 ahd->sc_child = config_found(ahd->sc_dev, &ahd->sc_channel, scsiprint,
104 CFARGS_NONE);
105
106 ahd_intr_enable(ahd, TRUE);
107
108 if (ahd->flags & AHD_RESET_BUS_A)
109 ahd_reset_channel(ahd, 'A', TRUE);
110
111 if (!pmf_device_register1(ahd->sc_dev,
112 ahd_pmf_suspend, ahd_pmf_resume, ahd_pmf_shutdown))
113 aprint_error_dev(ahd->sc_dev,
114 "couldn't establish power handler\n");
115
116 ahd_unlock(ahd, &s);
117
118 return (1);
119 }
120
121 static bool
122 ahd_pmf_suspend(device_t dev, const pmf_qual_t *qual)
123 {
124 struct ahd_softc *sc = device_private(dev);
125 #if 0
126 return (ahd_suspend(sc) == 0);
127 #else
128 ahd_shutdown(sc);
129 return true;
130 #endif
131 }
132
133 static bool
134 ahd_pmf_resume(device_t dev, const pmf_qual_t *qual)
135 {
136 #if 0
137 struct ahd_softc *sc = device_private(dev);
138
139 return (ahd_resume(sc) == 0);
140 #else
141 return true;
142 #endif
143 }
144
145 static bool
146 ahd_pmf_shutdown(device_t dev, int howto)
147 {
148 struct ahd_softc *sc = device_private(dev);
149
150 /* Disable all interrupt sources by resetting the controller */
151 ahd_shutdown(sc);
152
153 return true;
154 }
155
156 static int
157 ahd_ioctl(struct scsipi_channel *channel, u_long cmd,
158 void *addr, int flag, struct proc *p)
159 {
160 struct ahd_softc *ahd;
161 int s, ret = ENOTTY;
162
163 ahd = device_private(channel->chan_adapter->adapt_dev);
164
165 switch (cmd) {
166 case SCBUSIORESET:
167 s = splbio();
168 ahd_reset_channel(ahd, channel->chan_channel == 1 ? 'B' : 'A', TRUE);
169 splx(s);
170 ret = 0;
171 break;
172 default:
173 break;
174 }
175
176 return ret;
177 }
178
179 /*
180 * Catch an interrupt from the adapter
181 */
182 void
183 ahd_platform_intr(void *arg)
184 {
185 struct ahd_softc *ahd;
186
187 ahd = arg;
188
189 printf("%s; ahd_platform_intr\n", ahd_name(ahd));
190
191 ahd_intr(ahd);
192 }
193
194 /*
195 * We have an scb which has been processed by the
196 * adaptor, now we look to see how the operation * went.
197 */
198 void
199 ahd_done(struct ahd_softc *ahd, struct scb *scb)
200 {
201 struct scsipi_xfer *xs;
202 struct scsipi_periph *periph;
203 int s;
204
205 LIST_REMOVE(scb, pending_links);
206
207 xs = scb->xs;
208 periph = xs->xs_periph;
209
210 callout_stop(&scb->xs->xs_callout);
211
212 if (xs->datalen) {
213 int op;
214
215 if (xs->xs_control & XS_CTL_DATA_IN)
216 op = BUS_DMASYNC_POSTREAD;
217 else
218 op = BUS_DMASYNC_POSTWRITE;
219
220 bus_dmamap_sync(ahd->parent_dmat, scb->dmamap, 0,
221 scb->dmamap->dm_mapsize, op);
222 bus_dmamap_unload(ahd->parent_dmat, scb->dmamap);
223 }
224
225 /*
226 * If the recovery SCB completes, we have to be
227 * out of our timeout.
228 */
229 if ((scb->flags & SCB_RECOVERY_SCB) != 0) {
230 struct scb *list_scb;
231
232 /*
233 * We were able to complete the command successfully,
234 * so reinstate the timeouts for all other pending
235 * commands.
236 */
237 LIST_FOREACH(list_scb, &ahd->pending_scbs, pending_links) {
238 struct scsipi_xfer *txs = list_scb->xs;
239
240 if (!(txs->xs_control & XS_CTL_POLL)) {
241 callout_reset(&txs->xs_callout,
242 (txs->timeout > 1000000) ?
243 (txs->timeout / 1000) * hz :
244 (txs->timeout * hz) / 1000,
245 ahd_timeout, list_scb);
246 }
247 }
248
249 if (ahd_get_transaction_status(scb) != XS_NOERROR)
250 ahd_set_transaction_status(scb, XS_TIMEOUT);
251 scsipi_printaddr(xs->xs_periph);
252 printf("%s: no longer in timeout, status = %x\n",
253 ahd_name(ahd), xs->status);
254 }
255
256 if (xs->error != XS_NOERROR) {
257 /* Don't clobber any existing error state */
258 } else if ((xs->status == SCSI_STATUS_BUSY) ||
259 (xs->status == SCSI_STATUS_QUEUE_FULL)) {
260 ahd_set_transaction_status(scb, XS_BUSY);
261 printf("%s: drive (ID %d, LUN %d) queue full (SCB 0x%x)\n",
262 ahd_name(ahd), SCB_GET_TARGET(ahd,scb), SCB_GET_LUN(scb), SCB_GET_TAG(scb));
263 } else if ((scb->flags & SCB_SENSE) != 0) {
264 /*
265 * We performed autosense retrieval.
266 *
267 * zero the sense data before having
268 * the drive fill it. The SCSI spec mandates
269 * that any untransferred data should be
270 * assumed to be zero. Complete the 'bounce'
271 * of sense information through buffers accessible
272 * via bus-space by copying it into the clients
273 * csio.
274 */
275 memset(&xs->sense.scsi_sense, 0, sizeof(xs->sense.scsi_sense));
276 memcpy(&xs->sense.scsi_sense, ahd_get_sense_buf(ahd, scb),
277 sizeof(struct scsi_sense_data));
278
279 ahd_set_transaction_status(scb, XS_SENSE);
280 } else if ((scb->flags & SCB_PKT_SENSE) != 0) {
281 struct scsi_status_iu_header *siu;
282 u_int sense_len;
283 #ifdef AHD_DEBUG
284 int i;
285 #endif
286 /*
287 * Copy only the sense data into the provided buffer.
288 */
289 siu = (struct scsi_status_iu_header *)scb->sense_data;
290 sense_len = MIN(scsi_4btoul(siu->sense_length),
291 sizeof(xs->sense.scsi_sense));
292 memset(&xs->sense.scsi_sense, 0, sizeof(xs->sense.scsi_sense));
293 memcpy(&xs->sense.scsi_sense,
294 scb->sense_data + SIU_SENSE_OFFSET(siu), sense_len);
295 #ifdef AHD_DEBUG
296 printf("Copied %d bytes of sense data offset %d:", sense_len,
297 SIU_SENSE_OFFSET(siu));
298 for (i = 0; i < sense_len; i++)
299 printf(" 0x%x", ((uint8_t *)&xs->sense.scsi_sense)[i]);
300 printf("\n");
301 #endif
302 ahd_set_transaction_status(scb, XS_SENSE);
303 }
304
305 if (scb->flags & SCB_FREEZE_QUEUE) {
306 scsipi_periph_thaw(periph, 1);
307 scb->flags &= ~SCB_FREEZE_QUEUE;
308 }
309
310 if (scb->flags & SCB_REQUEUE)
311 ahd_set_transaction_status(scb, XS_REQUEUE);
312
313 ahd_lock(ahd, &s);
314 ahd_free_scb(ahd, scb);
315 ahd_unlock(ahd, &s);
316
317 scsipi_done(xs);
318 }
319
320 static void
321 ahd_action(struct scsipi_channel *chan, scsipi_adapter_req_t req, void *arg)
322 {
323 struct ahd_softc *ahd;
324 struct ahd_initiator_tinfo *tinfo;
325 struct ahd_tmode_tstate *tstate;
326
327 ahd = device_private(chan->chan_adapter->adapt_dev);
328
329 switch(req) {
330
331 case ADAPTER_REQ_RUN_XFER:
332 {
333 struct scsipi_xfer *xs;
334 struct scsipi_periph *periph;
335 struct scb *scb;
336 struct hardware_scb *hscb;
337 u_int target_id;
338 u_int our_id;
339 u_int col_idx;
340 char channel;
341 int s;
342
343 xs = arg;
344 periph = xs->xs_periph;
345
346 SC_DEBUG(periph, SCSIPI_DB3, ("ahd_action\n"));
347
348 target_id = periph->periph_target;
349 our_id = ahd->our_id;
350 channel = (chan->chan_channel == 1) ? 'B' : 'A';
351
352 /*
353 * get an scb to use.
354 */
355 ahd_lock(ahd, &s);
356 tinfo = ahd_fetch_transinfo(ahd, channel, our_id,
357 target_id, &tstate);
358
359 if (xs->xs_tag_type != 0 ||
360 (tinfo->curr.ppr_options & MSG_EXT_PPR_IU_REQ) != 0)
361 col_idx = AHD_NEVER_COL_IDX;
362 else
363 col_idx = AHD_BUILD_COL_IDX(target_id,
364 periph->periph_lun);
365
366 if ((scb = ahd_get_scb(ahd, col_idx)) == NULL) {
367 xs->error = XS_RESOURCE_SHORTAGE;
368 ahd_unlock(ahd, &s);
369 scsipi_done(xs);
370 return;
371 }
372 ahd_unlock(ahd, &s);
373
374 hscb = scb->hscb;
375
376 SC_DEBUG(periph, SCSIPI_DB3, ("start scb(%p)\n", scb));
377 scb->xs = xs;
378
379 /*
380 * Put all the arguments for the xfer in the scb
381 */
382 hscb->control = 0;
383 hscb->scsiid = BUILD_SCSIID(ahd, sim, target_id, our_id);
384 hscb->lun = periph->periph_lun;
385 if (xs->xs_control & XS_CTL_RESET) {
386 hscb->cdb_len = 0;
387 scb->flags |= SCB_DEVICE_RESET;
388 hscb->control |= MK_MESSAGE;
389 hscb->task_management = SIU_TASKMGMT_LUN_RESET;
390 ahd_execute_scb(scb, NULL, 0);
391 } else {
392 hscb->task_management = 0;
393 }
394
395 ahd_setup_data(ahd, xs, scb);
396 break;
397 }
398
399 case ADAPTER_REQ_GROW_RESOURCES:
400 #ifdef AHC_DEBUG
401 printf("%s: ADAPTER_REQ_GROW_RESOURCES\n", ahd_name(ahd));
402 #endif
403 chan->chan_adapter->adapt_openings += ahd_alloc_scbs(ahd);
404 if (ahd->scb_data.numscbs >= AHD_SCB_MAX_ALLOC)
405 chan->chan_flags &= ~SCSIPI_CHAN_CANGROW;
406 break;
407
408 case ADAPTER_REQ_SET_XFER_MODE:
409 {
410 struct scsipi_xfer_mode *xm = arg;
411 struct ahd_devinfo devinfo;
412 int target_id, our_id, first;
413 u_int width;
414 int s;
415 char channel;
416 u_int ppr_options = 0, period, offset;
417 uint16_t old_autoneg;
418
419 target_id = xm->xm_target;
420 our_id = chan->chan_id;
421 channel = 'A';
422 s = splbio();
423 tinfo = ahd_fetch_transinfo(ahd, channel, our_id, target_id,
424 &tstate);
425 ahd_compile_devinfo(&devinfo, our_id, target_id,
426 0, channel, ROLE_INITIATOR);
427
428 old_autoneg = tstate->auto_negotiate;
429
430 /*
431 * XXX since the period and offset are not provided here,
432 * fake things by forcing a renegotiation using the user
433 * settings if this is called for the first time (i.e.
434 * during probe). Also, cap various values at the user
435 * values, assuming that the user set it up that way.
436 */
437 if (ahd->inited_target[target_id] == 0) {
438 period = tinfo->user.period;
439 offset = tinfo->user.offset;
440 ppr_options = tinfo->user.ppr_options;
441 width = tinfo->user.width;
442 tstate->tagenable |=
443 (ahd->user_tagenable & devinfo.target_mask);
444 tstate->discenable |=
445 (ahd->user_discenable & devinfo.target_mask);
446 ahd->inited_target[target_id] = 1;
447 first = 1;
448 } else
449 first = 0;
450
451 if (xm->xm_mode & (PERIPH_CAP_WIDE16 | PERIPH_CAP_DT))
452 width = MSG_EXT_WDTR_BUS_16_BIT;
453 else
454 width = MSG_EXT_WDTR_BUS_8_BIT;
455
456 ahd_validate_width(ahd, NULL, &width, ROLE_UNKNOWN);
457 if (width > tinfo->user.width)
458 width = tinfo->user.width;
459 ahd_set_width(ahd, &devinfo, width, AHD_TRANS_GOAL, FALSE);
460
461 if (!(xm->xm_mode & (PERIPH_CAP_SYNC | PERIPH_CAP_DT))) {
462 period = 0;
463 offset = 0;
464 ppr_options = 0;
465 }
466
467 if ((xm->xm_mode & PERIPH_CAP_DT) &&
468 (tinfo->user.ppr_options & MSG_EXT_PPR_DT_REQ))
469 ppr_options |= MSG_EXT_PPR_DT_REQ;
470 else
471 ppr_options &= ~MSG_EXT_PPR_DT_REQ;
472
473 if ((tstate->discenable & devinfo.target_mask) == 0 ||
474 (tstate->tagenable & devinfo.target_mask) == 0)
475 ppr_options &= ~MSG_EXT_PPR_IU_REQ;
476
477 if ((xm->xm_mode & PERIPH_CAP_TQING) &&
478 (ahd->user_tagenable & devinfo.target_mask))
479 tstate->tagenable |= devinfo.target_mask;
480 else
481 tstate->tagenable &= ~devinfo.target_mask;
482
483 ahd_find_syncrate(ahd, &period, &ppr_options, AHD_SYNCRATE_MAX);
484 ahd_validate_offset(ahd, NULL, period, &offset,
485 MSG_EXT_WDTR_BUS_8_BIT, ROLE_UNKNOWN);
486 if (offset == 0) {
487 period = 0;
488 ppr_options = 0;
489 }
490 if (ppr_options != 0
491 && tinfo->user.transport_version >= 3) {
492 tinfo->goal.transport_version =
493 tinfo->user.transport_version;
494 tinfo->curr.transport_version =
495 tinfo->user.transport_version;
496 }
497
498 ahd_set_syncrate(ahd, &devinfo, period, offset,
499 ppr_options, AHD_TRANS_GOAL, FALSE);
500
501 /*
502 * If this is the first request, and no negotiation is
503 * needed, just confirm the state to the scsipi layer,
504 * so that it can print a message.
505 */
506 if (old_autoneg == tstate->auto_negotiate && first) {
507 xm->xm_mode = 0;
508 xm->xm_period = tinfo->curr.period;
509 xm->xm_offset = tinfo->curr.offset;
510 if (tinfo->curr.width == MSG_EXT_WDTR_BUS_16_BIT)
511 xm->xm_mode |= PERIPH_CAP_WIDE16;
512 if (tinfo->curr.period)
513 xm->xm_mode |= PERIPH_CAP_SYNC;
514 if (tstate->tagenable & devinfo.target_mask)
515 xm->xm_mode |= PERIPH_CAP_TQING;
516 if (tinfo->curr.ppr_options & MSG_EXT_PPR_DT_REQ)
517 xm->xm_mode |= PERIPH_CAP_DT;
518 scsipi_async_event(chan, ASYNC_EVENT_XFER_MODE, xm);
519 }
520 splx(s);
521 }
522 }
523
524 return;
525 }
526
527 static void
528 ahd_execute_scb(void *arg, bus_dma_segment_t *dm_segs, int nsegments)
529 {
530 struct scb *scb;
531 struct scsipi_xfer *xs;
532 struct ahd_softc *ahd;
533 struct ahd_initiator_tinfo *tinfo;
534 struct ahd_tmode_tstate *tstate;
535 u_int mask;
536 int s;
537
538 scb = arg;
539 xs = scb->xs;
540 xs->error = 0;
541 xs->status = 0;
542 xs->xs_status = 0;
543 ahd = device_private(
544 xs->xs_periph->periph_channel->chan_adapter->adapt_dev);
545
546 scb->sg_count = 0;
547 if (nsegments != 0) {
548 void *sg;
549 int op;
550 u_int i;
551
552 ahd_setup_data_scb(ahd, scb);
553
554 /* Copy the segments into our SG list */
555 for (i = nsegments, sg = scb->sg_list; i > 0; i--) {
556
557 sg = ahd_sg_setup(ahd, scb, sg, dm_segs->ds_addr,
558 dm_segs->ds_len,
559 /*last*/i == 1);
560 dm_segs++;
561 }
562
563 if (xs->xs_control & XS_CTL_DATA_IN)
564 op = BUS_DMASYNC_PREREAD;
565 else
566 op = BUS_DMASYNC_PREWRITE;
567
568 bus_dmamap_sync(ahd->parent_dmat, scb->dmamap, 0,
569 scb->dmamap->dm_mapsize, op);
570 }
571
572 ahd_lock(ahd, &s);
573
574 /*
575 * Last time we need to check if this SCB needs to
576 * be aborted.
577 */
578 if (ahd_get_scsi_status(scb) == XS_STS_DONE) {
579 if (nsegments != 0)
580 bus_dmamap_unload(ahd->parent_dmat,
581 scb->dmamap);
582 ahd_free_scb(ahd, scb);
583 ahd_unlock(ahd, &s);
584 return;
585 }
586
587 tinfo = ahd_fetch_transinfo(ahd, SCSIID_CHANNEL(ahd, scb->hscb->scsiid),
588 SCSIID_OUR_ID(scb->hscb->scsiid),
589 SCSIID_TARGET(ahd, scb->hscb->scsiid),
590 &tstate);
591
592 mask = SCB_GET_TARGET_MASK(ahd, scb);
593
594 if ((tstate->discenable & mask) != 0)
595 scb->hscb->control |= DISCENB;
596
597 if ((tstate->tagenable & mask) != 0)
598 scb->hscb->control |= xs->xs_tag_type|TAG_ENB;
599
600 if ((tinfo->curr.ppr_options & MSG_EXT_PPR_IU) != 0) {
601 scb->flags |= SCB_PACKETIZED;
602 if (scb->hscb->task_management != 0)
603 scb->hscb->control &= ~MK_MESSAGE;
604 }
605
606 #if 0 /* This looks like it makes sense at first, but it can loop */
607 if ((xs->xs_control & XS_CTL_DISCOVERY) &&
608 (tinfo->goal.width != 0
609 || tinfo->goal.period != 0
610 || tinfo->goal.ppr_options != 0)) {
611 scb->flags |= SCB_NEGOTIATE;
612 scb->hscb->control |= MK_MESSAGE;
613 } else
614 #endif
615 if ((tstate->auto_negotiate & mask) != 0) {
616 scb->flags |= SCB_AUTO_NEGOTIATE;
617 scb->hscb->control |= MK_MESSAGE;
618 }
619
620 LIST_INSERT_HEAD(&ahd->pending_scbs, scb, pending_links);
621
622 scb->flags |= SCB_ACTIVE;
623
624 if (!(xs->xs_control & XS_CTL_POLL)) {
625 callout_reset(&scb->xs->xs_callout, xs->timeout > 1000000 ?
626 (xs->timeout / 1000) * hz : (xs->timeout * hz) / 1000,
627 ahd_timeout, scb);
628 }
629
630 if ((scb->flags & SCB_TARGET_IMMEDIATE) != 0) {
631 /* Define a mapping from our tag to the SCB. */
632 ahd->scb_data.scbindex[SCB_GET_TAG(scb)] = scb;
633 ahd_pause(ahd);
634 ahd_set_scbptr(ahd, SCB_GET_TAG(scb));
635 ahd_outb(ahd, RETURN_1, CONT_MSG_LOOP_TARG);
636 ahd_unpause(ahd);
637 } else {
638 ahd_queue_scb(ahd, scb);
639 }
640
641 if (!(xs->xs_control & XS_CTL_POLL)) {
642 ahd_unlock(ahd, &s);
643 return;
644 }
645 /*
646 * If we can't use interrupts, poll for completion
647 */
648 SC_DEBUG(xs->xs_periph, SCSIPI_DB3, ("cmd_poll\n"));
649 do {
650 if (ahd_poll(ahd, xs->timeout)) {
651 if (!(xs->xs_control & XS_CTL_SILENT))
652 printf("cmd fail\n");
653 ahd_timeout(scb);
654 break;
655 }
656 } while (!(xs->xs_status & XS_STS_DONE));
657
658 ahd_unlock(ahd, &s);
659 }
660
661 static int
662 ahd_poll(struct ahd_softc *ahd, int wait)
663 {
664
665 while (--wait) {
666 DELAY(1000);
667 if (ahd_inb(ahd, INTSTAT) & INT_PEND)
668 break;
669 }
670
671 if (wait == 0) {
672 printf("%s: board is not responding\n", ahd_name(ahd));
673 return (EIO);
674 }
675
676 ahd_intr(ahd);
677 return (0);
678 }
679
680
681 static void
682 ahd_setup_data(struct ahd_softc *ahd, struct scsipi_xfer *xs,
683 struct scb *scb)
684 {
685 struct hardware_scb *hscb;
686
687 hscb = scb->hscb;
688 xs->resid = xs->status = 0;
689
690 hscb->cdb_len = xs->cmdlen;
691 if (hscb->cdb_len > MAX_CDB_LEN) {
692 int s;
693 /*
694 * Should CAM start to support CDB sizes
695 * greater than 16 bytes, we could use
696 * the sense buffer to store the CDB.
697 */
698 ahd_set_transaction_status(scb,
699 XS_DRIVER_STUFFUP);
700
701 ahd_lock(ahd, &s);
702 ahd_free_scb(ahd, scb);
703 ahd_unlock(ahd, &s);
704 scsipi_done(xs);
705 }
706 memcpy(hscb->shared_data.idata.cdb, xs->cmd, hscb->cdb_len);
707
708 /* Only use S/G if there is a transfer */
709 if (xs->datalen) {
710 int error;
711
712 error = bus_dmamap_load(ahd->parent_dmat,
713 scb->dmamap, xs->data,
714 xs->datalen, NULL,
715 ((xs->xs_control & XS_CTL_NOSLEEP) ?
716 BUS_DMA_NOWAIT : BUS_DMA_WAITOK) |
717 BUS_DMA_STREAMING |
718 ((xs->xs_control & XS_CTL_DATA_IN) ?
719 BUS_DMA_READ : BUS_DMA_WRITE));
720 if (error) {
721 #ifdef AHD_DEBUG
722 printf("%s: in ahd_setup_data(): bus_dmamap_load() "
723 "= %d\n",
724 ahd_name(ahd), error);
725 #endif
726 xs->error = XS_RESOURCE_SHORTAGE;
727 scsipi_done(xs);
728 return;
729 }
730 ahd_execute_scb(scb,
731 scb->dmamap->dm_segs,
732 scb->dmamap->dm_nsegs);
733 } else {
734 ahd_execute_scb(scb, NULL, 0);
735 }
736 }
737
738 void
739 ahd_timeout(void *arg)
740 {
741 struct scb *scb;
742 struct ahd_softc *ahd;
743 int s;
744
745 scb = arg;
746 ahd = scb->ahd_softc;
747
748 printf("%s: ahd_timeout\n", ahd_name(ahd));
749
750 ahd_lock(ahd, &s);
751
752 ahd_pause_and_flushwork(ahd);
753 (void)ahd_save_modes(ahd);
754 #if 0
755 ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI);
756 ahd_outb(ahd, SCSISIGO, ACKO);
757 printf("set ACK\n");
758 ahd_outb(ahd, SCSISIGO, 0);
759 printf("clearing Ack\n");
760 ahd_restore_modes(ahd, saved_modes);
761 #endif
762 if ((scb->flags & SCB_ACTIVE) == 0) {
763 /* Previous timeout took care of me already */
764 printf("%s: Timedout SCB already complete. "
765 "Interrupts may not be functioning.\n", ahd_name(ahd));
766 ahd_unpause(ahd);
767 ahd_unlock(ahd, &s);
768 return;
769 }
770
771 ahd_print_path(ahd, scb);
772 printf("SCB 0x%x - timed out\n", SCB_GET_TAG(scb));
773 ahd_dump_card_state(ahd);
774 ahd_reset_channel(ahd, SIM_CHANNEL(ahd, sim),
775 /*initiate reset*/TRUE);
776 ahd_unlock(ahd, &s);
777 return;
778 }
779
780 int
781 ahd_platform_alloc(struct ahd_softc *ahd, void *platform_arg)
782 {
783 ahd->platform_data = malloc(sizeof(struct ahd_platform_data), M_DEVBUF,
784 M_WAITOK | M_ZERO);
785 return (0);
786 }
787
788 void
789 ahd_platform_free(struct ahd_softc *ahd)
790 {
791 free(ahd->platform_data, M_DEVBUF);
792 }
793
794 int
795 ahd_softc_comp(struct ahd_softc *lahd, struct ahd_softc *rahd)
796 {
797 /* We don't sort softcs under NetBSD so report equal always */
798 return (0);
799 }
800
801 int
802 ahd_detach(struct ahd_softc *ahd, int flags)
803 {
804 int rv = 0;
805
806 if (ahd->sc_child != NULL)
807 rv = config_detach(ahd->sc_child, flags);
808
809 pmf_device_deregister(ahd->sc_dev);
810
811 ahd_free(ahd);
812
813 return rv;
814 }
815
816 void
817 ahd_platform_set_tags(struct ahd_softc *ahd,
818 struct ahd_devinfo *devinfo, ahd_queue_alg alg)
819 {
820 struct ahd_tmode_tstate *tstate;
821
822 ahd_fetch_transinfo(ahd, devinfo->channel, devinfo->our_scsiid,
823 devinfo->target, &tstate);
824
825 if (alg != AHD_QUEUE_NONE)
826 tstate->tagenable |= devinfo->target_mask;
827 else
828 tstate->tagenable &= ~devinfo->target_mask;
829 }
830
831 void
832 ahd_send_async(struct ahd_softc *ahd, char channel, u_int target, u_int lun,
833 ac_code code, void *opt_arg)
834 {
835 struct ahd_tmode_tstate *tstate;
836 struct ahd_initiator_tinfo *tinfo;
837 struct ahd_devinfo devinfo;
838 struct scsipi_channel *chan;
839 struct scsipi_xfer_mode xm;
840
841 #ifdef DIAGNOSTIC
842 if (channel != 'A')
843 panic("ahd_send_async: not channel A");
844 #endif
845 chan = &ahd->sc_channel;
846 switch (code) {
847 case AC_TRANSFER_NEG:
848 tinfo = ahd_fetch_transinfo(ahd, channel, ahd->our_id, target,
849 &tstate);
850 ahd_compile_devinfo(&devinfo, ahd->our_id, target, lun,
851 channel, ROLE_UNKNOWN);
852 /*
853 * Don't bother if negotiating. XXX?
854 */
855 if (tinfo->curr.period != tinfo->goal.period
856 || tinfo->curr.width != tinfo->goal.width
857 || tinfo->curr.offset != tinfo->goal.offset
858 || tinfo->curr.ppr_options != tinfo->goal.ppr_options)
859 break;
860 xm.xm_target = target;
861 xm.xm_mode = 0;
862 xm.xm_period = tinfo->curr.period;
863 xm.xm_offset = tinfo->curr.offset;
864 if (tinfo->goal.ppr_options & MSG_EXT_PPR_DT_REQ)
865 xm.xm_mode |= PERIPH_CAP_DT;
866 if (tinfo->curr.width == MSG_EXT_WDTR_BUS_16_BIT)
867 xm.xm_mode |= PERIPH_CAP_WIDE16;
868 if (tinfo->curr.period)
869 xm.xm_mode |= PERIPH_CAP_SYNC;
870 if (tstate->tagenable & devinfo.target_mask)
871 xm.xm_mode |= PERIPH_CAP_TQING;
872 scsipi_async_event(chan, ASYNC_EVENT_XFER_MODE, &xm);
873 break;
874 case AC_BUS_RESET:
875 scsipi_async_event(chan, ASYNC_EVENT_RESET, NULL);
876 case AC_SENT_BDR:
877 default:
878 break;
879 }
880 }
Cache object: 16caf32cc88cebf5c1c2cc1fc502c6f5
|