/* $NetBSD: xy.c,v 1.82 2021/08/07 16:19:06 thorpej Exp $ */ /* * Copyright (c) 1995 Charles D. Cranor * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ /* * * x y . c x y l o g i c s 4 5 0 / 4 5 1 s m d d r i v e r * * author: Chuck Cranor * id: &Id: xy.c,v 1.1 1995/09/25 20:35:14 chuck Exp & * started: 14-Sep-95 * references: [1] Xylogics Model 753 User's Manual * part number: 166-753-001, Revision B, May 21, 1988. * "Your Partner For Performance" * [2] other NetBSD disk device drivers * [3] Xylogics Model 450 User's Manual * part number: 166-017-001, Revision B, 1983. * [4] Addendum to Xylogics Model 450 Disk Controller User's * Manual, Jan. 1985. * [5] The 451 Controller, Rev. B3, September 2, 1986. * [6] David Jones 's unfinished 450/451 driver * */ #include __KERNEL_RCSID(0, "$NetBSD: xy.c,v 1.82 2021/08/07 16:19:06 thorpej Exp $"); #undef XYC_DEBUG /* full debug */ #undef XYC_DIAG /* extra sanity checks */ #if defined(DIAGNOSTIC) && !defined(XYC_DIAG) #define XYC_DIAG /* link in with master DIAG option */ #endif #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "ioconf.h" #include "locators.h" /* * Print a complaint when no xy children were specified * in the config file. Better than a link error... * * XXX: Some folks say this driver should be split in two, * but that seems pointless with ONLY one type of child. */ #include "xy.h" #if NXY == 0 #error "xyc but no xy?" #endif /* * macros */ /* * XYC_GO: start iopb ADDR (DVMA addr in a u_long) on XYC */ #define XYC_GO(XYC, ADDR) \ do { \ (XYC)->xyc_addr_lo = ((ADDR) & 0xff); \ (ADDR) = ((ADDR) >> 8); \ (XYC)->xyc_addr_hi = ((ADDR) & 0xff); \ (ADDR) = ((ADDR) >> 8); \ (XYC)->xyc_reloc_lo = ((ADDR) & 0xff); \ (ADDR) = ((ADDR) >> 8); \ (XYC)->xyc_reloc_hi = (ADDR); \ (XYC)->xyc_csr = XYC_GBSY; /* go! */ \ } while (/* CONSTCOND */ 0) /* * XYC_DONE: don't need IORQ, get error code and free (done after xyc_cmd) */ #define XYC_DONE(SC,ER) \ do { \ if ((ER) == XY_ERR_AOK) { \ (ER) = (SC)->ciorq->errno; \ (SC)->ciorq->mode = XY_SUB_FREE; \ wakeup((SC)->ciorq); \ } \ } while (/* CONSTCOND */ 0) /* * XYC_ADVANCE: advance iorq's pointers by a number of sectors */ #define XYC_ADVANCE(IORQ, N) \ do { \ if (N) { \ (IORQ)->sectcnt -= (N); \ (IORQ)->blockno += (N); \ (IORQ)->dbuf += ((N) * XYFM_BPS); \ } \ } while (/* CONSTCOND */ 0) /* * note - addresses you can sleep on: * [1] & of xy_softc's "state" (waiting for a chance to attach a drive) * [2] & an iorq (waiting for an XY_SUB_WAIT iorq to finish) */ /* * function prototypes * "xyc_*" functions are internal, all others are external interfaces */ /* internals */ struct xy_iopb *xyc_chain(struct xyc_softc *, struct xy_iorq *); int xyc_cmd(struct xyc_softc *, int, int, int, int, int, char *, int); const char *xyc_e2str(int); int xyc_entoact(int); int xyc_error(struct xyc_softc *, struct xy_iorq *, struct xy_iopb *, int); int xyc_ioctlcmd(struct xy_softc *, dev_t dev, struct xd_iocmd *); void xyc_perror(struct xy_iorq *, struct xy_iopb *, int); int xyc_piodriver(struct xyc_softc *, struct xy_iorq *); int xyc_remove_iorq(struct xyc_softc *); int xyc_reset(struct xyc_softc *, int, struct xy_iorq *, int, struct xy_softc *); inline void xyc_rqinit(struct xy_iorq *, struct xyc_softc *, struct xy_softc *, int, u_long, int, void *, struct buf *); void xyc_rqtopb(struct xy_iorq *, struct xy_iopb *, int, int); void xyc_start(struct xyc_softc *, struct xy_iorq *); int xyc_startbuf(struct xyc_softc *, struct xy_softc *, struct buf *); int xyc_submit_iorq(struct xyc_softc *, struct xy_iorq *, int); void xyc_tick(void *); int xyc_unbusy(struct xyc *, int); void xyc_xyreset(struct xyc_softc *, struct xy_softc *); /* machine interrupt hook */ int xycintr(void *); /* autoconf */ static int xycmatch(device_t, cfdata_t, void *); static void xycattach(device_t, device_t, void *); static int xyc_print(void *, const char *); static int xymatch(device_t, cfdata_t, void *); static void xyattach(device_t, device_t, void *); static void xy_init(struct xy_softc *); static void xydummystrat(struct buf *); int xygetdisklabel(struct xy_softc *, void *); /* * cfattach's: device driver interface to autoconfig */ CFATTACH_DECL_NEW(xyc, sizeof(struct xyc_softc), xycmatch, xycattach, NULL, NULL); CFATTACH_DECL_NEW(xy, sizeof(struct xy_softc), xymatch, xyattach, NULL, NULL); struct xyc_attach_args { /* this is the "aux" args to xyattach */ int driveno; /* unit number */ }; dev_type_open(xyopen); dev_type_close(xyclose); dev_type_read(xyread); dev_type_write(xywrite); dev_type_ioctl(xyioctl); dev_type_strategy(xystrategy); dev_type_dump(xydump); dev_type_size(xysize); const struct bdevsw xy_bdevsw = { .d_open = xyopen, .d_close = xyclose, .d_strategy = xystrategy, .d_ioctl = xyioctl, .d_dump = xydump, .d_psize = xysize, .d_discard = nodiscard, .d_flag = D_DISK }; const struct cdevsw xy_cdevsw = { .d_open = xyopen, .d_close = xyclose, .d_read = xyread, .d_write = xywrite, .d_ioctl = xyioctl, .d_stop = nostop, .d_tty = notty, .d_poll = nopoll, .d_mmap = nommap, .d_kqfilter = nokqfilter, .d_discard = nodiscard, .d_flag = D_DISK }; /* * dkdriver */ struct dkdriver xydkdriver = { .d_strategy = xystrategy }; /* * start: disk label fix code (XXX) */ static void *xy_labeldata; static void xydummystrat(struct buf *bp) { if (bp->b_bcount != XYFM_BPS) panic("%s: b_bcount", __func__); memcpy(bp->b_data, xy_labeldata, XYFM_BPS); bp->b_oflags |= BO_DONE; bp->b_cflags &= ~BC_BUSY; } int xygetdisklabel(struct xy_softc *xy, void *b) { const char *err; struct sun_disklabel *sdl; /* We already have the label data in `b'; setup for dummy strategy */ xy_labeldata = b; /* Required parameter for readdisklabel() */ xy->sc_dk.dk_label->d_secsize = XYFM_BPS; err = readdisklabel(MAKEDISKDEV(0, device_unit(xy->sc_dev), RAW_PART), xydummystrat, xy->sc_dk.dk_label, xy->sc_dk.dk_cpulabel); if (err) { printf("%s: %s\n", device_xname(xy->sc_dev), err); return XY_ERR_FAIL; } /* Ok, we have the label; fill in `pcyl' if there's SunOS magic */ sdl = (struct sun_disklabel *)xy->sc_dk.dk_cpulabel->cd_block; if (sdl->sl_magic == SUN_DKMAGIC) xy->pcyl = sdl->sl_pcyl; else { printf("%s: WARNING: no `pcyl' in disk label.\n", device_xname(xy->sc_dev)); xy->pcyl = xy->sc_dk.dk_label->d_ncylinders + xy->sc_dk.dk_label->d_acylinders; printf("%s: WARNING: guessing pcyl=%d (ncyl+acyl)\n", device_xname(xy->sc_dev), xy->pcyl); } xy->ncyl = xy->sc_dk.dk_label->d_ncylinders; xy->acyl = xy->sc_dk.dk_label->d_acylinders; xy->nhead = xy->sc_dk.dk_label->d_ntracks; xy->nsect = xy->sc_dk.dk_label->d_nsectors; xy->sectpercyl = xy->nhead * xy->nsect; xy->sc_dk.dk_label->d_secsize = XYFM_BPS; /* not handled by * sun->bsd */ return XY_ERR_AOK; } /* * end: disk label fix code (XXX) */ /* * a u t o c o n f i g f u n c t i o n s */ /* * xycmatch: determine if xyc is present or not. we do a * soft reset to detect the xyc. */ static int xycmatch(device_t parent, cfdata_t cf, void *aux) { struct confargs *ca = aux; /* No default VME address. */ if (ca->ca_paddr == -1) return 0; /* Make sure something is there... */ if (bus_peek(ca->ca_bustype, ca->ca_paddr + 5, 1) == -1) return 0; /* Default interrupt priority. */ if (ca->ca_intpri == -1) ca->ca_intpri = 2; return 1; } /* * xycattach: attach controller */ static void xycattach(device_t parent, device_t self, void *aux) { struct xyc_softc *xyc = device_private(self); struct confargs *ca = aux; struct xyc_attach_args xa; int lcv, err, res, pbsz; void *tmp, *tmp2; u_long ultmp; /* get addressing and intr level stuff from autoconfig and load it * into our xyc_softc. */ xyc->sc_dev = self; xyc->xyc = (struct xyc *)bus_mapin(ca->ca_bustype, ca->ca_paddr, sizeof(struct xyc)); xyc->bustype = ca->ca_bustype; xyc->ipl = ca->ca_intpri; xyc->vector = ca->ca_intvec; xyc->no_ols = 0; /* XXX should be from config */ for (lcv = 0; lcv < XYC_MAXDEV; lcv++) xyc->sc_drives[lcv] = NULL; /* * allocate and zero buffers * check boundaries of the KVA's ... all IOPBs must reside in * the same 64K region. */ pbsz = XYC_MAXIOPB * sizeof(struct xy_iopb); tmp = tmp2 = (struct xy_iopb *)dvma_malloc(pbsz); /* KVA */ ultmp = (u_long)tmp; if ((ultmp & 0xffff0000) != ((ultmp + pbsz) & 0xffff0000)) { tmp = (struct xy_iopb *)dvma_malloc(pbsz); /* retry! */ dvma_free(tmp2, pbsz); ultmp = (u_long) tmp; if ((ultmp & 0xffff0000) != ((ultmp + pbsz) & 0xffff0000)) { aprint_error(": can't alloc IOPB mem in 64K\n"); return; } } memset(tmp, 0, pbsz); xyc->iopbase = tmp; xyc->dvmaiopb = (struct xy_iopb *)dvma_kvtopa(xyc->iopbase, xyc->bustype); xyc->reqs = kmem_zalloc(XYC_MAXIOPB * sizeof(struct xy_iorq), KM_SLEEP); /* * init iorq to iopb pointers, and non-zero fields in the * iopb which never change. */ for (lcv = 0; lcv < XYC_MAXIOPB; lcv++) { xyc->xy_chain[lcv] = NULL; xyc->reqs[lcv].iopb = &xyc->iopbase[lcv]; xyc->iopbase[lcv].asr = 1; /* always the same */ xyc->iopbase[lcv].eef = 1; /* always the same */ xyc->iopbase[lcv].ecm = XY_ECM; /* always the same */ xyc->iopbase[lcv].aud = 1; /* always the same */ xyc->iopbase[lcv].relo = 1; /* always the same */ xyc->iopbase[lcv].thro = XY_THRO;/* always the same */ } xyc->ciorq = &xyc->reqs[XYC_CTLIOPB]; /* short hand name */ xyc->ciopb = &xyc->iopbase[XYC_CTLIOPB]; /* short hand name */ xyc->xy_hand = 0; /* read controller parameters and insure we have a 450/451 */ err = xyc_cmd(xyc, XYCMD_ST, 0, 0, 0, 0, 0, XY_SUB_POLL); res = xyc->ciopb->ctyp; XYC_DONE(xyc, err); if (res != XYCT_450) { if (err) aprint_error(": %s: ", xyc_e2str(err)); aprint_error(": doesn't identify as a 450/451\n"); return; } aprint_normal(": Xylogics 450/451"); if (xyc->no_ols) /* 450 doesn't overlap seek right */ aprint_normal(" [OLS disabled]"); aprint_normal("\n"); if (err) { aprint_error_dev(self, "error: %s\n", xyc_e2str(err)); return; } if ((xyc->xyc->xyc_csr & XYC_ADRM) == 0) { aprint_error_dev(self, "24 bit addressing turned off\n"); printf("please set hardware jumpers JM1-JM2=in, JM3-JM4=out\n"); printf("to enable 24 bit mode and this driver\n"); return; } /* link in interrupt with higher level software */ isr_add_vectored(xycintr, xyc, ca->ca_intpri, ca->ca_intvec); evcnt_attach_dynamic(&xyc->sc_intrcnt, EVCNT_TYPE_INTR, NULL, device_xname(self), "intr"); callout_init(&xyc->sc_tick_ch, 0); /* now we must look for disks using autoconfig */ for (xa.driveno = 0; xa.driveno < XYC_MAXDEV; xa.driveno++) (void)config_found(self, (void *)&xa, xyc_print, CFARGS_NONE); /* start the watchdog clock */ callout_reset(&xyc->sc_tick_ch, XYC_TICKCNT, xyc_tick, xyc); } static int xyc_print(void *aux, const char *name) { struct xyc_attach_args *xa = aux; if (name != NULL) aprint_normal("%s: ", name); if (xa->driveno != -1) aprint_normal(" drive %d", xa->driveno); return UNCONF; } /* * xymatch: probe for disk. * * note: we almost always say disk is present. this allows us to * spin up and configure a disk after the system is booted (we can * call xyattach!). Also, wire down the relationship between the * xy* and xyc* devices, to simplify boot device identification. */ static int xymatch(device_t parent, cfdata_t cf, void *aux) { struct xyc_attach_args *xa = aux; int xy_unit; /* Match only on the "wired-down" controller+disk. */ xy_unit = device_unit(parent) * 2 + xa->driveno; if (cf->cf_unit != xy_unit) return 0; return 1; } /* * xyattach: attach a disk. */ static void xyattach(device_t parent, device_t self, void *aux) { struct xy_softc *xy = device_private(self); struct xyc_softc *xyc = device_private(parent); struct xyc_attach_args *xa = aux; xy->sc_dev = self; aprint_normal("\n"); /* * Always re-initialize the disk structure. We want statistics * to start with a clean slate. */ memset(&xy->sc_dk, 0, sizeof(xy->sc_dk)); disk_init(&xy->sc_dk, device_xname(self), &xydkdriver); xy->state = XY_DRIVE_UNKNOWN; /* to start */ xy->flags = 0; xy->parent = xyc; /* init queue of waiting bufs */ bufq_alloc(&xy->xyq, "disksort", BUFQ_SORT_RAWBLOCK); xy->xyrq = &xyc->reqs[xa->driveno]; xy->xy_drive = xa->driveno; xyc->sc_drives[xa->driveno] = xy; /* Do init work common to attach and open. */ xy_init(xy); } /* * end of autoconfig functions */ /* * Initialize a disk. This can be called from both autoconf and * also from xyopen/xystrategy. */ static void xy_init(struct xy_softc *xy) { struct xyc_softc *xyc; struct dkbad *dkb; void *dvmabuf; int err, spt, mb, blk, lcv, fullmode, newstate; xyc = xy->parent; xy->state = XY_DRIVE_ATTACHING; newstate = XY_DRIVE_UNKNOWN; fullmode = (cold) ? XY_SUB_POLL : XY_SUB_WAIT; dvmabuf = dvma_malloc(XYFM_BPS); /* first try and reset the drive */ err = xyc_cmd(xyc, XYCMD_RST, 0, xy->xy_drive, 0, 0, 0, fullmode); XYC_DONE(xyc, err); if (err == XY_ERR_DNRY) { printf("%s: drive %d: off-line\n", device_xname(xy->sc_dev), xy->xy_drive); goto done; } if (err) { printf("%s: ERROR 0x%02x (%s)\n", device_xname(xy->sc_dev), err, xyc_e2str(err)); goto done; } printf("%s: drive %d ready", device_xname(xy->sc_dev), xy->xy_drive); /* * now set drive parameters (to semi-bogus values) so we can read the * disk label. */ xy->pcyl = xy->ncyl = 1; xy->acyl = 0; xy->nhead = 1; xy->nsect = 1; xy->sectpercyl = 1; for (lcv = 0; lcv < 126; lcv++) /* init empty bad144 table */ xy->dkb.bt_bad[lcv].bt_cyl = xy->dkb.bt_bad[lcv].bt_trksec = 0xffff; /* read disk label */ for (xy->drive_type = 0; xy->drive_type <= XYC_MAXDT; xy->drive_type++) { err = xyc_cmd(xyc, XYCMD_RD, 0, xy->xy_drive, 0, 1, dvmabuf, fullmode); XYC_DONE(xyc, err); if (err == XY_ERR_AOK) break; } if (err != XY_ERR_AOK) { printf("%s: reading disk label failed: %s\n", device_xname(xy->sc_dev), xyc_e2str(err)); goto done; } printf("%s: drive type %d\n", device_xname(xy->sc_dev), xy->drive_type); newstate = XY_DRIVE_NOLABEL; xy->hw_spt = spt = 0; /* XXX needed ? */ /* Attach the disk: must be before getdisklabel to malloc label */ disk_attach(&xy->sc_dk); if (xygetdisklabel(xy, dvmabuf) != XY_ERR_AOK) goto done; /* inform the user of what is up */ printf("%s: <%s>, pcyl %d\n", device_xname(xy->sc_dev), (char *)dvmabuf, xy->pcyl); mb = xy->ncyl * (xy->nhead * xy->nsect) / (1048576 / XYFM_BPS); printf("%s: %dMB, %d cyl, %d head, %d sec\n", device_xname(xy->sc_dev), mb, xy->ncyl, xy->nhead, xy->nsect); /* * 450/451 stupidity: the drive type is encoded into the format * of the disk. the drive type in the IOPB must match the drive * type in the format, or you will not be able to do I/O to the * disk (you get header not found errors). if you have two drives * of different sizes that have the same drive type in their * formatting then you are out of luck. * * this problem was corrected in the 753/7053. */ for (lcv = 0 ; lcv < XYC_MAXDEV ; lcv++) { struct xy_softc *oxy; oxy = xyc->sc_drives[lcv]; if (oxy == NULL || oxy == xy) continue; if (oxy->drive_type != xy->drive_type) continue; if (xy->nsect != oxy->nsect || xy->pcyl != oxy->pcyl || xy->nhead != oxy->nhead) { printf("%s: %s and %s must be the same size!\n", device_xname(xyc->sc_dev), device_xname(xy->sc_dev), device_xname(oxy->sc_dev)); panic("xy drive size mismatch"); } } /* now set the real drive parameters! */ blk = (xy->nsect - 1) + ((xy->nhead - 1) * xy->nsect) + ((xy->pcyl - 1) * xy->nsect * xy->nhead); err = xyc_cmd(xyc, XYCMD_SDS, 0, xy->xy_drive, blk, 0, 0, fullmode); XYC_DONE(xyc, err); if (err) { printf("%s: write drive size failed: %s\n", device_xname(xy->sc_dev), xyc_e2str(err)); goto done; } newstate = XY_DRIVE_ONLINE; /* * read bad144 table. this table resides on the first sector of the * last track of the disk (i.e. second cyl of "acyl" area). */ blk = (xy->ncyl + xy->acyl - 1) * (xy->nhead * xy->nsect) + /* last cyl */ (xy->nhead - 1) * xy->nsect; /* last head */ err = xyc_cmd(xyc, XYCMD_RD, 0, xy->xy_drive, blk, 1, dvmabuf, fullmode); XYC_DONE(xyc, err); if (err) { printf("%s: reading bad144 failed: %s\n", device_xname(xy->sc_dev), xyc_e2str(err)); goto done; } /* check dkbad for sanity */ dkb = (struct dkbad *)dvmabuf; for (lcv = 0; lcv < 126; lcv++) { if ((dkb->bt_bad[lcv].bt_cyl == 0xffff || dkb->bt_bad[lcv].bt_cyl == 0) && dkb->bt_bad[lcv].bt_trksec == 0xffff) continue; /* blank */ if (dkb->bt_bad[lcv].bt_cyl >= xy->ncyl) break; if ((dkb->bt_bad[lcv].bt_trksec >> 8) >= xy->nhead) break; if ((dkb->bt_bad[lcv].bt_trksec & 0xff) >= xy->nsect) break; } if (lcv != 126) { printf("%s: warning: invalid bad144 sector!\n", device_xname(xy->sc_dev)); } else { memcpy(&xy->dkb, dvmabuf, XYFM_BPS); } done: xy->state = newstate; dvma_free(dvmabuf, XYFM_BPS); } /* * { b , c } d e v s w f u n c t i o n s */ /* * xyclose: close device */ int xyclose(dev_t dev, int flag, int fmt, struct lwp *l) { struct xy_softc *xy = device_lookup_private(&xy_cd, DISKUNIT(dev)); int part = DISKPART(dev); /* clear mask bits */ switch (fmt) { case S_IFCHR: xy->sc_dk.dk_copenmask &= ~(1 << part); break; case S_IFBLK: xy->sc_dk.dk_bopenmask &= ~(1 << part); break; } xy->sc_dk.dk_openmask = xy->sc_dk.dk_copenmask | xy->sc_dk.dk_bopenmask; return 0; } /* * xydump: crash dump system */ int xydump(dev_t dev, daddr_t blkno, void *va, size_t sz) { int unit, part; struct xy_softc *xy; unit = DISKUNIT(dev); part = DISKPART(dev); xy = device_lookup_private(&xy_cd, unit); if (xy == NULL) return ENXIO; printf("%s%c: crash dump not supported (yet)\n", device_xname(xy->sc_dev), 'a' + part); return ENXIO; /* outline: globals: "dumplo" == sector number of partition to start * dump at (convert to physical sector with partition table) * "dumpsize" == size of dump in clicks "physmem" == size of physical * memory (clicks, ctob() to get bytes) (normal case: dumpsize == * physmem) * * dump a copy of physical memory to the dump device starting at sector * "dumplo" in the swap partition (make sure > 0). map in pages as * we go. use polled I/O. * * XXX how to handle NON_CONTIG? */ } static enum kauth_device_req xy_getkauthreq(u_char cmd) { enum kauth_device_req req; switch (cmd) { case XYCMD_WR: case XYCMD_WTH: case XYCMD_WFM: case XYCMD_WRH: req = KAUTH_REQ_DEVICE_RAWIO_PASSTHRU_WRITE; break; case XYCMD_RD: case XYCMD_RTH: case XYCMD_RDH: req = KAUTH_REQ_DEVICE_RAWIO_PASSTHRU_READ; break; case XYCMD_RDS: case XYCMD_MBD: req = KAUTH_REQ_DEVICE_RAWIO_PASSTHRU_READCONF; break; case XYCMD_RST: case XYCMD_SDS: case XYCMD_MBL: req = KAUTH_REQ_DEVICE_RAWIO_PASSTHRU_WRITECONF; break; case XYCMD_NOP: case XYCMD_SK: case XYCMD_ST: case XYCMD_R: default: req = 0; break; } return req; } /* * xyioctl: ioctls on XY drives. based on ioctl's of other netbsd disks. */ int xyioctl(dev_t dev, u_long cmd, void *addr, int flag, struct lwp *l) { struct xy_softc *xy; struct xd_iocmd *xio; int error, s, unit; unit = DISKUNIT(dev); xy = device_lookup_private(&xy_cd, unit); if (xy == NULL) return ENXIO; error = disk_ioctl(&xy->sc_dk, dev, cmd, addr, flag, l); if (error != EPASSTHROUGH) return error; /* switch on ioctl type */ switch (cmd) { case DIOCSBAD: /* set bad144 info */ if ((flag & FWRITE) == 0) return EBADF; s = splbio(); memcpy(&xy->dkb, addr, sizeof(xy->dkb)); splx(s); return 0; case DIOCSDINFO: /* set disk label */ if ((flag & FWRITE) == 0) return EBADF; error = setdisklabel(xy->sc_dk.dk_label, (struct disklabel *)addr, /* xy->sc_dk.dk_openmask : */ 0, xy->sc_dk.dk_cpulabel); if (error == 0) { if (xy->state == XY_DRIVE_NOLABEL) xy->state = XY_DRIVE_ONLINE; } return error; case DIOCWLABEL: /* change write status of disk label */ if ((flag & FWRITE) == 0) return EBADF; if (*(int *)addr) xy->flags |= XY_WLABEL; else xy->flags &= ~XY_WLABEL; return 0; case DIOCWDINFO: /* write disk label */ if ((flag & FWRITE) == 0) return EBADF; error = setdisklabel(xy->sc_dk.dk_label, (struct disklabel *)addr, /* xy->sc_dk.dk_openmask : */ 0, xy->sc_dk.dk_cpulabel); if (error == 0) { if (xy->state == XY_DRIVE_NOLABEL) xy->state = XY_DRIVE_ONLINE; /* Simulate opening partition 0 so write succeeds. */ xy->sc_dk.dk_openmask |= (1 << 0); error = writedisklabel(MAKEDISKDEV(major(dev), DISKUNIT(dev), RAW_PART), xystrategy, xy->sc_dk.dk_label, xy->sc_dk.dk_cpulabel); xy->sc_dk.dk_openmask = xy->sc_dk.dk_copenmask | xy->sc_dk.dk_bopenmask; } return error; case DIOSXDCMD: { enum kauth_device_req req; xio = (struct xd_iocmd *)addr; req = xy_getkauthreq(xio->cmd); if ((error = kauth_authorize_device_passthru(l->l_cred, dev, req, xio)) != 0) return error; return xyc_ioctlcmd(xy, dev, xio); } default: return ENOTTY; } } /* * xyopen: open drive */ int xyopen(dev_t dev, int flag, int fmt, struct lwp *l) { int err, unit, part, s; struct xy_softc *xy; /* first, could it be a valid target? */ unit = DISKUNIT(dev); xy = device_lookup_private(&xy_cd, unit); if (xy == NULL) return ENXIO; part = DISKPART(dev); err = 0; /* * If some other processing is doing init, sleep. */ s = splbio(); while (xy->state == XY_DRIVE_ATTACHING) { if (tsleep(&xy->state, PRIBIO, "xyopen", 0)) { err = EINTR; goto done; } } /* Do we need to init the drive? */ if (xy->state == XY_DRIVE_UNKNOWN) { xy_init(xy); wakeup(&xy->state); } /* Was the init successful? */ if (xy->state == XY_DRIVE_UNKNOWN) { err = EIO; goto done; } /* check for partition */ if (part != RAW_PART && (part >= xy->sc_dk.dk_label->d_npartitions || xy->sc_dk.dk_label->d_partitions[part].p_fstype == FS_UNUSED)) { err = ENXIO; goto done; } /* set open masks */ switch (fmt) { case S_IFCHR: xy->sc_dk.dk_copenmask |= (1 << part); break; case S_IFBLK: xy->sc_dk.dk_bopenmask |= (1 << part); break; } xy->sc_dk.dk_openmask = xy->sc_dk.dk_copenmask | xy->sc_dk.dk_bopenmask; done: splx(s); return err; } int xyread(dev_t dev, struct uio *uio, int flags) { return physio(xystrategy, NULL, dev, B_READ, minphys, uio); } int xywrite(dev_t dev, struct uio *uio, int flags) { return physio(xystrategy, NULL, dev, B_WRITE, minphys, uio); } /* * xysize: return size of a partition for a dump */ int xysize(dev_t dev) { struct xy_softc *xysc; int unit, part, size, omask; /* valid unit? */ unit = DISKUNIT(dev); xysc = device_lookup_private(&xy_cd, unit); if (xysc == NULL) return -1; part = DISKPART(dev); omask = xysc->sc_dk.dk_openmask & (1 << part); if (omask == 0 && xyopen(dev, 0, S_IFBLK, NULL) != 0) return -1; /* do it */ if (xysc->sc_dk.dk_label->d_partitions[part].p_fstype != FS_SWAP) size = -1; /* only give valid size for swap partitions */ else size = xysc->sc_dk.dk_label->d_partitions[part].p_size * (xysc->sc_dk.dk_label->d_secsize / DEV_BSIZE); if (omask == 0 && xyclose(dev, 0, S_IFBLK, NULL) != 0) return -1; return size; } /* * xystrategy: buffering system interface to xy. */ void xystrategy(struct buf *bp) { struct xy_softc *xy; int s, unit; struct disklabel *lp; daddr_t blkno; unit = DISKUNIT(bp->b_dev); /* check for live device */ xy = device_lookup_private(&xy_cd, unit); if (xy == NULL || bp->b_blkno < 0 || (bp->b_bcount % xy->sc_dk.dk_label->d_secsize) != 0) { bp->b_error = EINVAL; goto done; } /* There should always be an open first. */ if (xy->state == XY_DRIVE_UNKNOWN) { bp->b_error = EIO; goto done; } if (xy->state != XY_DRIVE_ONLINE && DISKPART(bp->b_dev) != RAW_PART) { /* no I/O to unlabeled disks, unless raw partition */ bp->b_error = EIO; goto done; } /* short circuit zero length request */ if (bp->b_bcount == 0) goto done; /* check bounds with label (disksubr.c). Determine the size of the * transfer, and make sure it is within the boundaries of the * partition. Adjust transfer if needed, and signal errors or early * completion. */ lp = xy->sc_dk.dk_label; if (bounds_check_with_label(&xy->sc_dk, bp, (xy->flags & XY_WLABEL) != 0) <= 0) goto done; /* * Now convert the block number to absolute and put it in * terms of the device's logical block size. */ blkno = bp->b_blkno / (lp->d_secsize / DEV_BSIZE); if (DISKPART(bp->b_dev) != RAW_PART) blkno += lp->d_partitions[DISKPART(bp->b_dev)].p_offset; bp->b_rawblkno = blkno; /* * now we know we have a valid buf structure that we need to do I/O * on. */ s = splbio(); /* protect the queues */ bufq_put(xy->xyq, bp); /* XXX disksort_cylinder */ /* start 'em up */ xyc_start(xy->parent, NULL); /* done! */ splx(s); return; done: /* tells upper layers we are done with this buf */ bp->b_resid = bp->b_bcount; biodone(bp); } /* * end of {b,c}devsw functions */ /* * i n t e r r u p t f u n c t i o n * * xycintr: hardware interrupt. */ int xycintr(void *v) { struct xyc_softc *xycsc = v; /* kick the event counter */ xycsc->sc_intrcnt.ev_count++; /* remove as many done IOPBs as possible */ xyc_remove_iorq(xycsc); /* start any iorq's already waiting */ xyc_start(xycsc, NULL); return 1; } /* * end of interrupt function */ /* * i n t e r n a l f u n c t i o n s */ /* * xyc_rqinit: fill out the fields of an I/O request */ inline void xyc_rqinit(struct xy_iorq *rq, struct xyc_softc *xyc, struct xy_softc *xy, int md, u_long blk, int cnt, void *db, struct buf *bp) { rq->xyc = xyc; rq->xy = xy; rq->ttl = XYC_MAXTTL + 10; rq->mode = md; rq->tries = rq->errno = rq->lasterror = 0; rq->blockno = blk; rq->sectcnt = cnt; rq->dbuf = rq->dbufbase = db; rq->buf = bp; } /* * xyc_rqtopb: load up an IOPB based on an iorq */ void xyc_rqtopb(struct xy_iorq *iorq, struct xy_iopb *iopb, int cmd, int subfun) { u_long block, dp; /* normal IOPB case, standard stuff */ /* chain bit handled later */ iopb->ien = (XY_STATE(iorq->mode) == XY_SUB_POLL) ? 0 : 1; iopb->com = cmd; iopb->errno = 0; iopb->errs = 0; iopb->done = 0; if (iorq->xy) { iopb->unit = iorq->xy->xy_drive; iopb->dt = iorq->xy->drive_type; } else { iopb->unit = 0; iopb->dt = 0; } block = iorq->blockno; if (iorq->xy == NULL || block == 0) { iopb->sect = iopb->head = iopb->cyl = 0; } else { iopb->sect = block % iorq->xy->nsect; block = block / iorq->xy->nsect; iopb->head = block % iorq->xy->nhead; block = block / iorq->xy->nhead; iopb->cyl = block; } iopb->scnt = iorq->sectcnt; if (iorq->dbuf == NULL) { iopb->dataa = 0; iopb->datar = 0; } else { dp = dvma_kvtopa(iorq->dbuf, iorq->xyc->bustype); iopb->dataa = (dp & 0xffff); iopb->datar = ((dp & 0xff0000) >> 16); } iopb->subfn = subfun; } /* * xyc_unbusy: wait for the xyc to go unbusy, or timeout. */ int xyc_unbusy(struct xyc *xyc, int del) { while (del-- > 0) { if ((xyc->xyc_csr & XYC_GBSY) == 0) break; DELAY(1); } return del == 0 ? XY_ERR_FAIL : XY_ERR_AOK; } /* * xyc_cmd: front end for POLL'd and WAIT'd commands. Returns 0 or error. * note that NORM requests are handled separately. */ int xyc_cmd(struct xyc_softc *xycsc, int cmd, int subfn, int unit, int block, int scnt, char *dptr, int fullmode) { struct xy_iorq *iorq = xycsc->ciorq; struct xy_iopb *iopb = xycsc->ciopb; int submode = XY_STATE(fullmode); /* * is someone else using the control iopq wait for it if we can */ start: if (submode == XY_SUB_WAIT && XY_STATE(iorq->mode) != XY_SUB_FREE) { if (tsleep(iorq, PRIBIO, "xyc_cmd", 0)) return XY_ERR_FAIL; goto start; } if (XY_STATE(iorq->mode) != XY_SUB_FREE) { DELAY(1000000); /* XY_SUB_POLL: steal the iorq */ iorq->mode = XY_SUB_FREE; printf("%s: stole control iopb\n", device_xname(xycsc->sc_dev)); } /* init iorq/iopb */ xyc_rqinit(iorq, xycsc, (unit == XYC_NOUNIT) ? NULL : xycsc->sc_drives[unit], fullmode, block, scnt, dptr, NULL); /* load IOPB from iorq */ xyc_rqtopb(iorq, iopb, cmd, subfn); /* submit it for processing */ xyc_submit_iorq(xycsc, iorq, fullmode); /* error code will be in iorq */ return XY_ERR_AOK; } /* * xyc_startbuf * start a buffer for running */ int xyc_startbuf(struct xyc_softc *xycsc, struct xy_softc *xysc, struct buf *bp) { struct xy_iorq *iorq; struct xy_iopb *iopb; u_long block; void *dbuf; iorq = xysc->xyrq; iopb = iorq->iopb; /* get buf */ if (bp == NULL) panic("%s null buf", __func__); #ifdef XYC_DEBUG int partno = DISKPART(bp->b_dev); printf("%s: %s%c: %s block %d\n", __func__, device_xname(xysc->sc_dev), 'a' + partno, (bp->b_flags & B_READ) ? "read" : "write", (int)bp->b_blkno); printf("xyc_startbuf: b_bcount %d, b_data 0x%x\n", bp->b_bcount, bp->b_data); #endif /* * load request. * * also, note that there are two kinds of buf structures, those with * B_PHYS set and those without B_PHYS. if B_PHYS is set, then it is * a raw I/O (to a cdevsw) and we are doing I/O directly to the users' * buffer which has already been mapped into DVMA space. (Not on sun3) * However, if B_PHYS is not set, then the buffer is a normal system * buffer which does *not* live in DVMA space. In that case we call * dvma_mapin to map it into DVMA space so we can do the DMA to it. * * in cases where we do a dvma_mapin, note that iorq points to the * buffer as mapped into DVMA space, where as the bp->b_data points * to its non-DVMA mapping. * * XXX - On the sun3, B_PHYS does NOT mean the buffer is mapped * into dvma space, only that it was remapped into the kernel. * We ALWAYS have to remap the kernel buf into DVMA space. * (It is done inexpensively, using whole segments!) */ block = bp->b_rawblkno; dbuf = dvma_mapin(bp->b_data, bp->b_bcount, 0); if (dbuf == NULL) { /* out of DVMA space */ printf("%s: warning: out of DVMA space\n", device_xname(xycsc->sc_dev)); return XY_ERR_FAIL; /* XXX: need some sort of * call-back scheme here? */ } /* init iorq and load iopb from it */ xyc_rqinit(iorq, xycsc, xysc, XY_SUB_NORM | XY_MODE_VERBO, block, bp->b_bcount / XYFM_BPS, dbuf, bp); xyc_rqtopb(iorq, iopb, (bp->b_flags & B_READ) ? XYCMD_RD : XYCMD_WR, 0); /* Instrumentation. */ disk_busy(&xysc->sc_dk); return XY_ERR_AOK; } /* * xyc_submit_iorq: submit an iorq for processing. returns XY_ERR_AOK * if ok. if it fail returns an error code. type is XY_SUB_*. * * note: caller frees iorq in all cases except NORM * * return value: * NORM: XY_AOK (req pending), XY_FAIL (couldn't submit request) * WAIT: XY_AOK (success), (failed) * POLL: * NOQ : * * there are three sources for i/o requests: * [1] xystrategy: normal block I/O, using "struct buf" system. * [2] autoconfig/crash dump: these are polled I/O requests, no interrupts. * [3] open/ioctl: these are I/O requests done in the context of a process, * and the process should block until they are done. * * software state is stored in the iorq structure. each iorq has an * iopb structure. the hardware understands the iopb structure. * every command must go through an iopb. a 450 handles one iopb at a * time, where as a 451 can take them in chains. [the 450 claims it * can handle chains, but is appears to be buggy...] iopb are allocated * in DVMA space at boot up time. each disk gets one iopb, and the * controller gets one (for POLL and WAIT commands). what happens if * the iopb is busy? for i/o type [1], the buffers are queued at the * "buff" layer and * picked up later by the interrupt routine. for case * [2] we can only be blocked if there is a WAIT type I/O request being * run. since this can only happen when we are crashing, we wait a sec * and then steal the IOPB. for case [3] the process can sleep * on the iorq free list until some iopbs are available. */ int xyc_submit_iorq(struct xyc_softc *xycsc, struct xy_iorq *iorq, int type) { struct xy_iopb *iopb; u_long iopbaddr; #ifdef XYC_DEBUG printf("%s(%s, addr=0x%x, type=%d)\n", __func__, device_xname(xycsc->sc_dev), iorq, type); #endif /* first check and see if controller is busy */ if ((xycsc->xyc->xyc_csr & XYC_GBSY) != 0) { #ifdef XYC_DEBUG printf("%s: XYC not ready (BUSY)\n", __func__); #endif if (type == XY_SUB_NOQ) return XY_ERR_FAIL; /* failed */ switch (type) { case XY_SUB_NORM: return XY_ERR_AOK; /* success */ case XY_SUB_WAIT: while (iorq->iopb->done == 0) { (void)tsleep(iorq, PRIBIO, "xyciorq", 0); } return (iorq->errno); case XY_SUB_POLL: /* steal controller */ iopbaddr = xycsc->xyc->xyc_rsetup; /* RESET */ if (xyc_unbusy(xycsc->xyc, XYC_RESETUSEC) == XY_ERR_FAIL) panic("%s: stuck xyc", __func__); printf("%s: stole controller\n", device_xname(xycsc->sc_dev)); break; default: panic("%s adding", __func__); } } iopb = xyc_chain(xycsc, iorq); /* build chain */ if (iopb == NULL) { /* nothing doing? */ if (type == XY_SUB_NORM || type == XY_SUB_NOQ) return XY_ERR_AOK; panic("xyc_submit_iorq: xyc_chain failed!"); } iopbaddr = dvma_kvtopa(iopb, xycsc->bustype); XYC_GO(xycsc->xyc, iopbaddr); /* command now running, wrap it up */ switch (type) { case XY_SUB_NORM: case XY_SUB_NOQ: return XY_ERR_AOK; /* success */ case XY_SUB_WAIT: while (iorq->iopb->done == 0) { (void)tsleep(iorq, PRIBIO, "xyciorq", 0); } return iorq->errno; case XY_SUB_POLL: return xyc_piodriver(xycsc, iorq); default: panic("%s wrap up", __func__); } panic("%s impossible", __func__); return 0; /* not reached */ } /* * xyc_chain: build a chain. return dvma address of first element in * the chain. iorq != NULL: means we only want that item on the chain. */ struct xy_iopb * xyc_chain(struct xyc_softc *xycsc, struct xy_iorq *iorq) { int togo, chain, hand; struct xy_iopb *iopb, *prev_iopb; memset(xycsc->xy_chain, 0, sizeof(xycsc->xy_chain)); /* * promote control IOPB to the top */ if (iorq == NULL) { if ((XY_STATE(xycsc->reqs[XYC_CTLIOPB].mode) == XY_SUB_POLL || XY_STATE(xycsc->reqs[XYC_CTLIOPB].mode) == XY_SUB_WAIT) && xycsc->iopbase[XYC_CTLIOPB].done == 0) iorq = &xycsc->reqs[XYC_CTLIOPB]; } /* * special case: if iorq != NULL then we have a POLL or WAIT request. * we let these take priority and do them first. */ if (iorq) { xycsc->xy_chain[0] = iorq; iorq->iopb->chen = 0; return iorq->iopb; } /* * NORM case: do round robin and maybe chain (if allowed and possible) */ chain = 0; hand = xycsc->xy_hand; xycsc->xy_hand = (xycsc->xy_hand + 1) % XYC_MAXIOPB; for (togo = XYC_MAXIOPB ; togo > 0 ; togo--, hand = (hand + 1) % XYC_MAXIOPB) { if (XY_STATE(xycsc->reqs[hand].mode) != XY_SUB_NORM || xycsc->iopbase[hand].done) continue; /* not ready-for-i/o */ xycsc->xy_chain[chain] = &xycsc->reqs[hand]; iopb = xycsc->xy_chain[chain]->iopb; iopb->chen = 0; if (chain != 0) { /* adding a link to a chain? */ prev_iopb = xycsc->xy_chain[chain-1]->iopb; prev_iopb->chen = 1; prev_iopb->nxtiopb = 0xffff & dvma_kvtopa(iopb, xycsc->bustype); } else { /* head of chain */ iorq = xycsc->xy_chain[chain]; } chain++; if (xycsc->no_ols) break; /* quit if chaining dis-allowed */ } return iorq ? iorq->iopb : NULL; } /* * xyc_piodriver * * programmed i/o driver. this function takes over the computer * and drains off the polled i/o request. it returns the status of the iorq * the caller is interesting in. */ int xyc_piodriver(struct xyc_softc *xycsc, struct xy_iorq *iorq) { int nreset = 0; int retval = 0; u_long res; #ifdef XYC_DEBUG printf("%s(%s, 0x%x)\n", __func__, device_xname(xycsc->sc_dev), iorq); #endif while (iorq->iopb->done == 0) { res = xyc_unbusy(xycsc->xyc, XYC_MAXTIME); /* we expect some progress soon */ if (res == XY_ERR_FAIL && nreset >= 2) { xyc_reset(xycsc, 0, XY_RSET_ALL, XY_ERR_FAIL, 0); #ifdef XYC_DEBUG printf("%s: timeout\n", __func__); #endif return XY_ERR_FAIL; } if (res == XY_ERR_FAIL) { if (xyc_reset(xycsc, 0, (nreset++ == 0) ? XY_RSET_NONE : iorq, XY_ERR_FAIL, 0) == XY_ERR_FAIL) return XY_ERR_FAIL; /* flushes all but POLL * requests, resets */ continue; } xyc_remove_iorq(xycsc); /* may resubmit request */ if (iorq->iopb->done == 0) xyc_start(xycsc, iorq); } /* get return value */ retval = iorq->errno; #ifdef XYC_DEBUG printf("%s: done, retval = 0x%x (%s)\n", __func__, iorq->errno, xyc_e2str(iorq->errno)); #endif /* start up any bufs that have queued */ xyc_start(xycsc, NULL); return retval; } /* * xyc_xyreset: reset one drive. NOTE: assumes xyc was just reset. * we steal iopb[XYC_CTLIOPB] for this, but we put it back when we are done. */ void xyc_xyreset(struct xyc_softc *xycsc, struct xy_softc *xysc) { struct xy_iopb tmpiopb; u_long addr; int del; memcpy(&tmpiopb, xycsc->ciopb, sizeof(tmpiopb)); xycsc->ciopb->chen = xycsc->ciopb->done = xycsc->ciopb->errs = 0; xycsc->ciopb->ien = 0; xycsc->ciopb->com = XYCMD_RST; xycsc->ciopb->unit = xysc->xy_drive; addr = dvma_kvtopa(xycsc->ciopb, xycsc->bustype); XYC_GO(xycsc->xyc, addr); del = XYC_RESETUSEC; while (del > 0) { if ((xycsc->xyc->xyc_csr & XYC_GBSY) == 0) break; DELAY(1); del--; } if (del <= 0 || xycsc->ciopb->errs) { printf("%s: off-line: %s\n", device_xname(xycsc->sc_dev), xyc_e2str(xycsc->ciopb->errno)); del = xycsc->xyc->xyc_rsetup; if (xyc_unbusy(xycsc->xyc, XYC_RESETUSEC) == XY_ERR_FAIL) panic("%s", __func__); } else { xycsc->xyc->xyc_csr = XYC_IPND; /* clear IPND */ } memcpy(xycsc->ciopb, &tmpiopb, sizeof(tmpiopb)); } /* * xyc_reset: reset everything: requests are marked as errors except * a polled request (which is resubmitted) */ int xyc_reset(struct xyc_softc *xycsc, int quiet, struct xy_iorq *blastmode, int error, struct xy_softc *xysc) { int del = 0, lcv, retval = XY_ERR_AOK; struct xy_iorq *iorq; /* soft reset hardware */ if (quiet == 0) printf("%s: soft reset\n", device_xname(xycsc->sc_dev)); del = xycsc->xyc->xyc_rsetup; del = xyc_unbusy(xycsc->xyc, XYC_RESETUSEC); if (del == XY_ERR_FAIL) { blastmode = XY_RSET_ALL; /* dead, flush all requests */ retval = XY_ERR_FAIL; } if (xysc) xyc_xyreset(xycsc, xysc); /* fix queues based on "blast-mode" */ for (lcv = 0; lcv < XYC_MAXIOPB; lcv++) { iorq = &xycsc->reqs[lcv]; if (XY_STATE(iorq->mode) != XY_SUB_POLL && XY_STATE(iorq->mode) != XY_SUB_WAIT && XY_STATE(iorq->mode) != XY_SUB_NORM) /* is it active? */ continue; if (blastmode == XY_RSET_ALL || blastmode != iorq) { /* failed */ iorq->errno = error; xycsc->iopbase[lcv].done = xycsc->iopbase[lcv].errs = 1; switch (XY_STATE(iorq->mode)) { case XY_SUB_NORM: iorq->buf->b_error = EIO; iorq->buf->b_resid = iorq->sectcnt * XYFM_BPS; /* Sun3: map/unmap regardless of B_PHYS */ dvma_mapout(iorq->dbufbase, iorq->buf->b_bcount); (void)bufq_get(iorq->xy->xyq); disk_unbusy(&iorq->xy->sc_dk, (iorq->buf->b_bcount - iorq->buf->b_resid), (iorq->buf->b_flags & B_READ)); biodone(iorq->buf); iorq->mode = XY_SUB_FREE; break; case XY_SUB_WAIT: wakeup(iorq); case XY_SUB_POLL: iorq->mode = XY_NEWSTATE(iorq->mode, XY_SUB_DONE); break; } } else { /* resubmit, no need to do anything here */ } } /* * now, if stuff is waiting, start it. * since we just reset it should go */ xyc_start(xycsc, NULL); return retval; } /* * xyc_start: start waiting buffers */ void xyc_start(struct xyc_softc *xycsc, struct xy_iorq *iorq) { int lcv; struct xy_softc *xy; if (iorq == NULL) { for (lcv = 0; lcv < XYC_MAXDEV ; lcv++) { if ((xy = xycsc->sc_drives[lcv]) == NULL) continue; if (bufq_peek(xy->xyq) == NULL) continue; if (xy->xyrq->mode != XY_SUB_FREE) continue; xyc_startbuf(xycsc, xy, bufq_peek(xy->xyq)); } } xyc_submit_iorq(xycsc, iorq, XY_SUB_NOQ); } /* * xyc_remove_iorq: remove "done" IOPB's. */ int xyc_remove_iorq(struct xyc_softc *xycsc) { int errno, rq, comm, errs; struct xyc *xyc = xycsc->xyc; u_long addr; struct xy_iopb *iopb; struct xy_iorq *iorq; struct buf *bp; if (xyc->xyc_csr & XYC_DERR) { /* * DOUBLE ERROR: should never happen under normal use. This * error is so bad, you can't even tell which IOPB is bad, so * we dump them all. */ errno = XY_ERR_DERR; printf("%s: DOUBLE ERROR!\n", device_xname(xycsc->sc_dev)); if (xyc_reset(xycsc, 0, XY_RSET_ALL, errno, 0) != XY_ERR_AOK) { printf("%s: soft reset failed!\n", device_xname(xycsc->sc_dev)); panic("%s: controller DEAD", __func__); } return XY_ERR_AOK; } /* * get iopb that is done, loop down the chain */ if (xyc->xyc_csr & XYC_ERR) { xyc->xyc_csr = XYC_ERR; /* clear error condition */ } if (xyc->xyc_csr & XYC_IPND) { xyc->xyc_csr = XYC_IPND; /* clear interrupt */ } for (rq = 0; rq < XYC_MAXIOPB; rq++) { iorq = xycsc->xy_chain[rq]; if (iorq == NULL) break; /* done ! */ if (iorq->mode == 0 || XY_STATE(iorq->mode) == XY_SUB_DONE) continue; /* free, or done */ iopb = iorq->iopb; if (iopb->done == 0) continue; /* not done yet */ comm = iopb->com; errs = iopb->errs; if (errs) iorq->errno = iopb->errno; else iorq->errno = 0; /* handle non-fatal errors */ if (errs && xyc_error(xycsc, iorq, iopb, comm) == XY_ERR_AOK) continue; /* AOK: we resubmitted it */ /* this iorq is now done (hasn't been restarted or anything) */ if ((iorq->mode & XY_MODE_VERBO) && iorq->lasterror) xyc_perror(iorq, iopb, 0); /* now, if read/write check to make sure we got all the data * we needed. (this may not be the case if we got an error in * the middle of a multisector request). */ if ((iorq->mode & XY_MODE_B144) != 0 && errs == 0 && (comm == XYCMD_RD || comm == XYCMD_WR)) { /* we just successfully processed a bad144 sector * note: if we are in bad 144 mode, the pointers have * been advanced already (see above) and are pointing * at the bad144 sector. to exit bad144 mode, we * must advance the pointers 1 sector and issue a new * request if there are still sectors left to process * */ XYC_ADVANCE(iorq, 1); /* advance 1 sector */ /* exit b144 mode */ iorq->mode = iorq->mode & (~XY_MODE_B144); if (iorq->sectcnt) { /* more to go! */ iorq->lasterror = iorq->errno = iopb->errno = 0; iopb->errs = iopb->done = 0; iorq->tries = 0; iopb->scnt = iorq->sectcnt; iopb->cyl = iorq->blockno / iorq->xy->sectpercyl; iopb->head = (iorq->blockno / iorq->xy->nhead) % iorq->xy->nhead; iopb->sect = iorq->blockno % XYFM_BPS; addr = dvma_kvtopa(iorq->dbuf, xycsc->bustype); iopb->dataa = (addr & 0xffff); iopb->datar = ((addr & 0xff0000) >> 16); /* will resubit at end */ continue; } } /* final cleanup, totally done with this request */ switch (XY_STATE(iorq->mode)) { case XY_SUB_NORM: bp = iorq->buf; if (errs) { bp->b_error = EIO; bp->b_resid = iorq->sectcnt * XYFM_BPS; } else { bp->b_resid = 0; /* done */ } /* Sun3: map/unmap regardless of B_PHYS */ dvma_mapout(iorq->dbufbase, iorq->buf->b_bcount); (void)bufq_get(iorq->xy->xyq); disk_unbusy(&iorq->xy->sc_dk, (bp->b_bcount - bp->b_resid), (bp->b_flags & B_READ)); iorq->mode = XY_SUB_FREE; biodone(bp); break; case XY_SUB_WAIT: iorq->mode = XY_NEWSTATE(iorq->mode, XY_SUB_DONE); wakeup(iorq); break; case XY_SUB_POLL: iorq->mode = XY_NEWSTATE(iorq->mode, XY_SUB_DONE); break; } } return XY_ERR_AOK; } /* * xyc_perror: print error. * - if still_trying is true: we got an error, retried and got a * different error. in that case lasterror is the old error, * and errno is the new one. * - if still_trying is not true, then if we ever had an error it * is in lasterror. also, if iorq->errno == 0, then we recovered * from that error (otherwise iorq->errno == iorq->lasterror). */ void xyc_perror(struct xy_iorq *iorq, struct xy_iopb *iopb, int still_trying) { int error = iorq->lasterror; printf("%s", (iorq->xy) ? device_xname(iorq->xy->sc_dev) : device_xname(iorq->xyc->sc_dev)); if (iorq->buf) printf("%c: ", 'a' + (char)DISKPART(iorq->buf->b_dev)); if (iopb->com == XYCMD_RD || iopb->com == XYCMD_WR) printf("%s %d/%d/%d: ", (iopb->com == XYCMD_RD) ? "read" : "write", iopb->cyl, iopb->head, iopb->sect); printf("%s", xyc_e2str(error)); if (still_trying) printf(" [still trying, new error=%s]", xyc_e2str(iorq->errno)); else if (iorq->errno == 0) printf(" [recovered in %d tries]", iorq->tries); printf("\n"); } /* * xyc_error: non-fatal error encountered... recover. * return AOK if resubmitted, return FAIL if this iopb is done */ int xyc_error(struct xyc_softc *xycsc, struct xy_iorq *iorq, struct xy_iopb *iopb, int comm) { int errno = iorq->errno; int erract = xyc_entoact(errno); int oldmode, advance, i; if (erract == XY_ERA_RSET) { /* some errors require a reset */ oldmode = iorq->mode; iorq->mode = XY_SUB_DONE | (~XY_SUB_MASK & oldmode); /* make xyc_start ignore us */ xyc_reset(xycsc, 1, XY_RSET_NONE, errno, iorq->xy); iorq->mode = oldmode; } /* check for read/write to a sector in bad144 table if bad: redirect * request to bad144 area */ if ((comm == XYCMD_RD || comm == XYCMD_WR) && (iorq->mode & XY_MODE_B144) == 0) { advance = iorq->sectcnt - iopb->scnt; XYC_ADVANCE(iorq, advance); if ((i = isbad(&iorq->xy->dkb, iorq->blockno / iorq->xy->sectpercyl, (iorq->blockno / iorq->xy->nsect) % iorq->xy->nhead, iorq->blockno % iorq->xy->nsect)) != -1) { iorq->mode |= XY_MODE_B144; /* enter bad144 mode & * redirect */ iopb->errno = iopb->done = iopb->errs = 0; iopb->scnt = 1; iopb->cyl = (iorq->xy->ncyl + iorq->xy->acyl) - 2; /* second to last acyl */ i = iorq->xy->sectpercyl - 1 - i; /* follow bad144 * standard */ iopb->head = i / iorq->xy->nhead; iopb->sect = i % iorq->xy->nhead; /* will resubmit when we come out of remove_iorq */ return XY_ERR_AOK; /* recovered! */ } } /* * it isn't a bad144 sector, must be real error! see if we can retry * it? */ if ((iorq->mode & XY_MODE_VERBO) && iorq->lasterror) xyc_perror(iorq, iopb, 1); /* inform of error state * change */ iorq->lasterror = errno; if ((erract == XY_ERA_RSET || erract == XY_ERA_HARD) && iorq->tries < XYC_MAXTRIES) { /* retry? */ iorq->tries++; iorq->errno = iopb->errno = iopb->done = iopb->errs = 0; /* will resubmit at end of remove_iorq */ return XY_ERR_AOK; /* recovered! */ } /* failed to recover from this error */ return XY_ERR_FAIL; } /* * xyc_tick: make sure xy is still alive and ticking (err, kicking). */ void xyc_tick(void *arg) { struct xyc_softc *xycsc = arg; int lcv, s, reset = 0; /* reduce ttl for each request if one goes to zero, reset xyc */ s = splbio(); for (lcv = 0; lcv < XYC_MAXIOPB; lcv++) { if (xycsc->reqs[lcv].mode == 0 || XY_STATE(xycsc->reqs[lcv].mode) == XY_SUB_DONE) continue; xycsc->reqs[lcv].ttl--; if (xycsc->reqs[lcv].ttl == 0) reset = 1; } if (reset) { printf("%s: watchdog timeout\n", device_xname(xycsc->sc_dev)); xyc_reset(xycsc, 0, XY_RSET_NONE, XY_ERR_FAIL, NULL); } splx(s); /* until next time */ callout_reset(&xycsc->sc_tick_ch, XYC_TICKCNT, xyc_tick, xycsc); } /* * xyc_ioctlcmd: this function provides a user level interface to the * controller via ioctl. this allows "format" programs to be written * in user code, and is also useful for some debugging. we return * an error code. called at user priority. * * XXX missing a few commands (see the 7053 driver for ideas) */ int xyc_ioctlcmd(struct xy_softc *xy, dev_t dev, struct xd_iocmd *xio) { int s, err, rqno; void *dvmabuf = NULL; struct xyc_softc *xycsc; /* check sanity of requested command */ switch (xio->cmd) { case XYCMD_NOP: /* no op: everything should be zero */ if (xio->subfn || xio->dptr || xio->dlen || xio->block || xio->sectcnt) return EINVAL; break; case XYCMD_RD: /* read / write sectors (up to XD_IOCMD_MAXS) */ case XYCMD_WR: if (xio->subfn || xio->sectcnt > XD_IOCMD_MAXS || xio->sectcnt * XYFM_BPS != xio->dlen || xio->dptr == NULL) return EINVAL; break; case XYCMD_SK: /* seek: doesn't seem useful to export this */ return EINVAL; break; default: return EINVAL;/* ??? */ } /* create DVMA buffer for request if needed */ if (xio->dlen) { dvmabuf = dvma_malloc(xio->dlen); if (xio->cmd == XYCMD_WR) { err = copyin(xio->dptr, dvmabuf, xio->dlen); if (err) { dvma_free(dvmabuf, xio->dlen); return err; } } } /* do it! */ err = 0; xycsc = xy->parent; s = splbio(); rqno = xyc_cmd(xycsc, xio->cmd, xio->subfn, xy->xy_drive, xio->block, xio->sectcnt, dvmabuf, XY_SUB_WAIT); if (rqno == XY_ERR_FAIL) { err = EIO; goto done; } xio->errno = xycsc->ciorq->errno; xio->tries = xycsc->ciorq->tries; XYC_DONE(xycsc, err); if (xio->cmd == XYCMD_RD) err = copyout(dvmabuf, xio->dptr, xio->dlen); done: splx(s); if (dvmabuf) dvma_free(dvmabuf, xio->dlen); return err; } /* * xyc_e2str: convert error code number into an error string */ const char * xyc_e2str(int no) { switch (no) { case XY_ERR_FAIL: return "Software fatal error"; case XY_ERR_DERR: return "DOUBLE ERROR"; case XY_ERR_AOK: return "Successful completion"; case XY_ERR_IPEN: return "Interrupt pending"; case XY_ERR_BCFL: return "Busy conflict"; case XY_ERR_TIMO: return "Operation timeout"; case XY_ERR_NHDR: return "Header not found"; case XY_ERR_HARD: return "Hard ECC error"; case XY_ERR_ICYL: return "Illegal cylinder address"; case XY_ERR_ISEC: return "Illegal sector address"; case XY_ERR_SMAL: return "Last sector too small"; case XY_ERR_SACK: return "Slave ACK error (non-existent memory)"; case XY_ERR_CHER: return "Cylinder and head/header error"; case XY_ERR_SRTR: return "Auto-seek retry successful"; case XY_ERR_WPRO: return "Write-protect error"; case XY_ERR_UIMP: return "Unimplemented command"; case XY_ERR_DNRY: return "Drive not ready"; case XY_ERR_SZER: return "Sector count zero"; case XY_ERR_DFLT: return "Drive faulted"; case XY_ERR_ISSZ: return "Illegal sector size"; case XY_ERR_SLTA: return "Self test A"; case XY_ERR_SLTB: return "Self test B"; case XY_ERR_SLTC: return "Self test C"; case XY_ERR_SOFT: return "Soft ECC error"; case XY_ERR_SFOK: return "Soft ECC error recovered"; case XY_ERR_IHED: return "Illegal head"; case XY_ERR_DSEQ: return "Disk sequencer error"; case XY_ERR_SEEK: return "Seek error"; default: return "Unknown error"; } } int xyc_entoact(int errno) { switch (errno) { case XY_ERR_FAIL: case XY_ERR_DERR: case XY_ERR_IPEN: case XY_ERR_BCFL: case XY_ERR_ICYL: case XY_ERR_ISEC: case XY_ERR_UIMP: case XY_ERR_SZER: case XY_ERR_ISSZ: case XY_ERR_SLTA: case XY_ERR_SLTB: case XY_ERR_SLTC: case XY_ERR_IHED: case XY_ERR_SACK: case XY_ERR_SMAL: return XY_ERA_PROG; /* program error ! */ case XY_ERR_TIMO: case XY_ERR_NHDR: case XY_ERR_HARD: case XY_ERR_DNRY: case XY_ERR_CHER: case XY_ERR_SEEK: case XY_ERR_SOFT: return XY_ERA_HARD; /* hard error, retry */ case XY_ERR_DFLT: case XY_ERR_DSEQ: return XY_ERA_RSET; /* hard error reset */ case XY_ERR_SRTR: case XY_ERR_SFOK: case XY_ERR_AOK: return XY_ERA_SOFT; /* an FYI error */ case XY_ERR_WPRO: return XY_ERA_WPRO; /* write protect */ } return XY_ERA_PROG; /* ??? */ }