2 #include "../port/lib.h"
8 #include "../port/error.h"
10 #include "../port/sd.h"
12 typedef struct Vring Vring;
13 typedef struct Vdesc Vdesc;
14 typedef struct Vused Vused;
15 typedef struct Vqueue Vqueue;
16 typedef struct Vdev Vdev;
18 typedef struct ScsiCfg ScsiCfg;
48 /* descriptor flags */
118 void *cfg; /* device specific config (for scsi) */
134 u32int event_info_size;
149 q = malloc(sizeof(*q) + sizeof(void*)*size);
151 PGROUND(sizeof(Vdesc)*size +
153 sizeof(u16int)*size +
159 if(p == nil || q == nil){
160 print("virtio: no memory for Vqueue\n");
167 p += sizeof(Vdesc)*size;
170 q->availent = (void*)p;
171 p += sizeof(u16int)*size;
172 q->availevent = (void*)p;
175 p = (uchar*)PGROUND((uintptr)p);
178 q->usedent = (void*)p;
179 p += sizeof(Vused)*size;
180 q->usedevent = (void*)p;
183 q->nfree = q->size = size;
184 for(i=0; i<size; i++){
185 q->desc[i].next = q->free;
201 for(p = nil; p = pcimatch(p, 0x1AF4, 0);){
202 if((p->did < 0x1000) || (p->did > 0x103F))
206 if(pcicfgr16(p, 0x2E) != typ)
208 if((vd = malloc(sizeof(*vd))) == nil){
209 print("virtio: no memory for Vdev\n");
212 vd->port = p->mem[0].bar & ~0x1;
213 if(ioalloc(vd->port, p->mem[0].size, 0, "virtio") < 0){
214 print("virtio: port %lux in use\n", vd->port);
222 outb(vd->port+Status, 0);
224 vd->feat = inl(vd->port+Devfeat);
225 outb(vd->port+Status, Acknowledge|Driver);
226 for(i=0; i<nelem(vd->queue); i++){
227 outs(vd->port+Qselect, i);
228 n = ins(vd->port+Qsize);
229 if(n == 0 || (n & (n-1)) != 0)
231 if((q = mkvqueue(n)) == nil)
237 outl(vd->port+Qaddr, PADDR(vd->queue[i]->desc)/BY2PG);
257 vqinterrupt(Vqueue *q)
266 while((q->lastused ^ q->used->idx) & m){
267 id = q->usedent[q->lastused++ & m].id;
271 r->done = 1; /* hands off */
277 id = q->desc[free].next;
278 q->desc[free].next = q->free;
281 } while(q->desc[free].flags & Next);
287 viointerrupt(Ureg *, void *arg)
291 if(inb(vd->port+Isr) & 1)
292 vqinterrupt(vd->queue[vd->typ == TypSCSI ? 2 : 0]);
298 return ((struct Rock*)arg)->done;
302 vioblkreq(Vdev *vd, int typ, void *a, long count, long secsize, uvlong lba)
310 struct Vioblkreqhdr {
322 rock.sleep = &up->sleep;
330 tsleep(&up->sleep, return0, 0, 500);
336 head = free = q->free;
338 d = &q->desc[free]; free = d->next;
339 d->addr = PADDR(&req);
340 d->len = sizeof(req);
343 d = &q->desc[free]; free = d->next;
345 d->len = secsize*count;
346 d->flags = typ ? Next : (Write|Next);
348 d = &q->desc[free]; free = d->next;
349 d->addr = PADDR(&status);
350 d->len = sizeof(status);
356 q->rock[head] = &rock;
359 q->availent[q->avail->idx++ & (q->size-1)] = head;
361 outs(vd->port+Qnotify, q->idx);
367 tsleep(rock.sleep, viodone, &rock, 1000);
380 u8int resp[4+4+2+2+SENSESIZE];
381 u8int req[8+8+3+CDBSIZE];
395 memset(resp, 0, sizeof(resp));
396 memset(req, 0, sizeof(req));
400 req[3] = r->lun&0xFF;
401 *(u64int*)(&req[8]) = (uintptr)r;
403 memmove(&req[8+8+3], r->cmd, r->clen);
406 rock.sleep = &up->sleep;
414 tsleep(&up->sleep, return0, 0, 500);
420 head = free = q->free;
422 d = &q->desc[free]; free = d->next;
423 d->addr = PADDR(req);
424 d->len = 8+8+3+cfg->cdb_size;
427 if(r->write && r->dlen > 0){
428 d = &q->desc[free]; free = d->next;
429 d->addr = PADDR(r->data);
434 d = &q->desc[free]; free = d->next;
435 d->addr = PADDR(resp);
436 d->len = 4+4+2+2+cfg->sense_size;
439 if(!r->write && r->dlen > 0){
442 d = &q->desc[free]; free = d->next;
443 d->addr = PADDR(r->data);
449 q->nfree -= 2 + (r->dlen > 0);
451 q->rock[head] = &rock;
454 q->availent[q->avail->idx++ & (q->size-1)] = head;
456 outs(vd->port+Qnotify, q->idx);
462 tsleep(rock.sleep, viodone, &rock, 1000);
469 /* response+status */
470 r->status = resp[10];
475 len = *((u32int*)&resp[0]);
477 if(len > sizeof(r->sense))
478 len = sizeof(r->sense);
479 memmove(r->sense, &resp[4+4+2+2], len);
480 r->flags |= SDvalidsense;
484 len = *((u32int*)&resp[4]);
488 r->rlen = r->dlen - len;
495 viobio(SDunit *u, int lun, int write, void *a, long count, uvlong lba)
497 long ss, cc, max, ret;
501 if(vd->typ == TypSCSI)
502 return scsibio(u, lun, write, a, count, lba);
508 if((cc = count) > max)
510 if(vioblkreq(vd, write != 0, (uchar*)a + ret, cc, ss, lba) != 0)
529 if(vd->typ == TypSCSI)
530 return vioscsireq(r);
531 if(r->cmd[0] == 0x35 || r->cmd[0] == 0x91){
532 if(vioblkreq(vd, 4, nil, 0, 0, 0) != 0)
533 return sdsetsense(r, SDcheck, 3, 0xc, 2);
534 return sdsetsense(r, SDok, 0, 0, 0);
536 if((i = sdfakescsi(r)) != SDnostatus)
537 return r->status = i;
538 if((i = sdfakescsirw(r, &lba, &count, &rw)) != SDnostatus)
540 r->rlen = viobio(u, r->lun, rw == SDwrite, r->data, count, lba);
541 return r->status = SDok;
551 if(vd->typ == TypSCSI)
552 return scsionline(u);
554 cap = inl(vd->port+Devspec+4);
556 cap |= inl(vd->port+Devspec);
557 if(u->sectors != cap){
571 if(vd->typ == TypSCSI)
572 return scsiverify(u);
582 intrenable(vd->pci->intl, viointerrupt, vd, vd->pci->tbdf, "virtio");
583 outb(vd->port+Status, inb(vd->port+Status) | DriverOk);
596 for(vd = viopnpdevs(TypBlk); vd; vd = vd->next){
602 if((s = malloc(sizeof(*s))) == nil)
606 s->ifc = &sdvirtioifc;
616 for(vd = viopnpdevs(TypSCSI); vd; vd = vd->next){
622 if((cfg = malloc(sizeof(*cfg))) == nil)
624 cfg->num_queues = inl(vd->port+Devspec+4*0);
625 cfg->seg_max = inl(vd->port+Devspec+4*1);
626 cfg->max_sectors = inl(vd->port+Devspec+4*2);
627 cfg->cmd_per_lun = inl(vd->port+Devspec+4*3);
628 cfg->event_info_size = inl(vd->port+Devspec+4*4);
629 cfg->sense_size = inl(vd->port+Devspec+4*5);
630 cfg->cdb_size = inl(vd->port+Devspec+4*6);
631 cfg->max_channel = ins(vd->port+Devspec+4*7);
632 cfg->max_target = ins(vd->port+Devspec+4*7+2);
633 cfg->max_lun = inl(vd->port+Devspec+4*8);
635 if(cfg->max_target == 0){
639 if((cfg->cdb_size > CDBSIZE) || (cfg->sense_size > SENSESIZE)){
640 print("sdvirtio: cdb %ud or sense size %ud too big\n",
641 cfg->cdb_size, cfg->sense_size);
649 if((s = malloc(sizeof(*s))) == nil)
653 s->ifc = &sdvirtioifc;
654 s->nunit = cfg->max_target;
665 SDifc sdvirtioifc = {
673 vioverify, /* verify */
674 vioonline, /* online */