struct vring_used_elem *virt_queue__set_used_elem(struct virt_queue *queue, u32 head, u32 len);
bool virtio_queue__should_signal(struct virt_queue *vq);
-u16 virt_queue__get_iov(struct virt_queue *queue, struct iovec iov[], u16 *out, u16 *in, struct kvm *kvm);
+u16 virt_queue__get_iov(struct virt_queue *vq, struct iovec iov[], u16 *out, u16 *in, struct kvm *kvm);
+u16 virt_queue__get_head_iov(struct virt_queue *vq, struct iovec iov[], u16 *out, u16 *in, u16 head, struct kvm *kvm);
u16 virt_queue__get_inout_iov(struct kvm *kvm, struct virt_queue *queue,
struct iovec in_iov[], struct iovec out_iov[],
u16 *in, u16 *out);
#define NUM_VIRT_QUEUES 1
struct blk_dev_req {
- struct list_head list;
struct virt_queue *vq;
struct blk_dev *bdev;
struct iovec iov[VIRTIO_BLK_QUEUE_SIZE];
static LIST_HEAD(bdevs);
static int compat_id;
-static struct blk_dev_req *virtio_blk_req_pop(struct blk_dev *bdev)
-{
- struct blk_dev_req *req = NULL;
-
- mutex_lock(&bdev->req_mutex);
- if (!list_empty(&bdev->req_list)) {
- req = list_first_entry(&bdev->req_list, struct blk_dev_req, list);
- list_del_init(&req->list);
- }
- mutex_unlock(&bdev->req_mutex);
-
- return req;
-}
-
-static void virtio_blk_req_push(struct blk_dev *bdev, struct blk_dev_req *req)
-{
- mutex_lock(&bdev->req_mutex);
- list_add(&req->list, &bdev->req_list);
- mutex_unlock(&bdev->req_mutex);
-}
-
void virtio_blk_complete(void *param, long len)
{
struct blk_dev_req *req = param;
if (virtio_queue__should_signal(&bdev->vqs[queueid]))
bdev->vtrans.trans_ops->signal_vq(req->kvm, &bdev->vtrans, queueid);
-
- virtio_blk_req_push(req->bdev, req);
}
static void virtio_blk_do_io_request(struct kvm *kvm, struct blk_dev_req *req)
static void virtio_blk_do_io(struct kvm *kvm, struct virt_queue *vq, struct blk_dev *bdev)
{
- while (virt_queue__available(vq)) {
- struct blk_dev_req *req = virtio_blk_req_pop(bdev);
+ struct blk_dev_req *req;
+ u16 head;
- *req = (struct blk_dev_req) {
- .vq = vq,
- .bdev = bdev,
- .kvm = kvm,
- };
- req->head = virt_queue__get_iov(vq, req->iov, &req->out, &req->in, kvm);
+ while (virt_queue__available(vq)) {
+ head = virt_queue__pop(vq);
+ req = &bdev->reqs[head];
+ req->head = virt_queue__get_head_iov(vq, req->iov, &req->out, &req->in, head, kvm);
+ req->vq = vq;
virtio_blk_do_io_request(kvm, req);
}
void virtio_blk__init(struct kvm *kvm, struct disk_image *disk)
{
struct blk_dev *bdev;
- size_t i;
+ unsigned int i;
if (!disk)
return;
list_add_tail(&bdev->list, &bdevs);
- INIT_LIST_HEAD(&bdev->req_list);
- for (i = 0; i < ARRAY_SIZE(bdev->reqs); i++)
- list_add(&bdev->reqs[i].list, &bdev->req_list);
+ for (i = 0; i < ARRAY_SIZE(bdev->reqs); i++) {
+ bdev->reqs[i].bdev = bdev;
+ bdev->reqs[i].kvm = kvm;
+ }
disk_image__set_callback(bdev->disk, virtio_blk_complete);
return used_elem;
}
-u16 virt_queue__get_iov(struct virt_queue *queue, struct iovec iov[], u16 *out, u16 *in, struct kvm *kvm)
+u16 virt_queue__get_head_iov(struct virt_queue *vq, struct iovec iov[], u16 *out, u16 *in, u16 head, struct kvm *kvm)
{
struct vring_desc *desc;
- u16 head, idx;
+ u16 idx;
- idx = head = virt_queue__pop(queue);
+ idx = head;
*out = *in = 0;
do {
- desc = virt_queue__get_desc(queue, idx);
- iov[*out + *in].iov_base = guest_flat_to_host(kvm, desc->addr);
- iov[*out + *in].iov_len = desc->len;
+ desc = virt_queue__get_desc(vq, idx);
+ iov[*out + *in].iov_base = guest_flat_to_host(kvm, desc->addr);
+ iov[*out + *in].iov_len = desc->len;
if (desc->flags & VRING_DESC_F_WRITE)
(*in)++;
else
return head;
}
+u16 virt_queue__get_iov(struct virt_queue *vq, struct iovec iov[], u16 *out, u16 *in, struct kvm *kvm)
+{
+ u16 head;
+
+ head = virt_queue__pop(vq);
+
+ return virt_queue__get_head_iov(vq, iov, out, in, head, kvm);
+}
+
/* in and out are relative to guest */
u16 virt_queue__get_inout_iov(struct kvm *kvm, struct virt_queue *queue,
struct iovec in_iov[], struct iovec out_iov[],
u16 *in, u16 *out)
{
- u16 head, idx;
struct vring_desc *desc;
+ u16 head, idx;
idx = head = virt_queue__pop(queue);
*out = *in = 0;
else
break;
} while (1);
+
return head;
}