summaryrefslogtreecommitdiffstats
path: root/hw/dataplane/ioq.c
blob: 0c9f5c4d6082a42e89c6df972cb8aeafbe0d6864 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
/*
 * Linux AIO request queue
 *
 * Copyright 2012 IBM, Corp.
 * Copyright 2012 Red Hat, Inc. and/or its affiliates
 *
 * Authors:
 *   Stefan Hajnoczi <stefanha@redhat.com>
 *
 * This work is licensed under the terms of the GNU GPL, version 2 or later.
 * See the COPYING file in the top-level directory.
 *
 */

#include "hw/dataplane/ioq.h"

void ioq_init(IOQueue *ioq, int fd, unsigned int max_reqs)
{
    int rc;

    ioq->fd = fd;
    ioq->max_reqs = max_reqs;

    memset(&ioq->io_ctx, 0, sizeof ioq->io_ctx);
    rc = io_setup(max_reqs, &ioq->io_ctx);
    if (rc != 0) {
        fprintf(stderr, "ioq io_setup failed %d\n", rc);
        exit(1);
    }

    rc = event_notifier_init(&ioq->io_notifier, 0);
    if (rc != 0) {
        fprintf(stderr, "ioq io event notifier creation failed %d\n", rc);
        exit(1);
    }

    ioq->freelist = g_malloc0(sizeof ioq->freelist[0] * max_reqs);
    ioq->freelist_idx = 0;

    ioq->queue = g_malloc0(sizeof ioq->queue[0] * max_reqs);
    ioq->queue_idx = 0;
}

void ioq_cleanup(IOQueue *ioq)
{
    g_free(ioq->freelist);
    g_free(ioq->queue);

    event_notifier_cleanup(&ioq->io_notifier);
    io_destroy(ioq->io_ctx);
}

EventNotifier *ioq_get_notifier(IOQueue *ioq)
{
    return &ioq->io_notifier;
}

struct iocb *ioq_get_iocb(IOQueue *ioq)
{
    /* Underflow cannot happen since ioq is sized for max_reqs */
    assert(ioq->freelist_idx != 0);

    struct iocb *iocb = ioq->freelist[--ioq->freelist_idx];
    ioq->queue[ioq->queue_idx++] = iocb;
    return iocb;
}

void ioq_put_iocb(IOQueue *ioq, struct iocb *iocb)
{
    /* Overflow cannot happen since ioq is sized for max_reqs */
    assert(ioq->freelist_idx != ioq->max_reqs);

    ioq->freelist[ioq->freelist_idx++] = iocb;
}

struct iocb *ioq_rdwr(IOQueue *ioq, bool read, struct iovec *iov,
                      unsigned int count, long long offset)
{
    struct iocb *iocb = ioq_get_iocb(ioq);

    if (read) {
        io_prep_preadv(iocb, ioq->fd, iov, count, offset);
    } else {
        io_prep_pwritev(iocb, ioq->fd, iov, count, offset);
    }
    io_set_eventfd(iocb, event_notifier_get_fd(&ioq->io_notifier));
    return iocb;
}

int ioq_submit(IOQueue *ioq)
{
    int rc = io_submit(ioq->io_ctx, ioq->queue_idx, ioq->queue);
    ioq->queue_idx = 0; /* reset */
    return rc;
}

int ioq_run_completion(IOQueue *ioq, IOQueueCompletion *completion,
                       void *opaque)
{
    struct io_event events[ioq->max_reqs];
    int nevents, i;

    do {
        nevents = io_getevents(ioq->io_ctx, 0, ioq->max_reqs, events, NULL);
    } while (nevents < 0 && errno == EINTR);
    if (nevents < 0) {
        return nevents;
    }

    for (i = 0; i < nevents; i++) {
        ssize_t ret = ((uint64_t)events[i].res2 << 32) | events[i].res;

        completion(events[i].obj, ret, opaque);
        ioq_put_iocb(ioq, events[i].obj);
    }
    return nevents;
}