1 /* 2 * QEMU aio implementation 3 * 4 * Copyright IBM, Corp. 2008 5 * 6 * Authors: 7 * Anthony Liguori <aliguori (at) us.ibm.com> 8 * 9 * This work is licensed under the terms of the GNU GPL, version 2. See 10 * the COPYING file in the top-level directory. 11 * 12 */ 13 14 #include "qemu-common.h" 15 #include "block.h" 16 #include "qemu-queue.h" 17 #include "qemu_socket.h" 18 19 typedef struct AioHandler AioHandler; 20 21 /* The list of registered AIO handlers */ 22 static QLIST_HEAD(, AioHandler) aio_handlers; 23 24 /* This is a simple lock used to protect the aio_handlers list. Specifically, 25 * it's used to ensure that no callbacks are removed while we're walking and 26 * dispatching callbacks. 27 */ 28 static int walking_handlers; 29 30 struct AioHandler 31 { 32 int fd; 33 IOHandler *io_read; 34 IOHandler *io_write; 35 AioFlushHandler *io_flush; 36 AioProcessQueue *io_process_queue; 37 int deleted; 38 void *opaque; 39 QLIST_ENTRY(AioHandler) node; 40 }; 41 42 static AioHandler *find_aio_handler(int fd) 43 { 44 AioHandler *node; 45 46 QLIST_FOREACH(node, &aio_handlers, node) { 47 if (node->fd == fd) 48 if (!node->deleted) 49 return node; 50 } 51 52 return NULL; 53 } 54 55 int qemu_aio_set_fd_handler(int fd, 56 IOHandler *io_read, 57 IOHandler *io_write, 58 AioFlushHandler *io_flush, 59 AioProcessQueue *io_process_queue, 60 void *opaque) 61 { 62 AioHandler *node; 63 64 node = find_aio_handler(fd); 65 66 /* Are we deleting the fd handler? */ 67 if (!io_read && !io_write) { 68 if (node) { 69 /* If the lock is held, just mark the node as deleted */ 70 if (walking_handlers) 71 node->deleted = 1; 72 else { 73 /* Otherwise, delete it for real. We can't just mark it as 74 * deleted because deleted nodes are only cleaned up after 75 * releasing the walking_handlers lock. 76 */ 77 QLIST_REMOVE(node, node); 78 qemu_free(node); 79 } 80 } 81 } else { 82 if (node == NULL) { 83 /* Alloc and insert if it's not already there */ 84 node = qemu_mallocz(sizeof(AioHandler)); 85 node->fd = fd; 86 QLIST_INSERT_HEAD(&aio_handlers, node, node); 87 } 88 /* Update handler with latest information */ 89 node->io_read = io_read; 90 node->io_write = io_write; 91 node->io_flush = io_flush; 92 node->io_process_queue = io_process_queue; 93 node->opaque = opaque; 94 } 95 96 qemu_set_fd_handler2(fd, NULL, io_read, io_write, opaque); 97 98 return 0; 99 } 100 101 void qemu_aio_flush(void) 102 { 103 AioHandler *node; 104 int ret; 105 106 do { 107 ret = 0; 108 109 /* 110 * If there are pending emulated aio start them now so flush 111 * will be able to return 1. 112 */ 113 qemu_aio_wait(); 114 115 QLIST_FOREACH(node, &aio_handlers, node) { 116 if (node->io_flush) { 117 ret |= node->io_flush(node->opaque); 118 } 119 } 120 } while (qemu_bh_poll() || ret > 0); 121 } 122 123 int qemu_aio_process_queue(void) 124 { 125 AioHandler *node; 126 int ret = 0; 127 128 walking_handlers = 1; 129 130 QLIST_FOREACH(node, &aio_handlers, node) { 131 if (node->io_process_queue) { 132 if (node->io_process_queue(node->opaque)) { 133 ret = 1; 134 } 135 } 136 } 137 138 walking_handlers = 0; 139 140 return ret; 141 } 142 143 void qemu_aio_wait(void) 144 { 145 int ret; 146 147 if (qemu_bh_poll()) 148 return; 149 150 /* 151 * If there are callbacks left that have been queued, we need to call then. 152 * Return afterwards to avoid waiting needlessly in select(). 153 */ 154 if (qemu_aio_process_queue()) 155 return; 156 157 do { 158 AioHandler *node; 159 fd_set rdfds, wrfds; 160 int max_fd = -1; 161 162 walking_handlers = 1; 163 164 FD_ZERO(&rdfds); 165 FD_ZERO(&wrfds); 166 167 /* fill fd sets */ 168 QLIST_FOREACH(node, &aio_handlers, node) { 169 /* If there aren't pending AIO operations, don't invoke callbacks. 170 * Otherwise, if there are no AIO requests, qemu_aio_wait() would 171 * wait indefinitely. 172 */ 173 if (node->io_flush && node->io_flush(node->opaque) == 0) 174 continue; 175 176 if (!node->deleted && node->io_read) { 177 FD_SET(node->fd, &rdfds); 178 max_fd = MAX(max_fd, node->fd + 1); 179 } 180 if (!node->deleted && node->io_write) { 181 FD_SET(node->fd, &wrfds); 182 max_fd = MAX(max_fd, node->fd + 1); 183 } 184 } 185 186 walking_handlers = 0; 187 188 /* No AIO operations? Get us out of here */ 189 if (max_fd == -1) 190 break; 191 192 /* wait until next event */ 193 ret = select(max_fd, &rdfds, &wrfds, NULL, NULL); 194 if (ret == -1 && errno == EINTR) 195 continue; 196 197 /* if we have any readable fds, dispatch event */ 198 if (ret > 0) { 199 walking_handlers = 1; 200 201 /* we have to walk very carefully in case 202 * qemu_aio_set_fd_handler is called while we're walking */ 203 node = QLIST_FIRST(&aio_handlers); 204 while (node) { 205 AioHandler *tmp; 206 207 if (!node->deleted && 208 FD_ISSET(node->fd, &rdfds) && 209 node->io_read) { 210 node->io_read(node->opaque); 211 } 212 if (!node->deleted && 213 FD_ISSET(node->fd, &wrfds) && 214 node->io_write) { 215 node->io_write(node->opaque); 216 } 217 218 tmp = node; 219 node = QLIST_NEXT(node, node); 220 221 if (tmp->deleted) { 222 QLIST_REMOVE(tmp, node); 223 qemu_free(tmp); 224 } 225 } 226 227 walking_handlers = 0; 228 } 229 } while (ret == 0); 230 } 231