| [1c49dc5] | 1 | extern "C" { | 
|---|
|  | 2 | #include <errno.h> | 
|---|
|  | 3 | #include <stdio.h> | 
|---|
|  | 4 | #include <stdint.h> | 
|---|
|  | 5 | #include <stdlib.h> | 
|---|
|  | 6 | #include <string.h> | 
|---|
|  | 7 | #include <unistd.h> | 
|---|
|  | 8 | #include <sys/mman.h> | 
|---|
|  | 9 | #include <sys/syscall.h> | 
|---|
|  | 10 | #include <sys/uio.h> | 
|---|
|  | 11 | #include <fcntl.h> | 
|---|
|  | 12 |  | 
|---|
|  | 13 | #include <linux/io_uring.h> | 
|---|
|  | 14 | } | 
|---|
|  | 15 |  | 
|---|
|  | 16 | # ifndef __NR_io_uring_setup | 
|---|
|  | 17 | #  define __NR_io_uring_setup           425 | 
|---|
|  | 18 | # endif | 
|---|
|  | 19 | # ifndef __NR_io_uring_enter | 
|---|
|  | 20 | #  define __NR_io_uring_enter           426 | 
|---|
|  | 21 | # endif | 
|---|
|  | 22 | # ifndef __NR_io_uring_register | 
|---|
|  | 23 | #  define __NR_io_uring_register        427 | 
|---|
|  | 24 | # endif | 
|---|
|  | 25 |  | 
|---|
|  | 26 | struct io_uring_sq { | 
|---|
|  | 27 | // Head and tail of the ring (associated with array) | 
|---|
|  | 28 | volatile uint32_t * head; | 
|---|
|  | 29 | volatile uint32_t * tail; | 
|---|
|  | 30 |  | 
|---|
|  | 31 | // The actual kernel ring which uses head/tail | 
|---|
|  | 32 | // indexes into the sqes arrays | 
|---|
|  | 33 | uint32_t * array; | 
|---|
|  | 34 |  | 
|---|
|  | 35 | // number of entries and mask to go with it | 
|---|
|  | 36 | const uint32_t * num; | 
|---|
|  | 37 | const uint32_t * mask; | 
|---|
|  | 38 |  | 
|---|
|  | 39 | // Submission flags (Not sure what for) | 
|---|
|  | 40 | uint32_t * flags; | 
|---|
|  | 41 |  | 
|---|
|  | 42 | // number of sqes not submitted (whatever that means) | 
|---|
|  | 43 | uint32_t * dropped; | 
|---|
|  | 44 |  | 
|---|
|  | 45 | // Like head/tail but not seen by the kernel | 
|---|
|  | 46 | volatile uint32_t alloc; | 
|---|
|  | 47 |  | 
|---|
|  | 48 | // A buffer of sqes (not the actual ring) | 
|---|
|  | 49 | struct io_uring_sqe * sqes; | 
|---|
|  | 50 |  | 
|---|
|  | 51 | // The location and size of the mmaped area | 
|---|
|  | 52 | void * ring_ptr; | 
|---|
|  | 53 | size_t ring_sz; | 
|---|
|  | 54 | }; | 
|---|
|  | 55 |  | 
|---|
|  | 56 | struct io_uring_cq { | 
|---|
|  | 57 | // Head and tail of the ring | 
|---|
|  | 58 | volatile uint32_t * head; | 
|---|
|  | 59 | volatile uint32_t * tail; | 
|---|
|  | 60 |  | 
|---|
|  | 61 | // number of entries and mask to go with it | 
|---|
|  | 62 | const uint32_t * mask; | 
|---|
|  | 63 | const uint32_t * num; | 
|---|
|  | 64 |  | 
|---|
|  | 65 | // number of cqes not submitted (whatever that means) | 
|---|
|  | 66 | uint32_t * overflow; | 
|---|
|  | 67 |  | 
|---|
|  | 68 | // the kernel ring | 
|---|
|  | 69 | struct io_uring_cqe * cqes; | 
|---|
|  | 70 |  | 
|---|
|  | 71 | // The location and size of the mmaped area | 
|---|
|  | 72 | void * ring_ptr; | 
|---|
|  | 73 | size_t ring_sz; | 
|---|
|  | 74 | }; | 
|---|
|  | 75 |  | 
|---|
|  | 76 | struct io_ring { | 
|---|
|  | 77 | struct io_uring_sq submit_q; | 
|---|
|  | 78 | struct io_uring_cq completion_q; | 
|---|
|  | 79 | uint32_t flags; | 
|---|
|  | 80 | int fd; | 
|---|
|  | 81 | }; | 
|---|
|  | 82 |  | 
|---|
|  | 83 | struct IO_singleton { | 
|---|
|  | 84 | io_ring io; | 
|---|
|  | 85 | }; | 
|---|
|  | 86 |  | 
|---|
|  | 87 | IO_singleton self; | 
|---|
|  | 88 |  | 
|---|
|  | 89 | void init_uring(uint32_t nentries) { | 
|---|
|  | 90 | // Step 1 : call to setup | 
|---|
|  | 91 | struct io_uring_params params; | 
|---|
|  | 92 | memset(¶ms, 0, sizeof(params)); | 
|---|
|  | 93 | // params.flags = IORING_SETUP_SQPOLL; | 
|---|
|  | 94 |  | 
|---|
|  | 95 | int fd = syscall(__NR_io_uring_setup, nentries, ¶ms ); | 
|---|
|  | 96 | if(fd < 0) { | 
|---|
|  | 97 | fprintf(stderr, "KERNEL ERROR: IO_URING SETUP - %s\n", strerror(errno)); | 
|---|
|  | 98 | abort(); | 
|---|
|  | 99 | } | 
|---|
|  | 100 |  | 
|---|
|  | 101 | // Step 2 : mmap result | 
|---|
|  | 102 | memset(&self.io, 0, sizeof(struct io_ring)); | 
|---|
|  | 103 | struct io_uring_sq & sq = self.io.submit_q; | 
|---|
|  | 104 | struct io_uring_cq & cq = self.io.completion_q; | 
|---|
|  | 105 |  | 
|---|
|  | 106 | // calculate the right ring size | 
|---|
|  | 107 | sq.ring_sz = params.sq_off.array + (params.sq_entries * sizeof(unsigned)           ); | 
|---|
|  | 108 | cq.ring_sz = params.cq_off.cqes  + (params.cq_entries * sizeof(struct io_uring_cqe)); | 
|---|
|  | 109 |  | 
|---|
|  | 110 | // Requires features | 
|---|
|  | 111 | // // adjust the size according to the parameters | 
|---|
|  | 112 | // if ((params.features & IORING_FEAT_SINGLE_MMAP) != 0) { | 
|---|
|  | 113 | //      cq->ring_sz = sq->ring_sz = max(cq->ring_sz, sq->ring_sz); | 
|---|
|  | 114 | // } | 
|---|
|  | 115 |  | 
|---|
|  | 116 | // mmap the Submit Queue into existence | 
|---|
|  | 117 | sq.ring_ptr = mmap(0, sq.ring_sz, PROT_READ | PROT_WRITE, MAP_SHARED | MAP_POPULATE, fd, IORING_OFF_SQ_RING); | 
|---|
|  | 118 | if (sq.ring_ptr == (void*)MAP_FAILED) { | 
|---|
|  | 119 | fprintf(stderr, "KERNEL ERROR: IO_URING MMAP1 - %s\n", strerror(errno)); | 
|---|
|  | 120 | abort(); | 
|---|
|  | 121 | } | 
|---|
|  | 122 |  | 
|---|
|  | 123 | // mmap the Completion Queue into existence (may or may not be needed) | 
|---|
|  | 124 | // Requires features | 
|---|
|  | 125 | // if ((params.features & IORING_FEAT_SINGLE_MMAP) != 0) { | 
|---|
|  | 126 | //      cq->ring_ptr = sq->ring_ptr; | 
|---|
|  | 127 | // } | 
|---|
|  | 128 | // else { | 
|---|
|  | 129 | // We need multiple call to MMAP | 
|---|
|  | 130 | cq.ring_ptr = mmap(0, cq.ring_sz, PROT_READ | PROT_WRITE, MAP_SHARED | MAP_POPULATE, fd, IORING_OFF_CQ_RING); | 
|---|
|  | 131 | if (cq.ring_ptr == (void*)MAP_FAILED) { | 
|---|
|  | 132 | munmap(sq.ring_ptr, sq.ring_sz); | 
|---|
|  | 133 | fprintf(stderr, "KERNEL ERROR: IO_URING MMAP2 - %s\n", strerror(errno)); | 
|---|
|  | 134 | abort(); | 
|---|
|  | 135 | } | 
|---|
|  | 136 | // } | 
|---|
|  | 137 |  | 
|---|
|  | 138 | // mmap the submit queue entries | 
|---|
|  | 139 | size_t size = params.sq_entries * sizeof(struct io_uring_sqe); | 
|---|
|  | 140 | sq.sqes = (struct io_uring_sqe *)mmap(0, size, PROT_READ | PROT_WRITE, MAP_SHARED | MAP_POPULATE, fd, IORING_OFF_SQES); | 
|---|
|  | 141 | if (sq.sqes == (struct io_uring_sqe *)MAP_FAILED) { | 
|---|
|  | 142 | munmap(sq.ring_ptr, sq.ring_sz); | 
|---|
|  | 143 | if (cq.ring_ptr != sq.ring_ptr) munmap(cq.ring_ptr, cq.ring_sz); | 
|---|
|  | 144 | fprintf(stderr, "KERNEL ERROR: IO_URING MMAP3 - %s\n", strerror(errno)); | 
|---|
|  | 145 | abort(); | 
|---|
|  | 146 | } | 
|---|
|  | 147 |  | 
|---|
|  | 148 | // Get the pointers from the kernel to fill the structure | 
|---|
|  | 149 | // submit queue | 
|---|
|  | 150 | sq.head    = (volatile uint32_t *)(((intptr_t)sq.ring_ptr) + params.sq_off.head); | 
|---|
|  | 151 | sq.tail    = (volatile uint32_t *)(((intptr_t)sq.ring_ptr) + params.sq_off.tail); | 
|---|
|  | 152 | sq.mask    = (   const uint32_t *)(((intptr_t)sq.ring_ptr) + params.sq_off.ring_mask); | 
|---|
|  | 153 | sq.num     = (   const uint32_t *)(((intptr_t)sq.ring_ptr) + params.sq_off.ring_entries); | 
|---|
|  | 154 | sq.flags   = (         uint32_t *)(((intptr_t)sq.ring_ptr) + params.sq_off.flags); | 
|---|
|  | 155 | sq.dropped = (         uint32_t *)(((intptr_t)sq.ring_ptr) + params.sq_off.dropped); | 
|---|
|  | 156 | sq.array   = (         uint32_t *)(((intptr_t)sq.ring_ptr) + params.sq_off.array); | 
|---|
|  | 157 | sq.alloc = *sq.tail; | 
|---|
|  | 158 |  | 
|---|
|  | 159 | // completion queue | 
|---|
|  | 160 | cq.head     = (volatile uint32_t *)(((intptr_t)cq.ring_ptr) + params.cq_off.head); | 
|---|
|  | 161 | cq.tail     = (volatile uint32_t *)(((intptr_t)cq.ring_ptr) + params.cq_off.tail); | 
|---|
|  | 162 | cq.mask     = (   const uint32_t *)(((intptr_t)cq.ring_ptr) + params.cq_off.ring_mask); | 
|---|
|  | 163 | cq.num      = (   const uint32_t *)(((intptr_t)cq.ring_ptr) + params.cq_off.ring_entries); | 
|---|
|  | 164 | cq.overflow = (         uint32_t *)(((intptr_t)cq.ring_ptr) + params.cq_off.overflow); | 
|---|
|  | 165 | cq.cqes   = (struct io_uring_cqe *)(((intptr_t)cq.ring_ptr) + params.cq_off.cqes); | 
|---|
|  | 166 |  | 
|---|
|  | 167 | self.io.fd = fd; | 
|---|
|  | 168 | } | 
|---|