[d30fdbc] | 1 | // Mutex buffer is embedded in the nomutex executor to allow the executor to delete the workers without causing a
|
---|
| 2 | // deadlock. If the executor is the monitor and the buffer is class, the thread calling the executor's destructor
|
---|
| 3 | // (which is mutex) blocks when deleting the workers, preventing outstanding workers from calling remove to drain the
|
---|
| 4 | // buffer.
|
---|
| 5 |
|
---|
[8d462e5] | 6 | #include <containers/list.hfa>
|
---|
[d30fdbc] | 7 | #include <thread.hfa>
|
---|
| 8 | #include <stdio.h>
|
---|
| 9 |
|
---|
| 10 | struct WRequest { // client request, no return
|
---|
| 11 | void (* action)( void );
|
---|
[8d462e5] | 12 | DLISTED_MGD_IMPL_IN(WRequest)
|
---|
[d30fdbc] | 13 | }; // WRequest
|
---|
[8d462e5] | 14 | DLISTED_MGD_IMPL_OUT(WRequest)
|
---|
[d30fdbc] | 15 |
|
---|
[8d462e5] | 16 | void ?{}( WRequest & req ) with(req) { action = 0; }
|
---|
| 17 | void ?{}( WRequest & req, void (* action)( void ) ) with(req) { req.action = action; }
|
---|
[d30fdbc] | 18 | bool stop( WRequest & req ) { return req.action == 0; }
|
---|
| 19 | void doit( WRequest & req ) { req.action(); }
|
---|
| 20 |
|
---|
[8d462e5] | 21 | monitor WRBuffer { // unbounded buffer
|
---|
| 22 | dlist( WRequest, WRequest ) queue; // unbounded list of work requests
|
---|
| 23 | condition delay;
|
---|
| 24 | }; // WRBuffer
|
---|
| 25 |
|
---|
| 26 | void insert( WRBuffer & mutex buf, WRequest * elem ) with(buf) {
|
---|
| 27 | insert_last( queue, *elem ); // insert element into buffer
|
---|
| 28 | signal( delay ); // restart
|
---|
| 29 | } // insert
|
---|
| 30 |
|
---|
| 31 | WRequest * remove( WRBuffer & mutex buf ) with(buf) {
|
---|
| 32 | if ( queue`is_empty ) wait( delay ); // no request to process ? => wait
|
---|
| 33 | return & pop_first( queue );
|
---|
| 34 | } // remove
|
---|
| 35 |
|
---|
[d30fdbc] | 36 | // Each worker has its own work buffer to reduce contention between client and server. Hence, work requests arrive and
|
---|
| 37 | // are distributed into buffers in a roughly round-robin order.
|
---|
| 38 |
|
---|
| 39 | thread Worker {
|
---|
[8d462e5] | 40 | WRBuffer * requests;
|
---|
[d30fdbc] | 41 | unsigned int start, range;
|
---|
| 42 | }; // Worker
|
---|
| 43 |
|
---|
| 44 | void main( Worker & w ) with(w) {
|
---|
| 45 | for ( int i = 0;; i = (i + 1) % range ) {
|
---|
| 46 | WRequest * request = remove( requests[i + start] );
|
---|
| 47 | if ( ! request ) { yield(); continue; }
|
---|
| 48 | if ( stop( *request ) ) break;
|
---|
| 49 | doit( *request );
|
---|
| 50 | delete( request );
|
---|
| 51 | } // for
|
---|
| 52 | } // Worker::main
|
---|
| 53 |
|
---|
[8d462e5] | 54 | void ?{}( Worker & worker, cluster * wc, WRBuffer * requests, unsigned int start, unsigned int range ) {
|
---|
[d30fdbc] | 55 | (*get_thread(worker)){ *wc }; // create on given cluster
|
---|
| 56 | worker.[requests, start, range] = [requests, start, range];
|
---|
| 57 | } // ?{}
|
---|
| 58 |
|
---|
| 59 | struct Executor {
|
---|
| 60 | cluster * cluster; // if workers execute on separate cluster
|
---|
| 61 | processor ** processors; // array of virtual processors adding parallelism for workers
|
---|
[8d462e5] | 62 | WRBuffer * requests; // list of work requests
|
---|
[d30fdbc] | 63 | Worker ** workers; // array of workers executing work requests
|
---|
| 64 | unsigned int nprocessors, nworkers, nmailboxes; // number of mailboxes/workers/processor tasks
|
---|
| 65 | bool sepClus; // use same or separate cluster for executor
|
---|
| 66 | }; // Executor
|
---|
| 67 |
|
---|
| 68 | static thread_local unsigned int next; // demultiplexed across worker buffers
|
---|
| 69 | unsigned int tickets( Executor & ex ) with(ex) {
|
---|
| 70 | //return uFetchAdd( next, 1 ) % nmailboxes;
|
---|
| 71 | return next++ % nmailboxes; // no locking, interference randomizes
|
---|
| 72 | } // tickets
|
---|
| 73 |
|
---|
| 74 | void ?{}( Executor & ex, unsigned int np, unsigned int nw, unsigned int nm, bool sc = false ) with(ex) {
|
---|
| 75 | [nprocessors, nworkers, nmailboxes, sepClus] = [np, nw, nm, sc];
|
---|
| 76 | assert( nmailboxes >= nworkers );
|
---|
| 77 | cluster = sepClus ? new( "Executor" ) : active_cluster();
|
---|
| 78 | processors = (processor **)anew( nprocessors );
|
---|
[8d462e5] | 79 | requests = (WRBuffer *)anew( nmailboxes );
|
---|
[d30fdbc] | 80 | workers = (Worker **)anew( nworkers );
|
---|
| 81 |
|
---|
| 82 | for ( i; nprocessors ) {
|
---|
| 83 | processors[ i ] = new( *cluster );
|
---|
| 84 | } // for
|
---|
| 85 |
|
---|
| 86 | unsigned int reqPerWorker = nmailboxes / nworkers, extras = nmailboxes % nworkers;
|
---|
| 87 | for ( unsigned int i = 0, step = 0; i < nworkers; i += 1, step += reqPerWorker + ( i < extras ? 1 : 0 ) ) {
|
---|
| 88 | workers[ i ] = new( cluster, requests, step, reqPerWorker + ( i < extras ? 1 : 0 ) );
|
---|
| 89 | } // for
|
---|
| 90 | } // ?{}
|
---|
| 91 |
|
---|
| 92 | void ?{}( Executor & ex, unsigned int nprocessors, unsigned int nworkers, bool sepClus = false ) {
|
---|
| 93 | ex{ nprocessors, nworkers, nworkers, sepClus };
|
---|
| 94 | }
|
---|
| 95 | void ?{}( Executor & ex, unsigned int nprocessors, bool sepClus = false ) {
|
---|
| 96 | ex{ nprocessors, nprocessors, nprocessors, sepClus };
|
---|
| 97 | }
|
---|
| 98 | void ?{}( Executor & ex ) { // special for current cluster
|
---|
| 99 | ex{ 0, active_cluster()->nprocessors, false };
|
---|
| 100 | }
|
---|
| 101 | void ^?{}( Executor & ex ) with(ex) {
|
---|
| 102 | // Add one sentinel per worker to stop them. Since in destructor, no new work should be queued. Cannot combine next
|
---|
| 103 | // two loops and only have a single sentinel because workers arrive in arbitrary order, so worker1 may take the
|
---|
| 104 | // single sentinel while waiting for worker 0 to end.
|
---|
| 105 |
|
---|
| 106 | WRequest sentinel[nworkers];
|
---|
| 107 | unsigned int reqPerWorker = nmailboxes / nworkers;
|
---|
| 108 | for ( unsigned int i = 0, step = 0; i < nworkers; i += 1, step += reqPerWorker ) {
|
---|
| 109 | insert( requests[step], &sentinel[i] ); // force eventually termination
|
---|
| 110 | } // for
|
---|
| 111 | for ( i; nworkers ) {
|
---|
| 112 | delete( workers[ i ] );
|
---|
| 113 | } // for
|
---|
| 114 | for ( i; nprocessors ) {
|
---|
| 115 | delete( processors[ i ] );
|
---|
| 116 | } // for
|
---|
| 117 |
|
---|
| 118 | delete( workers );
|
---|
| 119 | delete( requests );
|
---|
| 120 | delete( processors );
|
---|
| 121 | if ( sepClus ) { delete( cluster ); }
|
---|
| 122 | } // ^?{}
|
---|
| 123 |
|
---|
| 124 | void send( Executor & ex, void (* action)( void ) ) { // asynchronous call, no return value
|
---|
| 125 | WRequest * node = new( action );
|
---|
| 126 | insert( ex.requests[tickets( ex )], node );
|
---|
| 127 | } // send
|
---|
| 128 |
|
---|
| 129 | int counter = 0;
|
---|
| 130 |
|
---|
| 131 | void workie( void ) {
|
---|
| 132 | __atomic_add_fetch( &counter, 1, __ATOMIC_SEQ_CST );
|
---|
| 133 | // fprintf( stderr, "workie\n" );
|
---|
| 134 | }
|
---|
| 135 |
|
---|
| 136 | int main() {
|
---|
| 137 | {
|
---|
| 138 | Executor exector;
|
---|
| 139 | for ( i; 3000 ) {
|
---|
| 140 | send( exector, workie );
|
---|
[8d462e5] | 141 | if ( i % 100 == 0 ) {
|
---|
| 142 | // fprintf( stderr, "%d\n", i );
|
---|
| 143 | yield();
|
---|
| 144 | }
|
---|
[d30fdbc] | 145 | } // for
|
---|
| 146 | }
|
---|
| 147 | printf( "%d\n", counter );
|
---|
| 148 | }
|
---|
| 149 |
|
---|
| 150 | // Local Variables: //
|
---|
| 151 | // compile-command: "cfa executor.cfa" //
|
---|
| 152 | // End: //
|
---|