[d30fdbc] | 1 | // Mutex buffer is embedded in the nomutex executor to allow the executor to delete the workers without causing a |
---|
| 2 | // deadlock. If the executor is the monitor and the buffer is class, the thread calling the executor's destructor |
---|
| 3 | // (which is mutex) blocks when deleting the workers, preventing outstanding workers from calling remove to drain the |
---|
| 4 | // buffer. |
---|
| 5 | |
---|
| 6 | #include <bits/containers.hfa> |
---|
| 7 | #include <thread.hfa> |
---|
| 8 | #include <stdio.h> |
---|
| 9 | |
---|
| 10 | forall( otype T | is_node(T) | is_monitor(T) ) { |
---|
| 11 | monitor Buffer { // unbounded buffer |
---|
| 12 | __queue_t( T ) queue; // unbounded list of work requests |
---|
| 13 | condition delay; |
---|
| 14 | }; // Buffer |
---|
| 15 | |
---|
| 16 | void insert( Buffer( T ) & mutex buf, T * elem ) with(buf) { |
---|
| 17 | append( queue, elem ); // insert element into buffer |
---|
| 18 | signal( delay ); // restart |
---|
| 19 | } // insert |
---|
| 20 | |
---|
| 21 | T * remove( Buffer( T ) & mutex buf ) with(buf) { |
---|
| 22 | if ( ! queue ) wait( delay ); // no request to process ? => wait |
---|
| 23 | return pop_head( queue ); |
---|
| 24 | } // remove |
---|
| 25 | } // distribution |
---|
| 26 | |
---|
| 27 | struct WRequest { // client request, no return |
---|
| 28 | void (* action)( void ); |
---|
| 29 | WRequest * next; // intrusive queue field |
---|
| 30 | }; // WRequest |
---|
| 31 | |
---|
| 32 | WRequest *& get_next( WRequest & this ) { return this.next; } |
---|
| 33 | void ?{}( WRequest & req ) with(req) { action = 0; next = 0; } |
---|
| 34 | void ?{}( WRequest & req, void (* action)( void ) ) with(req) { req.action = action; next = 0; } |
---|
| 35 | bool stop( WRequest & req ) { return req.action == 0; } |
---|
| 36 | void doit( WRequest & req ) { req.action(); } |
---|
| 37 | |
---|
| 38 | // Each worker has its own work buffer to reduce contention between client and server. Hence, work requests arrive and |
---|
| 39 | // are distributed into buffers in a roughly round-robin order. |
---|
| 40 | |
---|
| 41 | thread Worker { |
---|
| 42 | Buffer( WRequest ) * requests; |
---|
| 43 | unsigned int start, range; |
---|
| 44 | }; // Worker |
---|
| 45 | |
---|
| 46 | void main( Worker & w ) with(w) { |
---|
| 47 | for ( int i = 0;; i = (i + 1) % range ) { |
---|
| 48 | WRequest * request = remove( requests[i + start] ); |
---|
| 49 | if ( ! request ) { yield(); continue; } |
---|
| 50 | if ( stop( *request ) ) break; |
---|
| 51 | doit( *request ); |
---|
| 52 | delete( request ); |
---|
| 53 | } // for |
---|
| 54 | } // Worker::main |
---|
| 55 | |
---|
| 56 | void ?{}( Worker & worker, cluster * wc, Buffer( WRequest ) * requests, unsigned int start, unsigned int range ) { |
---|
| 57 | (*get_thread(worker)){ *wc }; // create on given cluster |
---|
| 58 | worker.[requests, start, range] = [requests, start, range]; |
---|
| 59 | } // ?{} |
---|
| 60 | |
---|
| 61 | struct Executor { |
---|
| 62 | cluster * cluster; // if workers execute on separate cluster |
---|
| 63 | processor ** processors; // array of virtual processors adding parallelism for workers |
---|
| 64 | Buffer( WRequest ) * requests; // list of work requests |
---|
| 65 | Worker ** workers; // array of workers executing work requests |
---|
| 66 | unsigned int nprocessors, nworkers, nmailboxes; // number of mailboxes/workers/processor tasks |
---|
| 67 | bool sepClus; // use same or separate cluster for executor |
---|
| 68 | }; // Executor |
---|
| 69 | |
---|
| 70 | static thread_local unsigned int next; // demultiplexed across worker buffers |
---|
| 71 | unsigned int tickets( Executor & ex ) with(ex) { |
---|
| 72 | //return uFetchAdd( next, 1 ) % nmailboxes; |
---|
| 73 | return next++ % nmailboxes; // no locking, interference randomizes |
---|
| 74 | } // tickets |
---|
| 75 | |
---|
| 76 | void ?{}( Executor & ex, unsigned int np, unsigned int nw, unsigned int nm, bool sc = false ) with(ex) { |
---|
| 77 | [nprocessors, nworkers, nmailboxes, sepClus] = [np, nw, nm, sc]; |
---|
| 78 | assert( nmailboxes >= nworkers ); |
---|
| 79 | cluster = sepClus ? new( "Executor" ) : active_cluster(); |
---|
| 80 | processors = (processor **)anew( nprocessors ); |
---|
| 81 | requests = anew( nmailboxes ); |
---|
| 82 | workers = (Worker **)anew( nworkers ); |
---|
| 83 | |
---|
| 84 | for ( i; nprocessors ) { |
---|
| 85 | processors[ i ] = new( *cluster ); |
---|
| 86 | } // for |
---|
| 87 | |
---|
| 88 | unsigned int reqPerWorker = nmailboxes / nworkers, extras = nmailboxes % nworkers; |
---|
| 89 | for ( unsigned int i = 0, step = 0; i < nworkers; i += 1, step += reqPerWorker + ( i < extras ? 1 : 0 ) ) { |
---|
| 90 | workers[ i ] = new( cluster, requests, step, reqPerWorker + ( i < extras ? 1 : 0 ) ); |
---|
| 91 | } // for |
---|
| 92 | } // ?{} |
---|
| 93 | |
---|
| 94 | void ?{}( Executor & ex, unsigned int nprocessors, unsigned int nworkers, bool sepClus = false ) { |
---|
| 95 | ex{ nprocessors, nworkers, nworkers, sepClus }; |
---|
| 96 | } |
---|
| 97 | void ?{}( Executor & ex, unsigned int nprocessors, bool sepClus = false ) { |
---|
| 98 | ex{ nprocessors, nprocessors, nprocessors, sepClus }; |
---|
| 99 | } |
---|
| 100 | void ?{}( Executor & ex ) { // special for current cluster |
---|
| 101 | ex{ 0, active_cluster()->nprocessors, false }; |
---|
| 102 | } |
---|
| 103 | void ^?{}( Executor & ex ) with(ex) { |
---|
| 104 | // Add one sentinel per worker to stop them. Since in destructor, no new work should be queued. Cannot combine next |
---|
| 105 | // two loops and only have a single sentinel because workers arrive in arbitrary order, so worker1 may take the |
---|
| 106 | // single sentinel while waiting for worker 0 to end. |
---|
| 107 | |
---|
| 108 | WRequest sentinel[nworkers]; |
---|
| 109 | unsigned int reqPerWorker = nmailboxes / nworkers; |
---|
| 110 | for ( unsigned int i = 0, step = 0; i < nworkers; i += 1, step += reqPerWorker ) { |
---|
| 111 | insert( requests[step], &sentinel[i] ); // force eventually termination |
---|
| 112 | } // for |
---|
| 113 | for ( i; nworkers ) { |
---|
| 114 | delete( workers[ i ] ); |
---|
| 115 | } // for |
---|
| 116 | for ( i; nprocessors ) { |
---|
| 117 | delete( processors[ i ] ); |
---|
| 118 | } // for |
---|
| 119 | |
---|
| 120 | delete( workers ); |
---|
| 121 | delete( requests ); |
---|
| 122 | delete( processors ); |
---|
| 123 | if ( sepClus ) { delete( cluster ); } |
---|
| 124 | } // ^?{} |
---|
| 125 | |
---|
| 126 | void send( Executor & ex, void (* action)( void ) ) { // asynchronous call, no return value |
---|
| 127 | WRequest * node = new( action ); |
---|
| 128 | insert( ex.requests[tickets( ex )], node ); |
---|
| 129 | } // send |
---|
| 130 | |
---|
| 131 | int counter = 0; |
---|
| 132 | |
---|
| 133 | void workie( void ) { |
---|
| 134 | __atomic_add_fetch( &counter, 1, __ATOMIC_SEQ_CST ); |
---|
| 135 | // fprintf( stderr, "workie\n" ); |
---|
| 136 | } |
---|
| 137 | |
---|
| 138 | int main() { |
---|
| 139 | { |
---|
| 140 | Executor exector; |
---|
| 141 | for ( i; 3000 ) { |
---|
| 142 | send( exector, workie ); |
---|
| 143 | if ( i % 100 ) yield(); |
---|
| 144 | } // for |
---|
| 145 | } |
---|
| 146 | printf( "%d\n", counter ); |
---|
| 147 | } |
---|
| 148 | |
---|
| 149 | // Local Variables: // |
---|
| 150 | // compile-command: "cfa executor.cfa" // |
---|
| 151 | // End: // |
---|