source: doc/theses/colby_parsons_MMAth/text/waituntil.tex @ 210c737

Last change on this file since 210c737 was 210c737, checked in by caparsons <caparson@…>, 14 months ago

Merge branch 'master' of plg.uwaterloo.ca:software/cfa/cfa-cc

  • Property mode set to 100644
File size: 58.9 KB
Line 
1% ======================================================================
2% ======================================================================
3\chapter{Waituntil}\label{s:waituntil}
4% ======================================================================
5% ======================================================================
6
7Consider the following motivating problem.
8There are $N$ stalls (resources) in a bathroom and there are $M$ people (threads) using the bathroom.
9Each stall has its own lock since only one person may occupy a stall at a time.
10Humans solve this problem in the following way.
11They check if all of the stalls are occupied.
12If not, they enter and claim an available stall.
13If they are all occupied, people queue and watch the stalls until one is free, and then enter and lock the stall.
14This solution can be implemented on a computer easily, if all threads are waiting on all stalls and agree to queue.
15
16Now the problem is extended.
17Some stalls are wheelchair accessible and some stalls have gender identification.
18Each person (thread) may be limited to only one kind of stall or may choose among different kinds of stalls that match their criteria.
19Immediately, the problem becomes more difficult.
20A single queue no longer solves the problem.
21What happens when there is a stall available that the person at the front of the queue cannot choose?
22The na\"ive solution has each thread spin indefinitely continually checking every matching kind of stall(s) until a suitable one is free.
23This approach is insufficient since it wastes cycles and results in unfairness among waiting threads as a thread can acquire the first matching stall without regard to the waiting time of other threads.
24Waiting for the first appropriate stall (resource) that becomes available without spinning is an example of \gls{synch_multiplex}: the ability to wait synchronously for one or more resources based on some selection criteria.
25
26\section{History of Synchronous Multiplexing}\label{s:History}
27
28There is a history of tools that provide \gls{synch_multiplex}.
29Some well known \gls{synch_multiplex} tools include Unix system utilities: @select@~\cite{linux:select}, @poll@~\cite{linux:poll}, and @epoll@~\cite{linux:epoll}, and the @select@ statement provided by Go~\cite{go:selectref}, Ada~\cite[\S~9.7]{Ada16}, and \uC~\cite[\S~3.3.1]{uC++}.
30The concept and theory surrounding \gls{synch_multiplex} was introduced by Hoare in his 1985 book, Communicating Sequential Processes (CSP)~\cite{Hoare85},
31\begin{quote}
32A communication is an event that is described by a pair $c.v$ where $c$ is the name of the channel on which the communication takes place and $v$ is the value of the message which passes.~\cite[p.~113]{Hoare85}
33\end{quote}
34The ideas in CSP were implemented by Roscoe and Hoare in the language Occam~\cite{Roscoe88}.
35
36Both CSP and Occam include the ability to wait for a \Newterm{choice} among receiver channels and \Newterm{guards} to toggle which receives are valid.
37For example,
38\begin{cfa}[mathescape]
39(@G1@(x) $\rightarrow$ P @|@ @G2@(y) $\rightarrow$ Q )
40\end{cfa}
41waits for either channel @x@ or @y@ to have a value, if and only if guards @G1@ and @G2@ are true;
42if only one guard is true, only one channel receives, and if both guards are false, no receive occurs.
43% extended CSP with a \gls{synch_multiplex} construct @ALT@, which waits for one resource to be available and then executes a corresponding block of code.
44In detail, waiting for one resource out of a set of resources can be thought of as a logical exclusive-or over the set of resources.
45Guards are a conditional operator similar to an @if@, except they apply to the resource being waited on.
46If a guard is false, then the resource it guards is not in the set of resources being waited on.
47If all guards are false, the ALT, Occam's \gls{synch_multiplex} statement, does nothing and the thread continues.
48Guards can be simulated using @if@ statements as shown in~\cite[rule~2.4, p~183]{Roscoe88}
49\begin{lstlisting}[basicstyle=\rm,mathescape]
50ALT( $b$ & $g$ $P$, $G$ ) = IF ( $b$ ALT($\,g$ $P$, $G$ ), $\neg\,$b ALT( $G$ ) )                        (boolean guard elim).
51\end{lstlisting}
52but require $2^N-1$ @if@ statements, where $N$ is the number of guards.
53The exponential blowup comes from applying rule 2.4 repeatedly, since it works on one guard at a time.
54Figure~\ref{f:wu_if} shows in \CFA an example of applying rule 2.4 for three guards.
55Also, notice the additional code duplication for statements @S1@, @S2@, and @S3@.
56
57\begin{figure}
58\centering
59\begin{lrbox}{\myboxA}
60\begin{cfa}
61when( G1 )
62        waituntil( R1 ) S1
63or when( G2 )
64        waituntil( R2 ) S2
65or when( G3 )
66        waituntil( R3 ) S3
67
68
69
70
71
72
73
74\end{cfa}
75\end{lrbox}
76
77\begin{lrbox}{\myboxB}
78\begin{cfa}
79if ( G1 )
80        if ( G2 )
81                if ( G3 ) waituntil( R1 ) S1 or waituntil( R2 ) S2 or waituntil( R3 ) S3
82                else waituntil( R1 ) S1 or waituntil( R2 ) S2
83        else
84                if ( G3 ) waituntil( R1 ) S1 or waituntil( R3 ) S3
85                else waituntil( R1 ) S1
86else
87        if ( G2 )
88                if ( G3 ) waituntil( R2 ) S2 or waituntil( R3 ) S3
89                else waituntil( R2 ) S2
90        else
91                if ( G3 ) waituntil( R3 ) S3
92\end{cfa}
93\end{lrbox}
94
95\subfloat[Guards]{\label{l:guards}\usebox\myboxA}
96\hspace*{5pt}
97\vrule
98\hspace*{5pt}
99\subfloat[Simulated Guards]{\label{l:simulated_guards}\usebox\myboxB}
100\caption{\CFA guard simulated with \lstinline{if} statement.}
101\label{f:wu_if}
102\end{figure}
103
104When discussing \gls{synch_multiplex} implementations, the resource being multiplexed is important.
105While CSP waits on channels, the earliest known implementation of synch\-ronous multiplexing is Unix's @select@~\cite{linux:select}, multiplexing over file descriptors.
106The @select@ system-call is passed three sets of file descriptors (read, write, exceptional) to wait on and an optional timeout.
107@select@ blocks until either some subset of file descriptors are available or the timeout expires.
108All file descriptors that are ready are returned by modifying the argument sets to only contain the ready descriptors.
109
110This early implementation differs from the theory presented in CSP: when the call from @select@ returns it may provide more than one ready file descriptor.
111As such, @select@ has logical-or multiplexing semantics, whereas the theory described exclusive-or semantics.
112It is possible to achieve exclusive-or semantics with @select@ by arbitrarily operating on only one of the returned descriptors.
113@select@ passes the interest set of file descriptors between application and kernel in the form of a worst-case sized bit-mask, where the worst-case is the largest numbered file descriptor.
114@poll@ reduces the size of the interest sets changing from a bit mask to a linked data structure, independent of the file-descriptor values.
115@epoll@ further reduces the data passed per call by keeping the interest set in the kernel, rather than supplying it on every call.
116
117These early \gls{synch_multiplex} tools interact directly with the operating system and others are used to communicate among processes.
118Later, \gls{synch_multiplex} started to appear in applications, via programming languages, to support fast multiplexed concurrent communication among threads.
119An early example of \gls{synch_multiplex} is the @select@ statement in Ada~\cite[\S~9.7]{Ichbiah79}.
120This @select@ allows a task object, with their own threads, to multiplex over a subset of asynchronous calls to its methods.
121The Ada @select@ has the same exclusive-or semantics and guards as Occam ALT;
122however, it multiplexes over methods rather than channels.
123
124\begin{figure}
125\begin{lstlisting}[language=ada,literate=,{moredelim={**[is][\color{red}]{@}{@}}}]
126task type buffer is -- thread type
127        ... -- buffer declarations
128        count : integer := 0;
129begin -- thread starts here
130        loop
131                select
132                        @when count < Size@ => -- guard
133                        @accept insert( elem : in ElemType )@ do  -- method
134                                ... -- add to buffer
135                                count := count + 1;
136                        end;
137                        -- executed if this accept called
138                or
139                        @when count > 0@ => -- guard
140                        @accept remove( elem : out ElemType )@ do  -- method
141                                ... --remove and return from buffer via parameter
142                                count := count - 1;
143                        end;
144                        -- executed if this accept called
145                or @delay 10.0@;  -- unblock after 10 seconds without call
146                or @else@ -- do not block, cannot appear with delay
147                end select;
148        end loop;
149end buffer;
150var buf : buffer; -- create task object and start thread in task body
151\end{lstlisting}
152\caption{Ada Bounded Buffer}
153\label{f:BB_Ada}
154\end{figure}
155
156Figure~\ref{f:BB_Ada} shows the outline of a bounded buffer implemented with an Ada task.
157Note, a task method is associated with the \lstinline[language=ada]{accept} clause of the \lstinline[language=ada]{select} statement, rather than being a separate routine.
158The thread executing the loop in the task body blocks at the \lstinline[language=ada]{select} until a call occurs to @insert@ or @remove@.
159Then the appropriate \lstinline[language=ada]{accept} method is run with the called arguments.
160Hence, the \lstinline[language=ada]{select} statement provides rendezvous points for threads, rather than providing channels with message passing.
161The \lstinline[language=ada]{select} statement also provides a timeout and @else@ (nonblocking), which changes synchronous multiplexing to asynchronous.
162Now the thread polls rather than blocks.
163
164Another example of programming-language \gls{synch_multiplex} is Go using a @select@ statement with channels~\cite{go:selectref}.
165Figure~\ref{l:BB_Go} shows the outline of a bounded buffer administrator implemented with a Go routine.
166Here two channels are used for inserting and removing by client producers and consumers, respectively.
167(The @done@ channel is used to synchronize with the program main.)
168Go's @select@ has the same exclusive-or semantics as the ALT primitive from Occam and associated code blocks for each clause like ALT and Ada.
169However, unlike Ada and ALT, Go does not provide guards for the \lstinline[language=go]{case} clauses of the \lstinline[language=go]{select}.
170As such, the exponential blowup can be seen comparing Go and \uC in Figure~\label{f:AdaMultiplexing}.
171Go also provides a timeout via a channel and a @default@ clause like Ada @else@ for asynchronous multiplexing.
172
173\begin{figure}
174\centering
175
176\begin{lrbox}{\myboxA}
177\begin{lstlisting}[language=go,literate=]
178insert := make( chan int )
179remove := make( chan * int )
180buffer := make( chan int Size )
181done := make( chan int )
182count := 0
183func in_buf( int val ) {
184    buffer <- val
185    count++
186}
187func out_buf( int * ptr ) {
188    *ptr := <-buffer
189    count--
190}
191func BoundedBuffer {
192    L: for {
193        if ( count < Size && count > 0 ) {
194            select { // wait for message
195                case i := <- insert: in_buf( i )
196                case p := <- remove: out_buf( p )
197                case <- done: break L
198            }
199        } else if ( count < Size ) {
200            select { // wait for message
201                case i := <- insert: in_buf( i )
202                case <- done: break L
203            }
204        } else ( count > 0 ) {
205            select { // wait for message
206                case p := <- remove: out_buf( p )
207                case <- done: break L;
208            }
209        }
210    }
211    done <- 0
212}
213func main() {
214        go BoundedBuffer() // start administrator
215}
216
217
218
219
220\end{lstlisting}
221\end{lrbox}
222
223\begin{lrbox}{\myboxB}
224\begin{lstlisting}[language=uC++,{moredelim={**[is][\color{red}]{@}{@}}}]
225_Task BoundedBuffer {
226        int * buffer;
227        int front = back = count = 0;
228  public:
229    // ... constructor implementation
230        void insert( int elem ) {
231                buffer[front] = elem;
232        front = ( front + 1 ) % Size;
233                count += 1;
234        }
235        int remove() {
236                int ret = buffer[back];
237        back = ( back + 1 ) % Size;
238                count -= 1;
239        return ret;
240        }
241  private:
242        void main() {
243                for ( ;; ) {
244                        _Accept( ~buffer ) break;
245                        @or _When ( count < Size ) _Accept( insert )@;
246                        @or _When ( count > 0 ) _Accept( remove )@;
247                }
248        }
249};
250buffer buf; // start thread in main method
251\end{lstlisting}
252\end{lrbox}
253
254\subfloat[Go]{\label{l:BB_Go}\usebox\myboxA}
255\hspace*{5pt}
256\vrule
257\hspace*{5pt}
258\subfloat[\uC]{\label{l:BB_uC++}\usebox\myboxB}
259
260\caption{Bounded Buffer}
261\label{f:AdaMultiplexing}
262\end{figure}
263
264Finally, \uC provides \gls{synch_multiplex} with Ada-style @select@ over monitor and task methods with the @_Accept@ statement~\cite[\S~2.9.2.1]{uC++}, and over futures with the @_Select@ statement~\cite[\S~3.3.1]{uC++}.
265The @_Select@ statement extends the ALT/Go @select@ by offering both @and@ and @or@ semantics, which can be used together in the same statement.
266Both @_Accept@ and @_Select@ statements provide guards for multiplexing clauses, as well as, timeout, and @else@ clauses.
267Figure~\ref{l:BB_uC++} shows the outline of a bounded buffer administrator implemented with \uC @_Accept@ statements.
268
269There are other languages that provide \gls{synch_multiplex}, including Rust's @select!@ over futures~\cite{rust:select}, Java's @allof@/@anyof@ over futures~\cite{java:allof:anyof}, OCaml's @select@ over channels~\cite{ocaml:channel}, and C++14's @when_any@ over futures~\cite{cpp:whenany}.
270Note that while C++14 and Rust provide \gls{synch_multiplex}, the implementations leave much to be desired as both rely on polling to wait on multiple resources.
271
272\section{Other Approaches to Synchronous Multiplexing}
273
274To avoid the need for \gls{synch_multiplex}, all communication among threads/processes must come from a single source.
275For example, in Erlang each process has a single heterogeneous mailbox that is the sole source of concurrent communication, removing the need for \gls{synch_multiplex} as there is only one place to wait on resources.
276Similar, actor systems circumvent the \gls{synch_multiplex} problem as actors only block when waiting for the next message never in a behaviour.
277While these approaches solve the \gls{synch_multiplex} problem, they introduce other issues.
278Consider the case where a thread has a single source of communication and it wants a set of $N$ resources.
279It must sequentially request the $N$ resources and wait for each response.
280During the receives for the $N$ resources, it can receive other communication, and has to save and postpone these communications, or discard them.
281% If the requests for the other resources need to be retracted, the burden falls on the programmer to determine how to synchronize appropriately to ensure that only one resource is delivered.
282
283\section{\CFA's Waituntil Statement}
284
285The new \CFA \gls{synch_multiplex} utility introduced in this work is the @waituntil@ statement.
286There already exists a @waitfor@ statement in \CFA that supports Ada-style \gls{synch_multiplex} over monitor methods~\cite{Delisle21}, so this @waituntil@ focuses on synchronizing over other resources.
287All of the \gls{synch_multiplex} features mentioned so far are monomorphic, only waiting on one kind of resource: Unix @select@ supports file descriptors, Go's @select@ supports channel operations, \uC's @select@ supports futures, and Ada's @select@ supports monitor method calls.
288The \CFA @waituntil@ is polymorphic and provides \gls{synch_multiplex} over any objects that satisfy the trait in Figure~\ref{f:wu_trait}.
289No other language provides a synchronous multiplexing tool polymorphic over resources like \CFA's @waituntil@.
290
291\begin{figure}
292\begin{cfa}
293forall(T & | sized(T))
294trait is_selectable {
295        // For registering a waituntil stmt on a selectable type
296        bool register_select( T &, select_node & );
297
298        // For unregistering a waituntil stmt from a selectable type
299        bool unregister_select( T &, select_node & );
300
301        // on_selected is run on the selecting thread prior to executing
302        // the statement associated with the select_node
303        bool on_selected( T &, select_node & );
304};
305\end{cfa}
306\caption{Trait for types that can be passed into \CFA's \lstinline{waituntil} statement.}
307\label{f:wu_trait}
308\end{figure}
309
310Currently locks, channels, futures and timeouts are supported by the @waituntil@ statement, and this set can be expanded through the @is_selectable@ trait as other use-cases arise.
311The @waituntil@ statement supports guard clauses, both @or@ and @and@ semantics, and timeout and @else@ for asynchronous multiplexing.
312Figure~\ref{f:wu_example} shows a \CFA @waituntil@ usage, which is waiting for either @Lock@ to be available \emph{or} for a value to be read from @Channel@ into @i@ \emph{and} for @Future@ to be fulfilled \emph{or} a timeout of one second.
313Note, the expression inside a @waituntil@ clause is evaluated once at the start of the @waituntil@ algorithm.
314
315\begin{figure}
316\begin{cfa}
317future(int) Future;
318channel(int) Channel;
319owner_lock Lock;
320int i = 0;
321
322waituntil( @Lock@ ) { ... }
323or when( i == 0 ) waituntil( i << @Channel@ ) { ... }
324and waituntil( @Future@ ) { ... }
325or waituntil( @timeout( 1`s )@ ) { ... }
326// else { ... }
327\end{cfa}
328\caption{Example of \CFA's waituntil statement}
329\label{f:wu_example}
330\end{figure}
331
332\section{Waituntil Semantics}
333
334The @waituntil@ semantics has two parts: the semantics of the statement itself, \ie @and@, @or@, @when@ guards, and @else@ semantics, and the semantics of how the @waituntil@ interacts with types like locks, channels, and futures.
335
336\subsection{Statement Semantics}
337
338The @or@ semantics are the most straightforward and nearly match those laid out in the ALT statement from Occam.
339The clauses have an exclusive-or relationship where the first available one is run and only one clause is run.
340\CFA's @or@ semantics differ from ALT semantics: instead of randomly picking a clause when multiple are available, the first clause in the @waituntil@ that is available is executed.
341For example, in the following example, if @foo@ and @bar@ are both available, @foo@ is always selected since it comes first in the order of @waituntil@ clauses.
342\begin{cfa}
343future(int) bar, foo;
344waituntil( foo ) { ... } or waituntil( bar ) { ... } // prioritize foo
345\end{cfa}
346The reason for this semantics is that prioritizing resources can be useful in certain problems, such as shutdown.
347In the rare case where there is a starvation problem with the ordering, it possible to follow a @waituntil@ with its reverse form, alternating which resource has the highest priority:
348\begin{cfa}
349waituntil( foo ) { ... } or waituntil( bar ) { ... } // prioritize foo
350waituntil( bar ) { ... } or waituntil( foo ) { ... } // prioritize bar
351\end{cfa}
352While this approach is not general for many resources, it handles many basic cases.
353
354The \CFA @and@ semantics match the @and@ semantics of \uC \lstinline[language=uC++]{_Select}.
355When multiple clauses are joined by @and@, the @waituntil@ makes a thread wait for all to be available, but still runs the corresponding code blocks \emph{as they become available}.
356When an @and@ clause becomes available, the waiting thread unblocks and runs that clause's code-block, and then the thread waits again for the next available clause or the @waituntil@ statement is now satisfied.
357This semantics allows work to be done in parallel while synchronizing over a set of resources, and furthermore, gives a good reason to use the @and@ operator.
358If the @and@ operator waited for all clauses to be available before running, it is the same as just acquiring those resources consecutively by a sequence of @waituntil@ statements.
359
360As for normal C expressions, the @and@ operator binds more tightly than the @or@.
361To give an @or@ operator higher precedence, parenthesis are used.
362For example, the following @waituntil@ unconditionally waits for @C@ and one of either @A@ or @B@, since the @or@ is given higher precedence via parenthesis.
363\begin{cfa}
364@(@ waituntil( A ) { ... }              // bind tightly to or
365or waituntil( B ) { ... } @)@
366and waituntil( C ) { ... }
367\end{cfa}
368
369The guards in the @waituntil@ statement are called @when@ clauses.
370Each boolean expression inside a @when@ is evaluated \emph{once} before the @waituntil@ statement is run.
371Like Occam's ALT, the guards toggle clauses on and off, where a @waituntil@ clause is only evaluated and waited on if the corresponding guard is @true@.
372In addition, the @waituntil@ guards require some nuance since both @and@ and @or@ operators are supported \see{Section~\ref{s:wu_guards}}.
373When a guard is false and a clause is removed, it can be thought of as removing that clause and its preceding operation from the statement.
374For example, in the following, the two @waituntil@ statements are semantically equivalent.
375
376\begin{lrbox}{\myboxA}
377\begin{cfa}
378when( true ) waituntil( A ) { ... }
379or when( false ) waituntil( B ) { ... }
380and waituntil( C ) { ... }
381\end{cfa}
382\end{lrbox}
383
384\begin{lrbox}{\myboxB}
385\begin{cfa}
386waituntil( A ) { ... }
387and waituntil( C ) { ... }
388
389\end{cfa}
390\end{lrbox}
391
392\begin{tabular}{@{}lcl@{}}
393\usebox\myboxA & $\equiv$ & \usebox\myboxB
394\end{tabular}
395
396The @else@ clause on the @waituntil@ has identical semantics to the @else@ clause in Ada.
397If all resources are not immediately available and there is an @else@ clause, the @else@ clause is run and the thread continues.
398
399\subsection{Type Semantics}
400
401As mentioned, to support interaction with the @waituntil@ statement a type must support the trait in Figure~\ref{f:wu_trait}.
402The @waituntil@ statement expects types to register and unregister themselves via calls to @register_select@ and @unregister_select@, respectively.
403When a resource becomes available, @on_selected@ is run, and if it returns false, the corresponding code block is not run.
404Many types do not need @on_selected@, but it is provided if a type needs to perform work or checks before the resource can be accessed in the code block.
405The register/unregister routines in the trait also return booleans.
406The return value of @register_select@ is @true@, if the resource is immediately available and @false@ otherwise.
407The return value of @unregister_select@ is @true@, if the corresponding code block should be run after unregistration and @false@ otherwise.
408The routine @on_selected@ and the return value of @unregister_select@ are needed to support channels as a resource.
409More detail on channels and their interaction with @waituntil@ appear in Section~\ref{s:wu_chans}.
410
411The trait can be used directly by having a blocking object support the @is_selectable@ trait, or it can be used indirectly through routines that take the object as an argument.
412When used indirectly, the object's routine returns a type that supports the @is_selectable@ trait.
413This feature leverages \CFA's ability to overload on return type to select the correct overloaded routine for the @waituntil@ context.
414Indirect support through routines is needed for types that want to support multiple operations such as channels that allow both reading and writing.
415
416\section{\lstinline{waituntil} Implementation}
417
418The @waituntil@ statement is not inherently complex, and Figure~\ref{f:WU_Impl} shows the basic outline of the @waituntil@ algorithm.
419Complexity comes from the consideration of race conditions and synchronization needed when supporting various primitives.
420The following sections use examples to fill in complexity details missing in Figure~\ref{f:WU_Impl}.
421After which, the full pseudocode for the @waituntil@ algorithm is presented in Figure~\ref{f:WU_Full_Impl}.
422
423\begin{figure}
424\begin{cfa}
425select_nodes s[N];                                                              $\C[3.25in]{// declare N select nodes}$
426for ( node in s )                                                               $\C{// register nodes}$
427        register_select( resource, node );
428while ( statement predicate not satisfied ) {   $\C{// check predicate}$
429        // block until clause(s) satisfied
430        for ( resource in waituntil statement ) {       $\C{// run true code blocks}$
431                if ( resource is avail ) run code block
432                if ( statement predicate is satisfied ) break;
433        }
434}
435for ( node in s )                                                               $\C{// deregister nodes}\CRT$
436        if ( unregister_select( resource, node ) ) run code block
437\end{cfa}
438\caption{\lstinline{waituntil} Implementation}
439\label{f:WU_Impl}
440\end{figure}
441
442The basic steps of the algorithm are:
443\begin{enumerate}
444\item
445The @waituntil@ statement declares $N$ @select_node@s, one per resource that is being waited on, which stores any @waituntil@ data pertaining to that resource.
446
447\item
448Each @select_node@ is then registered with the corresponding resource.
449
450\item
451The thread executing the @waituntil@ then loops until the statement's predicate is satisfied.
452In each iteration, if the predicate is unsatisfied, the @waituntil@ thread blocks.
453When another thread satisfies a resource clause (\eg sends to a  channel), it unblocks the @waituntil@ thread.
454This thread checks all clauses for completion, and any completed clauses have their code blocks run.
455While checking clause completion, if enough clauses have been run such that the statement predicate is satisfied, the loop exits early.
456
457\item
458Once the thread escapes the loop, the @select_nodes@ are unregistered from the resources.
459\end{enumerate}
460These steps give a basic overview of how the statement works.
461The following sections shed light on the specific changes and provide more implementation detail.
462
463\subsection{Locks}\label{s:wu_locks}
464
465The \CFA runtime supports a number of spinning and blocking locks, \eg semaphore, MCS, futex, Go mutex, spinlock, owner, \etc.
466Many of these locks satisfy the @is_selectable@ trait, and hence, are resources supported by the @waituntil@ statement.
467For example, the following waits until the thread has acquired lock @l1@ or locks @l2@ and @l3@.
468\begin{cfa}
469owner_lock l1, l2, l3;
470waituntil ( l1 ) { ... }
471or waituntil( l2 ) { ... }
472and waituntil( l3 ) { ... }
473\end{cfa}
474Implicitly, the @waituntil@ is calling the lock acquire for each of these locks to establish a position in the lock's queue of waiting threads.
475When the lock schedules this thread, it unblocks and runs the code block associated with the lock and then releases the lock.
476
477In detail, when a thread waits on multiple locks via a @waituntil@, it enqueues a @select_node@ in each of the lock's waiting queues.
478When a @select_node@ reaches the front of the lock's queue and gains ownership, the thread blocked on the @waituntil@ is unblocked.
479Now, the lock is held by the @waituntil@ thread until the code block is executed, and then the node is unregistered, during which the lock is released.
480Immediately releasing the lock after the code block prevents the waiting thread from holding multiple locks and potentially introducing a deadlock.
481As such, the only unregistered nodes associated with locks are the ones that have not run.
482
483\subsection{Timeouts}
484
485A timeout for the @waituntil@ statement is a duration passed to routine \lstinline[deletekeywords={timeout}]{timeout}\footnote{\lstinline{timeout} is a quasi-keyword in \CFA, allowing it to be used an identifier.}, \eg:
486\begin{cquote}
487\begin{tabular}{@{}l|l@{}}
488\multicolumn{2}{@{}l@{}}{\lstinline{Duration D1\{ 1`ms \}, D2\{ 2`ms \}, D3\{ 3`ms \};}} \\
489\begin{cfa}[deletekeywords={timeout}]
490waituntil( i << C1 ) {}
491or waituntil( i << C2 ) {}
492or waituntil( i << C3 ) {}
493or waituntil( timeout( D1 ) ) {}
494or waituntil( timeout( D2 ) ) {}
495or waituntil( timeout( D3 ) ) {}
496\end{cfa}
497&
498\begin{cfa}[deletekeywords={timeout}]
499waituntil( i << C1 ) {}
500or waituntil( i << C2 ) {}
501or waituntil( i << C3 ) {}
502or waituntil( timeout( min( D1, D2, D3 ) ) ) {}
503
504
505\end{cfa}
506\end{tabular}
507\end{cquote}
508These two examples are basically equivalent.
509Here, the multiple timeouts are useful because the durations can change during execution and the separate clauses provide different code blocks if a timeout triggers.
510Multiple timeouts can also be used with @and@ to provide a minimal delay before proceeding.
511In following example, either channel @C1@ or @C2@ must be satisfied but nothing can be done for at least 1 or 3 seconds after the channel read, respectively.
512\begin{cfa}[deletekeywords={timeout}]
513waituntil( i << C1 ); and waituntil( timeout( 1`s ) );
514or waituntil( i << C2 ); and waituntil( timeout( 3`s ) );
515\end{cfa}
516If only @C2@ is satisfied, \emph{both} timeout code-blocks trigger because 1 second occurs before 3 seconds.
517Note, the \CFA @waitfor@ statement only provides a single @timeout@ clause because it only supports @or@ semantics.
518
519The \lstinline[deletekeywords={timeout}]{timeout} routine is different from UNIX @sleep@, which blocks for the specified duration and returns the amount of time elapsed since the call started.
520Instead, \lstinline[deletekeywords={timeout}]{timeout} returns a type that supports the @is_selectable@ trait, allowing the type system to select the correct overloaded routine for this context.
521For the @waituntil@, it is more idiomatic for the \lstinline[deletekeywords={timeout}]{timeout} to use the same syntax as other blocking operations instead of having a special language clause.
522
523\subsection{Channels}\label{s:wu_chans}
524
525Channels require more complexity to allow synchronous multiplexing.
526For locks, when an outside thread releases a lock and unblocks the @waituntil@ thread (WUT), the lock's MX property is passed to the WUT (no spinning locks).
527For futures, the outside thread deliveries a value to the future and unblocks any waiting threads, including WUTs.
528In either case, after the WUT unblocks, it is safe to execute its corresponding code block knowing access to the resource is protected by the lock or the read-only state of the future.
529Similarly, for channels, when an outside thread inserts a value into a channel, it must unblock the WUT.
530However, for channels, there is a race condition that poses an issue.
531If the outside thread inserts into the channel and unblocks the WUT, there is a race where another thread can remove the channel data, so after the WUT unblocks and attempts to remove from the buffer, it fails, and the WUT must reblock (busy waiting).
532This scenario is a \gls{toctou} race that needs to be consolidated.
533To close the race, the outside thread must detect this case and insert directly into the left-hand side of the channel expression (@i << chan@) rather than into the channel, and then unblock the WUT.
534Now the unblocked WUT is guaranteed to have a satisfied resource and its code block can safely executed.
535The insertion circumvents the channel buffer via the wait-morphing in the \CFA channel implementation \see{Section~\ref{s:chan_impl}}, allowing @waituntil@ channel unblocking to not be special-cased.
536
537Furthermore, if both @and@ and @or@ operators are used, the @or@ operations stop behaving like exclusive-or due to the race among channel operations, \eg:
538\begin{cfa}
539int i;
540waituntil( i << A ) {} and waituntil( i << B ) {}
541or waituntil( i << C ) {} and waituntil( i << D ) {}
542\end{cfa}
543If exclusive-or semantics are followed, only the code blocks for @A@ and @B@ are run, or the code blocks for @C@ and @D@.
544However, four outside threads inserting into each channel can simultaneously put values into @i@ and attempt to unblock the WUT to run the four code-blocks.
545This case introduces a race with complexity that increases with the size of the @waituntil@ statement.
546However, due to TOCTOU issues, it is impossible to know if all resources are available without acquiring all the internal locks of channels in the subtree of the @waituntil@ clauses.
547This approach is a poor solution for two reasons.
548It is possible that once all the locks are acquired, the subtree is not satisfied and the locks must be released.
549This work incurs a high cost for signalling threads and heavily increase contention on internal channel locks.
550Furthermore, the @waituntil@ statement is polymorphic and can support resources that do not have internal locks, which also makes this approach infeasible.
551As such, the exclusive-or semantics are lost when using both @and@ and @or@ operators since it cannot be supported without significant complexity and significantly affects @waituntil@ performance.
552Therefore, the example of reusing variable @i@ by multiple output channels is considered a user error without exclusive-or semantics.
553Given aliasing in C, it is impossible to even warn of the potential race.
554In the future, it would be interesting to support Go-like syntax, \lstinline[language=Go]{case i := <- ...}, defining a new scoped @i@ variable for each clause.
555
556It was deemed important that exclusive-or semantics are maintained when only @or@ operators are used, so this situation has been special-cased, and is handled by having all clauses race to set a value \emph{before} operating on the channel.
557Consider the following example where thread 1 is reading and threads 2 and 3 are writing to channels @A@ and @B@ concurrently.
558\begin{cquote}
559\begin{tabular}{@{}l|l|l@{}}
560\multicolumn{3}{@{}l@{}}{\lstinline{channel(int) A, B; // zero size channels}} \\
561thread 1 & thread 2 & thread 3 \\
562\begin{cfa}
563waituntil( i << A ) {}
564or waituntil( i << B ) {}
565\end{cfa}
566&
567\begin{cfa}
568A << 1;
569
570\end{cfa}
571&
572\begin{cfa}
573B << 2;
574
575\end{cfa}
576\end{tabular}
577\end{cquote}
578For thread 1 to have exclusive-or semantics, it must only consume from exactly one of @A@ or @B@.
579As such, thread 2 and 3 must race to establish the winning clause of the @waituntil@ in thread 1.
580This race is consolidated by thread 2 and 3 each attempting to set a pointer to the winning clause's @select_node@ address using \gls{cas}.
581The winner bypasses the channel and inserts into the WUT's left-hand, and signals thread 1.
582The loser continues checking if there is space in the channel, and if so performs the channel insert operation with a possible signal of a waiting remove thread;
583otherwise, if there is no space, the loser blocks.
584It is important the race occurs \emph{before} operating on the channel, because channel actions are different with respect to each thread.
585If the race was consolidated after the operation, both thread 2 and 3 could potentially write into @i@ concurrently.
586
587Channels introduce another interesting implementation issue.
588Supporting both reading and writing to a channel in a @waituntil@ means that one @waituntil@ clause may be the notifier of another @waituntil@ clause.
589This conjunction poses a problem when dealing with the special-cased @or@ where the clauses need to win a race to operate on a channel.
590Consider the following example, alongside a described ordering of events to highlight the race.
591\begin{cquote}
592\begin{tabular}{@{}l|l@{}}
593\multicolumn{2}{@{}l@{}}{\lstinline{channel(int) A, B; // zero size channels}} \\
594thread 1 & thread 2 \\
595\begin{cfa}[moredelim={**[is][\color{blue}]{\#}{\#}}]
596waituntil( @i << A@ ) {}
597or waituntil( #i << B# ) {}
598\end{cfa}
599&
600\begin{cfa}[moredelim={**[is][\color{blue}]{\#}{\#}}]
601waituntil( #B << 2# ) {}
602or waituntil( @A << 1@ ) {}
603\end{cfa}
604\end{tabular}
605\end{cquote}
606Assume thread 1 executes first, registers with channel @A@ and proceeds in the @waituntil@.
607Since @A@ is empty, thread 1 cannot remove, and then thread 1 is interrupted before registering with @B@.
608Thread 2 similarly registers with channel @B@, and proceeds in the @waituntil@.
609Since @B@ is zero size there is no space to insert, and then thread 2 is interrupted before registering with @A@.
610At this point, thread 1 and 2 resume execution.
611Remember from above, each exclusive-or @waituntil@ holds a race to set the winning clause of the statement.
612The issue that arises is that these two @waituntil@ statements must have matching winning clauses (both @A@ clauses or both @B@ clauses) to preserve the exclusive-or semantics, since a zero-sized channel needs an insert/remove pair for an operation to occur.
613If threads 1 and 2 race to set a winner only in their own @waituntil@, thread 1 can think it successfully removed from @B@, and thread 2 can think it successfully inserted into @A@, which is an error.
614Hence, there is a race on two fronts.
615If thread 1 wins the race and sees that @B@ has a waiting insertion, then thread 2 must execute the first clause of its @waituntil@ and thread 1 execute its second.
616Correspondingly, if thread 2 wins the race and sees that @A@ has a waiting removal, then thread 1 must execute the first clause of its @waituntil@ and thread 2 execute its second.
617Any other execution scenario is incorrect for exclusive-or semantics.
618Note, that priority execution of multiple satisfied @waituntil@ causes (\ie top to bottom) is not violated because, in this scenario, there is only one satisfied clause for either thread.
619
620The Go @select@ solves this problem by acquiring all the internal locks of the channels before registering the @select@ on the channels.
621This approach eliminates the race shown above since thread 1 and 2 cannot both be registering at the same time.
622However, this approach cannot be used in \CFA, since the @waituntil@ is polymorphic.
623Not all types in a @waituntil@ have an internal lock, and when using non-channel types, acquiring all the locks incurs extra unneeded overhead.
624Instead, this race is consolidated in \CFA in two phases by having an intermediate pending status value for the race.
625This race case is detectable, and if detected, each thread first races to set its own @waituntil@ race pointer to be pending.
626If it succeeds, it then attempts to set the other thread's @waituntil@ race pointer to its success value.
627If either thread successfully sets the the other thread's @waituntil@ race pointer, then the operation can proceed, if not the signalling threads set its own race pointer back to the initial value and repeats.
628This retry mechanism can potentially introduce a livelock, but in practice a livelock here is highly unlikely.
629Furthermore, the likelihood of a livelock here is zero unless the program is in the niche case of having two or more exclusive-or @waituntil@s with two or more clauses in reverse order of priority.
630This livelock case can be fully eliminated using locks like Go, or if a \gls{dcas} instruction is available.
631If any other threads attempt to set a WUT's race pointer and see a pending value, they wait until the value changes before proceeding to ensure that, in the case the WUT fails, the signal is not lost.
632This protocol ensures that signals cannot be lost and that the two races can be resolved in a safe manner.
633The implementation of this protocol is shown in Figure~\ref{f:WU_DeadlockAvoidance}.
634
635\begin{figure}
636\begin{cfa}
637bool pending_set_other( select_node & other, select_node & mine ) {
638    unsigned long int cmp_status = UNSAT;
639
640    // Try to set other status, if we succeed break and return true
641    while( ! CAS( other.clause_status, &cmp_status, SAT ) ) {
642        if ( cmp_status == SAT )
643            return false; // If other status is SAT we lost so return false
644
645        // Toggle own status flag to allow other thread to potentially win
646        mine.status = UNSAT;
647
648        // Reset compare flag
649        cmp_status = UNSAT;
650
651        // Attempt to set own status flag back to PENDING to retry
652        if ( ! CAS( mine.clause_status, &cmp_status, PENDING ) )
653            return false; // If we fail then we lost so return false
654       
655        // Reset compare flag
656        cmp_status = UNSAT;
657    }
658    return true;
659}
660\end{cfa}
661\caption{Exclusive-or \lstinline{waituntil} channel deadlock avoidance protocol}
662\label{f:WU_DeadlockAvoidance}
663\end{figure}
664
665Channels in \CFA have exception-based shutdown mechanisms that the @waituntil@ statement needs to support.
666These exception mechanisms are supported through the @on_selected@ routine.
667This routine is needed by channels to detect if they are closed after unblocking in a @waituntil@ statement, to ensure the appropriate behaviour is taken and an exception is thrown.
668Hence, the channel close-down mechanism is handled correctly.
669
670\subsection{Guards and Statement Predicate}\label{s:wu_guards}
671
672It is trivial to check when a synchronous multiplexing utility is done for the or/xor relationship, since any resource becoming available means that the blocked thread can proceed and the @waituntil@ statement is finished.
673In \uC and \CFA, the \gls{synch_multiplex} mechanism have both an and/or relationship, which along with guards, make the problem of checking for completion of the statement difficult.
674Consider the following @waituntil@.
675\begin{cfa}
676when( GA ) waituntil( A ) {}
677and when( GB ) waituntil( B ) {}
678or when( GC ) waituntil( C ) {}
679\end{cfa}
680When the @waituntil@ thread wakes up, the following predicate represents satisfaction:
681\begin{cfa}
682A && B || C || ! GA && B || ! GB && A || ! GA && ! GB && ! GC
683\end{cfa}
684which can be simplified to:
685\begin{cfa}
686( A || ! GA ) && ( B || ! GB ) || C || ! GA && ! GB && ! GC
687\end{cfa}
688Checking the complete predicate on each iteration of the pending @waituntil@ is expensive so \uC and \CFA both take steps to simplify checking for statement completion.
689
690In the \uC @_Select@ statement, this problem is solved by constructing a tree of the resources, where the internal nodes are operators and the leaves are booleans storing the state of each resource.
691A diagram of the tree for the complete predicate above is shown in Figure~\ref{f:uC_select_tree}, alongside the modification of the tree that occurs when @GA@ is @false@.
692Each internal node stores the statuses of the two subtrees beneath it.
693When resources become available, their corresponding leaf node status is modified, which percolates up the tree to update the state of the statement.
694Once the root of the tree has both subtrees marked as @true@ then the statement is complete.
695As an optimization, when the internal nodes are updated, the subtrees marked as @true@ are pruned and not examined again.
696To support statement guards in \uC, the tree is modified to remove an internal node if a guard is false to maintain the appropriate predicate representation.
697
698\begin{figure}
699\begin{center}
700\input{diagrams/uCpp_select_tree.tikz}
701\end{center}
702\caption{\uC \lstinline[language=uC++]{select} tree modification}
703\label{f:uC_select_tree}
704\end{figure}
705
706The \CFA @waituntil@ statement blocks a thread until a set of resources have become available that satisfy the complete predicate of a @waituntil@.
707The waiting condition of the @waituntil@ statement is implemented as the complete predicate over the resources, joined by the @waituntil@ operators, where a resource is @true@ if it is available, and @false@ otherwise.
708This complete predicate is used as the mechanism to check if a thread is done waiting on a @waituntil@.
709Leveraging the compiler, a predicate routine is generated per @waituntil@ that when passes the statuses of the resources, returns @true@ when the @waituntil@ is done, and false otherwise.
710To support guards on the \CFA @waituntil@ statement, the status of a resource disabled by a guard is set to a boolean value that ensures that the predicate function behaves as if that resource is no longer part of the predicate.
711The generated code allows the predicate that is checked with each iteration to be simplified to not check guard values.
712For example, the following is generated for the complete predicate above:
713\begin{cfa}
714// statement completion predicate
715bool check_completion( select_node * nodes ) {
716    return nodes[0].status && nodes[1].status || nodes[2].status;
717}
718if ( GA || GB || GC ) {                         $\C{// skip statement if all guards false}$
719    select_node nodes[3];
720    nodes[0].status = ! GA && GB;       $\C{// A's status}$
721    nodes[1].status = ! GB && GA;       $\C{// B's status}$
722    nodes[2].status = ! GC;                     $\C{// C's status}$
723    // ... rest of waituntil codegen ...
724}
725\end{cfa}
726
727\uC's @_Select@, supports operators both inside and outside of the \lstinline[language=uC++]{_Select} clauses.
728In the following example, the code blocks run once their corresponding predicate inside the round braces is satisfied.
729\begin{lstlisting}[language=uC++,{moredelim=**[is][\color{red}]{@}{@}}]
730Future_ISM<int> A, B, C, D;
731_Select( @A || B && C@ ) { ... }
732and _Select( @D && E@ ) { ... }
733\end{lstlisting}
734This feature is more expressive than the @waituntil@ statement in \CFA, allowing the code block for @&&@ to only run after \emph{both} resources are available.
735
736In \CFA, since the @waituntil@ statement supports more resources than just futures, implementing operators inside clauses is avoided for a few reasons.
737As a motivating example, suppose \CFA supported operators inside clauses as in:
738\begin{cfa}
739owner_lock A, B, C, D;
740waituntil( A && B ) { ... }
741or waituntil( C && D ) { ... }
742\end{cfa}
743If the @waituntil@ acquires each lock as it becomes available, there is a possible deadlock since it is in a hold-and-wait situation.
744Other semantics are needed to ensure this operation is safe.
745One possibility is to use \CC's @scoped_lock@ approach described in Section~\ref{s:DeadlockAvoidance};
746however, that opens the potential for livelock.
747Another possibility is to use resource ordering similar to \CFA's @mutex@ statement, but that alone is insufficient, if the resource ordering is not used universally.
748One other way this could be implemented is to wait until all resources for a given clause are available before proceeding to acquire them, but this also quickly becomes a poor approach.
749This approach does not work due to \gls{toctou} issues, \ie it is impossible to ensure that the full set of resources are available without holding them all first.
750Operators inside clauses in \CFA could potentially be implemented with careful circumvention of the problems.
751Finally, the problem of operators inside clauses is also a difficult to handle when supporting channels.
752It would require some way to ensure channels used with internal operators are modified, if and only if, the corresponding code block is run.
753However, that is not feasible due to reasons described in the exclusive-or portion of Section~\ref{s:wu_chans}.
754Ultimately, I did not deemed this feature to be crucial enough when taking into account the complexity and runtime cost.
755
756\subsection{The full \lstinline{waituntil} picture}
757
758Given all the complex details handled by the @waituntil@, its full pseudocode is presented in Figure \ref{f:WU_Full_Impl}.
759Some things to note are as follows.
760The @finally@ blocks provide exception-safe \gls{raii} unregistering of nodes, and in particular, the @finally@ inside the innermost loop performs the immediate unregistering required for deadlock-freedom mentioned in Section~\ref{s:wu_locks}.
761The @when_conditions@ array is used to store the boolean result of evaluating each guard at the beginning of the @waituntil@, and it is used to conditionally omit operations on resources with @false@ guards.
762As discussed in Section~\ref{s:wu_chans}, this pseudocode includes conditional code-block execution based on the result of both @on_selected@ and @unregister_select@, which allows the channel implementation to ensure all available channel resources have their corresponding code block run.
763
764\begin{figure}
765\begin{cfa}
766bool when_conditions[N];
767for ( node in nodes )                                                                   $\C[3.75in]{// evaluate guards}$
768        if ( node has guard )
769                when_conditions[node] = node_guard;
770        else
771                when_conditions[node] = true;
772
773if ( any when_conditions[node] are true ) {
774    select_nodes nodes[N];                                                                      $\C{// declare N select nodes}$
775    try {
776        // ... set statuses for nodes with when_conditions[node] == false ...
777
778        for ( node in nodes )                                                           $\C{// register nodes}$
779            if ( when_conditions[node] )
780                register_select( resource, node );
781
782        while ( !check_completion( nodes ) ) {  $\C{// check predicate}$
783            // block
784            for ( resource in waituntil statement ) {   $\C{// run true code blocks}$
785                if ( check_completion( nodes ) ) break;
786                if ( resource is avail )
787                    try {
788                        if( on_selected( resource ) )   $\C{// conditionally run block}$
789                            run code block
790                    } finally                                                   $\C{// for exception safety}$
791                        unregister_select( resource, node ); $\C{// immediate unregister}$
792            }
793        }
794    } finally {                                                                                 $\C{// for exception safety}$
795        for ( registered nodes in nodes )                                       $\C{// deregister nodes}$
796            if ( when_conditions[node] && unregister_select( resource, node )
797                    && on_selected( resource ) )
798                run code block                                                  $\C{// run code block upon unregister}\CRT$
799    }
800}
801\end{cfa}
802\caption{Full \lstinline{waituntil} Pseudocode Implementation}
803\label{f:WU_Full_Impl}
804\end{figure}
805
806\section{\lstinline{waituntil} Performance}
807
808Similar facilities to @waituntil@ are discussed in Section~\ref{s:History} covering C, Ada, Rust, \CC, and OCaml.
809However, these facilities are either not meaningful or feasible to benchmark against.
810The UNIX @select@ and related utilities are not comparable since they are system calls that go into the kernel and operate on file descriptors, whereas the @waituntil@ exists solely in user space.
811Ada's \lstinline[language=Ada]{select} and \uC's \lstinline[language=uC++]{_Accept} only operate on method calls, which is done in \CFA via the @waitfor@ statement.
812Rust and \CC only offer a busy-wait approach, which is not comparable to a blocking approach.
813OCaml's @select@ waits on channels that are not comparable with \CFA and Go channels, so OCaml @select@ is not benchmarked against Go's @select@ and \CFA's @waituntil@.
814
815The two \gls{synch_multiplex} utilities that are in the realm of comparability with the \CFA @waituntil@ statement are the Go \lstinline[language=Go]{select} statement and the \uC \lstinline[language=uC++]{_Select} statement.
816As such, two microbenchmarks are presented, one for Go and one for \uC to contrast this feature.
817Given the differences in features, polymorphism, and expressibility between @waituntil@ and \lstinline[language=Go]{select}, and \uC \lstinline[language=uC++]{_Select}, the aim of the microbenchmarking in this chapter is to show that these implementations lie in the same realm of performance, not to pick a winner.
818
819\subsection{Channel Benchmark}
820
821The channel multiplexing benchmarks compare \CFA's @waituntil@ and Go's \lstinline[language=Go]{select}, where the resource being waited on is a set of channels.
822The basic structure of the benchmark has the number of cores split evenly between producer and consumer threads, \ie, with 8 cores there are 4 producer and 4 consumer threads.
823The number of resource clauses $C$ is also varied across 2, 4, and 8 clauses, where each clause has a different channel that it waits on.
824Each producer and consumer repeatedly waits to either produce or consume from one of the $C$ clauses and respective channels.
825For example, in \CFA syntax, the work loop in the consumer main with $C = 4$ clauses is:
826\begin{cfa}
827for ()
828        waituntil( val << chans[0] ); or waituntil( val << chans[1] );
829        or waituntil( val << chans[2] ); or waituntil( val << chans[3] );
830\end{cfa}
831A successful consumption is counted as a channel operation, and the throughput of these operations is measured over 10 seconds.
832The first benchmark measures throughput of the producers and consumer synchronously waiting on the channels and the second has the threads asynchronously wait on the channels using the Go @default@ and \CFA @else@ clause.
833The results are shown in Figures~\ref{f:select_contend_bench} and~\ref{f:select_spin_bench} respectively.
834
835\begin{figure}
836        \centering
837        \captionsetup[subfloat]{labelfont=footnotesize,textfont=footnotesize}
838        \subfloat[AMD]{
839                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Contend_2.pgf}}
840        }
841        \subfloat[Intel]{
842                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Contend_2.pgf}}
843        }
844        \bigskip
845
846        \subfloat[AMD]{
847                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Contend_4.pgf}}
848        }
849        \subfloat[Intel]{
850                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Contend_4.pgf}}
851        }
852        \bigskip
853
854        \subfloat[AMD]{
855                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Contend_8.pgf}}
856        }
857        \subfloat[Intel]{
858                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Contend_8.pgf}}
859        }
860        \caption{The channel synchronous multiplexing benchmark comparing Go select and \CFA \lstinline{waituntil} statement throughput (higher is better).}
861        \label{f:select_contend_bench}
862\end{figure}
863
864\begin{figure}
865        \centering
866        \captionsetup[subfloat]{labelfont=footnotesize,textfont=footnotesize}
867        \subfloat[AMD]{
868                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Spin_2.pgf}}
869        }
870        \subfloat[Intel]{
871                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Spin_2.pgf}}
872        }
873        \bigskip
874
875        \subfloat[AMD]{
876                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Spin_4.pgf}}
877        }
878        \subfloat[Intel]{
879                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Spin_4.pgf}}
880        }
881        \bigskip
882
883        \subfloat[AMD]{
884                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Spin_8.pgf}}
885        }
886        \subfloat[Intel]{
887                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Spin_8.pgf}}
888        }
889        \caption{The asynchronous multiplexing channel benchmark comparing Go select and \CFA \lstinline{waituntil} statement throughput (higher is better).}
890        \label{f:select_spin_bench}
891\end{figure}
892
893Both Figures~\ref{f:select_contend_bench} and~\ref{f:select_spin_bench} have similar results when comparing \lstinline[language=Go]{select} and @waituntil@.
894In the AMD benchmarks (left column), the performance is very similar as the number of cores scale.
895The AMD machine has a high-caching contention cost because of its \emph{chicklet} L3 cache (\ie many L3 caches servicing a small number of cores), which creates a bottleneck on the channel locks and dominates the shape of the performance curve for both \CFA and Go.
896Hence, it is difficult to detect differences in the \gls{synch_multiplex}, except at low cores, where Go has significantly better performance, due to an optimization in its scheduler.
897Go heavily optimizes thread handoffs on the local run-queue, which can result in very good performance for low numbers of threads parking/unparking each other~\cite{go:sched}.
898In the Intel benchmarks (right column), \CFA performs better than Go as the number of cores scales past 2/4 and as the number of clauses increase.
899This difference is due to Go's acquiring all channel locks when registering and unregistering channels on a \lstinline[language=Go]{select}.
900Go then is holding a lock for every channel, resulting in worse performance as the number of channels increase.
901In \CFA, since races are consolidated without holding all locks, it scales much better both with cores and clauses since more work can occur in parallel.
902This scalability difference is more significant on the Intel machine than the AMD machine since the Intel has lower cache-contention costs.
903
904The Go approach of holding all internal channel-locks in the \lstinline[language=Go]{select} has additional drawbacks.
905There are pathological cases where Go's throughput has significant jitter.
906Consider a producer and consumer thread, @P1@ and @C1@, selecting from both channels @A@ and @B@.
907\begin{cquote}
908\begin{tabular}{@{}ll@{}}
909@P1@ & @C1@ \\
910\begin{cfa}
911waituntil( A << i ); or waituntil( B << i );
912\end{cfa}
913&
914\begin{cfa}
915waituntil( val << A ); or waituntil( val << B );
916\end{cfa}
917\end{tabular}
918\end{cquote}
919Additionally, there is another producer and consumer thread, @P2@ and @C2@, operating solely on @B@.
920\begin{cquote}
921\begin{tabular}{@{}ll@{}}
922@P2@ & @C2@ \\
923\begin{cfa}
924B << val;
925\end{cfa}
926&
927\begin{cfa}
928val << B;
929\end{cfa}
930\end{tabular}
931\end{cquote}
932In Go, this setup results in significantly worse performance since @P2@ and @C2@ cannot operate in parallel with @P1@ and @C1@ due to all locks being acquired.
933Interesting, this case may not be as pathological as it seems.
934If the set of channels belonging to a \lstinline[language=Go]{select} have channels that overlap with the set of another \lstinline[language=Go]{select}, these statements lose the ability to operate in parallel.
935The implementation in \CFA only holds a single lock at a time, resulting in better locking granularity, and hence, more parallelism.
936Comparison of this pathological case is shown in Table~\ref{t:pathGo}.
937The AMD results highlight the worst-case scenario for Go since contention is more costly on this machine than the Intel machine.
938
939\begin{table}[t]
940\centering
941\setlength{\extrarowheight}{2pt}
942\setlength{\tabcolsep}{5pt}
943
944\caption{Throughput (channel operations per second) of \CFA and Go in a pathological case for contention in Go's select implementation}
945\label{t:pathGo}
946\begin{tabular}{r|r|r}
947                        & \multicolumn{1}{c|}{\CFA} & \multicolumn{1}{c}{Go} \\
948        \hline
949        AMD             & \input{data/nasus_Order} \\
950        \hline
951        Intel   & \input{data/pyke_Order}
952\end{tabular}
953\end{table}
954
955Another difference between Go and \CFA is the order of clause selection when multiple clauses are available.
956Go \emph{randomly} selects a clause~\cite{go:select}, but \CFA chooses in the order clauses are listed.
957This \CFA design decision allows users to set implicit priorities, which can result in more predictable behaviour and even better performance.
958In the previous example, threads @P1@ and @C1@ prioritize channel @A@ in the @waituntil@, which can reduce contention for threads @P2@ and @C2@ accessing channel @B@.
959If \CFA did not have priorities, the performance difference in Table~\ref{t:pathGo} would be significant less due to extra contention on channel @B@.
960
961\subsection{Future Benchmark}
962
963The future benchmark compares \CFA's @waituntil@ with \uC's \lstinline[language=uC++]{_Select}, with both utilities waiting on futures.
964While both statements have very similar semantics, supporting @and@ and @or@ operators, \lstinline[language=uC++]{_Select} can only wait on futures, whereas the @waituntil@ is polymorphic.
965As such, the underlying implementation of the operators differs between @waituntil@ and \lstinline[language=uC++]{_Select}.
966The @waituntil@ statement checks for statement completion using a predicate function, whereas the \lstinline[language=uC++]{_Select} statement maintains a tree that represents the state of the internal predicate.
967
968This benchmark aims to indirectly measure the impact of various predicates on the performance of the @waituntil@ and \lstinline[language=uC++]{_Select} statements.
969The benchmark is indirect since the performance of futures in \CFA and \uC differ by a significant margin.
970The experiment has a server, which cycles fulfilling three futures, @A@, @B@, and @C@, and a client, which waits for these futures to be fulfilled using four different kinds of predicates given in \CFA:
971\begin{cquote}
972\begin{tabular}{@{}l|l@{}}
973OR & AND \\
974\hline
975\begin{cfa}
976waituntil( A ) { get( A ); }
977or waituntil( B ) { get( B ); }
978or waituntil( C ) { get( C ); }
979\end{cfa}
980&
981\begin{cfa}
982waituntil( A ) { get( A ); }
983and waituntil( B ) { get( B ); }
984and waituntil( C ) { get( C ); }
985\end{cfa}
986\\
987\multicolumn{2}{@{}c@{}}{} \\
988AND-OR & OR-AND \\
989\hline
990\begin{cfa}
991waituntil( A ) { get( A ); }
992and waituntil( B ) { get( B ); }
993or waituntil( C ) { get( C ); }
994\end{cfa}
995&
996\begin{cfa}
997@(@ waituntil( A ) { get( A ); }
998or waituntil( B ) { get( B ); } @)@
999and waituntil( C ) { get( C ); }
1000\end{cfa}
1001\end{tabular}
1002\end{cquote}
1003The server and client use a low cost synchronize after each fulfillment, so the server does not race ahead of the client.
1004
1005Results of this benchmark are shown in Figure~\ref{f:futurePerf}.
1006Each pair of bars is marked with the predicate name for that experiment and the value at the top of each bar is the standard deviation.
1007In detail, \uC results are lower in all cases due to the performance difference between futures and the more complex \gls{synch_multiplex} implementation.
1008However, the bars for both systems have similar height patterns across the experiments.
1009The @OR@ column for \CFA is more performant than the other \CFA predicates, due to the special-casing of @waituntil@ statements with only @or@ operators.
1010For both \uC and \CFA, the @AND@ experiment is the least performant, which is expected since all three futures need to be fulfilled for each statement completion.
1011Interestingly, \CFA has lower variation across predicates on the AMD (excluding the special OR case), whereas \uC has lower variation on the Intel.
1012Given the differences in semantics and implementation between \uC and \CFA, this test only illustrates the overall costs among the different kinds of predicates.
1013
1014\begin{figure}
1015        \centering
1016        \subfloat[AMD Future Synchronization Benchmark]{
1017                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Future.pgf}}
1018                \label{f:futureAMD}
1019        }
1020        \subfloat[Intel Future Synchronization Benchmark]{
1021                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Future.pgf}}
1022                \label{f:futureIntel}
1023        }
1024        \caption{\CFA \lstinline{waituntil} and \uC \lstinline{_Select} statement throughput synchronizing on a set of futures with varying wait predicates (higher is better).}
1025        \label{f:futurePerf}
1026%%%%%%%%%%%%%%%
1027\vspace*{5.5in}
1028\end{figure}
Note: See TracBrowser for help on using the repository browser.