source: doc/theses/colby_parsons_MMAth/text/channels.tex

Last change on this file was 9509d67a, checked in by caparsons <caparson@…>, 8 months ago

Incorporated changes in response to Trevor's comments.

  • Property mode set to 100644
File size: 31.6 KB
Line 
1% ======================================================================
2% ======================================================================
3\chapter{Channels}\label{s:channels}
4% ======================================================================
5% ======================================================================
6
7Most modern concurrent programming languages do not subscribe to just one style of communication among threads and provide features that support multiple approaches.
8Channels are a concurrent-language feature used to perform \Newterm{message-passing concurrency}: a model of concurrency where threads communicate by sending data as messages (mostly non\-blocking) and synchronizing by receiving sent messages (blocking).
9This model is an alternative to shared-memory concurrency, where threads communicate directly by changing shared state.
10
11Channels were first introduced by Kahn~\cite{Kahn74} and extended by Hoare~\cite{CSP} (CSP).
12Both papers present a pseudo (unimplemented) concurrent language where processes communicate using input/output channels to send data.
13Both languages are highly restrictive.
14Kahn's language restricts a reading process to only wait for data on a single channel at a time and different writing processes cannot send data on the same channel.
15Hoare's language restricts both the sender and receiver to explicitly name the process that is the destination of a channel send or the source of a channel receive.
16These channel semantics remove the ability to have an anonymous sender or receiver.
17Additionally all channel operations in CSP are synchronous (no buffering).
18Advanced channels as a programming language feature has been popularized in recent years by the language Go~\cite{Go}, which encourages the use of channels as its fundamental concurrent feature.
19It was the popularity of Go channels that lead to their implementation in \CFA.
20Neither Go nor \CFA channels have the restrictions of the early channel-based concurrent systems.
21
22Other popular languages and libraries that provide channels include C++ Boost~\cite{boost:channel}, Rust~\cite{rust:channel}, Haskell~\cite{haskell:channel}, OCaml~\cite{ocaml:channel}, and Kotlin~\cite{kotlin:channel}.
23Boost channels only support asynchronous (non-blocking) operations, and Rust channels are limited to only having one consumer per channel.
24Haskell channels are unbounded in size, and OCaml channels are zero-size.
25These restrictions in Haskell and OCaml are likely due to their functional approach, which results in them both using a list as the underlying data structure for their channel.
26These languages and libraries are not discussed further, as their channel implementation is not comparable to the bounded-buffer style channels present in Go and \CFA.
27Kotlin channels are comparable to Go and \CFA, but unfortunately they were not identified as a comparator until after presentation of this thesis and are omitted due to time constraints.
28
29\section{Producer-Consumer Problem}
30A channel is an abstraction for a shared-memory buffer, which turns the implementation of a channel into the producer-consumer problem.
31The producer-consumer problem, also known as the bounded-buffer problem, was introduced by Dijkstra~\cite[\S~4.1]{Dijkstra65}.
32In the problem, threads interact with a buffer in two ways: producing threads insert values into the buffer and consuming threads remove values from the buffer.
33In general, a buffer needs protection to ensure a producer only inserts into a non-full buffer and a consumer only removes from a non-empty buffer (synchronization).
34As well, a buffer needs protection from concurrent access by multiple producers or consumers attempting to insert or remove simultaneously, which is often provided by MX.
35
36\section{Channel Size}\label{s:ChannelSize}
37Channels come in three flavours of buffers:
38\begin{enumerate}
39\item
40Zero sized implies the communication is synchronous, \ie the producer must wait for the consumer to arrive or vice versa for a value to be communicated.
41\item
42Fixed sized (bounded) implies the communication is mostly asynchronous, \ie the producer can proceed up to the buffer size and vice versa for the consumer with respect to removal, at which point the producer/consumer would wait.
43\item
44Infinite sized (unbounded) implies the communication is asymmetrically asynchronous, \ie the producer never waits but the consumer waits when the buffer is empty.
45\end{enumerate}
46
47In general, the order values are processed by the consumer does not affect the correctness of the producer-consumer problem.
48For example, the buffer can be \gls{lifo}, \gls{fifo}, or prioritized with respect to insertion and removal.
49However, like MX, a buffer should ensure every value is eventually removed after some reasonable bounded time (no long-term starvation).
50The simplest way to prevent starvation is to implement the buffer as a queue, either with a cyclic array or linked nodes.
51While \gls{fifo} is not required for producer-consumer problem correctness, it is a desired property in channels as it provides predictable and often relied upon channel ordering behaviour to users.
52
53\section{First-Come First-Served}
54As pointed out, a bounded buffer implementation often provides MX among multiple producers or consumers.
55This MX should be fair among threads, independent of the \gls{fifo} buffer being fair among values.
56Fairness among threads is called \gls{fcfs} and was defined by Lamport~\cite[p.~454]{Lamport74}.
57\gls{fcfs} is defined in relation to a doorway~\cite[p.~330]{Lamport86II}, which is the point at which an ordering among threads can be established.
58Given this doorway, a CS is said to be \gls{fcfs}, if threads access the shared resource in the order they proceed through the doorway.
59A consequence of \gls{fcfs} execution is the elimination of \Newterm{barging}, where barging means a thread arrives at a CS with waiting threads, and the MX protecting the CS allows the arriving thread to enter the CS ahead of one or more of the waiting threads.
60
61\gls{fcfs} is a fairness property that prevents unequal access to the shared resource and prevents starvation, however it comes at a cost.
62Implementing an algorithm with \gls{fcfs} can lead to \Newterm{double blocking}, where arriving threads block outside the doorway waiting for a thread in the lock entry-protocol and inside the doorway waiting for a thread in the CS.
63An analogue is boarding an airplane: first you wait to get through security to the departure gates (short term), and then wait again at the departure gate for the airplane (long term).
64As such, algorithms that are not \gls{fcfs} (barging) can be more performant by skipping the wait for the CS and entering directly;
65however, this performance gain comes by introducing unfairness with possible starvation for waiting threads.
66
67\section{Channel Implementation}\label{s:chan_impl}
68The programming languages Go, Kotlin, and Erlang provide user-level threading where the primary communication mechanism is channels.
69These languages have user-level threading and preemptive scheduling, and both use channels for communication.
70Go and Kotlin provide multiple homogeneous channels; each have a single associated type.
71Erlang, which is closely related to actor systems, provides one heterogeneous channel per thread (mailbox) with a typed receive pattern.
72Go and Kotlin encourage users to communicate via channels, but provides them as an optional language feature.
73On the other hand, Erlang's single heterogeneous channel is a fundamental part of the threading system design; using it is unavoidable.
74Similar to Go and Kotlin, \CFA's channels are offered as an optional language feature.
75
76While iterating on channel implementation, experiments were conducted that varied the producer-consumer algorithm and lock type used inside the channel.
77With the exception of non-\gls{fcfs} or non-\gls{fifo} algorithms, no algorithm or lock usage in the channel implementation was found to be consistently more performant that Go's choice of algorithm and lock implementation.
78Performance of channels can be improved by sharding the underlying buffer \cite{Dice11}.
79However, the \gls{fifo} property is lost, which is undesirable for user-facing channels.
80Therefore, the low-level channel implementation in \CFA is largely copied from the Go implementation, but adapted to the \CFA type and runtime systems.
81As such the research contributions added by \CFA's channel implementation lie in the realm of safety and productivity features.
82
83The Go channel implementation utilizes cooperation among threads to achieve good performance~\cite{go:chan}.
84This cooperation only occurs when producers or consumers need to block due to the buffer being full or empty.
85After a producer blocks it must wait for a consumer to signal it and vice versa.
86The consumer or producer that signals a blocked thread is called the signalling thread.
87In these cases, a blocking thread stores their relevant data in a shared location and the signalling thread completes the blocking thread's operation before waking them;
88\ie the blocking thread has no work to perform after it unblocks because the signalling threads has done this work.
89This approach is similar to wait morphing for locks~\cite[p.~82]{Butenhof97} and improves performance in a few ways.
90First, each thread interacting with the channel only acquires and releases the internal channel lock once.
91As a result, contention on the internal lock is decreased; only entering threads compete for the lock since unblocking threads do not reacquire the lock.
92The other advantage of Go's wait-morphing approach is that it eliminates the need to wait for signalled threads to run.
93Note that the property of acquiring/releasing the lock only once can also be achieved with a different form of cooperation, called \Newterm{baton passing}.
94Baton passing occurs when one thread acquires a lock but does not release it, and instead signals a thread inside the critical section, conceptually ``passing'' the mutual exclusion from the signalling thread to the signalled thread.
95The baton-passing approach has threads cooperate to pass mutual exclusion without additional lock acquires or releases;
96the wait-morphing approach has threads cooperate by completing the signalled thread's operation, thus removing a signalled thread's need for mutual exclusion after unblocking.
97While baton passing is useful in some algorithms, it results in worse channel performance than the Go approach.
98In the baton-passing approach, all threads need to wait for the signalled thread to unblock and run before other operations on the channel can proceed, since the signalled thread holds mutual exclusion;
99in the wait-morphing approach, since the operation is completed before the signal, other threads can continue to operate on the channel without waiting for the signalled thread to run.
100
101In this work, all channel sizes \see{Sections~\ref{s:ChannelSize}} are implemented with bounded buffers.
102However, only non-zero-sized buffers are analysed because of their complexity and higher usage.
103
104\section{Safety and Productivity}
105Channels in \CFA come with safety and productivity features to aid users.
106The features include the following.
107
108\begin{itemize}
109\item Toggle-able statistic collection on channel behaviour that count channel and blocking operations.
110Tracking blocking operations helps illustrate usage for tuning the channel size, where the aim is to reduce blocking.
111
112\item Deadlock detection on channel deallocation.
113If threads are blocked inside a channel when it terminates, this case is detected and the user is informed, as this can cause a deadlock.
114
115\item A @flush@ routine that delivers copies of an element to all waiting consumers, flushing the buffer.
116Programmers use this mechanism to broadcast a sentinel value to multiple consumers.
117Additionally, the @flush@ routine is more performant than looping around the @insert@ operation since it can deliver the elements without having to reacquire mutual exclusion for each element sent.
118
119\item Go-style @?<<?@ shorthand operator for inserting and removing.
120\begin{cfa}
121channel(int) chan;
122int i = 2;
123chan << i;                      $\C{// insert i into chan}$
124i << chan;              $\C{// remove element from chan into i}$
125\end{cfa}
126\end{itemize}
127
128\subsection{Toggle-able Statistics}
129As discussed, a channel is a concurrent layer over a bounded buffer.
130To achieve efficient buffering, users should aim for as few blocking operations on a channel as possible.
131Mechanisms to reduce blocking are: change the buffer size, shard a channel into multiple channels, or tweak the number of producer and consumer threads.
132For users to be able to make informed decisions when tuning channel usage, toggle-able channel statistics are provided.
133The statistics are toggled on during the \CFA build by defining the @CHAN_STATS@ macro, which guarantees zero cost when not using this feature.
134When statistics are turned on, four counters are maintained per channel, two for inserting (producers) and two for removing (consumers).
135The two counters per type of operation track the number of blocking operations and total operations.
136In the channel destructor, the counters are printed out aggregated and also per type of operation.
137An example use case is noting that producer inserts are blocking often while consumer removes do not block often.
138This information can be used to increase the number of consumers to decrease the blocking producer operations, thus increasing the channel throughput.
139Whereas, increasing the channel size in this scenario is unlikely to produce a benefit because the consumers can never keep up with the producers.
140
141\subsection{Deadlock Detection}
142The deadlock detection in the \CFA channels is fairly basic but detects a very common channel mistake during termination.
143That is, it detects the case where threads are blocked on the channel during channel deallocation.
144This case is guaranteed to deadlock since there are no other threads to supply or consume values needed by the waiting threads.
145Only if a user maintained a separate reference to the blocked threads and manually unblocks them outside the channel could the deadlock be avoid.
146However, without special semantics, this unblocking would generate other runtime errors where the unblocked thread attempts to access non-existing channel data or even a deallocated channel.
147More robust deadlock detection needs to be implemented separate from channels since it requires knowledge about the threading system and other channel/thread state.
148
149\subsection{Program Shutdown}
150Terminating concurrent programs is often one of the most difficult parts of writing concurrent code, particularly if graceful termination is needed.
151Graceful termination can be difficult to achieve with synchronization primitives that need to be handled carefully during shutdown.
152It is easy to deadlock during termination if threads are left behind on synchronization primitives.
153Additionally, most synchronization primitives are prone to \gls{toctou} issues where there is race between one thread checking the state of a concurrent object and another thread changing the state.
154\gls{toctou} issues with synchronization primitives often involve a race between one thread checking the primitive for blocked threads and another thread blocking on it.
155Channels are a particularly hard synchronization primitive to terminate since both sending and receiving to/from a channel can block.
156Thus, improperly handled \gls{toctou} issues with channels often result in deadlocks as threads performing the termination may end up unexpectedly blocking in their attempt to help other threads exit the system.
157
158\subsubsection{Go Channel Close}
159Go channels provide a set of tools to help with concurrent shutdown~\cite{go:chan} using a @close@ operation in conjunction with the \Go{select} statement.
160The \Go{select} statement is discussed in \ref{s:waituntil}, where \CFA's @waituntil@ statement is compared with the Go \Go{select} statement.
161
162The @close@ operation on a channel in Go changes the state of the channel.
163When a channel is closed, sends to the channel panic along with additional calls to @close@.
164Receives are handled differently.
165Receivers (consumers) never block on a closed channel and continue to remove elements from the channel.
166Once a channel is empty, receivers can continue to remove elements, but receive the zero-value version of the element type.
167To avoid unwanted zero-value elements, Go provides the ability to iterate over a closed channel to remove the remaining elements.
168These Go design choices enforce a specific interaction style with channels during termination: careful thought is needed to ensure additional @close@ calls do not occur and no sends occur after a channel is closed.
169These design choices fit Go's paradigm of error management, where users are expected to explicitly check for errors, rather than letting errors occur and catching them.
170If errors need to occur in Go, return codes are used to pass error information up call levels.
171Note that panics in Go can be caught, but it is not the idiomatic way to write Go programs.
172
173While Go's channel-closing semantics are powerful enough to perform any concurrent termination needed by a program, their lack of ease of use leaves much to be desired.
174Since both closing and sending panic once a channel is closed, a user often has to synchronize the senders (producers) before the channel can be closed to avoid panics.
175However, in doing so it renders the @close@ operation nearly useless, as the only utilities it provides are the ability to ensure receivers no longer block on the channel and receive zero-valued elements.
176This functionality is only useful if the zero-typed element is recognized as a sentinel value, but if another sentinel value is necessary, then @close@ only provides the non-blocking feature.
177To avoid \gls{toctou} issues during shutdown, a busy wait with a \Go{select} statement is often used to add or remove elements from a channel.
178Hence, due to Go's asymmetric approach to channel shutdown, separate synchronization between producers and consumers of a channel has to occur during shutdown.
179
180\subsubsection{\CFA Channel Close}
181\CFA channels have access to an extensive exception handling mechanism~\cite{Beach21}.
182As such \CFA uses an exception-based approach to channel shutdown that is symmetric for both producers and consumers, and supports graceful shutdown.
183
184Exceptions in \CFA support both termination and resumption.
185\Newterm{Termination exception}s perform a dynamic call that unwinds the stack preventing the exception handler from returning to the raise point, such as in \CC, Python and Java.
186\Newterm{Resumption exception}s perform a dynamic call that does not unwind the stack allowing the exception handler to return to the raise point.
187In \CFA, if a resumption exception is not handled, it is reraised as a termination exception.
188This mechanism is used to create a flexible and robust termination system for channels.
189
190When a channel in \CFA is closed, all subsequent calls to the channel raise a resumption exception at the caller.
191If the resumption is handled, the caller attempts to complete the channel operation.
192However, if the channel operation would block, a termination exception is thrown.
193If the resumption is not handled, the exception is rethrown as a termination.
194These termination exceptions allow for non-local transfer that is used to great effect to eagerly and gracefully shut down a thread.
195When a channel is closed, if there are any blocked producers or consumers inside the channel, they are woken up and also have a resumption thrown at them.
196The resumption exception, @channel_closed@, has internal fields to aid in handling the exception.
197The exception contains a pointer to the channel it is thrown from and a pointer to a buffer element.
198For exceptions thrown from @remove@, the buffer element pointer is null.
199For exceptions thrown from @insert@, the element pointer points to the buffer element that the thread attempted to insert.
200Utility routines @bool is_insert( channel_closed & e );@ and @bool is_remove( channel_closed & e );@ are provided for convenient checking of the element pointer.
201This element pointer allows the handler to know which operation failed and also allows the element to not be lost on a failed insert since it can be moved elsewhere in the handler.
202Furthermore, due to \CFA's powerful exception system, this data can be used to choose handlers based on which channel and operation failed.
203For example, exception handlers in \CFA have an optional predicate which can be used to trigger or skip handlers based on the content of the matching exception.
204It is worth mentioning that using exceptions for termination may incur a larger performance cost than the Go approach.
205However, this should not be an issue, since termination is rarely on the fast-path of an application.
206In contrast, ensuring termination can be easily implemented correctly is the aim of the exception approach.
207
208\section{\CFA / Go channel Examples}
209To highlight the differences between \CFA's and Go's close semantics, two examples are presented.
210The first example is a simple shutdown case, where there are producer threads and consumer threads operating on a channel for a fixed duration.
211Once the duration ends, producers and consumers terminate immediately leaving unprocessed elements in the channel.
212The second example extends the first by requiring the channel to be empty after shutdown.
213Both the first and second example are shown in Figure~\ref{f:ChannelTermination}.
214
215\begin{figure}
216\centering
217
218\begin{lrbox}{\myboxA}
219\begin{Golang}[aboveskip=0pt,belowskip=0pt]
220var channel chan int = make( chan int, 128 )
221var prodJoin chan int = make( chan int, 4 )
222var consJoin chan int = make( chan int, 4 )
223var cons_done, prod_done bool = false, false;
224func producer() {
225        for {
226                if prod_done { break }
227                channel <- 5
228        }
229        prodJoin <- 0 // synch with main thd
230}
231
232func consumer() {
233        for {
234                if cons_done { break }
235                <- channel
236        }
237        consJoin <- 0 // synch with main thd
238}
239
240
241func main() {
242        for j := 0; j < 4; j++ { go consumer() }
243        for j := 0; j < 4; j++ { go producer() }
244        time.Sleep( time.Second * 10 )
245        prod_done = true
246        for j := 0; j < 4 ; j++ { <- prodJoin }
247        cons_done = true
248        close(channel) // ensure no cons deadlock
249        @for elem := range channel {@
250                // process leftover values
251        @}@
252        for j := 0; j < 4; j++ { <- consJoin }
253}
254\end{Golang}
255\end{lrbox}
256
257\begin{lrbox}{\myboxB}
258\begin{cfa}[aboveskip=0pt,belowskip=0pt]
259channel( int ) chan{ 128 };
260thread Consumer {};
261thread Producer {};
262
263void main( Producer & this ) {
264        try {
265                for ()
266                        chan << 5;
267        } catch( channel_closed * ) {
268                // unhandled resume or full
269        }
270}
271void main( Consumer & this ) {
272        int i;
273        try {
274                for () { i << chan; }
275        @} catchResume( channel_closed * ) {@
276                // handled resume => consume from chan
277        } catch( channel_closed * ) {
278                // empty or unhandled resume
279        }
280}
281int main() {
282        Consumer c[4];
283        Producer p[4];
284        sleep( 10`s );
285        close( chan );
286}
287
288
289
290
291
292
293\end{cfa}
294\end{lrbox}
295
296\subfloat[Go style]{\label{l:go_chan_term}\usebox\myboxA}
297\hspace*{3pt}
298\vrule
299\hspace*{3pt}
300\subfloat[\CFA style]{\label{l:cfa_chan_term}\usebox\myboxB}
301\caption{Channel Termination Examples 1 and 2. Code specific to example 2 is highlighted.}
302\label{f:ChannelTermination}
303\end{figure}
304
305Figure~\ref{l:go_chan_term} shows the Go solution.
306Since some of the elements being passed through the channel are zero-valued, closing the channel in Go does not aid in communicating shutdown.
307Instead, a different mechanism to communicate with the consumers and producers needs to be used.
308Flag variables are common in Go-channel shutdown-code to avoid panics on a channel, meaning the channel shutdown has to be communicated with threads before it occurs.
309Hence, the two flags @cons_done@ and @prod_done@ are used to communicate with the producers and consumers, respectively.
310Furthermore, producers and consumers need to shutdown separately to ensure that producers terminate before the channel is closed to avoid panicking, and to avoid the case where all the consumers terminate first, which can result in a deadlock for producers if the channel is full.
311The producer flag is set first;
312then after all producers terminate, the consumer flag is set and the channel is closed leaving elements in the buffer.
313To purge the buffer, a loop is added (red) that iterates over the closed channel to process any remaining values.
314
315Figure~\ref{l:cfa_chan_term} shows the \CFA solution.
316Here, shutdown is communicated directly to both producers and consumers via the @close@ call.
317A @Producer@ thread knows to stop producing when the @insert@ call on a closed channel raises exception @channel_closed@.
318If a @Consumer@ thread ignores the first resumption exception from the @close@, the exception is reraised as a termination exception and elements are left in the buffer.
319If a @Consumer@ thread handles the resumptions exceptions (red), control returns to complete the remove.
320A @Consumer@ thread knows to stop consuming after all elements of a closed channel are removed and the consumer would block, which causes a termination raise of @channel_closed@.
321The \CFA semantics allow users to communicate channel shutdown directly through the channel, without having to share extra state between threads.
322Additionally, when the channel needs to be drained, \CFA provides users with easy options for processing the leftover channel values in the main thread or in the consumer threads.
323
324Figure~\ref{f:ChannelBarrierTermination} shows a final shutdown example using channels to implement a barrier.
325A Go and \CFA style solution are presented but both are implemented using \CFA syntax so they can be easily compared.
326Implementing a barrier is interesting because threads are both producers and consumers on the barrier-internal channels, @entryWait@ and @barWait@.
327The outline for the barrier implementation starts by initially filling the @entryWait@ channel with $N$ tickets in the barrier constructor, allowing $N$ arriving threads to remove these values and enter the barrier.
328After @entryWait@ is empty, arriving threads block when removing.
329However, the arriving threads that entered the barrier cannot leave the barrier until $N$ threads have arrived.
330Hence, the entering threads block on the empty @barWait@ channel until the $N$th arriving thread inserts $N-1$ elements into @barWait@ to unblock the $N-1$ threads calling @remove@.
331The race between these arriving threads blocking on @barWait@ and the $N$th thread inserting values into @barWait@ does not affect correctness;
332\ie an arriving thread may or may not block on channel @barWait@ to get its value.
333Finally, the last thread to remove from @barWait@ with ticket $N-2$, refills channel @entryWait@ with $N$ values to start the next group into the barrier.
334
335Now, the two channels makes termination synchronization between producers and consumers difficult.
336Interestingly, the shutdown details for this problem are also applicable to other problems with threads producing and consuming from the same channel.
337The Go-style solution cannot use the Go @close@ call since all threads are both potentially producers and consumers, causing panics on close to be unavoidable without complex synchronization.
338As such in Figure \ref{l:go_chan_bar}, a flush routine is needed to insert a sentinel value, @-1@, to inform threads waiting in the buffer they need to leave the barrier.
339This sentinel value has to be checked at two points along the fast-path and sentinel values daisy-chained into the buffers.
340Furthermore, an additional flag @done@ is needed to communicate to threads once they have left the barrier that they are done.
341Also note that in the Go version~\ref{l:go_chan_bar}, the size of the barrier channels has to be larger than in the \CFA version to ensure that the main thread does not block when attempting to clear the barrier.
342For The \CFA solution~\ref{l:cfa_chan_bar}, the barrier shutdown results in an exception being thrown at threads operating on it, to inform waiting threads they must leave the barrier.
343This avoids the need to use a separate communication method other than the barrier, and avoids extra conditional checks on the fast path of the barrier implementation.
344
345\begin{figure}
346\centering
347
348\begin{lrbox}{\myboxA}
349\begin{cfa}[aboveskip=0pt,belowskip=0pt]
350struct barrier {
351        channel( int ) barWait, entryWait;
352        int size;
353};
354void ?{}( barrier & this, int size ) with(this) {
355        barWait{size + 1};   entryWait{size + 1};
356        this.size = size;
357        for ( i; size )
358                insert( entryWait, i );
359}
360void wait( barrier & this ) with(this) {
361        int ticket = remove( entryWait );
362        @if ( ticket == -1 ) { insert( entryWait, -1 ); return; }@
363        if ( ticket == size - 1 ) {
364                for ( i; size - 1 )
365                        insert( barWait, i );
366                return;
367        }
368        ticket = remove( barWait );
369        @if ( ticket == -1 ) { insert( barWait, -1 ); return; }@
370        if ( size == 1 || ticket == size - 2 ) { // last ?
371                for ( i; size )
372                        insert( entryWait, i );
373        }
374}
375void flush(barrier & this) with(this) {
376        @insert( entryWait, -1 );   insert( barWait, -1 );@
377}
378enum { Threads = 4 };
379barrier b{Threads};
380@bool done = false;@
381thread Thread {};
382void main( Thread & this ) {
383        for () {
384          @if ( done ) break;@
385                wait( b );
386        }
387}
388int main() {
389        Thread t[Threads];
390        sleep(10`s);
391        done = true;
392        flush( b );
393} // wait for threads to terminate
394\end{cfa}
395\end{lrbox}
396
397\begin{lrbox}{\myboxB}
398\begin{cfa}[aboveskip=0pt,belowskip=0pt]
399struct barrier {
400        channel( int ) barWait, entryWait;
401        int size;
402};
403void ?{}( barrier & this, int size ) with(this) {
404        barWait{size};   entryWait{size};
405        this.size = size;
406        for ( i; size )
407                insert( entryWait, i );
408}
409void wait( barrier & this ) with(this) {
410        int ticket = remove( entryWait );
411
412        if ( ticket == size - 1 ) {
413                for ( i; size - 1 )
414                        insert( barWait, i );
415                return;
416        }
417        ticket = remove( barWait );
418
419        if ( size == 1 || ticket == size - 2 ) { // last ?
420                for ( i; size )
421                        insert( entryWait, i );
422        }
423}
424void flush(barrier & this) with(this) {
425        @close( barWait );   close( entryWait );@
426}
427enum { Threads = 4 };
428barrier b{Threads};
429
430thread Thread {};
431void main( Thread & this ) {
432        @try {@
433                for ()
434                        wait( b );
435        @} catch ( channel_closed * ) {}@
436}
437int main() {
438        Thread t[Threads];
439        sleep(10`s);
440
441        flush( b );
442} // wait for threads to terminate
443\end{cfa}
444\end{lrbox}
445
446\subfloat[Go style]{\label{l:go_chan_bar}\usebox\myboxA}
447\hspace*{3pt}
448\vrule
449\hspace*{3pt}
450\subfloat[\CFA style]{\label{l:cfa_chan_bar}\usebox\myboxB}
451\caption{Channel Barrier Termination}
452\label{f:ChannelBarrierTermination}
453\end{figure}
454
455\section{Performance}
456
457Given that the base implementation of the \CFA channels is very similar to the Go implementation, this section aims to show the performance of the two implementations are comparable.
458The microbenchmark for the channel comparison is similar to Figure~\ref{f:ChannelTermination}, where the number of threads and processors is set from the command line.
459The processors are divided equally between producers and consumers, with one producer or consumer owning each core.
460The number of cores is varied to measure how throughput scales.
461
462The results of the benchmark are shown in Figure~\ref{f:chanPerf}.
463The performance of Go and \CFA channels on this microbenchmark is comparable.
464Note that the performance should decline as the number of cores increases as the channel operations occur in a critical section, so increasing cores results in higher contention with no increase in parallelism.
465
466The performance of \CFA and Go's shutdown mechanisms is not measured, as shutdown is an exceptional case that does not occur frequently in most programs. Additionally, it is difficult to measure channel shutdown performance; threads need to be synchronized between each subsequent shutdown, which is likely more expensive than the shutdown mechanism itself.
467
468\begin{figure}
469        \centering
470        \subfloat[AMD Channel Benchmark]{
471                \resizebox{0.5\textwidth}{!}{\input{figures/nasus_Channel_Contention.pgf}}
472                \label{f:chanAMD}
473        }
474        \subfloat[Intel Channel Benchmark]{
475                \resizebox{0.5\textwidth}{!}{\input{figures/pyke_Channel_Contention.pgf}}
476                \label{f:chanIntel}
477        }
478        \caption{The channel contention benchmark comparing \CFA and Go channel throughput (higher is better).}
479        \label{f:chanPerf}
480\end{figure}
481
482% Local Variables: %
483% tab-width: 4 %
484% End: %
Note: See TracBrowser for help on using the repository browser.