source: doc/theses/colby_parsons_MMAth/text/actors.tex @ 60c3d87e

Last change on this file since 60c3d87e was 60c3d87e, checked in by caparsons <caparson@…>, 13 months ago

intermediate commit to pull actor changes

  • Property mode set to 100644
File size: 74.3 KB
Line 
1% ======================================================================
2% ======================================================================
3\chapter{Actors}\label{s:actors}
4% ======================================================================
5% ======================================================================
6
7% C_TODO: add citations throughout chapter
8Actors are an indirect concurrent feature that abstracts threading away from a programmer, and instead provides \gls{actor}s and messages as building blocks for concurrency, where message passing means there is no shared data to protect, making actors amenable in a distributed environment.
9Actors are another message passing concurrency feature, similar to channels but with more abstraction, and are in the realm of \gls{impl_concurrency}, where programmers write concurrent code without dealing with explicit thread creation or interaction.
10The study of actors can be broken into two concepts, the \gls{actor_model}, which describes the model of computation and the \gls{actor_system}, which refers to the implementation of the model.
11Before discussing \CFA's actor system in detail, it is important to first describe the actor model, and the classic approach to implementing an actor system.
12
13\section{Actor Model}
14The actor model is a concurrent paradigm where computation is broken into units of work called actors, and the data for computation is distributed to actors in the form of messages~\cite{Hewitt73}.
15An actor is composed of a \Newterm{mailbox} (message queue) and a set of \Newterm{behaviours} that receive from the mailbox to perform work.
16Actors execute asynchronously upon receiving a message and can modify their own state, make decisions, spawn more actors, and send messages to other actors.
17Because the actor model is implicit concurrency, its strength is that it abstracts away many details and concerns needed in other concurrent paradigms.
18For example, mutual exclusion and locking are rarely relevant concepts in an actor model, as actors typically only operate on local state.
19
20An actor does not have a thread.
21An actor is executed by an underlying \Newterm{executor} (kernel thread-pool) that fairly invokes each actor, where an actor invocation processes one or more messages from its mailbox.
22The default number of executor threads is often proportional to the number of computer cores to achieve good performance.
23An executor is often tunable with respect to the number of kernel threads and its scheduling algorithm, which optimize for specific actor applications and workloads \see{end of Section~\ref{s:CFAActorSyntax}}.
24
25\section{Classic Actor System}
26An implementation of the actor model with a community of actors is called an actor system.
27Actor systems largely follow the actor model, but can differ in some ways.
28While the semantics of message \emph{send} is asynchronous, the implementation may be synchronous or a combination.
29The default semantics for message \emph{receive} is FIFO, so an actor receives messages from its mailbox in temporal (arrival) order;
30however, messages sent among actors arrive in any order.
31Some actor systems provide priority-based mailboxes and/or priority-based message-selection within a mailbox, where custom message dispatchers search among or within a mailbox(es) with a predicate for specific kinds of actors and/or messages.
32Some actor systems provide a shared mailbox where multiple actors receive from a common mailbox~\cite{Akka}, which is contrary to the no-sharing design of the basic actor-model (and requires additional locking).
33For non-FIFO service, some notion of fairness (eventual progress) must exist, otherwise messages have a high latency or starve, \ie never received.
34Finally, some actor systems provide multiple typed-mailboxes, which then lose the actor-\lstinline{become} mechanism (see Section~\ref{s:SafetyProductivity}).
35%While the definition of the actor model provides no restrictions on message ordering, actor systems tend to guarantee that messages sent from a given actor $i$ to actor $j$ will arrive at actor $j$ in the order they were sent.
36Another way an actor system varies from the model is allowing access to shared global-state.
37When this occurs, it complicates the implementation as this breaks any implicit mutual-exclusion guarantees when only accessing local-state.
38
39\begin{figure}
40\begin{tabular}{l|l}
41\subfloat[Actor-centric system]{\label{f:standard_actor}\input{diagrams/standard_actor.tikz}} &
42\subfloat[Message-centric system]{\label{f:inverted_actor}\raisebox{.1\height}{\input{diagrams/inverted_actor.tikz}}}
43\end{tabular}
44\caption{Classic and inverted actor implementation approaches with sharded queues.}
45\end{figure}
46
47\section{\CFA Actors}
48Figure~\ref{f:standard_actor} shows an actor system designed as \Newterm{actor-centric}, where a set of actors are scheduled and run on underlying executor threads~\cite{CAF,Akka,ProtoActor}.
49The simplest design has a single global queue of actors accessed by the executor threads, but this approach results in high contention as both ends of the queue by the executor threads.
50The more common design is to \Newterm{shard} the single queue among the executor threads, where actors are permanently assigned or can float among the queues.
51Sharding significantly decreases contention among executor threads adding and removing actors to/from a queue.
52Finally, each actor has a receive queue of messages (mailbox), which is a single consumer, multi-producer queue, \ie only the actor removes from the mailbox but multiple actors can attach messages.
53When an actor receives a message in its mailbox, the actor is marked ready and scheduled by a thread to run the actor's current work unit on the message(s).
54
55% cite parallel theatre and our paper
56Figure \ref{f:inverted_actor} shows an actor system designed as \Newterm{message-centric}, where a set of messages are scheduled and run on underlying executor threads~\cite{uC++,Nigro21}.
57Again, the simplest design has a single global queue of messages accessed by the executor threads, but this approach has the same contention problem by the executor threads.
58Therefore, the messages (mailboxes) are sharded and executor threads schedule each message, which points to its corresponding actor.
59Here, an actor's messages are permanently assigned to one queue to ensure FIFO receiving and/or reduce searching for specific actor/messages.
60Since multiple actors belong to each message queue, actor messages are interleaved on a queue.
61This design is \Newterm{inverted} because actors belong to a message queue, whereas in the classic approach a message queue belongs to each actor.
62% In this inverted actor system instead of each executor threads owning a queue of actors, they each own a queue of messages.
63% In this scheme work is consumed from their queue and executed by underlying threads.
64The inverted model can be taken a step further by sharding the message queues for each executor threads, so each executor thread owns a set of queues and cycles through them.
65Again, this extra level of sharding is to reduce queue contention.
66% The arrows from the message queues to the actors in the diagram indicate interleaved messages addressed to each actor.
67
68The actor system in \CFA uses a message-centric design, adopts several features from my prior actor work in \uC~\cite{}, and adds the following contributions related to \CFA:
69\begin{enumerate}[topsep=5pt,itemsep=3pt,parsep=0pt]
70\item
71Provide insight into the impact of envelope allocation in actor systems.
72In all actor systems, dynamic allocation is needed to ensure the lifetime of a unit of work persists from its creation until the unit of work is executed.
73This allocation is often called an \Newterm{envelope} as it ``packages'' the information needed to run the unit of work, alongside any other information needed to send the unit of work, such as an actor's address or link fields.
74This dynamic allocation occurs once per message sent.
75Unfortunately, the high rate of message sends in an actor system results in significant contention on the memory allocator.
76A novel data structure is introduced to consolidate allocations to improve performance by minimizing allocator contention.
77
78\item
79Improve performance of the inverted actor system using multiple approaches to minimize contention on queues, such as queue gulping and avoiding atomic operations.
80
81\item
82Introduce work stealing in the inverted actor system.
83Work stealing in an actor-centric system involves stealing one or more actors among executor threads.
84In the inverted system, the notion of stealing message queues is introduced.
85The queue stealing is implemented such that the act of stealing work does not contend with non-stealing executor threads running actors.
86
87\item
88Introduce and evaluate a timestamp-based work-stealing heuristic with the goal of maintaining non-workstealing performance in work-saturated workloads and improving performance on unbalanced workloads.
89
90\item
91Provide a suite of safety and productivity features including static-typing, detection of erroneous message sends, statistics tracking, and more.
92\end{enumerate}
93
94\section{\CFA Actor Syntax}\label{s:CFAActorSyntax}
95\CFA is not an object oriented language and it does not have \gls{rtti}.
96As such, all message sends and receives among actors can only occur using static type-matching, as in Typed-Akka~\cite{AkkaTyped}.
97Figure~\ref{f:BehaviourStyles} contrasts dynamic and static type-matching.
98Figure~\ref{l:dynamic_style} shows the dynamic style with a heterogeneous message receive and an indirect dynamic type-discrimination for message processing.
99Figure~\ref{l:static_style} shows the static style with a homogeneous message receive and a direct static type-discrimination for message processing.
100The static-typing style is safer because of the static check and faster because there is no dynamic type-discrimination.
101The dynamic-typing style is more flexible because multiple kinds of messages can be handled in a behaviour condensing the processing code.
102
103\begin{figure}
104\centering
105
106\begin{lrbox}{\myboxA}
107\begin{cfa}[morekeywords=case]
108allocation receive( message & msg ) {
109        case( @msg_type1@, msg ) {      // discriminate type
110                ... msg_d-> ...;        // msg_type1 msg_d
111        } else case( @msg_type2@, msg ) {
112                ... msg_d-> ...;        // msg_type2 msg_d
113        ...
114}
115\end{cfa}
116\end{lrbox}
117
118\begin{lrbox}{\myboxB}
119\begin{cfa}
120allocation receive( @msg_type1@ & msg ) {
121        ... msg ...;
122}
123allocation receive( @msg_type2@ & msg ) {
124        ... msg ...;
125}
126...
127\end{cfa}
128\end{lrbox}
129\subfloat[dynamic typing]{\label{l:dynamic_style}\usebox\myboxA}
130\hspace*{10pt}
131\vrule
132\hspace*{10pt}
133\subfloat[static typing]{\label{l:static_style}\usebox\myboxB}
134\caption{Behaviour Styles}
135\label{f:BehaviourStyles}
136\end{figure}
137
138\begin{figure}
139\centering
140
141\begin{cfa}
142// actor
143struct my_actor {
144        @inline actor;@                                                 $\C[3.25in]{// Plan-9 C inheritance}$
145};
146// messages
147struct str_msg {
148        @inline message;@                                               $\C{// Plan-9 C nominal inheritance}$
149        char str[12];
150};
151void ?{}( str_msg & this, char * str ) { strcpy( this.str, str ); }  $\C{// constructor}$
152struct int_msg {
153        @inline message;@                                               $\C{// Plan-9 C nominal inheritance}$
154        int i;
155};
156void ?{}( int_msg & this, int i ) { this.i = i; }       $\C{// constructor}$
157// behaviours
158allocation receive( my_actor &, @str_msg & msg@ ) {
159        sout | "string message \"" | msg.str | "\"";
160        return Nodelete;                                                $\C{// actor not finished}$
161}
162allocation receive( my_actor &, @int_msg & msg@ ) {
163        sout | "integer message" | msg.i;
164        return Nodelete;                                                $\C{// actor not finished}$
165}
166int main() {
167        start_actor_system();                                   $\C{// sets up executor}$
168        my_actor actor;                                                 $\C{// default constructor call}$
169        str_msg str_msg{ "Hello World" };               $\C{// constructor call}$
170        int_msg int_msg{ 42 };                                  $\C{// constructor call}$
171        @actor << str_msg << int_msg;@                  $\C{// cascade sends}$
172        @actor << int_msg;@                                             $\C{// send}$
173        @actor << finished_msg;@                                $\C{// send => terminate actor (deallocation deferred)}$
174        stop_actor_system();                                    $\C{// waits until actors finish}\CRT$
175} // deallocate int_msg, str_msg, actor
176\end{cfa}
177\caption{\CFA Actor Syntax}
178\label{f:CFAActorSyntax}
179\end{figure}
180
181Figure~\ref{f:CFAActorSyntax} shows a complete \CFA actor example starting with the actor type @my_actor@ created by defining a @struct@ that inherits from the base @actor@ @struct@ via the @inline@ keyword.
182This inheritance style is the Plan-9 C-style nominal inheritance discussed in Section~\ref{s:Inheritance}.
183Similarly, the message types @str_msg@ and @int_msg@ are created by defining a @struct@ that inherits from the base @message@ @struct@ via the @inline@ keyword.
184Both message types have constructors to set the message value.
185There are two matching @receive@ (behaviour) routines that process the corresponding typed messages.
186The program main begins by calling @start_actor_system@ to start the actor implementation, including executor threads to run the actors.
187An actor and two messages are created on the stack, and four messages are sent to the actor using operator @<<@.
188The last message is the builtin @finish_msg@, which returns @Finished@ to an executor thread, which removes the actor from the actor system \see{Section~\ref{s:ActorBehaviours}}.
189The call to @stop_actor_system@ blocks program main until all actors are finished and removed from the actor system.
190The program main ends by deleting the actor and messages from the stack.
191The output for the program is:
192\begin{cfa}
193string message "Hello World"
194integer message 42
195integer message 42
196\end{cfa}
197
198\subsection{Actor Behaviours}\label{s:ActorBehaviours}
199In general, a behaviour for some derived actor and derived message type is defined with following signature:
200\begin{cfa}
201allocation receive( my_actor & receiver, my_msg & msg )
202\end{cfa}
203where @my_actor@ and @my_msg@ inherit from types @actor@ and @message@, respectively.
204The return value of @receive@ must be a value from enumerated type, @allocation@:
205\begin{cfa}
206enum allocation { Nodelete, Delete, Destroy, Finished };
207\end{cfa}
208The values represent a set of actions that dictate what the executor does with an actor or message after a given behaviour returns.
209For actors, the @receive@ routine returns the @allocation@ status to the executor, which takes the appropriate action.
210For messages, either the default allocation, @Nodelete@, or any changed value in the message is examined by the executor, which takes the appropriate action.
211Message state is updated via a call to:
212\begin{cfa}
213void set_allocation( message & this, allocation state )
214\end{cfa}
215
216In detail, the actions taken by an executor for each of the @allocation@ values are:
217
218\noindent@Nodelete@
219tells the executor that no action is to be taken with regard to an actor or message.
220This status is used when an actor continues receiving messages or a message may be reused.
221
222\noindent@Delete@
223tells the executor to call the object's destructor and deallocate (delete) the object.
224This status is used with dynamically allocated actors and messages when they are not reused.
225
226\noindent@Destroy@
227tells the executor to call the object's destructor, but not deallocate the object.
228This status is used with dynamically allocated actors and messages whose storage is reused.
229
230\noindent@Finished@
231tells the executor to mark the respective actor as finished executing, but not call the object's destructor nor deallocate the object.
232This status is used when actors or messages are global or stack allocated, or a programmer wants to manage deallocation themselves.
233
234For the actor system to terminate, all actors must have returned a status other than @Nodelete@.
235After an actor is terminated, it is erroneous to send messages to it.
236Similarly,  after a message is terminated, it cannot be sent to an actor.
237Note, it is safe to construct an actor or message with a status other than @Nodelete@, since the executor only examines the allocation action after a behaviour returns.
238
239\subsection{Actor System}\label{s:ActorSystem}
240The calls to @start_actor_system@, and @stop_actor_system@ mark the start and end of a \CFA actor system.
241The call to @start_actor_system@ sets up an executor and executor threads for the actor system.
242It is possible to have multiple start/stop scenarios in a program.
243
244@start_actor_system@ has three overloaded signatures that vary the executor's configuration:
245
246\noindent@void start_actor_system()@
247configures the executor to implicitly use all preallocated kernel-threads (processors), \ie the processors created by the program main prior to starting the actor system.
248When the number of processors is greater than 1, each executor's message queue is sharded by a factor of 16 to reduce contention, \ie for 4 executor threads (processors), there is a total of 4 $\times$ 16 message queues evenly distributed across the executor threads.
249
250\noindent@void start_actor_system( size_t num_thds )@
251configures the number of executor threads to @num_thds@, with the same message queue sharding.
252
253\noindent@void start_actor_system( executor & this )@
254allows the programmer to explicitly create and configure an executor for use by the actor system.
255Executor configuration options include are discussed in Section~\ref{s:executor}.
256
257\noindent
258All actors must be created \emph{after} calling @start_actor_system@ so the executor can keep track of the number of actors that have entered the system but not yet terminated.
259
260% All message sends are done using the left-shift operator, @<<@, similar to the syntax of \CC's stream output.
261% \begin{cfa}
262% allocation ?<<?( my_actor & receiver, my_msg & msg )
263% \end{cfa}
264% Notice this signature is the same as the @receive@ routine, which is no coincidence.
265% The \CFA compiler generates a @?<<?@ routine definition and forward declaration for each @receive@ routine that has the appropriate signature.
266% The generated routine packages the message and actor in an \hyperref[s:envelope]{envelope} and adds it to the executor's queues via an executor routine.
267% As part of packaging the envelope, the @?<<?@ routine sets a routine pointer in the envelope to point to the appropriate receive routine for given actor and message types.
268
269\subsection{Actor Send}\label{s:ActorSend} % C_TODO: rework this paragraph based on discussion with Mike, see ~/cfa-cc/actor_poly.tex for notes (and fangren's resolver changes)
270All message sends are done using the left-shift operator, @<<@, similar to the syntax of \CC's stream output.
271As stated, \CFA does not have named inheritance with RTTI.
272\CFA does have a preliminary form of virtual routines, but it is not mature enough for use in this work.
273Therefore, there is no mechanism to write a generic @<<@ routine taking a base actor and message type, and then dynamically selecting the @receive@ routine from the actor argument.
274(For messages, the Plan-9 inheritance is sufficient because only the inherited fields are needed during the message send.)
275Hence, programmers must write a matching @<<@ routine for each @receive@ routine, which is awkward and generates a maintenance problem.
276Therefore, I chose to use a template-like approach, where the compiler generates a matching @<<@ routine for each @receive@ routine it finds with an actor/message type-signature.
277Then, \CFA uses the type from the left-hand side of an assignment to select the matching receive routine.
278(When the \CFA virtual routines mature, it should be possible to seamlessly transition to it from the template approach.)
279
280% Funneling all message sends through a single @allocation ?<<?(actor &, message &)@ routine is not feasible since the type of the actor and message would be erased, making it impossible to acquire a pointer to the correct @receive@.
281% As such a @?<<?@ routine per @receive@ provides type information needed to write the correct "address" on the envelope.
282
283% The left-shift operator routines are generated by the compiler.
284An example of a receive routine and its corresponding generated operator routine is shown in Figure~\ref{f:actor_gen}.
285Notice the parameter signature of @?<<?@ is the same as the @receive@ routine.
286A @?<<?@ routine is generated per @receive@ routine with a matching signature.
287The @?<<?@ routine packages the message and actor in an \hyperref[s:envelope]{envelope} and adds it to the executor's queues via the executor routine @send@.
288The envelope is conceptually "addressed" to a behaviour, which is stored in the envelope as a function pointer to a @receive@ routine.
289The @?<<?@ routines ensure that messages are sent to the right address, \ie sent to the right @receive@ routine based on actor and message type.
290
291\begin{figure}
292\begin{cfa}
293$\LstCommentStyle{// from Figure~\ref{f:CFAActorSyntax}}$
294struct my_actor { inline actor; };                                      $\C[3.5in]{// actor}$
295struct int_msg { inline message; int i; };                      $\C{// message}$
296void ?{}( int_msg & this, int i ) { this.i = i; }       $\C{// constructor}$
297allocation receive( @my_actor &, int_msg & msg@ ) {     $\C{// receiver}$
298        sout | "integer message" | msg.i;
299        return Nodelete;
300}
301
302// compiler generated operator
303#define RECEIVER( A, M ) (allocation (*)(actor &, message &))(allocation (*)( A &, M & ))receive
304my_actor & ?<<?( @my_actor & receiver, int_msg & msg@ ) {
305        send( receiver, (request){ &receiver, &msg, RECEIVER( my_actor, int_msg ) } );
306        return receiver;
307}
308\end{cfa}
309\caption{Generated Send Operator}
310\label{f:actor_gen}
311\end{figure}
312
313\section{\CFA Executor}\label{s:executor}
314This section describes the basic architecture of the \CFA executor.
315An executor of an actor system is the scheduler that organizes where actor behaviours are run and how messages are sent and delivered.
316In \CFA, the executor is message-centric \see{Figure~\ref{f:inverted_actor}}, but extended by over sharding of a message queue \see{left side of Figure~\ref{f:gulp}}, \ie there are $M$ message queues where $M$ is greater than the number of executor threads $N$ (usually a multiple of $N$).
317This approach reduces contention by spreading message delivery among the $M$ queues rather than $N$, while still maintaining actor FIFO message-delivery semantics.
318The only extra overhead is each executor cycling (usually round-robin) through its $M$/$N$ queues.
319The goal is to achieve better performance and scalability for certain kinds of actor applications by reducing executor locking.
320Note, lock-free queues do not help because busy waiting on any atomic instruction is the source of the slowdown whether it is a lock or lock-free.
321Work steal now becomes queue stealing, where an entire actor/message queue is stolen, which trivially preserves message ordering in a queue \see{Section~\ref{s:steal}}.
322
323\begin{figure}
324\begin{center}
325\input{diagrams/gulp.tikz}
326\end{center}
327\caption{Queue Gulping Mechanism}
328\label{f:gulp}
329\end{figure}
330
331Each executor thread iterates over its own message queues until it finds one with messages.
332At this point, the executor thread atomically \gls{gulp}s the queue, meaning move the contents of message queue to a local queue of the executor thread using a single atomic instruction.
333This step allows the executor threads to process the local queue without any atomics until the next gulp, while other executor threads are adding in parallel to the end of one of the message queues.
334In detail, an executor thread performs a test-and-gulp, non-atomically checking if a queue is non-empty before gulping it.
335If a test fails during a message add, the worst-case is cycling through all the message queues.
336However, the gain is minimizing costly lock acquisitions.
337An example of the queue gulping operation is shown in the right side of Figure \ref{f:gulp}, where a executor threads gulps queue 0 and begins to process it locally.
338
339Processing a local queue involves removing a unit of work from the queue and executing it.
340Since all messages to a given actor are in the same queue, this guarantees atomicity across behaviours of that actor since it can only execute on one thread at a time.
341After running a behaviour, the executor thread examines the returned allocation status from the @receive@ routine for the actor and internal status in the delivered message, and takes the appropriate action.
342Once all actors have marked themselves as being finished the executor initiates shutdown by inserting a sentinel value into the message queues. % C_TODO: potentially change if I keep shutdown flag change
343Once a executor threads sees a sentinel it stops running.
344After all executors stop running the actor system shutdown is complete.
345
346\section{Envelopes}\label{s:envelope}
347As stated, each message, regardless of where it is allocated, can be sent to an arbitrary number of actors, and hence, appear on an arbitrary number of message queues.
348Because a C program manages message lifetime, messages cannot be copied for each send, otherwise who manages the copies.
349Therefore, it up to the actor program to manage message life-time across receives.
350However, for a message to appear on multiple message queues, it needs an arbitrary number of associated destination behaviours.
351Hence, there is the concept of an envelop, which is dynamically allocated on each send, that wraps a message with any extra implementation fields needed to persist between send and receive.
352Managing the envelop is straightforward because it is created at the send and deleted after the receive, \ie there is 1:1 relationship for an envelop and a many to one relationship for a message.
353
354Unfortunately, this frequent allocation of envelopes for each send results in heavy contention on the memory allocator.
355As such, a way to alleviate contention on the memory allocator would result in a performance improvement.
356Contention is reduced using a novel data structure, called a \Newterm{copy queue}.
357
358\subsection{Copy Queue}\label{s:copyQueue}
359The copy queue is a thin layer over a dynamically sized array that is designed with the envelope use case in mind.
360A copy queue supports the typical queue operations of push/pop but in a different way than a typical array based queue.
361The copy queue is designed to take advantage of the \gls{gulp}ing pattern.
362As such, the amortized runtime cost of each push/pop operation for the copy queue is $O(1)$.
363In contrast, a na\"ive array-based queue often has either push or pop cost $O(n)$ and the other cost $O(1)$ since one of the operations requires shifting the elements of the queue.
364Since the executor threads gulp a queue to operate on it locally, this creates a usage pattern where all elements are popped from the copy queue without any interleaved pushes.
365As such, during pop operations there is no need to shift array elements.
366Instead, an index is stored in the copy-queue data-structure that keeps track of which element to pop next allowing pop to be $O(1)$.
367Push operations are amortized $O(1)$ since pushes may cause doubling reallocations of the underlying dynamic-sized array (like \CC @vector@).
368
369% C_TODO: maybe make copy_queue diagram
370
371Since the copy queue is an array, envelopes are allocated first on the stack and then copied into the copy queue to persist until they are no longer needed.
372For many workload, the copy queues grow in size to facilitate the average number of messages in flight and there is no further dynamic allocations.
373One downside of this approach that more storage is allocated than needed, \ie each copy queue is only partially full.
374Comparatively, the individual envelope allocations of a list based queue mean that the actor system always uses the minimum amount of heap space and cleans up eagerly.
375Additionally, bursty workloads can cause the copy queues to allocate a large amounts of space to accommodate the peaks of the throughput, even if most of that storage is not needed for the rest of the workload's execution.
376
377To mitigate memory wastage, a reclamation scheme is introduced.
378Initially, the memory reclamation na\"ively reclaims one index of the array per \gls{gulp}, if the array size is above a low fixed threshold.
379However, this approach has a problem.
380The high memory usage watermark nearly doubled!
381The issue can easily be highlighted with an example.
382Assume a fixed throughput workload, where a queue never has more than 19 messages at a time.
383If the copy queue starts with a size of 10, it ends up doubling at some point to size 20 to accommodate 19 messages.
384However, after 2 gulps and subsequent reclamations the array size is 18.
385The next time 19 messages are enqueued, the array size is doubled to 36!
386To avoid this issue a second check is added.
387Each copy queue now tracks the utilization of its array size.
388Reclamation only occurs if less than half of the array is utilized.
389In doing this, the reclamation scheme is able to achieve a lower high-watermark and a lower overall memory utilization compared to the non-reclamation copy queues.
390However, the use of copy queues still incurs a higher memory cost than list-based queueing.
391With the inclusion of a memory reclamation scheme the increase in memory usage is reasonable considering the performance gains and is discussed further in Section~\ref{s:actor_perf}.
392
393\section{Work Stealing}\label{s:steal}
394Work stealing is a scheduling strategy that attempts to load balance, and increase resource utilization by having idle threads steal work.
395There are many parts that make up a work stealing actor scheduler, but the two that will be highlighted in this work are the stealing mechanism and victim selection.
396
397% C_TODO enter citation for langs
398\subsection{Stealing Mechanism}
399In this discussion of work stealing the worker being stolen from will be referred to as the \textbf{victim} and the worker stealing work will be called the \textbf{thief}.
400The stealing mechanism presented here differs from existing work stealing actor systems due the inverted actor system.
401Other actor systems such as Akka \cite{} and CAF \cite{} have work stealing, but since they use an classic actor system that is actor-centric, stealing work is the act of stealing an actor from a dequeue.
402As an example, in CAF, the sharded actor queue is a set of double ended queues (dequeues).
403Whenever an actor is moved to a ready queue, it is inserted into a worker's dequeue.
404Workers then consume actors from the dequeue and execute their behaviours.
405To steal work, thieves take one or more actors from a victim's dequeue.
406This action creates contention on the dequeue, which can slow down the throughput of the victim.
407The notion of which end of the dequeue is used for stealing, consuming, and inserting is not discussed since it isn't relevant.
408By the pigeon hole principle there are three dequeue operations (push/victim pop/thief pop) that can occur concurrently and only two ends to a dequeue, so work stealing being present in a dequeue based system will always result in a potential increase in contention on the dequeues.
409
410% C_TODO: maybe insert stealing diagram
411
412In \CFA, the actor work stealing implementation is unique.
413While other systems are concerned with stealing actors, the \CFA actor system steals queues.
414This is a result of \CFA's use of the inverted actor system.
415The goal of the \CFA actor work stealing mechanism is to have a zero-victim-cost stealing mechanism.
416This does not means that stealing has no cost.
417This goal is to ensure that stealing work does not impact the performance of victim workers.
418This means that thieves can not contend with victims, and that victims should perform no stealing related work unless they become a thief.
419In theory this goal is not achieved, but results will be presented that show the goal is achieved in practice.
420In \CFA's actor system workers own a set of sharded queues which they iterate over and gulp.
421If a worker has iterated over the queues they own twice without finding any work, they try to steal a queue from another worker.
422Stealing a queue is done wait-free with a few atomic instructions that can only create contention with other stealing workers.
423To steal a queue a worker does the following:
424\begin{enumerate}[topsep=5pt,itemsep=3pt,parsep=0pt]
425\item
426The thief chooses a victim.
427
428\item
429The thief starts at a random index in the array of the victim's queues and searches for a candidate queue.
430A candidate queue is any queue that is not empty, is not being stolen by another thief, and is not being processed by the victim.
431These are not strictly enforced rules.
432The candidate is identified non-atomically and as such queues that do not satisfy these rules may be stolen.
433However, steals that do not meet these requirements do not affect correctness so they are allowed and do not constitute failed steals as the queues will still be swapped.
434
435
436\item
437Once a candidate queue is chosen, the thief attempts a wait-free swap of the victim's queue and a random on of the thief's queues.
438This swap can fail.
439If the swap is successful the thief swaps the two queues.
440If the swap fails, another thief must have attempted to steal one of the two queues being swapped.
441Failing to steal is good in this case since stealing a queue that was just swapped would likely result in stealing an empty queue.
442\end{enumerate}
443
444Once a thief fails or succeeds in stealing a queue, it goes back to its own set of queues and iterates over them again.
445It will only try to steal again once it has completed two consecutive iterations over its owned queues without finding any work.
446The key to the stealing mechanism is that the queues can still be operated on while they are being swapped.
447This eliminates any contention between thieves and victims.
448The first key to this is that actors and workers maintain two distinct arrays of references to queues.
449Actors will always receive messages via the same queues.
450Workers, on the other hand will swap the pointers to queues in their shared array and operate on queues in the range of that array that they own.
451Swapping queues is a matter of atomically swapping two pointers in the worker array.
452As such pushes to the queues can happen concurrently during the swap since pushes happen via the actor queue references.
453
454Gulping can also occur during queue swapping, but the implementation requires more nuance than the pushes.
455When a worker is not stealing it iterates across its own range of queues and gulps them one by one.
456When a worker operates on a queue it first copies the current pointer from the worker array of references to a local variable.
457It then uses that local variable for all queue operations until it moves to the next index of its range of the queue array.
458This ensures that any swaps do not interrupt gulping operations, however this introduces a correctness issue.
459If any behaviours from a queue are run by two workers at a time it violates both mutual exclusion and the actor ordering guarantees.
460As such this must be avoided.
461To avoid this each queue has a @being_processed@ flag that is atomically set to @true@ when a queue is gulped.
462The flag indicates that a queue is being processed locally and is set back to @false@ once the local processing is finished.
463If a worker attempts to gulp a queue and finds that the @being_processed@ flag is @true@, it does not gulp the queue and moves on to the next queue in its range.
464This is a source of contention between victims and thieves since a thief may steal a queue and set @being_processed@ to @true@ between a victim saving a pointer to a queue and gulping it.
465However, the window for this race is very small, making this contention rare.
466This is why the claim is made that this mechanism is zero-victim-cost in practice but not in theory.
467By collecting statistics on failed gulps due to the @being_processed@ flag, it is found that this contention occurs ~0.05\% of the time when a gulp occurs.
468Hence, the claim is made that this stealing mechanism has zero-victim-cost in practice.
469
470
471\subsection{Queue Swap Correctness}
472Given the wait-free swap used is novel, it is important to show that it is correct.
473Firstly, it is clear to show that the swap is wait-free since all workers will fail or succeed in swapping the queues in a finite number of steps since there are no locks or looping.
474There is no retry mechanism in the case of a failed swap, since a failed swap either means the work was already stolen, or that work was stolen from the thief.
475In both cases it is apropos for a thief to given up on stealing.
476\CFA-style pseudocode for the queue swap is presented below.
477The swap uses compare-and-swap (@CAS@) which is just pseudocode for C's @__atomic_compare_exchange_n@.
478A pseudocode implementation of @CAS@ is also shown below.
479The correctness of the wait-free swap will now be discussed in detail.
480To first verify sequential correctness, consider the equivalent sequential swap below:
481
482\begin{cfa}
483void swap( uint victim_idx, uint my_idx ) {
484        // Step 0:
485        work_queue * my_queue = request_queues[my_idx];
486        work_queue * vic_queue = request_queues[victim_idx];
487        // Step 2:
488        request_queues[my_idx] = 0p;
489        // Step 3:
490        request_queues[victim_idx] = my_queue;
491        // Step 4:
492        request_queues[my_idx] = vic_queue;
493}
494\end{cfa}
495
496Step 1 is missing in the sequential example since in only matter in the concurrent context presented later.
497By looking at the sequential swap it is easy to see that it is correct.
498Temporary copies of each pointer being swapped are stored, and then the original values of each pointer are set using the copy of the other pointer.
499
500\begin{cfa}
501// This routine is atomic
502bool CAS( work_queue ** ptr, work_queue ** old, work_queue * new ) {
503        if ( *ptr != *old )
504                return false;
505        *ptr = new;
506        return true;
507}
508
509bool try_swap_queues( worker & this, uint victim_idx, uint my_idx ) with(this) {
510        // Step 0:
511        // request_queues is the shared array of all sharded queues
512        work_queue * my_queue = request_queues[my_idx];
513        work_queue * vic_queue = request_queues[victim_idx];
514
515        // Step 1:
516        // If either queue is 0p then they are in the process of being stolen
517        // 0p is Cforall's equivalent of C++'s nullptr
518        if ( vic_queue == 0p ) return false;
519
520        // Step 2:
521        // Try to set thief's queue ptr to be 0p.
522        // If this CAS fails someone stole thief's queue so return false
523        if ( !CAS( &request_queues[my_idx], &my_queue, 0p ) )
524                return false;
525
526        // Step 3:
527        // Try to set victim queue ptr to be thief's queue ptr.
528        // If it fails someone stole the other queue, so fix up then return false
529        if ( !CAS( &request_queues[victim_idx], &vic_queue, my_queue ) ) {
530                request_queues[my_idx] = my_queue; // reset queue ptr back to prev val
531                return false;
532        }
533
534        // Step 4:
535        // Successfully swapped.
536        // Thief's ptr is 0p so no one will touch it
537        // Write back without CAS is safe
538        request_queues[my_idx] = vic_queue;
539        return true;
540}
541\end{cfa}\label{c:swap}
542
543Now consider the concurrent implementation of the swap.
544\begin{enumerate}[topsep=5pt,itemsep=3pt,parsep=0pt]
545\item
546Step 0 is the same as the sequential example, and the thief stores local copies of the two pointers to be swapped.
547\item
548Step 1 verifies that the stored copy of the victim queue pointer, @vic_queue@, is valid.
549If @vic_queue@ is equal to @0p@, then the victim queue is part of another swap so the operation fails.
550No state has changed at this point so no fixups are needed.
551Note, @my_queue@ can never be equal to @0p@ at this point since thieves only set their own queues pointers to @0p@ when stealing.
552At no other point will a queue pointer be set to @0p@.
553Since each worker owns a disjoint range of the queue array, it is impossible for @my_queue@ to be @0p@.
554\item
555Step 2 attempts to set the thief's queue pointer to @0p@ via @CAS@.
556The @CAS@ will only fail if the thief's queue pointer is no longer equal to @my_queue@, which implies that this thief has become a victim and its queue has been stolen.
557At this point the thief-turned-victim will fail and since it has not changed any state it just fails and returns false.
558If the @CAS@ succeeds then the thief's queue pointer will now be @0p@.
559Nulling the pointer is safe since only thieves look at other worker's queue ranges, and whenever thieves need to dereference a queue pointer they check for @0p@.
560\item
561Step 3 attempts to set the victim's queue pointer to be @my_queue@ via @CAS@.
562If the @CAS@ succeeds then the victim's queue pointer has been set and swap can no longer fail.
563If the @CAS@ fails then the thief's queue pointer must be restored to its previous value before returning.
564\item
565Step 4 sets the thief's queue pointer to be @vic_queue@ completing the swap.
566\end{enumerate}
567
568\begin{theorem}
569The presented swap is correct and concurrently safe in both the success and failure cases.
570\end{theorem}
571
572Correctness of the swap is shown through the existence of an invariant.
573The invariant is that when a queue pointer is set to @0p@ by a thief, then the next write to the pointer can only be performed by the same thief.
574To show that this invariant holds, it is shown that it is true at each step of the swap.
575Step 0 and 1 do not write and as such they cannot invalidate the invariant of any other thieves.
576In step 2 a thief attempts to write @0p@ to one of their queue pointers.
577This queue pointer cannot be @0p@.
578As stated above, @my_queue@ is never equal to @0p@ since thieves will only write @0p@ to queue pointers from their own queue range and all worker's queue ranges are disjoint.
579As such step 2 upholds the invariant since in a failure case no write occurs, and in the success case, the value of the queue pointer is guaranteed to not be 0p.
580In step 3 the thief attempts to write @my_queue@ to the victim's queue pointer.
581If the current value of the victim's queue pointer is @0p@, then the CAS will fail since @vic_queue@ cannot be equal to @0p@ because of the check in step 1.
582Therefore in the success case where the @CAS@ succeeds, the value of the victim's queue pointer must not be @0p@.
583As such, the write will never overwrite a value of @0p@, hence the invariant is held in the @CAS@ of step 3.
584The write back to the thief's queue pointer that happens in the failure case of step three and in step 4 hold the invariant since they are the subsequent write to a @0p@ queue pointer and they are being set by the same thief that set the pointer to @0p@.
585
586Given this informal proof of invariance it can be shown that the successful swap is correct.
587Once a thief atomically sets their queue pointer to be @0p@ in step 2, the invariant guarantees that pointer will not change.
588As such, in the success case step 3 it is known that the value of the victim's queue pointer that was overwritten must be @vic_queue@ due to the use of @CAS@.
589Given that pointers all have unique memory locations, this first write of the successful swap is correct since it can only occur when the pointer has not changed.
590By the invariant the write back in the successful case is correct since no other worker can write to the @0p@ pointer.
591
592In the failed case the outcome is correct in steps 1 and 2 since no writes have occurred so the program state is unchanged.
593In the failed case of step 3 the program state is safely restored to its state it had prior to the @0p@ write in step 2, thanks to the invariant that makes the write back to the @0p@ pointer safe.
594
595\subsection{Stealing Guarantees}
596
597% C_TODO insert graphs for each proof
598Given that the stealing operation can potentially fail, it is important to discuss the guarantees provided by the stealing implementation.
599Given a set of $N$ swaps a set of connected directed graphs can be constructed where each vertex is a queue and each edge is a swap directed from a thief queue to a victim queue.
600Since each thief can only steal from one victim at a time, each vertex can only have at most one outgoing edge.
601A corollary that can be drawn from this, is that there are at most $V$ edges in this constructed set of connected directed graphs, where $V$ is the total number of vertices.
602
603\begin{figure}
604\begin{center}
605\input{diagrams/M_to_one_swap.tikz}
606\end{center}
607\caption{Graph of $M$ thieves swapping with one victim.}
608\label{f:M_one_swap}
609\end{figure}
610
611\begin{theorem}
612Given $M$ thieves queues all attempting to swap with one victim queue, and no other swaps occurring that involve these queues, at least one swap is guaranteed to succeed.
613\end{theorem}\label{t:one_vic}
614A graph of the $M$ thieves swapping with one victim discussed in this theorem is presented in Figure~\ref{f:M_one_swap}.
615\\
616First it is important to state that a thief will not attempt to steal from themselves.
617As such, the victim here is not also a thief.
618Stepping through the code in \ref{c:swap}, for all thieves steps 0-1 succeed since the victim is not stealing and will have no queue pointers set to be @0p@.
619Similarly for all thieves step 2 will succeed since no one is stealing from any of the thieves.
620In step 3 the first thief to @CAS@ will win the race and successfully swap the queue pointer.
621Since it is the first one to @CAS@ and @CAS@ is atomic, there is no way for the @CAS@ to fail since no other thief could have written to the victim's queue pointer and the victim did not write to the pointer since they aren't stealing.
622Hence at least one swap is guaranteed to succeed in this case.
623
624\begin{figure}
625\begin{center}
626\input{diagrams/chain_swap.tikz}
627\end{center}
628\caption{Graph of a chain of swaps.}
629\label{f:chain_swap}
630\end{figure}
631
632\begin{theorem}
633Given $M$ > 1, ordered queues pointers all attempting to swap with the queue in front of them in the ordering, except the first queue, and no other swaps occurring that involve these queues, at least one swap is guaranteed to succeed.
634\end{theorem}\label{t:vic_chain}
635A graph of the chain of swaps discussed in this theorem is presented in Figure~\ref{f:chain_swap}.
636\\
637This is a proof by contradiction.
638Assume no swaps occur.
639Then all thieves must have failed at step 1, step 2 or step 3.
640For a given thief $b$ to fail at step 1, thief $b + 1$ must have succeeded at step 2 before $b$ executes step 0.
641Hence, not all thieves can fail at step 1.
642Furthermore if a thief $b$ fails at step 1 it logically splits the chain into two subchains $0 <- b$ and $b + 1 <- M - 1$, where $b$ has become solely a victim since its swap has failed and it did not modify any state.
643There must exist at least one chain containing two or more queues after since it is impossible for a split to occur both before and after a thief, since that requires failing at step 1 and succeeding at step 2.
644Hence, without loss of generality, whether thieves succeed or fail at step 1, this proof can proceed inductively.
645
646For a given thief $i$ to fail at step 2, it means that another thief $j$ had to have written to $i$'s queue pointer between $i$'s step 0 and step 2.
647The only way for $j$ to write to $i$'s queue pointer would be if $j$ was stealing from $i$ and had successfully finished step 3.
648If $j$ finished step 3 then the at least one swap was successful.
649Therefore all thieves did not fail at step 2.
650Hence all thieves must successfully complete step 2 and fail at step 3.
651However, since the first worker, thief $0$, is solely a victim and not a thief, it does not change the state of any of its queue pointers.
652Hence, in this case thief $1$ will always succeed in step 3 if all thieves succeed in step 2.
653Thus, by contradiction with the earlier assumption that no swaps occur, at least one swap must succeed.
654
655% \raisebox{.1\height}{}
656\begin{figure}
657\centering
658\begin{tabular}{l|l}
659\subfloat[Cyclic Swap Graph]{\label{f:cyclic_swap}\input{diagrams/cyclic_swap.tikz}} &
660\subfloat[Acyclic Swap Graph]{\label{f:acyclic_swap}\input{diagrams/acyclic_swap.tikz}}
661\end{tabular}
662\caption{Illustrations of cyclic and acyclic swap graphs.}
663\end{figure}
664
665\begin{theorem}
666Given a set of $M > 1$ swaps occurring that form a single directed connected graph.
667At least one swap is guaranteed to succeed if and only if the graph does not contain a cycle.
668\end{theorem}\label{t:vic_cycle}
669Representations of cyclic and acyclic swap graphs discussed in this theorem are presented in Figures~\ref{f:cyclic_swap} and \ref{f:acyclic_swap}.
670\\
671First the reverse direction is proven.
672If the graph does not contain a cycle, then there must be at least one successful swap.
673Since the graph contains no cycles and is finite in size, then there must be a vertex $A$ with no outgoing edges.
674The graph can then be formulated as a tree with $A$ at the top since each node only has at most one outgoing edge and there are no cycles.
675The forward direction is proven by contradiction in a similar fashion to \ref{t:vic_chain}.
676Assume no swaps occur.
677Similar to \ref{t:vic_chain}, this graph can be inductively split into subgraphs of the same type by failure at step 1, so the proof proceeds without loss of generality.
678Similar to \ref{t:vic_chain} the conclusion is drawn that all thieves must successfully complete step 2 for no swaps to occur, since for step 2 to fail, a different thief has to successfully complete step 3, which would imply a successful swap.
679Hence, the only way forward is to assume all thieves successfully complete step 2.
680Hence for there to be no swaps all thieves must fail step 3.
681However, since $A$ has no outgoing edges, since the graph is connected there must be some $K$ such that $K < M - 1$ thieves are attempting to swap with $A$.
682Since all $K$ thieves have passed step 2, similar to \ref{t:one_vic} the first one of the $K$ thieves to attempt step 3 is guaranteed to succeed.
683Thus, by contradiction with the earlier assumption that no swaps occur, if the graph does not contain a cycle, at least one swap must succeed.
684
685The forward direction is proven by contrapositive.
686If the graph contains a cycle then there exists a situation where no swaps occur.
687This situation is constructed.
688Since all vertices have at most one outgoing edge the cycle must be directed.
689Furthermore, since the graph contains a cycle all vertices in the graph must have exactly one outgoing edge.
690This is shown through construction of an arbitrary cyclic graph.
691The graph contains a directed cycle by definition, so the construction starts with $T$ vertices in a directed cycle.
692Since the graph is connected, and each vertex has at most one outgoing edge, none of the vertices in the cycle have available outgoing edges to accommodate new vertices with no outgoing edges.
693Any vertices added to the graph must have an outgoing edge to connect, leaving the resulting graph with no available outgoing edges.
694Thus, by induction all vertices in the graph must have exactly one outgoing edge.
695Hence all vertices are thief queues.
696Now consider the case where all thieves successfully complete step 0-1, and then they all complete step 2.
697At this point all thieves are attempting to swap with a queue pointer whose value has changed to @0p@.
698If all thieves attempt the @CAS@ before any write backs, then they will all fail.
699Thus, by contrapositive, if the graph contains a cycle then there exists a situation where no swaps occur.
700Hence, at least one swap is guaranteed to succeed if and only if the graph does not contain a cycle.
701
702% C_TODO: go through and use \paragraph to format to make it look nicer
703\subsection{Victim Selection}\label{s:victimSelect}
704In any work stealing algorithm thieves have some heuristic to determine which victim to choose from.
705Choosing this algorithm is difficult and can have implications on performance.
706There is no one selection heuristic that is known to be the best on all workloads.
707Recent work focuses on locality aware scheduling in actor systems\cite{barghi18}\cite{wolke17}.
708However, while locality aware scheduling provides good performance on some workloads, something as simple as randomized selection performs better on other workloads\cite{barghi18}.
709Since locality aware scheduling has been explored recently, this work introduces a heuristic called \textbf{longest victim} and compares it to randomized work stealing.
710The longest victim heuristic maintains a timestamp per executor threads that is updated every time a worker attempts to steal work.
711Thieves then attempt to steal from the thread with the oldest timestamp.
712This means that if two thieves look to steal at the same time, they likely will attempt to steal from the same victim.
713This does increase the chance at contention between thieves, however given that workers have multiple queues under them, often in the tens or hundreds of queues per worker it is rare for two queues to attempt so steal the same queue.
714Furthermore in the case they attempt to steal the same queue at least one of them is guaranteed to successfully steal the queue as shown in Theorem \ref{t:one_vic}.
715Additionally, the longest victim heuristic makes it very improbable that the no swap scenario presented in Theorem \ref{t:vic_cycle} manifests.
716Given the longest victim heuristic, for a cycle to manifest it would require all workers to attempt to steal in a short timeframe.
717This is the only way that more than one thief could choose another thief as a victim, since timestamps are only updated upon attempts to steal.
718In this case, the probability of lack of any successful swaps is a non issue, since it is likely that these steals were not important if all workers are trying to steal.
719
720\section{Safety and Productivity}\label{s:SafetyProductivity}
721\CFA's actor system comes with a suite of safety and productivity features.
722Most of these features are present in \CFA's debug mode, but are removed when code is compiled in nodebug mode.
723The suit of features include the following.
724
725\begin{itemize}
726\item Static-typed message sends.
727If an actor does not support receiving a given message type, the actor program is rejected at compile time, allowing unsupported messages to never be sent to actors.
728\item Detection of message sends to Finished/Destroyed/Deleted actors.
729All actors have a ticket that assigns them to a respective queue.
730The maximum integer value of the ticket is reserved to indicate that an actor is dead, and subsequent message sends result in an error.
731\item Actors made before the executor can result in undefined behaviour since an executor needs to be created beforehand so it can give out the tickets to actors.
732As such, this is detected and an error is printed.
733\item When an executor is created, the queues are handed out to executor threads in round robin order.
734If there are fewer queues than executor threads, then some workers will spin and never do any work.
735There is no reasonable use case for this behaviour so an error is printed if the number of queues is fewer than the number of executor threads.
736\item A warning is printed when messages are deallocated without being sent.
737Since the @Finished@ allocation status is unused for messages, it is used internally to detect if a message has been sent.
738Deallocating a message without sending it could indicate to a user that they are touching freed memory later, or it could point out extra allocations that could be removed.
739\end{itemize}
740
741In addition to these features, \CFA's actor system comes with a suite of statistics that can be toggled on and off.
742These statistics have minimal impact on the actor system's performance since they are counted on a per executor threads basis.
743During shutdown of the actor system they are aggregated, ensuring that the only atomic instructions used by the statistics counting happen at shutdown.
744The statistics measured are as follows.
745
746\begin{description}
747\item[\LstBasicStyle{\textbf{Actors Created}}]
748Actors created.
749Includes both actors made by the main and ones made by other actors.
750\item[\LstBasicStyle{\textbf{Messages Sent}}]
751Messages sent and received.
752Includes termination messages send to the executor threads.
753\item[\LstBasicStyle{\textbf{Gulps}}]
754Gulps that occurred across the executor threads.
755\item[\LstBasicStyle{\textbf{Average Gulp Size}}]
756Average number of messages in a gulped queue.
757\item[\LstBasicStyle{\textbf{Missed gulps}}]
758Occurrences where a worker missed a gulp due to the concurrent queue processing by another worker.
759\item[\LstBasicStyle{\textbf{Steal attempts}}]
760Worker threads attempts to steal work.
761
762\item[\LstBasicStyle{\textbf{Steal failures (no candidates)}}]
763Work stealing failures due to selected victim not having any non empty or non-being-processed queues.
764\item[\LstBasicStyle{\textbf{Steal failures (failed swaps)}}]
765Work stealing failures due to the two stage atomic swap failing.
766\item[\LstBasicStyle{\textbf{Messages stolen}}]
767Aggregate of the number of messages in queues as they were stolen.
768\item[\LstBasicStyle{\textbf{Average steal size}}]
769Average number of messages in a stolen queue.
770\end{description}
771
772These statistics enable a user of \CFA's actor system to make informed choices about how to configure their executor, or how to structure their actor program.
773For example, if there is a lot of messages being stolen relative to the number of messages sent, it could indicate to a user that their workload is heavily imbalanced across executor threads.
774In another example, if the average gulp size is very high, it could indicate that the executor could use more queue sharding.
775
776% C_TODO cite poison pill messages and add languages
777Another productivity feature that is included is a group of poison-pill messages.
778Poison-pill messages are common across actor systems, including Akka and ProtoActor \cite{}.
779Poison-pill messages inform an actor to terminate.
780In \CFA, due to the allocation of actors and lack of garbage collection, there needs to be a suite of poison-pills.
781The messages that \CFA provides are @DeleteMsg@, @DestroyMsg@, and @FinishedMsg@.
782These messages are supported on all actor types via inheritance and when sent to an actor, the actor takes the corresponding allocation action after receiving the message.
783Note that any pending messages to the actor will still be sent.
784It is still the user's responsibility to ensure that an actor does not receive any messages after termination.
785
786\section{Performance}\label{s:actor_perf}
787\CAP{I will update the figures to have the larger font size and different line markers once we start editing this chapter.}
788The performance of \CFA's actor system is tested using a suite of microbenchmarks, and compared with other actor systems.
789Most of the benchmarks are the same as those presented in \ref{}, with a few additions.
790% C_TODO cite actor paper
791At the time of this work the versions of the actor systems are as follows.
792\CFA 1.0, \uC 7.0.0, Akka Typed 2.7.0, CAF 0.18.6, and ProtoActor-Go v0.0.0-20220528090104-f567b547ea07.
793Akka Classic is omitted as Akka Typed is their newest version and seems to be the direction they are headed in.
794The experiments are run on
795\begin{list}{\arabic{enumi}.}{\usecounter{enumi}\topsep=5pt\parsep=5pt\itemsep=0pt}
796\item
797Supermicro SYS--6029U--TR4 Intel Xeon Gold 5220R 24--core socket, hyper-threading $\times$ 2 sockets (48 process\-ing units) 2.2GHz, running Linux v5.8.0--59--generic
798\item
799Supermicro AS--1123US--TR4 AMD EPYC 7662 64--core socket, hyper-threading $\times$ 2 sockets (256 processing units) 2.0 GHz, running Linux v5.8.0--55--generic
800\end{list}
801
802The benchmarks are run on up to 48 cores.
803On the Intel, when going beyond 24 cores there is the choice to either hop sockets or to use hyperthreads.
804Either choice will cause a blip in performance trends, which can be seen in the following performance figures.
805On the Intel the choice was made to hyperthread instead of hopping sockets for experiments with more than 24 cores.
806
807All benchmarks presented are run 5 times and the median is taken.
808Error bars showing the 95\% confidence intervals are drawn on each point on the graphs.
809If the confidence bars are small enough, they may be obscured by the point.
810In this section \uC will be compared to \CFA frequently, as the actor system in \CFA was heavily based off \uC's actor system.
811As such the performance differences that arise are largely due to the contributions of this work.
812
813\begin{table}[t]
814\centering
815\setlength{\extrarowheight}{2pt}
816\setlength{\tabcolsep}{5pt}
817
818\caption{Static Actor/Message Performance: message send, program memory}
819\label{t:StaticActorMessagePerformance}
820\begin{tabular}{*{5}{r|}r}
821        & \multicolumn{1}{c|}{\CFA (100M)} & \multicolumn{1}{c|}{CAF (10M)} & \multicolumn{1}{c|}{Akka (100M)} & \multicolumn{1}{c|}{\uC (100M)} & \multicolumn{1}{c@{}}{ProtoActor (100M)} \\
822        \hline
823        AMD             & \input{data/pykeSendStatic} \\
824        \hline
825        Intel   & \input{data/nasusSendStatic}
826\end{tabular}
827
828\bigskip
829
830\caption{Dynamic Actor/Message Performance: message send, program memory}
831\label{t:DynamicActorMessagePerformance}
832
833\begin{tabular}{*{5}{r|}r}
834        & \multicolumn{1}{c|}{\CFA (20M)} & \multicolumn{1}{c|}{CAF (2M)} & \multicolumn{1}{c|}{Akka (2M)} & \multicolumn{1}{c|}{\uC (20M)} & \multicolumn{1}{c@{}}{ProtoActor (2M)} \\
835        \hline
836        AMD             & \input{data/pykeSendDynamic} \\
837        \hline
838        Intel   & \input{data/nasusSendDynamic}
839\end{tabular}
840\end{table}
841
842\subsection{Message Sends}
843Message sending is the key component of actor communication.
844As such latency of a single message send is the fundamental unit of fast-path performance for an actor system.
845The following two microbenchmarks evaluate the average latency for a static actor/message send and a dynamic actor/message send.
846Static and dynamic refer to the allocation of the message and actor.
847In the static send benchmark a message and actor are allocated once and then the message is sent to the same actor repeatedly until it has been sent 100 million (100M) times.
848The average latency per message send is then calculated by dividing the duration by the number of sends.
849This benchmark evaluates the cost of message sends in the actor use case where all actors and messages are allocated ahead of time and do not need to be created dynamically during execution.
850The CAF static send benchmark only sends a message 10M times to avoid extensively long run times.
851
852In the dynamic send benchmark the same experiment is performed, with the change that with each send a new actor and message is allocated.
853This evaluates the cost of message sends in the other common actor pattern where actors and message are created on the fly as the actor program tackles a workload of variable or unknown size.
854Since dynamic sends are more expensive, this benchmark repeats the actor/message creation and send 20M times (\uC, \CFA), or 2M times (Akka, CAF, ProtoActor), to give an appropriate benchmark duration.
855
856The results from the static/dynamic send benchmarks are shown in Figures~\ref{t:StaticActorMessagePerformance} and \ref{t:DynamicActorMessagePerformance} respectively.
857\CFA leads the charts in both benchmarks, largely due to the copy queue removing the majority of the envelope allocations.
858In the static send benchmark all systems except CAF have static send costs that are in the same ballpark, only varying by ~70ns.
859In the dynamic send benchmark all systems experience slower message sends, as expected due to the extra allocations.
860However, Akka and ProtoActor, slow down by a more significant margin than the \uC and \CFA.
861This is likely a result of Akka and ProtoActor's garbage collection, which can suffer from hits in performance for allocation heavy workloads, whereas \uC and \CFA have explicit allocation/deallocation.
862
863\subsection{Work Stealing}
864\CFA's actor system has a work stealing mechanism which uses the longest victim heuristic, introduced in Section~ref{s:victimSelect}.
865In this performance section, \CFA with the longest victim heuristic is compared with other actor systems on the benchmark suite, and is separately compared with vanilla non-stealing \CFA and \CFA with randomized work stealing.
866
867\begin{figure}
868        \centering
869        \subfloat[AMD \CFA Balance-One Benchmark]{
870                \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFABalance-One.pgf}}
871                \label{f:BalanceOneAMD}
872        }
873        \subfloat[Intel \CFA Balance-One Benchmark]{
874                \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFABalance-One.pgf}}
875                \label{f:BalanceOneIntel}
876        }
877        \caption{The balance-one benchmark comparing stealing heuristics (lower is better).}
878\end{figure}
879
880\begin{figure}
881        \centering
882        \subfloat[AMD \CFA Balance-Multi Benchmark]{
883                \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFABalance-Multi.pgf}}
884                \label{f:BalanceMultiAMD}
885        }
886        \subfloat[Intel \CFA Balance-Multi Benchmark]{
887                \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFABalance-Multi.pgf}}
888                \label{f:BalanceMultiIntel}
889        }
890        \caption{The balance-multi benchmark comparing stealing heuristics (lower is better).}
891\end{figure}
892
893There are two benchmarks in which \CFA's work stealing is solely evaluated.
894The main goal of introducing work stealing to \CFA's actor system is to eliminate the pathological unbalanced cases that can present themselves in a system without some form of load balancing.
895The following two microbenchmarks construct two such pathological cases, and compare the work stealing variations of \CFA.
896The balance benchmarks adversarially takes advantage of the round robin assignment of actors to load all actors that will do work on specific cores and create 'dummy' actors that terminate after a single message send on all other cores.
897The workload on the loaded cores is the same as the executor benchmark described in \ref{s:executorPerf}, but with fewer rounds.
898The balance-one benchmark loads all the work on a single core, whereas the balance-multi loads all the work on half the cores (every other core).
899Given this layout, one expects the ideal speedup of work stealing in the balance-one case to be $N / N - 1$ where $N$ is the number of threads.
900In the balance-multi case the ideal speedup is 0.5.
901Note that in the balance-one benchmark the workload is fixed so decreasing runtime is expected.
902In the balance-multi experiment, the workload increases with the number of cores so an increasing or constant runtime is expected.
903
904On both balance microbenchmarks slightly less than ideal speedup compared to the non stealing variation is achieved by both the random and longest victim stealing heuristics.
905On the balance-multi benchmark \ref{f:BalanceMultiAMD},\ref{f:BalanceMultiIntel} the random heuristic outperforms the longest victim.
906This is likely a result of the longest victim heuristic having a higher stealing cost as it needs to maintain timestamps and look at all timestamps before stealing.
907Additionally, a performance cost can be observed when hyperthreading kicks in in Figure~\ref{f:BalanceMultiIntel}.
908
909In the balance-one benchmark on AMD \ref{f:BalanceOneAMD}, the performance bottoms out at 32 cores onwards likely due to the amount of work becoming less than the cost to steal it and move it across cores and cache.
910On Intel \ref{f:BalanceOneIntel}, above 32 cores the performance gets worse for all variants due to hyperthreading.
911Note that the non stealing variation of balance-one will slow down marginally as the cores increase due to having to create more dummy actors on the inactive cores during startup.
912
913\subsection{Executor}\label{s:executorPerf}
914The microbenchmarks in this section are designed to stress the executor.
915The executor is the scheduler of an actor system and is responsible for organizing the interaction of executor threads to service the needs of a workload.
916In the executor benchmark, 40'000 actors are created and assigned a group.
917Each group of actors is a group of 100 actors who send and receive 100 messages from all other actors in their group.
918Each time an actor completes all their sends and receives, they are done a round.
919After all groups have completed 400 rounds the system terminates.
920This microbenchmark is designed to flood the executor with a large number of messages flowing between actors.
921Given there is no work associated with each message, other than sending more messages, the intended bottleneck of this experiment is the executor message send process.
922
923\begin{figure}
924        \centering
925        \subfloat[AMD Executor Benchmark]{
926                \resizebox{0.5\textwidth}{!}{\input{figures/nasusExecutor.pgf}}
927                \label{f:ExecutorAMD}
928        }
929        \subfloat[Intel Executor Benchmark]{
930                \resizebox{0.5\textwidth}{!}{\input{figures/pykeExecutor.pgf}}
931                \label{f:ExecutorIntel}
932        }
933        \caption{The executor benchmark comparing actor systems (lower is better).}
934\end{figure}
935
936The results of the executor benchmark in Figures~\ref{f:ExecutorIntel} and \ref{f:ExecutorAMD} show \CFA with the lowest runtime relative to its peers.
937The difference in runtime between \uC and \CFA is largely due to the usage of the copy queue described in Section~\ref{s:copyQueue}.
938The copy queue both reduces and consolidates allocations, heavily reducing contention on the memory allocator.
939Additionally, due to the static typing in \CFA's actor system, it is able to get rid of expensive dynamic casts that occur in \uC to discriminate messages by type.
940Note that dynamic casts are usually not very expensive, but relative to the high performance of the rest of the implementation of the \uC actor system, the cost is significant.
941
942\begin{figure}
943        \centering
944        \subfloat[AMD \CFA Executor Benchmark]{
945                \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFAExecutor.pgf}}
946                \label{f:cfaExecutorAMD}
947        }
948        \subfloat[Intel \CFA Executor Benchmark]{
949                \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFAExecutor.pgf}}
950                \label{f:cfaExecutorIntel}
951        }
952        \caption{Executor benchmark comparing \CFA stealing heuristics (lower is better).}
953\end{figure}
954
955When comparing the \CFA stealing heuristics in Figure~\ref{f:cfaExecutorAMD} it can be seen that the random heuristic falls slightly behind the other two, but in Figure~\ref{f:cfaExecutorIntel} the runtime of all heuristics are nearly identical to each other.
956
957\begin{figure}
958        \centering
959        \subfloat[AMD Repeat Benchmark]{
960                \resizebox{0.5\textwidth}{!}{\input{figures/nasusRepeat.pgf}}
961                \label{f:RepeatAMD}
962        }
963        \subfloat[Intel Repeat Benchmark]{
964                \resizebox{0.5\textwidth}{!}{\input{figures/pykeRepeat.pgf}}
965                \label{f:RepeatIntel}
966        }
967        \caption{The repeat benchmark comparing actor systems (lower is better).}
968\end{figure}
969
970The repeat microbenchmark also evaluates the executor.
971It stresses the executor's ability to withstand contention on queues, as it repeatedly fans out messages from a single client to 100000 servers who then all respond to the client.
972After this scatter and gather repeats 200 times the benchmark terminates.
973The messages from the servers to the client will likely all come in on the same queue, resulting in high contention.
974As such this benchmark will not scale with the number of processors, since more processors will result in higher contention.
975In Figure~\ref{f:RepeatAMD} we can see that \CFA performs well compared to \uC, however by less of a margin than the executor benchmark.
976One factor in this result is that the contention on the queues poses a significant bottleneck.
977As such the gains from using the copy queue are much less apparent.
978
979\begin{figure}
980        \centering
981        \subfloat[AMD \CFA Repeat Benchmark]{
982                \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFARepeat.pgf}}
983                \label{f:cfaRepeatAMD}
984        }
985        \subfloat[Intel \CFA Repeat Benchmark]{
986                \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFARepeat.pgf}}
987                \label{f:cfaRepeatIntel}
988        }
989        \caption{The repeat benchmark comparing \CFA stealing heuristics (lower is better).}
990\end{figure}
991
992In Figure~\ref{f:RepeatIntel} \uC and \CFA are very comparable.
993In comparison with the other systems \uC does well on the repeat benchmark since it does not have work stealing.
994The client of this experiment is long running and maintains a lot of state, as it needs to know the handles of all the servers.
995When stealing the client or its respective queue (in \CFA's inverted model), moving the client incurs a high cost due to cache invalidation.
996As such stealing the client can result in a hit in performance.
997
998This result is shown in Figure~\ref{f:cfaRepeatAMD} and \ref{f:cfaRepeatIntel} where the no-stealing version of \CFA performs better than both stealing variations.
999In particular on the Intel machine in Figure~\ref{f:cfaRepeatIntel}, the cost of stealing is higher, which can be seen in the vertical shift of Akka, CAF and CFA results in Figure~\ref{f:RepeatIntel} (\uC and ProtoActor do not have work stealing).
1000The shift for CAF is particularly large, which further supports the hypothesis that CAF's work stealing is particularly eager.
1001In both the executor and the repeat benchmark CAF performs poorly.
1002It is hypothesized that CAF has an aggressive work stealing algorithm, that eagerly attempts to steal.
1003This results in poor performance in benchmarks with small messages containing little work per message.
1004On the other hand, in \ref{f:MatrixAMD} CAF performs much better since each message has a large amount of work, and few messages are sent, so the eager work stealing allows for the clean up of loose ends to occur faster.
1005This hypothesis stems from experimentation with \CFA.
1006CAF uses a randomized work stealing heuristic.
1007In \CFA if the system is tuned so that it steals work much more eagerly with a randomized it was able to replicate the results that CAF achieves in the matrix benchmark, but this tuning performed much worse on all other microbenchmarks that we present, since they all perform a small amount of work per message.
1008
1009\begin{table}[t]
1010        \centering
1011        \setlength{\extrarowheight}{2pt}
1012        \setlength{\tabcolsep}{5pt}
1013
1014        \caption{Executor Program Memory High Watermark}
1015        \label{t:ExecutorMemory}
1016        \begin{tabular}{*{5}{r|}r}
1017                & \multicolumn{1}{c|}{\CFA} & \multicolumn{1}{c|}{CAF} & \multicolumn{1}{c|}{Akka} & \multicolumn{1}{c|}{\uC} & \multicolumn{1}{c@{}}{ProtoActor} \\
1018                \hline
1019                AMD             & \input{data/pykeExecutorMem} \\
1020                \hline
1021                Intel   & \input{data/nasusExecutorMem}
1022        \end{tabular}
1023\end{table}
1024
1025Figure~\ref{t:ExecutorMemory} shows the high memory watermark of the actor systems when running the executor benchmark on 48 cores.
1026\CFA has a high watermark relative to the other non-garbage collected systems \uC, and CAF.
1027This is a result of the copy queue data structure, as it will over-allocate storage and not clean up eagerly, whereas the per envelope allocations will always allocate exactly the amount of storage needed.
1028
1029\subsection{Matrix Multiply}
1030The matrix benchmark evaluates the actor systems in a practical application, where actors concurrently multiplies two matrices.
1031The majority of the computation in this benchmark involves computing the final matrix, so this benchmark stresses the actor systems' ability to have actors run work, rather than stressing the executor or message sending system.
1032
1033Given $Z_{m,r} = X_{m,n} \cdot Y_{n,r}$, the matrix multiply is defined as:
1034\begin{displaymath}
1035X_{i,j} \cdot Y_{j,k} = \left( \sum_{c=1}^{j} X_{row,c}Y_{c,column} \right)_{i,k}
1036\end{displaymath}
1037
1038The benchmark uses input matrices $X$ and $Y$ that are both $3072$ by $3072$ in size.
1039An actor is made for each row of $X$ and is passed via message the information needed to calculate a row of the result matrix $Z$.
1040
1041
1042Given that the bottleneck of the benchmark is the computation of the result matrix, it follows that the results in Figures~\ref{f:MatrixAMD} and \ref{f:MatrixIntel} are clustered closer than other experiments.
1043In Figure~\ref{f:MatrixAMD} \uC and \CFA have identical performance and in Figure~\ref{f:MatrixIntel} \uC pulls ahead of \CFA after 24 cores likely due to costs associated with work stealing while hyperthreading.
1044As mentioned in \ref{s:executorPerf}, it is hypothesized that CAF performs better in this benchmark compared to others due to its eager work stealing implementation.
1045In Figures~\ref{f:cfaMatrixAMD} and \ref{f:cfaMatrixIntel} there is little negligible performance difference across \CFA stealing heuristics.
1046
1047\begin{figure}
1048        \centering
1049        \subfloat[AMD Matrix Benchmark]{
1050                \resizebox{0.5\textwidth}{!}{\input{figures/nasusMatrix.pgf}}
1051                \label{f:MatrixAMD}
1052        }
1053        \subfloat[Intel Matrix Benchmark]{
1054                \resizebox{0.5\textwidth}{!}{\input{figures/pykeMatrix.pgf}}
1055                \label{f:MatrixIntel}
1056        }
1057        \caption{The matrix benchmark comparing actor systems (lower is better).}
1058\end{figure}
1059
1060\begin{figure}
1061        \centering
1062        \subfloat[AMD \CFA Matrix Benchmark]{
1063                \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFAMatrix.pgf}}
1064                \label{f:cfaMatrixAMD}
1065        }
1066        \subfloat[Intel \CFA Matrix Benchmark]{
1067                \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFAMatrix.pgf}}
1068                \label{f:cfaMatrixIntel}
1069        }
1070        \caption{The matrix benchmark comparing \CFA stealing heuristics (lower is better).}
1071\end{figure}
1072
1073% Local Variables: %
1074% tab-width: 4 %
1075% End: %
Note: See TracBrowser for help on using the repository browser.