% ====================================================================== % ====================================================================== \chapter{Actors}\label{s:actors} % ====================================================================== % ====================================================================== % C_TODO: add citations throughout chapter Actors are an indirect concurrent feature that abstracts threading away from a programmer, and instead provides \gls{actor}s and messages as building blocks for concurrency, where message passing means there is no shared data to protect, making actors amenable in a distributed environment. Actors are another message passing concurrency feature, similar to channels but with more abstraction, and are in the realm of \gls{impl_concurrency}, where programmers write concurrent code without dealing with explicit thread creation or interaction. The study of actors can be broken into two concepts, the \gls{actor_model}, which describes the model of computation and the \gls{actor_system}, which refers to the implementation of the model. Before discussing \CFA's actor system in detail, it is important to first describe the actor model, and the classic approach to implementing an actor system. \section{Actor Model} The actor model is a concurrent paradigm where computation is broken into units of work called actors, and the data for computation is distributed to actors in the form of messages~\cite{Hewitt73}. An actor is composed of a \Newterm{mailbox} (message queue) and a set of \Newterm{behaviours} that receive from the mailbox to perform work. Actors execute asynchronously upon receiving a message and can modify their own state, make decisions, spawn more actors, and send messages to other actors. Because the actor model is implicit concurrency, its strength is that it abstracts away many details and concerns needed in other concurrent paradigms. For example, mutual exclusion and locking are rarely relevant concepts in an actor model, as actors typically only operate on local state. An actor does not have a thread. An actor is executed by an underlying \Newterm{executor} (kernel thread-pool) that fairly invokes each actor, where an actor invocation processes one or more messages from its mailbox. The default number of executor threads is often proportional to the number of computer cores to achieve good performance. An executor is often tunable with respect to the number of kernel threads and its scheduling algorithm, which optimize for specific actor applications and workloads \see{end of Section~\ref{s:CFAActorSyntax}}. \section{Classic Actor System} An implementation of the actor model with a community of actors is called an actor system. Actor systems largely follow the actor model, but can differ in some ways. While the semantics of message \emph{send} is asynchronous, the implementation may be synchronous or a combination. The default semantics for message \emph{receive} is FIFO, so an actor receives messages from its mailbox in temporal (arrival) order; however, messages sent among actors arrive in any order. Some actor systems provide priority-based mailboxes and/or priority-based message-selection within a mailbox, where custom message dispatchers search among or within a mailbox(es) with a predicate for specific kinds of actors and/or messages. Some actor systems provide a shared mailbox where multiple actors receive from a common mailbox~\cite{Akka}, which is contrary to the no-sharing design of the basic actor-model (and requires additional locking). For non-FIFO service, some notion of fairness (eventual progress) must exist, otherwise messages have a high latency or starve, \ie never received. Finally, some actor systems provide multiple typed-mailboxes, which then lose the actor-\lstinline{become} mechanism (see Section~\ref{s:SafetyProductivity}). %While the definition of the actor model provides no restrictions on message ordering, actor systems tend to guarantee that messages sent from a given actor $i$ to actor $j$ will arrive at actor $j$ in the order they were sent. Another way an actor system varies from the model is allowing access to shared global-state. When this occurs, it complicates the implementation as this breaks any implicit mutual-exclusion guarantees when only accessing local-state. \begin{figure} \begin{tabular}{l|l} \subfloat[Actor-centric system]{\label{f:standard_actor}\input{diagrams/standard_actor.tikz}} & \subfloat[Message-centric system]{\label{f:inverted_actor}\raisebox{.1\height}{\input{diagrams/inverted_actor.tikz}}} \end{tabular} \caption{Classic and inverted actor implementation approaches with sharded queues.} \end{figure} \section{\CFA Actors} Figure~\ref{f:standard_actor} shows an actor system designed as \Newterm{actor-centric}, where a set of actors are scheduled and run on underlying executor threads~\cite{CAF,Akka,ProtoActor}. The simplest design has a single global queue of actors accessed by the executor threads, but this approach results in high contention as both ends of the queue by the executor threads. The more common design is to \Newterm{shard} the single queue among the executor threads, where actors are permanently assigned or can float among the queues. Sharding significantly decreases contention among executor threads adding and removing actors to/from a queue. Finally, each actor has a receive queue of messages (mailbox), which is a single consumer, multi-producer queue, \ie only the actor removes from the mailbox but multiple actors can attach messages. When an actor receives a message in its mailbox, the actor is marked ready and scheduled by a thread to run the actor's current work unit on the message(s). % cite parallel theatre and our paper Figure \ref{f:inverted_actor} shows an actor system designed as \Newterm{message-centric}, where a set of messages are scheduled and run on underlying executor threads~\cite{uC++,Nigro21}. Again, the simplest design has a single global queue of messages accessed by the executor threads, but this approach has the same contention problem by the executor threads. Therefore, the messages (mailboxes) are sharded and executor threads schedule each message, which points to its corresponding actor. Here, an actor's messages are permanently assigned to one queue to ensure FIFO receiving and/or reduce searching for specific actor/messages. Since multiple actors belong to each message queue, actor messages are interleaved on a queue. This design is \Newterm{inverted} because actors belong to a message queue, whereas in the classic approach a message queue belongs to each actor. % In this inverted actor system instead of each executor threads owning a queue of actors, they each own a queue of messages. % In this scheme work is consumed from their queue and executed by underlying threads. The inverted model can be taken a step further by sharding the message queues for each executor threads, so each executor thread owns a set of queues and cycles through them. Again, this extra level of sharding is to reduce queue contention. % The arrows from the message queues to the actors in the diagram indicate interleaved messages addressed to each actor. The actor system in \CFA uses a message-centric design, adopts several features from my prior actor work in \uC~\cite{}, and adds the following contributions related to \CFA: \begin{enumerate}[topsep=5pt,itemsep=3pt,parsep=0pt] \item Provide insight into the impact of envelope allocation in actor systems. In all actor systems, dynamic allocation is needed to ensure the lifetime of a unit of work persists from its creation until the unit of work is executed. This allocation is often called an \Newterm{envelope} as it ``packages'' the information needed to run the unit of work, alongside any other information needed to send the unit of work, such as an actor's address or link fields. This dynamic allocation occurs once per message sent. Unfortunately, the high rate of message sends in an actor system results in significant contention on the memory allocator. A novel data structure is introduced to consolidate allocations to improve performance by minimizing allocator contention. \item Improve performance of the inverted actor system using multiple approaches to minimize contention on queues, such as queue gulping and avoiding atomic operations. \item Introduce work stealing in the inverted actor system. Work stealing in an actor-centric system involves stealing one or more actors among executor threads. In the inverted system, the notion of stealing message queues is introduced. The queue stealing is implemented such that the act of stealing work does not contend with non-stealing executor threads running actors. \item Introduce and evaluate a timestamp-based work-stealing heuristic with the goal of maintaining non-workstealing performance in work-saturated workloads and improving performance on unbalanced workloads. \item Provide a suite of safety and productivity features including static-typing, detection of erroneous message sends, statistics tracking, and more. \end{enumerate} \section{\CFA Actor Syntax}\label{s:CFAActorSyntax} \CFA is not an object oriented language and it does not have \gls{rtti}. As such, all message sends and receives among actors can only occur using static type-matching, as in Typed-Akka~\cite{AkkaTyped}. Figure~\ref{f:BehaviourStyles} contrasts dynamic and static type-matching. Figure~\ref{l:dynamic_style} shows the dynamic style with a heterogeneous message receive and an indirect dynamic type-discrimination for message processing. Figure~\ref{l:static_style} shows the static style with a homogeneous message receive and a direct static type-discrimination for message processing. The static-typing style is safer because of the static check and faster because there is no dynamic type-discrimination. The dynamic-typing style is more flexible because multiple kinds of messages can be handled in a behaviour condensing the processing code. \begin{figure} \centering \begin{lrbox}{\myboxA} \begin{cfa}[morekeywords=case] allocation receive( message & msg ) { case( @msg_type1@, msg ) { // discriminate type ... msg_d-> ...; // msg_type1 msg_d } else case( @msg_type2@, msg ) { ... msg_d-> ...; // msg_type2 msg_d ... } \end{cfa} \end{lrbox} \begin{lrbox}{\myboxB} \begin{cfa} allocation receive( @msg_type1@ & msg ) { ... msg ...; } allocation receive( @msg_type2@ & msg ) { ... msg ...; } ... \end{cfa} \end{lrbox} \subfloat[dynamic typing]{\label{l:dynamic_style}\usebox\myboxA} \hspace*{10pt} \vrule \hspace*{10pt} \subfloat[static typing]{\label{l:static_style}\usebox\myboxB} \caption{Behaviour Styles} \label{f:BehaviourStyles} \end{figure} \begin{figure} \centering \begin{cfa} // actor struct my_actor { @inline actor;@ $\C[3.25in]{// Plan-9 C inheritance}$ }; // messages struct str_msg { @inline message;@ $\C{// Plan-9 C nominal inheritance}$ char str[12]; }; void ?{}( str_msg & this, char * str ) { strcpy( this.str, str ); } $\C{// constructor}$ struct int_msg { @inline message;@ $\C{// Plan-9 C nominal inheritance}$ int i; }; void ?{}( int_msg & this, int i ) { this.i = i; } $\C{// constructor}$ // behaviours allocation receive( my_actor &, @str_msg & msg@ ) { sout | "string message \"" | msg.str | "\""; return Nodelete; $\C{// actor not finished}$ } allocation receive( my_actor &, @int_msg & msg@ ) { sout | "integer message" | msg.i; return Nodelete; $\C{// actor not finished}$ } int main() { start_actor_system(); $\C{// sets up executor}$ my_actor actor; $\C{// default constructor call}$ str_msg str_msg{ "Hello World" }; $\C{// constructor call}$ int_msg int_msg{ 42 }; $\C{// constructor call}$ @actor << str_msg << int_msg;@ $\C{// cascade sends}$ @actor << int_msg;@ $\C{// send}$ @actor << finished_msg;@ $\C{// send => terminate actor (deallocation deferred)}$ stop_actor_system(); $\C{// waits until actors finish}\CRT$ } // deallocate int_msg, str_msg, actor \end{cfa} \caption{\CFA Actor Syntax} \label{f:CFAActorSyntax} \end{figure} Figure~\ref{f:CFAActorSyntax} shows a complete \CFA actor example starting with the actor type @my_actor@ created by defining a @struct@ that inherits from the base @actor@ @struct@ via the @inline@ keyword. This inheritance style is the Plan-9 C-style nominal inheritance discussed in Section~\ref{s:Inheritance}. Similarly, the message types @str_msg@ and @int_msg@ are created by defining a @struct@ that inherits from the base @message@ @struct@ via the @inline@ keyword. Both message types have constructors to set the message value. There are two matching @receive@ (behaviour) routines that process the corresponding typed messages. The program main begins by calling @start_actor_system@ to start the actor implementation, including executor threads to run the actors. An actor and two messages are created on the stack, and four messages are sent to the actor using operator @<<@. The last message is the builtin @finish_msg@, which returns @Finished@ to an executor thread, which removes the actor from the actor system \see{Section~\ref{s:ActorBehaviours}}. The call to @stop_actor_system@ blocks program main until all actors are finished and removed from the actor system. The program main ends by deleting the actor and messages from the stack. The output for the program is: \begin{cfa} string message "Hello World" integer message 42 integer message 42 \end{cfa} \subsection{Actor Behaviours}\label{s:ActorBehaviours} In general, a behaviour for some derived actor and derived message type is defined with following signature: \begin{cfa} allocation receive( my_actor & receiver, my_msg & msg ) \end{cfa} where @my_actor@ and @my_msg@ inherit from types @actor@ and @message@, respectively. The return value of @receive@ must be a value from enumerated type, @allocation@: \begin{cfa} enum allocation { Nodelete, Delete, Destroy, Finished }; \end{cfa} The values represent a set of actions that dictate what the executor does with an actor or message after a given behaviour returns. For actors, the @receive@ routine returns the @allocation@ status to the executor, which takes the appropriate action. For messages, either the default allocation, @Nodelete@, or any changed value in the message is examined by the executor, which takes the appropriate action. Message state is updated via a call to: \begin{cfa} void set_allocation( message & this, allocation state ) \end{cfa} In detail, the actions taken by an executor for each of the @allocation@ values are: \noindent@Nodelete@ tells the executor that no action is to be taken with regard to an actor or message. This status is used when an actor continues receiving messages or a message may be reused. \noindent@Delete@ tells the executor to call the object's destructor and deallocate (delete) the object. This status is used with dynamically allocated actors and messages when they are not reused. \noindent@Destroy@ tells the executor to call the object's destructor, but not deallocate the object. This status is used with dynamically allocated actors and messages whose storage is reused. \noindent@Finished@ tells the executor to mark the respective actor as finished executing, but not call the object's destructor nor deallocate the object. This status is used when actors or messages are global or stack allocated, or a programmer wants to manage deallocation themselves. For the actor system to terminate, all actors must have returned a status other than @Nodelete@. After an actor is terminated, it is erroneous to send messages to it. Similarly, after a message is terminated, it cannot be sent to an actor. Note, it is safe to construct an actor or message with a status other than @Nodelete@, since the executor only examines the allocation action after a behaviour returns. \subsection{Actor System}\label{s:ActorSystem} The calls to @start_actor_system@, and @stop_actor_system@ mark the start and end of a \CFA actor system. The call to @start_actor_system@ sets up an executor and executor threads for the actor system. It is possible to have multiple start/stop scenarios in a program. @start_actor_system@ has three overloaded signatures that vary the executor's configuration: \noindent@void start_actor_system()@ configures the executor to implicitly use all preallocated kernel-threads (processors), \ie the processors created by the program main prior to starting the actor system. When the number of processors is greater than 1, each executor's message queue is sharded by a factor of 16 to reduce contention, \ie for 4 executor threads (processors), there is a total of 4 $\times$ 16 message queues evenly distributed across the executor threads. \noindent@void start_actor_system( size_t num_thds )@ configures the number of executor threads to @num_thds@, with the same message queue sharding. \noindent@void start_actor_system( executor & this )@ allows the programmer to explicitly create and configure an executor for use by the actor system. Executor configuration options include are discussed in Section~\ref{s:executor}. \noindent All actors must be created \emph{after} calling @start_actor_system@ so the executor can keep track of the number of actors that have entered the system but not yet terminated. % All message sends are done using the left-shift operator, @<<@, similar to the syntax of \CC's stream output. % \begin{cfa} % allocation ?< 1, ordered queues pointers all attempting to swap with the queue in front of them in the ordering, except the first queue, and no other swaps occurring that involve these queues, at least one swap is guaranteed to succeed. \end{theorem}\label{t:vic_chain} A graph of the chain of swaps discussed in this theorem is presented in Figure~\ref{f:chain_swap}. \\ This is a proof by contradiction. Assume no swaps occur. Then all thieves must have failed at step 1, step 2 or step 3. For a given thief $b$ to fail at step 1, thief $b + 1$ must have succeeded at step 2 before $b$ executes step 0. Hence, not all thieves can fail at step 1. Furthermore if a thief $b$ fails at step 1 it logically splits the chain into two subchains $0 <- b$ and $b + 1 <- M - 1$, where $b$ has become solely a victim since its swap has failed and it did not modify any state. There must exist at least one chain containing two or more queues after since it is impossible for a split to occur both before and after a thief, since that requires failing at step 1 and succeeding at step 2. Hence, without loss of generality, whether thieves succeed or fail at step 1, this proof can proceed inductively. For a given thief $i$ to fail at step 2, it means that another thief $j$ had to have written to $i$'s queue pointer between $i$'s step 0 and step 2. The only way for $j$ to write to $i$'s queue pointer would be if $j$ was stealing from $i$ and had successfully finished step 3. If $j$ finished step 3 then the at least one swap was successful. Therefore all thieves did not fail at step 2. Hence all thieves must successfully complete step 2 and fail at step 3. However, since the first worker, thief $0$, is solely a victim and not a thief, it does not change the state of any of its queue pointers. Hence, in this case thief $1$ will always succeed in step 3 if all thieves succeed in step 2. Thus, by contradiction with the earlier assumption that no swaps occur, at least one swap must succeed. % \raisebox{.1\height}{} \begin{figure} \centering \begin{tabular}{l|l} \subfloat[Cyclic Swap Graph]{\label{f:cyclic_swap}\input{diagrams/cyclic_swap.tikz}} & \subfloat[Acyclic Swap Graph]{\label{f:acyclic_swap}\input{diagrams/acyclic_swap.tikz}} \end{tabular} \caption{Illustrations of cyclic and acyclic swap graphs.} \end{figure} \begin{theorem} Given a set of $M > 1$ swaps occurring that form a single directed connected graph. At least one swap is guaranteed to succeed if and only if the graph does not contain a cycle. \end{theorem}\label{t:vic_cycle} Representations of cyclic and acyclic swap graphs discussed in this theorem are presented in Figures~\ref{f:cyclic_swap} and \ref{f:acyclic_swap}. \\ First the reverse direction is proven. If the graph does not contain a cycle, then there must be at least one successful swap. Since the graph contains no cycles and is finite in size, then there must be a vertex $A$ with no outgoing edges. The graph can then be formulated as a tree with $A$ at the top since each node only has at most one outgoing edge and there are no cycles. The forward direction is proven by contradiction in a similar fashion to \ref{t:vic_chain}. Assume no swaps occur. Similar to \ref{t:vic_chain}, this graph can be inductively split into subgraphs of the same type by failure at step 1, so the proof proceeds without loss of generality. Similar to \ref{t:vic_chain} the conclusion is drawn that all thieves must successfully complete step 2 for no swaps to occur, since for step 2 to fail, a different thief has to successfully complete step 3, which would imply a successful swap. Hence, the only way forward is to assume all thieves successfully complete step 2. Hence for there to be no swaps all thieves must fail step 3. However, since $A$ has no outgoing edges, since the graph is connected there must be some $K$ such that $K < M - 1$ thieves are attempting to swap with $A$. Since all $K$ thieves have passed step 2, similar to \ref{t:one_vic} the first one of the $K$ thieves to attempt step 3 is guaranteed to succeed. Thus, by contradiction with the earlier assumption that no swaps occur, if the graph does not contain a cycle, at least one swap must succeed. The forward direction is proven by contrapositive. If the graph contains a cycle then there exists a situation where no swaps occur. This situation is constructed. Since all vertices have at most one outgoing edge the cycle must be directed. Furthermore, since the graph contains a cycle all vertices in the graph must have exactly one outgoing edge. This is shown through construction of an arbitrary cyclic graph. The graph contains a directed cycle by definition, so the construction starts with $T$ vertices in a directed cycle. Since the graph is connected, and each vertex has at most one outgoing edge, none of the vertices in the cycle have available outgoing edges to accommodate new vertices with no outgoing edges. Any vertices added to the graph must have an outgoing edge to connect, leaving the resulting graph with no available outgoing edges. Thus, by induction all vertices in the graph must have exactly one outgoing edge. Hence all vertices are thief queues. Now consider the case where all thieves successfully complete step 0-1, and then they all complete step 2. At this point all thieves are attempting to swap with a queue pointer whose value has changed to @0p@. If all thieves attempt the @CAS@ before any write backs, then they will all fail. Thus, by contrapositive, if the graph contains a cycle then there exists a situation where no swaps occur. Hence, at least one swap is guaranteed to succeed if and only if the graph does not contain a cycle. % C_TODO: go through and use \paragraph to format to make it look nicer \subsection{Victim Selection}\label{s:victimSelect} In any work stealing algorithm thieves have some heuristic to determine which victim to choose from. Choosing this algorithm is difficult and can have implications on performance. There is no one selection heuristic that is known to be the best on all workloads. Recent work focuses on locality aware scheduling in actor systems\cite{barghi18}\cite{wolke17}. However, while locality aware scheduling provides good performance on some workloads, something as simple as randomized selection performs better on other workloads\cite{barghi18}. Since locality aware scheduling has been explored recently, this work introduces a heuristic called \textbf{longest victim} and compares it to randomized work stealing. The longest victim heuristic maintains a timestamp per executor threads that is updated every time a worker attempts to steal work. Thieves then attempt to steal from the thread with the oldest timestamp. This means that if two thieves look to steal at the same time, they likely will attempt to steal from the same victim. This does increase the chance at contention between thieves, however given that workers have multiple queues under them, often in the tens or hundreds of queues per worker it is rare for two queues to attempt so steal the same queue. Furthermore in the case they attempt to steal the same queue at least one of them is guaranteed to successfully steal the queue as shown in Theorem \ref{t:one_vic}. Additionally, the longest victim heuristic makes it very improbable that the no swap scenario presented in Theorem \ref{t:vic_cycle} manifests. Given the longest victim heuristic, for a cycle to manifest it would require all workers to attempt to steal in a short timeframe. This is the only way that more than one thief could choose another thief as a victim, since timestamps are only updated upon attempts to steal. In this case, the probability of lack of any successful swaps is a non issue, since it is likely that these steals were not important if all workers are trying to steal. \section{Safety and Productivity}\label{s:SafetyProductivity} \CFA's actor system comes with a suite of safety and productivity features. Most of these features are present in \CFA's debug mode, but are removed when code is compiled in nodebug mode. The suit of features include the following. \begin{itemize} \item Static-typed message sends. If an actor does not support receiving a given message type, the actor program is rejected at compile time, allowing unsupported messages to never be sent to actors. \item Detection of message sends to Finished/Destroyed/Deleted actors. All actors have a ticket that assigns them to a respective queue. The maximum integer value of the ticket is reserved to indicate that an actor is dead, and subsequent message sends result in an error. \item Actors made before the executor can result in undefined behaviour since an executor needs to be created beforehand so it can give out the tickets to actors. As such, this is detected and an error is printed. \item When an executor is created, the queues are handed out to executor threads in round robin order. If there are fewer queues than executor threads, then some workers will spin and never do any work. There is no reasonable use case for this behaviour so an error is printed if the number of queues is fewer than the number of executor threads. \item A warning is printed when messages are deallocated without being sent. Since the @Finished@ allocation status is unused for messages, it is used internally to detect if a message has been sent. Deallocating a message without sending it could indicate to a user that they are touching freed memory later, or it could point out extra allocations that could be removed. \end{itemize} In addition to these features, \CFA's actor system comes with a suite of statistics that can be toggled on and off. These statistics have minimal impact on the actor system's performance since they are counted on a per executor threads basis. During shutdown of the actor system they are aggregated, ensuring that the only atomic instructions used by the statistics counting happen at shutdown. The statistics measured are as follows. \begin{description} \item[\LstBasicStyle{\textbf{Actors Created}}] Actors created. Includes both actors made by the main and ones made by other actors. \item[\LstBasicStyle{\textbf{Messages Sent}}] Messages sent and received. Includes termination messages send to the executor threads. \item[\LstBasicStyle{\textbf{Gulps}}] Gulps that occurred across the executor threads. \item[\LstBasicStyle{\textbf{Average Gulp Size}}] Average number of messages in a gulped queue. \item[\LstBasicStyle{\textbf{Missed gulps}}] Occurrences where a worker missed a gulp due to the concurrent queue processing by another worker. \item[\LstBasicStyle{\textbf{Steal attempts}}] Worker threads attempts to steal work. \item[\LstBasicStyle{\textbf{Steal failures (no candidates)}}] Work stealing failures due to selected victim not having any non empty or non-being-processed queues. \item[\LstBasicStyle{\textbf{Steal failures (failed swaps)}}] Work stealing failures due to the two stage atomic swap failing. \item[\LstBasicStyle{\textbf{Messages stolen}}] Aggregate of the number of messages in queues as they were stolen. \item[\LstBasicStyle{\textbf{Average steal size}}] Average number of messages in a stolen queue. \end{description} These statistics enable a user of \CFA's actor system to make informed choices about how to configure their executor, or how to structure their actor program. For example, if there is a lot of messages being stolen relative to the number of messages sent, it could indicate to a user that their workload is heavily imbalanced across executor threads. In another example, if the average gulp size is very high, it could indicate that the executor could use more queue sharding. % C_TODO cite poison pill messages and add languages Another productivity feature that is included is a group of poison-pill messages. Poison-pill messages are common across actor systems, including Akka and ProtoActor \cite{}. Poison-pill messages inform an actor to terminate. In \CFA, due to the allocation of actors and lack of garbage collection, there needs to be a suite of poison-pills. The messages that \CFA provides are @DeleteMsg@, @DestroyMsg@, and @FinishedMsg@. These messages are supported on all actor types via inheritance and when sent to an actor, the actor takes the corresponding allocation action after receiving the message. Note that any pending messages to the actor will still be sent. It is still the user's responsibility to ensure that an actor does not receive any messages after termination. \section{Performance}\label{s:actor_perf} \CAP{I will update the figures to have the larger font size and different line markers once we start editing this chapter.} The performance of \CFA's actor system is tested using a suite of microbenchmarks, and compared with other actor systems. Most of the benchmarks are the same as those presented in \ref{}, with a few additions. % C_TODO cite actor paper At the time of this work the versions of the actor systems are as follows. \CFA 1.0, \uC 7.0.0, Akka Typed 2.7.0, CAF 0.18.6, and ProtoActor-Go v0.0.0-20220528090104-f567b547ea07. Akka Classic is omitted as Akka Typed is their newest version and seems to be the direction they are headed in. The experiments are run on \begin{list}{\arabic{enumi}.}{\usecounter{enumi}\topsep=5pt\parsep=5pt\itemsep=0pt} \item Supermicro SYS--6029U--TR4 Intel Xeon Gold 5220R 24--core socket, hyper-threading $\times$ 2 sockets (48 process\-ing units) 2.2GHz, running Linux v5.8.0--59--generic \item Supermicro AS--1123US--TR4 AMD EPYC 7662 64--core socket, hyper-threading $\times$ 2 sockets (256 processing units) 2.0 GHz, running Linux v5.8.0--55--generic \end{list} The benchmarks are run on up to 48 cores. On the Intel, when going beyond 24 cores there is the choice to either hop sockets or to use hyperthreads. Either choice will cause a blip in performance trends, which can be seen in the following performance figures. On the Intel the choice was made to hyperthread instead of hopping sockets for experiments with more than 24 cores. All benchmarks presented are run 5 times and the median is taken. Error bars showing the 95\% confidence intervals are drawn on each point on the graphs. If the confidence bars are small enough, they may be obscured by the point. In this section \uC will be compared to \CFA frequently, as the actor system in \CFA was heavily based off \uC's actor system. As such the performance differences that arise are largely due to the contributions of this work. \begin{table}[t] \centering \setlength{\extrarowheight}{2pt} \setlength{\tabcolsep}{5pt} \caption{Static Actor/Message Performance: message send, program memory} \label{t:StaticActorMessagePerformance} \begin{tabular}{*{5}{r|}r} & \multicolumn{1}{c|}{\CFA (100M)} & \multicolumn{1}{c|}{CAF (10M)} & \multicolumn{1}{c|}{Akka (100M)} & \multicolumn{1}{c|}{\uC (100M)} & \multicolumn{1}{c@{}}{ProtoActor (100M)} \\ \hline AMD & \input{data/pykeSendStatic} \\ \hline Intel & \input{data/nasusSendStatic} \end{tabular} \bigskip \caption{Dynamic Actor/Message Performance: message send, program memory} \label{t:DynamicActorMessagePerformance} \begin{tabular}{*{5}{r|}r} & \multicolumn{1}{c|}{\CFA (20M)} & \multicolumn{1}{c|}{CAF (2M)} & \multicolumn{1}{c|}{Akka (2M)} & \multicolumn{1}{c|}{\uC (20M)} & \multicolumn{1}{c@{}}{ProtoActor (2M)} \\ \hline AMD & \input{data/pykeSendDynamic} \\ \hline Intel & \input{data/nasusSendDynamic} \end{tabular} \end{table} \subsection{Message Sends} Message sending is the key component of actor communication. As such latency of a single message send is the fundamental unit of fast-path performance for an actor system. The following two microbenchmarks evaluate the average latency for a static actor/message send and a dynamic actor/message send. Static and dynamic refer to the allocation of the message and actor. In the static send benchmark a message and actor are allocated once and then the message is sent to the same actor repeatedly until it has been sent 100 million (100M) times. The average latency per message send is then calculated by dividing the duration by the number of sends. This benchmark evaluates the cost of message sends in the actor use case where all actors and messages are allocated ahead of time and do not need to be created dynamically during execution. The CAF static send benchmark only sends a message 10M times to avoid extensively long run times. In the dynamic send benchmark the same experiment is performed, with the change that with each send a new actor and message is allocated. This evaluates the cost of message sends in the other common actor pattern where actors and message are created on the fly as the actor program tackles a workload of variable or unknown size. Since dynamic sends are more expensive, this benchmark repeats the actor/message creation and send 20M times (\uC, \CFA), or 2M times (Akka, CAF, ProtoActor), to give an appropriate benchmark duration. The results from the static/dynamic send benchmarks are shown in Figures~\ref{t:StaticActorMessagePerformance} and \ref{t:DynamicActorMessagePerformance} respectively. \CFA leads the charts in both benchmarks, largely due to the copy queue removing the majority of the envelope allocations. In the static send benchmark all systems except CAF have static send costs that are in the same ballpark, only varying by ~70ns. In the dynamic send benchmark all systems experience slower message sends, as expected due to the extra allocations. However, Akka and ProtoActor, slow down by a more significant margin than the \uC and \CFA. This is likely a result of Akka and ProtoActor's garbage collection, which can suffer from hits in performance for allocation heavy workloads, whereas \uC and \CFA have explicit allocation/deallocation. \subsection{Work Stealing} \CFA's actor system has a work stealing mechanism which uses the longest victim heuristic, introduced in Section~ref{s:victimSelect}. In this performance section, \CFA with the longest victim heuristic is compared with other actor systems on the benchmark suite, and is separately compared with vanilla non-stealing \CFA and \CFA with randomized work stealing. \begin{figure} \centering \subfloat[AMD \CFA Balance-One Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFABalance-One.pgf}} \label{f:BalanceOneAMD} } \subfloat[Intel \CFA Balance-One Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFABalance-One.pgf}} \label{f:BalanceOneIntel} } \caption{The balance-one benchmark comparing stealing heuristics (lower is better).} \end{figure} \begin{figure} \centering \subfloat[AMD \CFA Balance-Multi Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFABalance-Multi.pgf}} \label{f:BalanceMultiAMD} } \subfloat[Intel \CFA Balance-Multi Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFABalance-Multi.pgf}} \label{f:BalanceMultiIntel} } \caption{The balance-multi benchmark comparing stealing heuristics (lower is better).} \end{figure} There are two benchmarks in which \CFA's work stealing is solely evaluated. The main goal of introducing work stealing to \CFA's actor system is to eliminate the pathological unbalanced cases that can present themselves in a system without some form of load balancing. The following two microbenchmarks construct two such pathological cases, and compare the work stealing variations of \CFA. The balance benchmarks adversarially takes advantage of the round robin assignment of actors to load all actors that will do work on specific cores and create 'dummy' actors that terminate after a single message send on all other cores. The workload on the loaded cores is the same as the executor benchmark described in \ref{s:executorPerf}, but with fewer rounds. The balance-one benchmark loads all the work on a single core, whereas the balance-multi loads all the work on half the cores (every other core). Given this layout, one expects the ideal speedup of work stealing in the balance-one case to be $N / N - 1$ where $N$ is the number of threads. In the balance-multi case the ideal speedup is 0.5. Note that in the balance-one benchmark the workload is fixed so decreasing runtime is expected. In the balance-multi experiment, the workload increases with the number of cores so an increasing or constant runtime is expected. On both balance microbenchmarks slightly less than ideal speedup compared to the non stealing variation is achieved by both the random and longest victim stealing heuristics. On the balance-multi benchmark \ref{f:BalanceMultiAMD},\ref{f:BalanceMultiIntel} the random heuristic outperforms the longest victim. This is likely a result of the longest victim heuristic having a higher stealing cost as it needs to maintain timestamps and look at all timestamps before stealing. Additionally, a performance cost can be observed when hyperthreading kicks in in Figure~\ref{f:BalanceMultiIntel}. In the balance-one benchmark on AMD \ref{f:BalanceOneAMD}, the performance bottoms out at 32 cores onwards likely due to the amount of work becoming less than the cost to steal it and move it across cores and cache. On Intel \ref{f:BalanceOneIntel}, above 32 cores the performance gets worse for all variants due to hyperthreading. Note that the non stealing variation of balance-one will slow down marginally as the cores increase due to having to create more dummy actors on the inactive cores during startup. \subsection{Executor}\label{s:executorPerf} The microbenchmarks in this section are designed to stress the executor. The executor is the scheduler of an actor system and is responsible for organizing the interaction of executor threads to service the needs of a workload. In the executor benchmark, 40'000 actors are created and assigned a group. Each group of actors is a group of 100 actors who send and receive 100 messages from all other actors in their group. Each time an actor completes all their sends and receives, they are done a round. After all groups have completed 400 rounds the system terminates. This microbenchmark is designed to flood the executor with a large number of messages flowing between actors. Given there is no work associated with each message, other than sending more messages, the intended bottleneck of this experiment is the executor message send process. \begin{figure} \centering \subfloat[AMD Executor Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusExecutor.pgf}} \label{f:ExecutorAMD} } \subfloat[Intel Executor Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeExecutor.pgf}} \label{f:ExecutorIntel} } \caption{The executor benchmark comparing actor systems (lower is better).} \end{figure} The results of the executor benchmark in Figures~\ref{f:ExecutorIntel} and \ref{f:ExecutorAMD} show \CFA with the lowest runtime relative to its peers. The difference in runtime between \uC and \CFA is largely due to the usage of the copy queue described in Section~\ref{s:copyQueue}. The copy queue both reduces and consolidates allocations, heavily reducing contention on the memory allocator. Additionally, due to the static typing in \CFA's actor system, it is able to get rid of expensive dynamic casts that occur in \uC to discriminate messages by type. Note that dynamic casts are usually not very expensive, but relative to the high performance of the rest of the implementation of the \uC actor system, the cost is significant. \begin{figure} \centering \subfloat[AMD \CFA Executor Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFAExecutor.pgf}} \label{f:cfaExecutorAMD} } \subfloat[Intel \CFA Executor Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFAExecutor.pgf}} \label{f:cfaExecutorIntel} } \caption{Executor benchmark comparing \CFA stealing heuristics (lower is better).} \end{figure} When comparing the \CFA stealing heuristics in Figure~\ref{f:cfaExecutorAMD} it can be seen that the random heuristic falls slightly behind the other two, but in Figure~\ref{f:cfaExecutorIntel} the runtime of all heuristics are nearly identical to each other. \begin{figure} \centering \subfloat[AMD Repeat Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusRepeat.pgf}} \label{f:RepeatAMD} } \subfloat[Intel Repeat Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeRepeat.pgf}} \label{f:RepeatIntel} } \caption{The repeat benchmark comparing actor systems (lower is better).} \end{figure} The repeat microbenchmark also evaluates the executor. It stresses the executor's ability to withstand contention on queues, as it repeatedly fans out messages from a single client to 100000 servers who then all respond to the client. After this scatter and gather repeats 200 times the benchmark terminates. The messages from the servers to the client will likely all come in on the same queue, resulting in high contention. As such this benchmark will not scale with the number of processors, since more processors will result in higher contention. In Figure~\ref{f:RepeatAMD} we can see that \CFA performs well compared to \uC, however by less of a margin than the executor benchmark. One factor in this result is that the contention on the queues poses a significant bottleneck. As such the gains from using the copy queue are much less apparent. \begin{figure} \centering \subfloat[AMD \CFA Repeat Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFARepeat.pgf}} \label{f:cfaRepeatAMD} } \subfloat[Intel \CFA Repeat Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFARepeat.pgf}} \label{f:cfaRepeatIntel} } \caption{The repeat benchmark comparing \CFA stealing heuristics (lower is better).} \end{figure} In Figure~\ref{f:RepeatIntel} \uC and \CFA are very comparable. In comparison with the other systems \uC does well on the repeat benchmark since it does not have work stealing. The client of this experiment is long running and maintains a lot of state, as it needs to know the handles of all the servers. When stealing the client or its respective queue (in \CFA's inverted model), moving the client incurs a high cost due to cache invalidation. As such stealing the client can result in a hit in performance. This result is shown in Figure~\ref{f:cfaRepeatAMD} and \ref{f:cfaRepeatIntel} where the no-stealing version of \CFA performs better than both stealing variations. In particular on the Intel machine in Figure~\ref{f:cfaRepeatIntel}, the cost of stealing is higher, which can be seen in the vertical shift of Akka, CAF and CFA results in Figure~\ref{f:RepeatIntel} (\uC and ProtoActor do not have work stealing). The shift for CAF is particularly large, which further supports the hypothesis that CAF's work stealing is particularly eager. In both the executor and the repeat benchmark CAF performs poorly. It is hypothesized that CAF has an aggressive work stealing algorithm, that eagerly attempts to steal. This results in poor performance in benchmarks with small messages containing little work per message. On the other hand, in \ref{f:MatrixAMD} CAF performs much better since each message has a large amount of work, and few messages are sent, so the eager work stealing allows for the clean up of loose ends to occur faster. This hypothesis stems from experimentation with \CFA. CAF uses a randomized work stealing heuristic. In \CFA if the system is tuned so that it steals work much more eagerly with a randomized it was able to replicate the results that CAF achieves in the matrix benchmark, but this tuning performed much worse on all other microbenchmarks that we present, since they all perform a small amount of work per message. \begin{table}[t] \centering \setlength{\extrarowheight}{2pt} \setlength{\tabcolsep}{5pt} \caption{Executor Program Memory High Watermark} \label{t:ExecutorMemory} \begin{tabular}{*{5}{r|}r} & \multicolumn{1}{c|}{\CFA} & \multicolumn{1}{c|}{CAF} & \multicolumn{1}{c|}{Akka} & \multicolumn{1}{c|}{\uC} & \multicolumn{1}{c@{}}{ProtoActor} \\ \hline AMD & \input{data/pykeExecutorMem} \\ \hline Intel & \input{data/nasusExecutorMem} \end{tabular} \end{table} Figure~\ref{t:ExecutorMemory} shows the high memory watermark of the actor systems when running the executor benchmark on 48 cores. \CFA has a high watermark relative to the other non-garbage collected systems \uC, and CAF. This is a result of the copy queue data structure, as it will over-allocate storage and not clean up eagerly, whereas the per envelope allocations will always allocate exactly the amount of storage needed. \subsection{Matrix Multiply} The matrix benchmark evaluates the actor systems in a practical application, where actors concurrently multiplies two matrices. The majority of the computation in this benchmark involves computing the final matrix, so this benchmark stresses the actor systems' ability to have actors run work, rather than stressing the executor or message sending system. Given $Z_{m,r} = X_{m,n} \cdot Y_{n,r}$, the matrix multiply is defined as: \begin{displaymath} X_{i,j} \cdot Y_{j,k} = \left( \sum_{c=1}^{j} X_{row,c}Y_{c,column} \right)_{i,k} \end{displaymath} The benchmark uses input matrices $X$ and $Y$ that are both $3072$ by $3072$ in size. An actor is made for each row of $X$ and is passed via message the information needed to calculate a row of the result matrix $Z$. Given that the bottleneck of the benchmark is the computation of the result matrix, it follows that the results in Figures~\ref{f:MatrixAMD} and \ref{f:MatrixIntel} are clustered closer than other experiments. In Figure~\ref{f:MatrixAMD} \uC and \CFA have identical performance and in Figure~\ref{f:MatrixIntel} \uC pulls ahead of \CFA after 24 cores likely due to costs associated with work stealing while hyperthreading. As mentioned in \ref{s:executorPerf}, it is hypothesized that CAF performs better in this benchmark compared to others due to its eager work stealing implementation. In Figures~\ref{f:cfaMatrixAMD} and \ref{f:cfaMatrixIntel} there is little negligible performance difference across \CFA stealing heuristics. \begin{figure} \centering \subfloat[AMD Matrix Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusMatrix.pgf}} \label{f:MatrixAMD} } \subfloat[Intel Matrix Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeMatrix.pgf}} \label{f:MatrixIntel} } \caption{The matrix benchmark comparing actor systems (lower is better).} \end{figure} \begin{figure} \centering \subfloat[AMD \CFA Matrix Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/nasusCFAMatrix.pgf}} \label{f:cfaMatrixAMD} } \subfloat[Intel \CFA Matrix Benchmark]{ \resizebox{0.5\textwidth}{!}{\input{figures/pykeCFAMatrix.pgf}} \label{f:cfaMatrixIntel} } \caption{The matrix benchmark comparing \CFA stealing heuristics (lower is better).} \end{figure} % Local Variables: % % tab-width: 4 % % End: %