[27dde72] | 1 | % ###### # ###### # # # ####### # ### ##### # # |
---|
| 2 | % # # # # # # # # # # # # # # # ## ## |
---|
| 3 | % # # # # # # # # # # # # # # # # # # |
---|
| 4 | % ###### # # ###### # # # # ##### # # ##### # # # |
---|
| 5 | % # ####### # # ####### # # # # # # # # |
---|
| 6 | % # # # # # # # # # # # # # # # # |
---|
| 7 | % # # # # # # # ####### ####### ####### ####### ### ##### # # |
---|
| 8 | \chapter{Parallelism} |
---|
| 9 | Historically, computer performance was about processor speeds and instructions count. However, with heat dissipation being a direct consequence of speed increase, parallelism has become the new source for increased performance~\cite{Sutter05, Sutter05b}. In this decade, it is not longer reasonnable to create a high-performance application without caring about parallelism. Indeed, parallelism is an important aspect of performance and more specifically throughput and hardware utilization. The lowest-level approach of parallelism is to use \glspl{kthread} in combination with semantics like \code{fork}, \code{join}, etc. However, since these have significant costs and limitations, \glspl{kthread} are now mostly used as an implementation tool rather than a user oriented one. There are several alternatives to solve these issues that all have strengths and weaknesses. While there are many variations of the presented paradigms, most of these variations do not actually change the guarantees or the semantics, they simply move costs in order to achieve better performance for certain workloads. |
---|
| 10 | |
---|
| 11 | \section{Paradigm} |
---|
| 12 | \subsection{User-level threads} |
---|
[d67cdb7] | 13 | A direct improvement on the \gls{kthread} approach is to use \glspl{uthread}. These threads offer most of the same features that the operating system already provide but can be used on a much larger scale. This approach is the most powerfull solution as it allows all the features of multi-threading, while removing several of the more expensive costs of kernel threads. The down side is that almost none of the low-level threading problems are hidden; users still have to think about data races, deadlocks and synchronization issues. These issues can be somewhat alleviated by a concurrency toolkit with strong garantees but the parallelism toolkit offers very little to reduce complexity in itself. |
---|
[27dde72] | 14 | |
---|
| 15 | Examples of languages that support \glspl{uthread} are Erlang~\cite{Erlang} and \uC~\cite{uC++book}. |
---|
| 16 | |
---|
| 17 | \subsection{Fibers : user-level threads without preemption} |
---|
[d67cdb7] | 18 | A popular varient of \glspl{uthread} is what is often refered to as \glspl{fiber}. However, \glspl{fiber} do not present meaningful semantical differences with \glspl{uthread}. Advocates of \glspl{fiber} list their high performance and ease of implementation as majors strenghts of \glspl{fiber} but the performance difference between \glspl{uthread} and \glspl{fiber} is controversial, and the ease of implementation, while true, is a weak argument in the context of language design. Therefore this proposal largely ignore fibers. |
---|
[27dde72] | 19 | |
---|
| 20 | An example of a language that uses fibers is Go~\cite{Go} |
---|
| 21 | |
---|
| 22 | \subsection{Jobs and thread pools} |
---|
[d67cdb7] | 23 | The approach on the opposite end of the spectrum is to base parallelism on \glspl{pool}. Indeed, \glspl{pool} offer limited flexibility but at the benefit of a simpler user interface. In \gls{pool} based systems, users express parallelism as units of work, called jobs, and a dependency graph (either explicit or implicit) that tie them together. This approach means users need not worry about concurrency but significantly limits the interaction that can occur among jobs. Indeed, any \gls{job} that blocks also blocks the underlying worker, which effectively means the CPU utilization, and therefore throughput, suffers noticeably. It can be argued that a solution to this problem is to use more workers than available cores. However, unless the number of jobs and the number of workers are comparable, having a significant amount of blocked jobs always results in idles cores. |
---|
[27dde72] | 24 | |
---|
| 25 | The gold standard of this implementation is Intel's TBB library~\cite{TBB}. |
---|
| 26 | |
---|
| 27 | \subsection{Paradigm performance} |
---|
[d67cdb7] | 28 | While the choice between the three paradigms listed above may have significant performance implication, it is difficult to pindown the performance implications of chosing a model at the language level. Indeed, in many situations one of these paradigms may show better performance but it all strongly depends on the workload. Having a large amount of mostly independent units of work to execute almost guarantess that the \gls{pool} based system has the best performance thanks to the lower memory overhead (i.e., not thread stack per job). However, interactions among jobs can easily exacerbate contention. User-level threads allow fine-grain context switching, which results in better resource utilisation, but context switches is more expansive and the extra control means users need to tweak more variables to get the desired performance. Finally, if the units of uninterrupted work are large enough the paradigm choice is largely amortised by the actual work done. |
---|
[27dde72] | 29 | |
---|
| 30 | \TODO |
---|
| 31 | |
---|
[d67cdb7] | 32 | \section{The \protect\CFA\ Kernel : Processors, Clusters and Threads}\label{kernel} |
---|
[27dde72] | 33 | |
---|
| 34 | |
---|
[d67cdb7] | 35 | \subsubsection{Future Work: Machine setup}\label{machine} |
---|
| 36 | While this was not done in the context of this proposal, another important aspect of clusters is affinity. While many common desktop and laptop PCs have homogeneous CPUs, other devices often have more heteregenous setups. For example, system using \acrshort{numa} configurations may benefit from users being able to tie clusters and/or kernel threads to certains CPU cores. OS support for CPU affinity is now common \cit, which means it is both possible and desirable for \CFA to offer an abstraction mechanism for portable CPU affinity. |
---|
| 37 | |
---|
| 38 | \subsection{Paradigms}\label{cfaparadigms} |
---|
| 39 | Given these building blocks, it is possible to reproduce all three of the popular paradigms. Indeed, \glspl{uthread} is the default paradigm in \CFA. However, disabling \glspl{preemption} on the \gls{cfacluster} means \glspl{cfathread} effectively become \glspl{fiber}. Since several \glspl{cfacluster} with different scheduling policy can coexist in the same application, this allows \glspl{fiber} and \glspl{uthread} to coexist in the runtime of an application. Finally, it is possible to build executors for thread pools from \glspl{uthread} or \glspl{fiber}. |
---|