[2aab69b] | 1 | A revised version of your manuscript that takes into account the comments |
---|
| 2 | of the referees will be reconsidered for publication. |
---|
| 3 | |
---|
| 4 | We have attempted to address all the referee's comments in the revised version |
---|
| 5 | of the paper, with notes below for each comment. |
---|
| 6 | |
---|
| 7 | ============================================================================= |
---|
| 8 | |
---|
| 9 | Reviewing: 1 |
---|
| 10 | |
---|
| 11 | As far as I can tell, the article contains three main ideas: an |
---|
| 12 | asynchronous execution / threading model; a model for monitors to provide |
---|
| 13 | mutual exclusion; and an implementation. The first two ideas are drawn |
---|
| 14 | together in Table 1: unfortunately this is on page 25 of 30 pages of |
---|
| 15 | text. Implementation choices and descriptions are scattered throughout the |
---|
| 16 | paper - and the sectioning of the paper seems almost arbitrary. |
---|
| 17 | |
---|
| 18 | Fixed, Table 1 is moved to the start and explained in detail. |
---|
| 19 | |
---|
| 20 | The article is about its contributions. Simply adding feature X to |
---|
| 21 | language Y isn't by itself a contribution, (when feature X isn't already a |
---|
| 22 | contribution). |
---|
| 23 | |
---|
| 24 | C++ (Y) added object-oriented programming (X) to C, where OO programming (X) |
---|
| 25 | was not a contribution. |
---|
| 26 | |
---|
| 27 | For example: why support two kinds of generators as well as user-level |
---|
| 28 | threads? Why support both low and high level synchronization constructs? |
---|
| 29 | |
---|
| 30 | Fixed, as part of discussing Table 1. |
---|
| 31 | |
---|
| 32 | Similarly I would have found the article easier to follow if it was written |
---|
| 33 | top down, presenting the design principles, present the space of language |
---|
| 34 | features, justify chosen language features (and rationale) and those |
---|
| 35 | excluded, and then present implementation, and performance. |
---|
| 36 | |
---|
| 37 | Fixed, the paper is now restructured in this form. |
---|
| 38 | |
---|
| 39 | Then the writing of the article is often hard to follow, to say the |
---|
| 40 | least. Two examples: section 3 "stateful functions" - I've some idea |
---|
| 41 | what that is (a function with Algol's "own" or C's "static" variables? |
---|
| 42 | but in fact the paper has a rather more specific idea than that. |
---|
| 43 | |
---|
| 44 | Fixed, at the start of this section. |
---|
| 45 | |
---|
| 46 | The top of page 3 throws a whole lot of definitions at the reader |
---|
| 47 | "generator" "coroutine" "stackful" "stackless" "symmetric" "asymmetric" |
---|
| 48 | without every stopping to define each one |
---|
| 49 | |
---|
| 50 | Hopefully fixed by moving Table 1 forward. |
---|
| 51 | |
---|
| 52 | --- but then in footnote "C" takes the time to explain what C's "main" |
---|
| 53 | function is? I cannot imagine a reader of this paper who doesn't know what |
---|
| 54 | "main" is in C; especially if they understand the other concepts already |
---|
| 55 | presented in the paper. |
---|
| 56 | |
---|
| 57 | Fixed by shortening. |
---|
| 58 | |
---|
| 59 | The start of section 3 then does the same |
---|
| 60 | thing: putting up a whole lot of definitions, making distinctions and |
---|
| 61 | comparisons, even talking about some runtime details, but the critical |
---|
| 62 | definition of a monitor doesn't appear until three pages later, at the |
---|
| 63 | start of section 5 on p15, lines 29-34 are a good, clear, description |
---|
| 64 | of what a monitor actually is. That needs to come first, rather than |
---|
| 65 | being buried again after two sections of comparisons, discussions, |
---|
| 66 | implementations, and options that are ungrounded because they haven't |
---|
| 67 | told the reader what they are actually talking about. First tell the |
---|
| 68 | reader what something is, then how they might use it (as programmers: |
---|
| 69 | what are the rules and restrictions) and only then start comparison |
---|
| 70 | with other things, other approaches, other languages, or |
---|
| 71 | implementations. |
---|
| 72 | |
---|
| 73 | Hopefully fixed by moving Table 1 forward. |
---|
| 74 | |
---|
| 75 | The description of the implementation is similarly lost in the trees |
---|
| 76 | without ever really seeing the wood. Figure 19 is crucial here, but |
---|
| 77 | it's pretty much at the end of the paper, and comments about |
---|
| 78 | implementations are threaded throughout the paper without the context |
---|
| 79 | (fig 19) to understand what's going on. |
---|
| 80 | |
---|
| 81 | We have to agree to disagree on the location of Fig 19. Early discussion about |
---|
| 82 | implementation for the various control structures are specific to that feature. |
---|
| 83 | Fig 19 shows the global runtime structure, which manages only the threading |
---|
| 84 | aspect of the control structures and their global organization. |
---|
| 85 | |
---|
| 86 | The protocol for performance testing may just about suffice for C (although |
---|
| 87 | is N constantly ten million, or does it vary for each benchmark) |
---|
| 88 | |
---|
| 89 | Fixed, the paper states N varies per language/benchmark so the benchmark runs |
---|
| 90 | long enough to get a good average per operation. |
---|
| 91 | |
---|
| 92 | but such evaluation isn't appropriate for garbage-collected or JITTed |
---|
| 93 | languages like Java or Go. |
---|
| 94 | |
---|
| 95 | Please explain. All the actions in the benchmarks occur independently of the |
---|
| 96 | storage-management scheme, e.g., acquiring a lock is an aspect of execution not |
---|
| 97 | storage. In fact, garbage-collected or JITTed languages cheat on benchmarks and |
---|
| 98 | we had to take great care to prevent cheating and measure the actual operation. |
---|
| 99 | |
---|
| 100 | p1 only a subset of C-forall extensions? |
---|
| 101 | |
---|
| 102 | Fixed, removed. |
---|
| 103 | |
---|
| 104 | p1 "has features often associated with object-oriented programming |
---|
| 105 | languages, such as constructors, destructors, virtuals and simple |
---|
| 106 | inheritance." There's no need to quibble about this. Once a language has |
---|
| 107 | inheritance, it's hard to claim it's not object-oriented. |
---|
| 108 | |
---|
| 109 | We have to agree to disagree. Object languages are defined by the notion of |
---|
| 110 | nested functions in a aggregate structure with a special receiver parameter |
---|
| 111 | "this", not by inheritance. Inheritance is a polymorphic mechanism, e.g, |
---|
| 112 | Plan-9 C has simple inheritance but is not object-oriented. Because Cforall |
---|
| 113 | does not have a specific receiver, it is possible to have multiple function |
---|
| 114 | parameters as receivers, which introduces new concepts like bulk acquire for |
---|
| 115 | monitors. |
---|
| 116 | |
---|
| 117 | p2 barging? signals-as-hints? |
---|
| 118 | |
---|
| 119 | Added a footnote for barging. We feel these terms are well known in the |
---|
| 120 | concurrency literature, especially in pthreads and Java, and both terms have |
---|
| 121 | citations with extensive explanations and further citations. |
---|
| 122 | |
---|
| 123 | p3 start your discussion of generations with a simple example of a |
---|
| 124 | C-forall generator. Fig 1(b) might do: but put it inline instead of |
---|
| 125 | the python example - and explain the key rules and restrictions on the |
---|
| 126 | construct. Then don't even start to compare with coroutines until |
---|
| 127 | you've presented, described and explained your coroutines... |
---|
| 128 | p3 I'd probably leave out the various "C" versions unless there are |
---|
| 129 | key points to make you can't make in C-forall. All the alternatives |
---|
| 130 | are just confusing. |
---|
| 131 | |
---|
| 132 | Hopefully fixed as this block of text has been rewritten. |
---|
| 133 | |
---|
| 134 | p4 but what's that "with" in Fig 1(B) |
---|
| 135 | |
---|
| 136 | Footnote D explains the semantic of "with", which is like unqualified access |
---|
| 137 | for the receiver to the fields of a class from member routines, i.e., no |
---|
| 138 | "this->". |
---|
| 139 | |
---|
| 140 | p5 start with the high level features of C-forall generators... |
---|
| 141 | |
---|
| 142 | Hopefully fixed by moving Table 1 forward. |
---|
| 143 | |
---|
| 144 | p5 why is the paper explaining networking protocols? |
---|
| 145 | |
---|
| 146 | Fixed, added discussion on this point. |
---|
| 147 | |
---|
| 148 | p7 lines 1-9 (transforming generator to coroutine - why would I do any of |
---|
| 149 | this? Why would I want one instead of the other (do not use "stack" in your |
---|
| 150 | answer!) |
---|
| 151 | |
---|
| 152 | As stated on line 1 because state declarations from the generator type can be |
---|
| 153 | moved out of the coroutine type into the coroutine main |
---|
| 154 | |
---|
| 155 | p10 last para "A coroutine must retain its last resumer to suspend back |
---|
| 156 | because the resumer is on a different stack. These reverse pointers allow |
---|
| 157 | suspend to cycle backwards, " I've no idea what is going on here? why |
---|
| 158 | should I care? Shouldn't I just be using threads instead? why not? |
---|
| 159 | |
---|
| 160 | Hopefully fixed by moving Table 1 forward. |
---|
| 161 | |
---|
| 162 | p16 for the same reasons - what reasons? |
---|
| 163 | |
---|
| 164 | Hopefully fixed by moving Table 1 forward. |
---|
| 165 | |
---|
| 166 | p17 if the multiple-monitor entry procedure really is novel, write a paper |
---|
| 167 | about that, and only about that. |
---|
| 168 | |
---|
| 169 | We do not believe this is a practical suggestion. |
---|
| 170 | |
---|
| 171 | p23 "Loose Object Definitions" - no idea what that means. in that |
---|
| 172 | section: you can't leave out JS-style dynamic properties. Even in |
---|
| 173 | OOLs that (one way or another) allow separate definitions of methods |
---|
| 174 | (like Objective-C, Swift, Ruby, C#) at any time a runtime class has a |
---|
| 175 | fixed definition. Quite why the detail about bit mask implementation |
---|
| 176 | is here anyway, I've no idea. |
---|
| 177 | |
---|
| 178 | Fixed by rewriting the section. |
---|
| 179 | |
---|
| 180 | p25 this cluster isn't a CLU cluster then? |
---|
| 181 | |
---|
| 182 | No. A CLU cluster is like a class in an object-oriented programming language. |
---|
| 183 | A CFA cluster is a runtime organizational mechanism. |
---|
| 184 | |
---|
| 185 | * conclusion should conclude the paper, not the related. |
---|
| 186 | |
---|
| 187 | We do not understand this comment. |
---|
| 188 | |
---|
| 189 | ============================================================================= |
---|
| 190 | |
---|
| 191 | Reviewing: 2 |
---|
| 192 | |
---|
| 193 | There is much description of the system and its details, but nothing about |
---|
| 194 | (non-artificial) uses of it. Although the microbenchmark data is |
---|
| 195 | encouraging, arguably not enough practical experience with the system has |
---|
| 196 | been reported here to say much about either its usability advantages or its |
---|
| 197 | performance. |
---|
| 198 | |
---|
| 199 | We have a Catch-22 problem. Without publicity, there is no user community; |
---|
| 200 | without a user community, there are no publications for publicity. |
---|
| 201 | |
---|
| 202 | p2: lines 4--9 are a little sloppy. It is not the languages but their |
---|
| 203 | popular implementations which "adopt" the 1:1 kernel threading model. |
---|
| 204 | |
---|
| 205 | Fixed. |
---|
| 206 | |
---|
| 207 | line 10: "medium work" -- "medium-sized work"? |
---|
| 208 | |
---|
| 209 | Fixed. |
---|
| 210 | |
---|
| 211 | line 18: "is all sequential to the compiler" -- not true in modern |
---|
| 212 | compilers, and in 2004 H-J Boehm wrote a tech report describing exactly why |
---|
| 213 | ("Threads cannot be implemented as a library", HP Labs). |
---|
| 214 | |
---|
| 215 | We will have to disagree on this point. First, I am aware of Hans's 2004 paper |
---|
| 216 | because in that paper Hans cites my seminal work on this topic from 1995, which |
---|
| 217 | we cite in this paper. Second, while modern memory-models have been added to |
---|
| 218 | languages like Java/C/C++ and new languages usually start with a memory model, |
---|
| 219 | it is still the programmer's responsibility to use them for racy code. Only |
---|
| 220 | when the programing language provides race-free constructs is the language |
---|
| 221 | aware of the concurrency; otherwise the code is sequential. Hans's paper "You |
---|
| 222 | Don't Know Jack About Shared Variables or Memory Models" talks about these |
---|
| 223 | issues, and is also cited in the paper. |
---|
| 224 | |
---|
| 225 | line 20: "knows the optimization boundaries" -- I found this vague. What's |
---|
| 226 | an example? |
---|
| 227 | |
---|
| 228 | Fixed. |
---|
| 229 | |
---|
| 230 | line 31: this paragraph has made a lot of claims. Perhaps forward-reference |
---|
| 231 | to the parts of the paper that discuss each one. |
---|
| 232 | |
---|
| 233 | Fixed by adding a road-map paragraph at the end of the introduction. |
---|
| 234 | |
---|
| 235 | line 33: "so the reader can judge if" -- this reads rather |
---|
| 236 | passive-aggressively. Perhaps better: "... to support our argument that..." |
---|
| 237 | |
---|
| 238 | Fixed. |
---|
| 239 | |
---|
| 240 | line 41: "a dynamic partitioning mechanism" -- I couldn't tell what this |
---|
| 241 | meant |
---|
| 242 | |
---|
| 243 | Fixed. |
---|
| 244 | |
---|
| 245 | p3. Presenting concept of a "stateful function" as a new language feature |
---|
| 246 | seems odd. In C, functions often have local state thanks to static local |
---|
| 247 | variables (or globals, indeed). Of course, that has several |
---|
| 248 | limitations. Can you perhaps present your contributions by enumerating |
---|
| 249 | these limitations? See also my suggestion below about a possible framing |
---|
| 250 | centred on a strawman. |
---|
| 251 | |
---|
| 252 | Fixed, at the start of this section. |
---|
| 253 | |
---|
| 254 | line 2: "an old idea that is new again" -- this is too oblique |
---|
| 255 | |
---|
| 256 | Fixed, removed. |
---|
| 257 | |
---|
| 258 | lines 2--15: I found this to be a word/concept soup. Stacks, closures, |
---|
| 259 | generators, stackless stackful, coroutine, symmetric, asymmetric, |
---|
| 260 | resume/suspend versus resume/resume... there needs to be a more gradual and |
---|
| 261 | structured way to introduce all this, and ideally one that minimises |
---|
| 262 | redundancy. Maybe present it as a series of "definitions" each with its own |
---|
| 263 | heading, e.g. "A closure is stackless if its local state has statically |
---|
| 264 | known fixed size"; "A generator simply means a stackless closure." And so |
---|
| 265 | on. Perhaps also strongly introduce the word "activate" as a direct |
---|
| 266 | contrast with resume and suspend. These are just a flavour of the sort of |
---|
| 267 | changes that might make this paragraph into something readable. |
---|
| 268 | |
---|
| 269 | Continuing the thought: I found it confusing that by these definitions, a |
---|
| 270 | stackful closure is not a stack, even though logically the stack *is* a |
---|
| 271 | kind of closure (it is a representation of the current thread's |
---|
| 272 | continuation). |
---|
| 273 | |
---|
| 274 | Fixed. Rewrote paragraph and moved Table 1 forward. |
---|
| 275 | |
---|
| 276 | lines 24--27: without explaining what the boost functor types mean, I don't |
---|
| 277 | think the point here comes across. |
---|
| 278 | |
---|
| 279 | Replaced with uC++ example because boost appears to have dropped symmetric |
---|
| 280 | coroutines. |
---|
| 281 | |
---|
| 282 | line 34: "semantically coupled" -- I wasn't sure what this meant |
---|
| 283 | |
---|
| 284 | Fixed. |
---|
| 285 | |
---|
| 286 | p4: the point of Figure 1 (C) was not immediately clear. It seem to be |
---|
| 287 | showing how one might "compile down" Figure 1 (B). Or is that Figure 1 (A)? |
---|
| 288 | |
---|
| 289 | Fixed. Rewrote sentence. |
---|
| 290 | |
---|
| 291 | It's right that the incidental language features of the system are not |
---|
| 292 | front-and-centre, but I'd appreciate some brief glossing of non-C languages |
---|
| 293 | features as they appear. Examples are the square bracket notation, the pipe |
---|
| 294 | notation and the constructor syntax. These explanations could go in the |
---|
| 295 | caption of the figure which first uses them, perhaps. Overall I found the |
---|
| 296 | figure captions to be terse, and a missed opportunity to explain clearly |
---|
| 297 | what was going on. |
---|
| 298 | |
---|
| 299 | Fixed, added descriptive footnote about Cforall. We prefer to put text in the |
---|
| 300 | body of the paper and keep captions short. |
---|
| 301 | |
---|
| 302 | p5 line 23: "This restriction is removed..." -- give us some up-front |
---|
| 303 | summary of your contributions and the elements of the language design that |
---|
| 304 | will be talked about, so that this isn't an aside. This will reduce the |
---|
| 305 | "twisty passages" feeling that characterises much of the paper. |
---|
| 306 | |
---|
| 307 | Fixed, remove parenthesis. |
---|
| 308 | |
---|
| 309 | line 40: "a killer asymmetric generator" -- this is stylistically odd, and |
---|
| 310 | the sentence about failures doesn't convincingly argue that C\/ will help |
---|
| 311 | with them. Have you any experience writing device drivers using C\/? Or any |
---|
| 312 | argument that the kinds of failures can be traced to the "stack-ripping" |
---|
| 313 | style that one is forced to use without coroutines ? |
---|
| 314 | |
---|
| 315 | Fixed, added new paragraph. |
---|
| 316 | |
---|
| 317 | Also, a typo on line |
---|
| 318 | 41: "device drives". And saying "Windows/Linux" is sloppy... what does the |
---|
| 319 | cited paper actually say? |
---|
| 320 | |
---|
| 321 | Fixed. |
---|
| 322 | |
---|
| 323 | p6 lines 13--23: this paragraph is difficult to understand. It seems to be |
---|
| 324 | talking about a control-flow pattern roughly equivalent to tail recursion. |
---|
| 325 | What is the high-level point, other than that this is possible? |
---|
| 326 | |
---|
| 327 | Fixed, rewrote start of the paragraph. |
---|
| 328 | |
---|
| 329 | line 34: "which they call coroutines" -- a better way to make this point is |
---|
| 330 | presumably that the C++20 proposal only provides a specialised kind of |
---|
| 331 | coroutine, namely generators, despite its use of the more general word. |
---|
| 332 | |
---|
| 333 | Fixed. |
---|
| 334 | |
---|
| 335 | line 47: "... due to dynamic stack allocation, execution..." -- this |
---|
| 336 | sentence doesn't scan. I suggest adding "and for" in the relevant places |
---|
| 337 | where currently there are only commas. |
---|
| 338 | |
---|
| 339 | Fixed. |
---|
| 340 | |
---|
| 341 | p8 / Figure 5 (B) -- the GNU C extension of unary "&&" needs to be |
---|
| 342 | explained. |
---|
| 343 | |
---|
| 344 | Fixed, added explanation at first usage in Figure 1(C) and reference. |
---|
| 345 | |
---|
| 346 | The whole figure needs a better explanation, in fact. |
---|
| 347 | |
---|
| 348 | Fixed, rewrote start of the paragraph. |
---|
| 349 | |
---|
| 350 | p9, lines 1--10: I wasn't sure this stepping-through really added much |
---|
| 351 | value. What are the truly important points to note about this code? |
---|
| 352 | |
---|
| 353 | Fixed, shortened and merged with previous paragraph. |
---|
| 354 | |
---|
| 355 | p10: similarly, lines 3--27 again are somewhere between tedious and |
---|
| 356 | confusing. I'm sure the motivation and details of "starter semantics" can |
---|
| 357 | both be stated much more pithily. |
---|
| 358 | |
---|
| 359 | Fixed, shortened these paragraphs. |
---|
| 360 | |
---|
| 361 | line 32: "a self-resume does not overwrite the last resumer" -- is this a |
---|
| 362 | hack or a defensible principled decision? |
---|
| 363 | |
---|
| 364 | Fixed, removed but it is a defensible principled decision. |
---|
| 365 | |
---|
| 366 | p11: "a common source of errors" -- among beginners or among production |
---|
| 367 | code? Presumably the former. |
---|
| 368 | |
---|
| 369 | Forgetting is not specific to beginners. |
---|
| 370 | |
---|
| 371 | line 23: "with builtin and library" -- not sure what this means |
---|
| 372 | |
---|
| 373 | Fixed. |
---|
| 374 | |
---|
| 375 | lines 31--36: these can be much briefer. The only important point here |
---|
| 376 | seems to be that coroutines cannot be copied. |
---|
| 377 | |
---|
| 378 | Fixed, shortened. |
---|
| 379 | |
---|
| 380 | p12: line 1: what is a "task"? Does it matter? |
---|
| 381 | |
---|
| 382 | Fixed, "task" has been changed to "thread" throughout the paper. |
---|
| 383 | |
---|
| 384 | line 7: calling it "heap stack" seems to be a recipe for |
---|
| 385 | confusion. "Stack-and-heap" might be better, and contrast with |
---|
| 386 | "stack-and-VLS" perhaps. When "VLS" is glossed, suggest actually expanding |
---|
| 387 | its initials: say "length" not "size". |
---|
| 388 | |
---|
| 389 | Fixed, make correction and rewrote some of the text. |
---|
| 390 | |
---|
| 391 | line 21: are you saying "cooperative threading" is the same as |
---|
| 392 | "non-preemptive scheduling", or that one is a special case (kind) of the |
---|
| 393 | other? Both are defensible, but be clear. |
---|
| 394 | |
---|
| 395 | Fixed, clarified the definitions. |
---|
| 396 | |
---|
| 397 | line 27: "mutual exclusion and synchronization" -- the former is a kind of |
---|
| 398 | the latter, so I suggest "and other forms of synchronization". |
---|
| 399 | |
---|
| 400 | We have to agree to disagree. Included a citation that explains the |
---|
| 401 | differences. |
---|
| 402 | |
---|
| 403 | line 30: "can either be a stackless or stackful" -- stray "a", but also, |
---|
| 404 | this seems to be switching from generic/background terminology to |
---|
| 405 | C\/-specific terminology. |
---|
| 406 | |
---|
| 407 | Fixed, but the terms stackless or stackful are not specific to Cforall; they |
---|
| 408 | are well known in the literature. |
---|
| 409 | |
---|
| 410 | An expositional idea occurs: start the paper with a strawman naive/limited |
---|
| 411 | realisation of coroutines -- say, Simon Tatham's popular "Coroutines in C" |
---|
| 412 | web page -- and identify point by point what the limitations are and how |
---|
| 413 | C\/ overcomes them. Currently the presentation is often flat (lacking |
---|
| 414 | motivating contrasts) and backwards (stating solutions before |
---|
| 415 | problems). The foregoing approach might fix both of these. |
---|
| 416 | |
---|
| 417 | We prefer the current structure of our paper and believe the paper does |
---|
| 418 | explain basic coding limitations and how they are overcome in using high-level |
---|
| 419 | control-floe mechanisms. |
---|
| 420 | |
---|
| 421 | page 13: line 23: it seems a distraction to mention the Python feature |
---|
| 422 | here. |
---|
| 423 | |
---|
| 424 | Why? It is the first location in the paper where dynamic allocation and |
---|
| 425 | initialization are mentioned. |
---|
| 426 | |
---|
| 427 | p14 line 5: it seems odd to describe these as "stateless" just because they |
---|
| 428 | lack shared mutable state. It means the code itself is even more |
---|
| 429 | stateful. Maybe the "stack ripping" argument could usefully be given here. |
---|
| 430 | |
---|
| 431 | Fixed, changed "stateless" to "non-shared". |
---|
| 432 | |
---|
| 433 | line 16: "too restrictive" -- would be good to have a reference to justify |
---|
| 434 | this, or at least give a sense of what the state-of-the-art performance in |
---|
| 435 | transactional memory systems is (both software and hardware) |
---|
| 436 | |
---|
| 437 | Fixed, added 2 citations. |
---|
| 438 | |
---|
| 439 | line 22: "simulate monitors" -- what about just *implementing* monitors? |
---|
| 440 | isn't that what these systems do? or is the point more about refining them |
---|
| 441 | somehow into something more specialised? |
---|
| 442 | |
---|
| 443 | Fixed, changed "simulate monitors" to "manually implement a monitor". |
---|
| 444 | |
---|
| 445 | p15: sections 4.1 and 4.2 seem adrift and misplaced. Split them into basic |
---|
| 446 | parts (which go earlier) and more advanced parts (e.g. barging, which can |
---|
| 447 | be explained later). |
---|
| 448 | |
---|
| 449 | Fixed, removed them by shortening and merging with previous section. |
---|
| 450 | |
---|
| 451 | line 31: "acquire/release" -- misses an opportunity to contrast the |
---|
| 452 | monitor's "enter/exit" abstraction with the less structured acquire/release |
---|
| 453 | of locks. |
---|
| 454 | |
---|
| 455 | Fixed, added "by call/return" in sentence. |
---|
| 456 | |
---|
| 457 | p16 line 12: the "implicit" versus "explicit" point is unclear. Is it |
---|
| 458 | perhaps about the contract between an opt-in *discipline* and a |
---|
| 459 | language-enforced *guarantee*? |
---|
| 460 | |
---|
| 461 | Fixed. |
---|
| 462 | |
---|
| 463 | line 28: no need to spend ages dithering about which one is default and |
---|
| 464 | which one is the explicit qualifier. Tell us what you decided, briefly |
---|
| 465 | justify it, and move on. |
---|
| 466 | |
---|
| 467 | Fixed, shortened paragraph. |
---|
| 468 | |
---|
| 469 | p17: Figure 11: since the main point seems to be to highlight bulk acquire, |
---|
| 470 | include a comment which identifies the line where this is happening. |
---|
| 471 | |
---|
| 472 | Fixed. |
---|
| 473 | |
---|
| 474 | line 2: "impossible to statically..." -- or dynamically. Doing it |
---|
| 475 | dynamically would be perfectly acceptable (locking is a dynamic operation |
---|
| 476 | after all) |
---|
| 477 | |
---|
| 478 | Fixed, clarified the "statically" applied to the unknown-sized pointer types. |
---|
| 479 | |
---|
| 480 | "guarantees acquisition order is consistent" -- assuming it's done in a |
---|
| 481 | single bulk acquire. |
---|
| 482 | |
---|
| 483 | Fixed. |
---|
| 484 | |
---|
| 485 | p18: section 5.3: the text here is a mess. The explanations of "internal" |
---|
| 486 | versus "external" scheduling are unclear, and "signals as hints" is not |
---|
| 487 | explained. "... can cause thread starvation" -- means including a while |
---|
| 488 | loop, or not doing so? "There are three signalling mechanisms.." but the |
---|
| 489 | text does not follow that by telling us what they are. My own scribbled |
---|
| 490 | attempt at unpicking the internal/external thing: "threads already in the |
---|
| 491 | monitor, albeit waiting, have priority over those trying to enter". |
---|
| 492 | |
---|
| 493 | Fixed, rewrote and shortened paragraphs. |
---|
| 494 | |
---|
| 495 | p19: line 3: "empty condition" -- explain that condition variables don't |
---|
| 496 | store anything. So being "empty" means that the queue of waiting threads |
---|
| 497 | (threads waiting to be signalled that the condition has become true) is |
---|
| 498 | empty. |
---|
| 499 | |
---|
| 500 | Fixed, changed condition variable to condition queue throughout the paper. |
---|
| 501 | |
---|
| 502 | line 6: "... can be transformed into external scheduling..." -- OK, but |
---|
| 503 | give some motivation. |
---|
| 504 | |
---|
| 505 | The paper states that it removes the condition queues and signal/wait. Changed |
---|
| 506 | "transform" to "simplified". |
---|
| 507 | |
---|
| 508 | p20: line 6: "mechnaism" |
---|
| 509 | |
---|
| 510 | Fixed. |
---|
| 511 | |
---|
| 512 | lines 16--20: this is dense and can probably only be made clear with an |
---|
| 513 | example |
---|
| 514 | |
---|
| 515 | Fixed, rewrote and added example. |
---|
| 516 | |
---|
| 517 | p21 line 21: clarify that nested monitor deadlock was describe earlier (in |
---|
| 518 | 5.2). (Is the repetition necessary?) |
---|
| 519 | |
---|
| 520 | Fixed, put in a forward reference, and the point bears repeating because |
---|
| 521 | releasing a subset of acquired monitors in unique to Cforall concurrency. |
---|
| 522 | |
---|
| 523 | line 27: "locks, and by extension monitors" -- this is true but the "by |
---|
| 524 | extension" argument is faulty. It is perfectly possible to use locks as a |
---|
| 525 | primitive and build a compositional mechanism out of them, |
---|
| 526 | e.g. transactions. |
---|
| 527 | |
---|
| 528 | True, but that is not what we said. Locks are not composable, monitors are |
---|
| 529 | built using locks not transactions, so by extension monitors are not composable. |
---|
| 530 | |
---|
| 531 | p22 line 2: should say "restructured" |
---|
| 532 | |
---|
| 533 | Fixed. |
---|
| 534 | |
---|
| 535 | line 33: "Implementing a fast subset check..." -- make clear that the |
---|
| 536 | following section explains how to do this. Restructuring the sections |
---|
| 537 | themselves could do this, or noting in the text. |
---|
| 538 | |
---|
| 539 | Fixed, added a forward reference to the following sections. |
---|
| 540 | |
---|
| 541 | p23: line 3: "dynamic member adding, eg, JavaScript" -- needs to say "as |
---|
| 542 | permitted in JavaScript", and "dynamically adding members" is stylistically |
---|
| 543 | better |
---|
| 544 | |
---|
| 545 | Fixed. |
---|
| 546 | |
---|
| 547 | p23: line 18: "urgent stack" -- back-reference to where this was explained |
---|
| 548 | before |
---|
| 549 | |
---|
| 550 | Fixed. |
---|
| 551 | |
---|
| 552 | p24 line 7: I did not understand what was more "direct" about "direct |
---|
| 553 | communication". Also, what is a "passive monitor" -- just a monitor, given |
---|
| 554 | that monitors are passive by design? |
---|
| 555 | |
---|
| 556 | The back half of line 7 defines "direct". For example, Go, Java, pthread |
---|
| 557 | threads cannot directly call/communicate with one another, where they can in |
---|
| 558 | Ada, uC++, and Cforall threads. Figure 18 show this exact difference. |
---|
| 559 | |
---|
| 560 | A monitor object is *passive* because it does not have a thread, while a Go, |
---|
| 561 | Java, Cforall "thread" object is *active* because it has a thread. |
---|
| 562 | |
---|
| 563 | line 14 / section 5.9: this table was useful and it (or something like it) |
---|
| 564 | could be used much earlier on to set the structure of the rest of the |
---|
| 565 | paper. |
---|
| 566 | |
---|
| 567 | Fixed, Table 1 is moved to the start and explained in detail. |
---|
| 568 | |
---|
| 569 | The explanation at present is too brief, e.g. I did not really understand |
---|
| 570 | the point about cases 7 and 8. Table 1: what does "No / Yes" mean? |
---|
| 571 | |
---|
| 572 | Fixed, expanded the explanation. |
---|
| 573 | |
---|
| 574 | p25 line 2: instead of casually dropping in a terse explanation for the |
---|
| 575 | newly introduced term "virtual processor", introduce it |
---|
| 576 | properly. Presumably the point is to give a less ambiguous meaning to |
---|
| 577 | "thread" by reserving it only for C\/'s green threads. |
---|
| 578 | |
---|
| 579 | Fixed. |
---|
| 580 | |
---|
| 581 | p26 line 15: "transforms user threads into fibres" -- a reference is needed |
---|
| 582 | to explain what "fibres" means... guessing it's in the sense of Adya et al. |
---|
| 583 | |
---|
| 584 | Fixed. In a prior correct, the term fibre from Adya is defined. |
---|
| 585 | |
---|
| 586 | line 20: "Microsoft runtime" -- means Windows? |
---|
| 587 | |
---|
| 588 | Fixed. |
---|
| 589 | |
---|
| 590 | lines 21--26: don't say "interrupt" to mean "signal", especially not |
---|
| 591 | without clear introduction. You can use "POSIX signal" to disambiguate from |
---|
| 592 | condition variables' "signal". |
---|
| 593 | |
---|
| 594 | We have to agree to disagree on this terminology. Interrupt is the action of |
---|
| 595 | stopping the CPU while a signal is a specific kind of interrupt. The two terms |
---|
| 596 | seem to be well understood in the literature. |
---|
| 597 | |
---|
| 598 | p27 line 3: "frequency is usually long" -- that's a "time period" or |
---|
| 599 | "interval", not a frequency |
---|
| 600 | |
---|
| 601 | Fixed. |
---|
| 602 | |
---|
| 603 | line 5: the lengthy quotation is not really necessary; just paraphrase the |
---|
| 604 | first sentence and move on. |
---|
| 605 | |
---|
| 606 | Fixed. |
---|
| 607 | |
---|
| 608 | line 20: "to verify the implementation" -- I don't think that means what is |
---|
| 609 | intended |
---|
| 610 | |
---|
| 611 | Fixed, changed "verify" to "test". |
---|
| 612 | |
---|
| 613 | Tables in section 7 -- too many significant figures. How many overall runs |
---|
| 614 | are described? What is N in each case? |
---|
| 615 | |
---|
| 616 | Fixed. As stated, N=31. |
---|
| 617 | |
---|
| 618 | p29 line 2: "to eliminate this cost" -- arguably confusing since nowadays |
---|
| 619 | on commodity CPUs most of the benefits of inlining are not to do with call |
---|
| 620 | overheads, but from later optimizations enabled as a consequence of the |
---|
| 621 | inlining |
---|
| 622 | |
---|
| 623 | Fixed. |
---|
| 624 | |
---|
| 625 | line 41: "a hierarchy" -- are they a hierarchy? If so, this could be |
---|
| 626 | explained earlier. Also, to say these make up "an integrated set... of |
---|
| 627 | control-flow features" verges on the tautologous. |
---|
| 628 | |
---|
| 629 | Fixed, rewrote sentence. |
---|
| 630 | |
---|
| 631 | p30 line 15: "a common case being web servers and XaaS" -- that's two cases |
---|
| 632 | |
---|
| 633 | Fixed. |
---|
| 634 | |
---|
| 635 | ============================================================================ |
---|
| 636 | |
---|
| 637 | Reviewing: 3 |
---|
| 638 | |
---|
| 639 | * Expand on the motivations for including both generator and coroutines, vs |
---|
| 640 | trying to build one atop the other |
---|
| 641 | |
---|
| 642 | Fixed, Table 1 is moved to the start and explained in detail. |
---|
| 643 | |
---|
| 644 | * Expand on the motivations for having both symmetric and asymmetric |
---|
| 645 | coroutines? |
---|
| 646 | |
---|
| 647 | A coroutine is not marked as symmetric or asymmetric, it is a coroutine. |
---|
| 648 | Symmetric or asymmetric is a stylistic use of a coroutine. By analogy, a |
---|
| 649 | function is not marked as recursive or non-recursive. Recursion is a style of |
---|
| 650 | programming with a function. So there is no notion of motivation for having |
---|
| 651 | both symmetric and asymmetric as they follow from how a programmer uses suspend |
---|
| 652 | and resume. |
---|
| 653 | |
---|
| 654 | * Comparison to async-await model adopted by other languages |
---|
| 655 | |
---|
| 656 | Fixed, added a new section on this topic. |
---|
| 657 | |
---|
| 658 | * Consider performance comparisons against node.js and Rust frameworks |
---|
| 659 | |
---|
| 660 | Fixed. |
---|
| 661 | |
---|
| 662 | * Discuss performance of monitors vs finer-grained memory models and atomic |
---|
| 663 | operations found in other languages |
---|
| 664 | |
---|
| 665 | The paper never suggested high-level concurrency constructs can or should |
---|
| 666 | replace race programming or hardware atomics. The paper suggests programmers |
---|
| 667 | use high-level constructs when and where is it feasible because they are easy |
---|
| 668 | and safer to use. The monitor example of an atomic counter is just that, an |
---|
| 669 | example, not the way it should be done if maximal performance is required. We |
---|
| 670 | have tried to make this point clear in the paper. |
---|
| 671 | |
---|
| 672 | * Why both internal/external scheduling for synchronization? |
---|
| 673 | |
---|
| 674 | Some additional motivation has been added. |
---|
| 675 | |
---|
| 676 | * Generators are not exposed as a "function" that returns a generator |
---|
| 677 | object, but rather as a kind of struct, with communication happening via |
---|
| 678 | mutable state instead of "return values". |
---|
| 679 | |
---|
| 680 | Yes, Cforall uses an object-style of coroutine, which allows multiple interface |
---|
| 681 | functions that pass and return values through a structure. This approach allows |
---|
| 682 | a generator function to have different kinds of return values and different |
---|
| 683 | kinds of parameters to produce those values. Our generators can provide this |
---|
| 684 | capability via multiple interface functions to the generator/coroutine state, |
---|
| 685 | which is discussed on page 5, lines 13-21. |
---|
| 686 | |
---|
| 687 | That is, the generator must be manually resumed and (if I understood) it |
---|
| 688 | is expected to store values that can then later be read (perhaps via |
---|
| 689 | methods), instead of having a `yield <Expr>` statement that yields up a |
---|
| 690 | value explicitly. |
---|
| 691 | |
---|
| 692 | All generators are manually resumed, e.g., Python/nodejs use "next" to resume a |
---|
| 693 | generator. Yes, yield <Expr> has a single interface with one input/return type, |
---|
| 694 | versus the Cforall approach allowing arbitrary number of interfaces of |
---|
| 695 | arbitrary types. |
---|
| 696 | |
---|
| 697 | * Both "symmetric" and "asymmetric" generators are supported, instead of |
---|
| 698 | only asymmetric. |
---|
| 699 | |
---|
| 700 | Yes, because they support different functionality as discussed in Chris |
---|
| 701 | Marlin's seminal work and both forms are implemented in Simula67. We did not |
---|
| 702 | invent symmetric and asymmetric generators/coroutines, we took them from the |
---|
| 703 | literature. |
---|
| 704 | |
---|
| 705 | * Coroutines (multi-frame generators) are an explicit mechanism. |
---|
| 706 | |
---|
| 707 | In most other languages, coroutines are rather built by layering |
---|
| 708 | single-frame generators atop one another (e.g., using a mechanism like |
---|
| 709 | async-await), |
---|
| 710 | |
---|
| 711 | We disagree. Node.js has async-await but has a separate coroutine feature. |
---|
| 712 | While there are claims that coroutines can be built from async-await and/or |
---|
| 713 | continuations, in actuality they cannot. |
---|
| 714 | |
---|
| 715 | and symmetric coroutines are basically not supported. I'd like to see a bit |
---|
| 716 | more justification for Cforall including all the above mechanisms -- it |
---|
| 717 | seemed like symmetric coroutines were a useful building block for some of |
---|
| 718 | the user-space threading and custom scheduler mechanisms that were briefly |
---|
| 719 | mentioned later in the paper. |
---|
| 720 | |
---|
| 721 | Hopefully fixed by moving Table 1 forward. |
---|
| 722 | |
---|
| 723 | In the discussion of coroutines, I would have expected a bit more of a |
---|
| 724 | comparison to the async-await mechanism offered in other languages. |
---|
| 725 | |
---|
| 726 | We added a new section at the start to point out there is no comparison between |
---|
| 727 | coroutines and async-await. |
---|
| 728 | |
---|
| 729 | Certainly the semantics of async-await in JavaScript implies |
---|
| 730 | significantly more overhead (because each async fn is a distinct heap |
---|
| 731 | object). [Rust's approach avoids this overhead][zc], however, and might be |
---|
| 732 | worthy of a comparison (see the Performance section). |
---|
| 733 | |
---|
| 734 | We could not get Rust async-await to work, and when reading the description of |
---|
| 735 | rust async-await, it appears to be Java-style executors with futures (possibly |
---|
| 736 | fast futures). |
---|
| 737 | |
---|
| 738 | There are several sections in the paper that compare against atomics -- for |
---|
| 739 | example, on page 15, the paper shows a simple monitor that encapsulates an |
---|
| 740 | integer and compares that to C++ atomics. Later, the paper compares the |
---|
| 741 | simplicity of monitors against the `volatile` quantifier from Java. The |
---|
| 742 | conclusion in section 8 also revisits this point. |
---|
| 743 | While I agree that monitors are simpler, they are obviously also |
---|
| 744 | significantly different from a performance perspective -- the paper doesn't |
---|
| 745 | seem to address this at all. It's plausible that (e.g.) the `Aint` monitor |
---|
| 746 | type described in the paper can be compiled and mapped to the specialized |
---|
| 747 | instructions offered by hardware, but I didn't see any mention of how this |
---|
| 748 | would be done. |
---|
| 749 | |
---|
| 750 | Fixed, see response above. |
---|
| 751 | |
---|
| 752 | There is also no mention of the more nuanced memory ordering |
---|
| 753 | relations offered by C++11 and how one might achieve similar performance |
---|
| 754 | characteristics in Cforall (perhaps the answer is that one simply doesn't |
---|
| 755 | need to; I think that's defensible, but worth stating explicitly). |
---|
| 756 | |
---|
| 757 | Cforall is built on C, and therefore has full access to all the gcc atomics, |
---|
| 758 | and automatically gets any gcc updates. Furthermore, section 6.9 states that |
---|
| 759 | Cforall provides the full panoply of low-level locks, as does Java, Go, C++, |
---|
| 760 | for performance programming. |
---|
| 761 | |
---|
| 762 | Cforall includes both internal and external scheduling; I found the |
---|
| 763 | explanation for the external scheduling mechanism to be lacking in |
---|
| 764 | justification. Why include both mechanisms when most languages seem to make |
---|
| 765 | do with only internal scheduling? It would be useful to show some scenarios |
---|
| 766 | where external scheduling is truly more powerful. |
---|
| 767 | |
---|
| 768 | Fixed. Pointed out external scheduling is simpler as part of rewriting in that |
---|
| 769 | section, and added additional examples. |
---|
| 770 | |
---|
| 771 | I would have liked to see some more discussion of external scheduling and |
---|
| 772 | how it interacts with software engineering best practices. It seems |
---|
| 773 | somewhat similar to AOP in certain regards. It seems to add a bit of "extra |
---|
| 774 | semantics" to monitor methods, in that any method may now also become a |
---|
| 775 | kind of synchronization point. |
---|
| 776 | |
---|
| 777 | Fixed somewhat. Pointed out that external scheduling has been around for a long |
---|
| 778 | time (40 years) in Ada, so there is a body of the software-engineering |
---|
| 779 | experience using it. As well, I have been teaching it for 30 years in the |
---|
| 780 | concurrency course at Waterloo. We don't know what software engineering best |
---|
| 781 | practices you imagine it interacting with. Yes, monitor functions are |
---|
| 782 | synchronization points with external scheduling. |
---|
| 783 | |
---|
| 784 | The "open-ended" nature of this feels like it could easily lead to subtle |
---|
| 785 | bugs, particularly when code refactoring occurs (which may e.g. split an |
---|
| 786 | existing method into two). |
---|
| 787 | |
---|
| 788 | Any time a public interface is refactored, it invalids existing calls, so there |
---|
| 789 | is always an issue. For mutex routines and external scheduling, the waitfor |
---|
| 790 | statements may have to be updated, but that update is part of the refactoring. |
---|
| 791 | |
---|
| 792 | This seems particularly true if external scheduling can occur across |
---|
| 793 | compilation units -- the paper suggested that this is true, but I wasn't |
---|
| 794 | entirely clear. |
---|
| 795 | |
---|
| 796 | Every aspect of Cforall allows separate compilation. The function prototypes |
---|
| 797 | necessary for separate compilation provide all the information necessary to |
---|
| 798 | compile any aspect of a program. |
---|
| 799 | |
---|
| 800 | I would have also appreciated a few more details on how external scheduling |
---|
| 801 | is implemented. It seems to me that there must be some sort of "hooks" on |
---|
| 802 | mutex methods so that they can detect whether some other function is |
---|
| 803 | waiting on them and awaken those blocked threads. I'm not sure how such |
---|
| 804 | hooks are inserted, particularly across compilation units. |
---|
| 805 | |
---|
| 806 | Hooks are inserted by the Cforall translator, in the same way that Java |
---|
| 807 | inserted hooks into a "synchronized" member of a monitor. As for Java, as long |
---|
| 808 | as the type information is consistent across compilation units, the correct |
---|
| 809 | code is inserted. |
---|
| 810 | |
---|
| 811 | The material in Section 5.6 didn't quite clarify the matter for me. For |
---|
| 812 | example, it left me somewhat confused about whether the `f` and `g` |
---|
| 813 | functions declared were meant to be local to a translation unit, or shared |
---|
| 814 | with other unit. |
---|
| 815 | |
---|
| 816 | There are no restrictions with respect to static or external mutex functions. |
---|
| 817 | Cforall is C. Any form of access or separate compilation in C applies to |
---|
| 818 | Cforall. As in C, function prototypes carry all necessary information to |
---|
| 819 | compile the code. |
---|
| 820 | |
---|
| 821 | To start, I did not realize that the `mutex_opt` notation was a keyword, I |
---|
| 822 | thought it was a type annotation. I think this could be called out more |
---|
| 823 | explicitly. |
---|
| 824 | |
---|
| 825 | Fixed, indicated "mutex" is a C-style parameter-only declaration type-qualifier. |
---|
| 826 | |
---|
| 827 | Later, in section 5.2, the paper discusses `nomutex` annotations, which |
---|
| 828 | initially threw me, as they had not been introduced (now I realize that |
---|
| 829 | this paragraph is there to justify why there is no such keyword). The |
---|
| 830 | paragraph might be rearranged to make that clearer, perhaps by leading with |
---|
| 831 | the choice that Cforall made. |
---|
| 832 | |
---|
| 833 | Fixed, rewrote paragraph removing nomutex. |
---|
| 834 | |
---|
| 835 | On page 17, the paper states that "acquiring multiple monitors is safe from |
---|
| 836 | deadlock", but this could be stated a bit more precisely: acquiring |
---|
| 837 | multiple monitors in a bulk-acquire is safe from deadlock (deadlock can |
---|
| 838 | still result from nested acquires). |
---|
| 839 | |
---|
| 840 | Fixed. |
---|
| 841 | |
---|
| 842 | On page 18, the paper states that wait states do not have to be enclosed in |
---|
| 843 | loops, as there is no concern of barging. This seems true but there are |
---|
| 844 | also other reasons to use loops (e.g., if there are multiple reasons to |
---|
| 845 | notify on the same condition). Thus the statement initially surprised me, |
---|
| 846 | as barging is only one of many reasons that I typically employ loops around |
---|
| 847 | waits. |
---|
| 848 | |
---|
| 849 | Fixed. Rewrote the sentence. Note, for all non-barging cases where you employ a |
---|
| 850 | loop around a wait, the unblocking task must change state before blocking |
---|
| 851 | again. In the barging case, the unblocking thread blocks again without |
---|
| 852 | changing state. |
---|
| 853 | |
---|
| 854 | I did not understand the diagram in Figure 12 for some time. Initially, I |
---|
| 855 | thought that it was generic to all monitors, and I could not understand the |
---|
| 856 | state space. It was only later that I realized it was specific to your |
---|
| 857 | example. Updating the caption from "Monitor scheduling to "Monitor |
---|
| 858 | scheduling in the example from Fig 13" might have helped me quite a bit. |
---|
| 859 | |
---|
| 860 | Fixed, updated text to clarify. Did not change the caption because the |
---|
| 861 | signal_block does not apply to Figure 13. |
---|
| 862 | |
---|
| 863 | I spent quite some time reading the boy/girl dating example (\*) and I |
---|
| 864 | admit I found it somewhat confusing. For example, I couldn't tell whether |
---|
| 865 | there were supposed to be many "girl" threads executing at once, or if |
---|
| 866 | there was only supposed to be one girl and one boy thread executing in a |
---|
| 867 | loop. |
---|
| 868 | |
---|
| 869 | The paper states: |
---|
| 870 | |
---|
| 871 | The dating service matches girl and boy threads with matching compatibility |
---|
| 872 | codes so they can exchange phone numbers. |
---|
| 873 | |
---|
| 874 | so there are many girl/boy threads. There is nothing preventing an individual |
---|
| 875 | girl/boy from arranging multiple dates. |
---|
| 876 | |
---|
| 877 | Are the girl/boy threads supposed to invoke the girl/boy methods or vice |
---|
| 878 | versa? |
---|
| 879 | |
---|
| 880 | As long as the girls/boys are consistent in the calls, it does not matter. The |
---|
| 881 | goal is to find a partner and exchange phone numbers. |
---|
| 882 | |
---|
| 883 | Surely there is some easier way to set this up? |
---|
| 884 | |
---|
| 885 | There are some other solutions using monitors but they all have a similar |
---|
| 886 | structure. |
---|
| 887 | |
---|
| 888 | The paper offered a number of comparisons to Go, C#, Scala, and so forth, |
---|
| 889 | but seems to have overlooked another recent language, Rust. In many ways, |
---|
| 890 | Rust seems to be closest in philosophy to Cforall, so it seems like an odd |
---|
| 891 | omission. I already mentioned above that Rust is in the process of shipping |
---|
| 892 | [async-await syntax][aa], which is definitely an alternative to the |
---|
| 893 | generator/coroutine approach in Cforall (though one with clear pros/cons). |
---|
| 894 | |
---|
| 895 | We cannot get rust async-await example programs to compile nor does the select! |
---|
| 896 | macro compile. |
---|
| 897 | |
---|
| 898 | @plg2[1]% rustc --version |
---|
| 899 | rustc 1.40.0 (73528e339 2019-12-16) |
---|
| 900 | |
---|
| 901 | @plg2[2]% cat future.rs |
---|
| 902 | use futures::executor::block_on; |
---|
| 903 | |
---|
| 904 | async fn hello_world() { |
---|
| 905 | println!("hello, world!"); |
---|
| 906 | } |
---|
| 907 | |
---|
| 908 | fn main() { |
---|
| 909 | let future = hello_world(); // Nothing is printed |
---|
| 910 | block_on(future); // `future` is run and "hello, world!" is printed |
---|
| 911 | } |
---|
| 912 | |
---|
| 913 | @plg2[3]% rustc -C opt-level=3 future.rs |
---|
| 914 | error[E0670]: `async fn` is not permitted in the 2015 edition |
---|
| 915 | --> future.rs:3:1 |
---|
| 916 | | |
---|
| 917 | 3 | async fn hello_world() { |
---|
| 918 | | ^^^^^ |
---|
| 919 | |
---|
| 920 | error[E0433]: failed to resolve: maybe a missing crate `futures`? |
---|
| 921 | --> future.rs:1:5 |
---|
| 922 | | |
---|
| 923 | 1 | use futures::executor::block_on; |
---|
| 924 | | ^^^^^^^ maybe a missing crate `futures`? |
---|
| 925 | |
---|
| 926 | error[E0425]: cannot find function `block_on` in this scope |
---|
| 927 | --> future.rs:9:5 |
---|
| 928 | | |
---|
| 929 | 9 | block_on(future); // `future` is run and "hello, world!" is printed |
---|
| 930 | | ^^^^^^^^ not found in this scope |
---|
| 931 | |
---|
| 932 | error: aborting due to 3 previous errors |
---|
| 933 | |
---|
| 934 | Some errors have detailed explanations: E0425, E0433, E0670. |
---|
| 935 | For more information about an error, try `rustc --explain E0425`. |
---|
| 936 | |
---|
| 937 | |
---|
| 938 | In the performance section in particular, you might consider comparing |
---|
| 939 | against some of the Rust web servers and threading systems. |
---|
| 940 | |
---|
| 941 | This paper is not about building web-servers. Nor are web-servers a reasonable |
---|
| 942 | benchmark for language concurrency. Web-servers are a benchmark for |
---|
| 943 | non-blocking I/O library efficiency accessed in the underlying operating |
---|
| 944 | system. Our prior work on web-server performance: |
---|
| 945 | |
---|
| 946 | @inproceedings{Pariag07, |
---|
| 947 | author = {David Pariag and Tim Brecht and Ashif Harji and Peter Buhr and Amol Shukla}, |
---|
| 948 | title = {Comparing the Performance of Web Server Architectures}, |
---|
| 949 | booktitle = {Proceedings of the 2007 Eurosys conference}, |
---|
| 950 | month = mar, |
---|
| 951 | year = 2007, |
---|
| 952 | pages = {231--243}, |
---|
| 953 | } |
---|
| 954 | |
---|
| 955 | @inproceedings{Harji12, |
---|
| 956 | author = {Ashif S. Harji and Peter A. Buhr and Tim Brecht}, |
---|
| 957 | title = {Comparing High-Performance Multi-core Web-Server Architectures}, |
---|
| 958 | booktitle = {Proceedings of the 5th Annual International Systems and Storage Conference}, |
---|
| 959 | series = {SYSTOR '12}, |
---|
| 960 | publisher = {ACM}, |
---|
| 961 | address = {New York, NY, USA}, |
---|
| 962 | location = {Haifa, Israel}, |
---|
| 963 | month = jun, |
---|
| 964 | year = 2012, |
---|
| 965 | articleno = 1, |
---|
| 966 | pages = {1:1--1:12}, |
---|
| 967 | } |
---|
| 968 | |
---|
| 969 | shows the steps to build a high-performance web-server, which are largely |
---|
| 970 | independent of the server architecture and programing language. |
---|
| 971 | |
---|
| 972 | It would seem worth trying to compare their "context switching" costs as |
---|
| 973 | well -- I believe both actix and tokio have a notion of threads that could |
---|
| 974 | be readily compared. |
---|
| 975 | |
---|
| 976 | Again, context-switching speed is largely irrelevant because the amount of code |
---|
| 977 | to process an http request is large enough to push any concurrency costs into |
---|
| 978 | the background. |
---|
| 979 | |
---|
| 980 | Another addition that might be worth considering is to compare against |
---|
| 981 | node.js promises, although I think the comparison to process creation is |
---|
| 982 | not as clean. |
---|
| 983 | |
---|
| 984 | Done. |
---|