Changeset 8e4aa05 for doc


Ignore:
Timestamp:
Mar 4, 2021, 7:40:25 PM (5 years ago)
Author:
Thierry Delisle <tdelisle@…>
Branches:
ADT, arm-eh, ast-experimental, enum, forall-pointer-decay, jacob/cs343-translation, master, new-ast-unique-expr, pthread-emulation, qualifiedEnum, stuck-waitfor-destruct
Children:
77d601f
Parents:
342af53 (diff), a5040fe (diff)
Note: this is a merge changeset, the changes displayed below correspond to the merge itself.
Use the (diff) links above to see all the changes relative to each parent.
Message:

Merge branch 'master' of plg.uwaterloo.ca:software/cfa/cfa-cc

Location:
doc
Files:
6 added
23 edited

Legend:

Unmodified
Added
Removed
  • doc/LaTeXmacros/common.tex

    r342af53 r8e4aa05  
    1111%% Created On       : Sat Apr  9 10:06:17 2016
    1212%% Last Modified By : Peter A. Buhr
    13 %% Last Modified On : Mon Oct  5 09:34:46 2020
    14 %% Update Count     : 464
     13%% Last Modified On : Sun Feb 14 15:52:46 2021
     14%% Update Count     : 524
    1515%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
    1616
     
    3232\setlist[enumerate]{listparindent=\parindent}% global
    3333\setlist[enumerate,2]{leftmargin=\parindent,labelsep=*,align=parleft,label=\alph*.}% local
    34 \setlist[description]{itemsep=0pt,listparindent=\parindent,leftmargin=\parindent,labelsep=1.5ex}
     34\setlist[description]{topsep=0.5ex,itemsep=0pt,listparindent=\parindent,leftmargin=\parindent,labelsep=1.5ex}
    3535
    3636% Names used in the document.
    3737
    3838\usepackage{xspace}
    39 \newcommand{\CFAIcon}{\textsf{C}\raisebox{\depth}{\rotatebox{180}{\textsf{A}}}\xspace} % Cforall symbolic name
    40 \newcommand{\CFA}{\protect\CFAIcon}             % safe for section/caption
    41 \newcommand{\CFL}{\textrm{Cforall}\xspace} % Cforall symbolic name
    42 \newcommand{\Celeven}{\textrm{C11}\xspace} % C11 symbolic name
    43 \newcommand{\CC}{\textrm{C}\kern-.1em\hbox{+\kern-.25em+}\xspace} % C++ symbolic name
    44 \newcommand{\CCeleven}{\textrm{C}\kern-.1em\hbox{+\kern-.25em+}11\xspace} % C++11 symbolic name
    45 \newcommand{\CCfourteen}{\textrm{C}\kern-.1em\hbox{+\kern-.25em+}14\xspace} % C++14 symbolic name
    46 \newcommand{\CCseventeen}{\textrm{C}\kern-.1em\hbox{+\kern-.25em+}17\xspace} % C++17 symbolic name
    47 \newcommand{\CCtwenty}{\textrm{C}\kern-.1em\hbox{+\kern-.25em+}20\xspace} % C++20 symbolic name
     39\newcommand{\CFAIcon}{\textsf{C}\raisebox{\depth}{\rotatebox{180}{\textsf{A}}}} % Cforall icon
     40\newcommand{\CFA}{\protect\CFAIcon\xspace}                      % CFA symbolic name
     41\newcommand{\CFL}{\textrm{Cforall}\xspace}                      % Cforall non-icon name
     42\newcommand{\Celeven}{\textrm{C11}\xspace}                      % C11 symbolic name
     43\newcommand{\CCIcon}{\textrm{C}\kern-.1em\hbox{+\kern-.25em+}} % C++ icon
     44\newcommand{\CC}{\protect\CCIcon\xspace}                        % C++ symbolic name
     45% numbers disallowed in latex variables names => use number names
     46\newcommand{\CCeleven}{\protect\CCIcon{11}\xspace}      % C++11 symbolic name
     47\newcommand{\CCfourteen}{\protect\CCIcon{14}\xspace} % C++14 symbolic name
     48\newcommand{\CCseventeen}{\protect\CCIcon{17}\xspace} % C++17 symbolic name
     49\newcommand{\CCtwenty}{\protect\CCIcon{20}\xspace}      % C++20 symbolic name
    4850\newcommand{\Csharp}{C\raisebox{-0.7ex}{\Large$^\sharp$}\xspace} % C# symbolic name
    4951
    5052%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
    5153
     54% remove special-character warning in PDF side-bar names
    5255\makeatletter
     56\@ifpackageloaded{hyperref}{
     57  \pdfstringdefDisableCommands{
     58  \def\CFA{\CFL}
     59  \def\Celeven{C11\xspace}
     60  \def\CC{C++\xspace}
     61  \def\CCeleven{C++11\xspace}
     62  \def\CCfourteen{C++14\xspace}
     63  \def\CCseventeen{C++17\xspace}
     64  \def\CCtwenty{C++20\xspace}
     65  \def\Csharp{C\#\xspace}
     66  \def\lstinline{\xspace}% must use {} as delimiters, e.g., \lstinline{...}
     67  }{}
     68}
     69
    5370% parindent is relative, i.e., toggled on/off in environments like itemize, so store the value for
    5471% use rather than use \parident directly.
     
    8198    \vskip 50\p@
    8299  }}
    83 \renewcommand\section{\@startsection{section}{1}{\z@}{-3.5ex \@plus -1ex \@minus -.2ex}{1.75ex \@plus .2ex}{\normalfont\large\bfseries}}
    84 \renewcommand\subsection{\@startsection{subsection}{2}{\z@}{-3.25ex \@plus -1ex \@minus -.2ex}{1.5ex \@plus .2ex}{\normalfont\normalsize\bfseries}}
     100\renewcommand\section{\@startsection{section}{1}{\z@}{-3.0ex \@plus -1ex \@minus -.2ex}{1.5ex \@plus .2ex}{\normalfont\large\bfseries}}
     101\renewcommand\subsection{\@startsection{subsection}{2}{\z@}{-2.75ex \@plus -1ex \@minus -.2ex}{1.25ex \@plus .2ex}{\normalfont\normalsize\bfseries}}
    85102\renewcommand\subsubsection{\@startsection{subsubsection}{3}{\z@}{-2.5ex \@plus -1ex \@minus -.2ex}{1.0ex \@plus .2ex}{\normalfont\normalsize\bfseries}}
    86103\renewcommand\paragraph{\@startsection{paragraph}{4}{\z@}{-2.0ex \@plus -1ex \@minus -.2ex}{-1em}{\normalfont\normalsize\bfseries}}
     
    89106\newcommand{\italic}[1]{\emph{\hyperpage{#1}}}
    90107\newcommand{\Definition}[1]{\textbf{\hyperpage{#1}}}
    91 \newcommand{\see}[1]{\emph{see}~#1}
     108\newcommand{\see}[1]{(see #1)}
    92109
    93110% Define some commands that produce formatted index entries suitable for cross-references.
     
    129146% The star version does not lowercase the index information, e.g., \newterm*{IBM}.
    130147\newcommand{\newtermFontInline}{\emph}
    131 \newcommand{\newterm}{\@ifstar\@snewterm\@newterm}
     148\newcommand{\newterm}{\protect\@ifstar\@snewterm\@newterm}
    132149\newcommand{\@newterm}[2][\@empty]{\lowercase{\def\temp{#2}}{\newtermFontInline{#2}}\ifx#1\@empty\index{\temp}\else\index{#1@{\protect#2}}\fi}
    133150\newcommand{\@snewterm}[2][\@empty]{{\newtermFontInline{#2}}\ifx#1\@empty\index{#2}\else\index{#1@{\protect#2}}\fi}
     
    229246\usepackage{listings}                                                                   % format program code
    230247\usepackage{lstlang}
     248\usepackage{calc}                                                                               % latex arithmetic
     249
    231250\makeatletter
    232 
    233251\newcommand{\LstBasicStyle}[1]{{\lst@basicstyle{#1}}}
    234252\newcommand{\LstKeywordStyle}[1]{{\lst@basicstyle{\lst@keywordstyle{#1}}}}
    235253\newcommand{\LstCommentStyle}[1]{{\lst@basicstyle{\lst@commentstyle{#1}}}}
     254\newcommand{\LstStringStyle}[1]{{\lst@basicstyle{\lst@stringstyle{#1}}}}
    236255
    237256\newlength{\gcolumnposn}                                % temporary hack because lstlisting does not handle tabs correctly
     
    259278xleftmargin=\parindentlnth,                             % indent code to paragraph indentation
    260279extendedchars=true,                                             % allow ASCII characters in the range 128-255
    261 escapechar=§,                                                   % LaTeX escape in CFA code §...§ (section symbol), emacs: C-q M-'
    262 mathescape=true,                                                % LaTeX math escape in CFA code $...$
     280escapechar=\$,                                                  % LaTeX escape in CFA code §...§ (section symbol), emacs: C-q M-'
     281mathescape=false,                                               % LaTeX math escape in CFA code $...$
    263282keepspaces=true,                                                %
    264283showstringspaces=false,                                 % do not show spaces with cup
    265284showlines=true,                                                 % show blank lines at end of code
    266285aboveskip=4pt,                                                  % spacing above/below code block
    267 belowskip=3pt,
     286belowskip=0pt,
     287numberstyle=\footnotesize\sf,                   % numbering style
    268288% replace/adjust listing characters that look bad in sanserif
    269289literate={-}{\makebox[1ex][c]{\raisebox{0.4ex}{\rule{0.75ex}{0.1ex}}}}1 {^}{\raisebox{0.6ex}{$\scriptscriptstyle\land\,$}}1
     
    274294
    275295\ifdefined\CFALatin% extra Latin-1 escape characters
    276 \lstnewenvironment{cfa}[1][]{
     296\lstnewenvironment{cfa}[1][]{% necessary
    277297\lstset{
    278298language=CFA,
    279 moredelim=**[is][\color{red}]{®}{®},    % red highlighting ®...® (registered trademark symbol) emacs: C-q M-.
    280 moredelim=**[is][\color{blue}]{ß}{ß},   % blue highlighting ß...ß (sharp s symbol) emacs: C-q M-_
    281 moredelim=**[is][\color{OliveGreen}]{¢}{¢}, % green highlighting ¢...¢ (cent symbol) emacs: C-q M-"
    282 moredelim=[is][\lstset{keywords={}}]{¶}{¶}, % keyword escape ¶...¶ (pilcrow symbol) emacs: C-q M-^
    283 % replace/adjust listing characters that look bad in sanserif
    284 add to literate={`}{\ttfamily\upshape\hspace*{-0.1ex}`}1
     299moredelim=**[is][\color{red}]{@}{@},    % red highlighting @...@
     300%moredelim=**[is][\color{red}]{®}{®},   % red highlighting ®...® (registered trademark symbol) emacs: C-q M-.
     301%moredelim=**[is][\color{blue}]{ß}{ß},  % blue highlighting ß...ß (sharp s symbol) emacs: C-q M-_
     302%moredelim=**[is][\color{OliveGreen}]{¢}{¢}, % green highlighting ¢...¢ (cent symbol) emacs: C-q M-"
     303%moredelim=[is][\lstset{keywords={}}]{¶}{¶}, % keyword escape ¶...¶ (pilcrow symbol) emacs: C-q M-^
    285304}% lstset
    286 \lstset{#1}
     305\lstset{#1}% necessary
    287306}{}
    288307% inline code ©...© (copyright symbol) emacs: C-q M-)
    289308\lstMakeShortInline©                                    % single-character for \lstinline
    290309\else% regular ASCI characters
    291 \lstnewenvironment{cfa}[1][]{
     310\lstnewenvironment{cfa}[1][]{% necessary
    292311\lstset{
    293312language=CFA,
    294313escapechar=\$,                                                  % LaTeX escape in CFA code
     314mathescape=false,                                               % LaTeX math escape in CFA code $...$
    295315moredelim=**[is][\color{red}]{@}{@},    % red highlighting @...@
    296316}% lstset
    297 \lstset{#1}
     317\lstset{#1}% necessary
    298318}{}
    299319% inline code @...@ (at symbol)
  • doc/bibliography/pl.bib

    r342af53 r8e4aa05  
    688688    title       = {Asynchronous Exception Propagation in Blocked Tasks},
    689689    booktitle   = {4th International Workshop on Exception Handling (WEH.08)},
    690     organization= {16th International Symposium on the Foundations of Software Engineering (FSE 16)},
     690    optorganization= {16th International Symposium on the Foundations of Software Engineering (FSE 16)},
    691691    address     = {Atlanta, U.S.A},
    692692    month       = nov,
     
    17971797}
    17981798
    1799 @article{Delisle19,
     1799@article{Delisle20,
    18001800    keywords    = {concurrency, Cforall},
    18011801    contributer = {pabuhr@plg},
    18021802    author      = {Thierry Delisle and Peter A. Buhr},
    18031803    title       = {Advanced Control-flow and Concurrency in \textsf{C}$\mathbf{\forall}$},
    1804     year        = 2019,
     1804    year        = 2020,
    18051805    journal     = spe,
    1806     pages       = {1-33},
    1807     note        = {submitted},
     1806    pages       = {1-38},
     1807    note        = {\href{https://doi-org.proxy.lib.uwaterloo.ca/10.1002/spe.2925}{https://\-doi-org.proxy.lib.uwaterloo.ca/\-10.1002/\-spe.2925}},
     1808    note        = {},
    18081809}
    18091810
     
    72467247
    72477248@inproceedings{Edelson92,
    7248     keywords    = {persistence, pointers},
     7249    keywords    = {persistence, smart pointers},
    72497250    contributer = {pabuhr@plg},
    72507251    author      = {Daniel R. Edelson},
     
    72567257    year        = 1992,
    72577258    pages       = {1-19},
     7259}
     7260
     7261@incollection{smartpointers,
     7262    keywords    = {smart pointers},
     7263    contributer = {pabuhr@plg},
     7264    author      = {Andrei Alexandrescu},
     7265    title       = {Smart Pointers},
     7266    booktitle   = {Modern C++ Design: Generic Programming and Design Patterns Applied},
     7267    publisher   = {Addison-Wesley},
     7268    year        = 2001,
     7269    chapter     = 7,
     7270    optpages    = {?-?},
    72587271}
    72597272
     
    82458258}
    82468259
     8260@misc{vistorpattern,
     8261    keywords    = {visitor pattern},
     8262    contributer = {pabuhr@plg},
     8263    key         = {vistor pattern},
     8264    title       = {vistor pattern},
     8265    year        = 2020,
     8266    note        = {WikipediA},
     8267    howpublished= {\href{https://en.wikipedia.org/wiki/Visitor\_pattern}
     8268                  {https://\-en.wikipedia.org/\-wiki/\-Visitor\_pattern}},
     8269}
     8270
    82478271% W
    82488272
  • doc/papers/concurrency/mail2

    r342af53 r8e4aa05  
    12881288
    12891289
     1290From: "Wiley Online Proofing" <onlineproofing@eproofing.in>
     1291To: pabuhr@uwaterloo.ca
     1292Reply-To: eproofing@wiley.com
     1293Date: 3 Nov 2020 08:25:06 +0000
     1294Subject: Action: Proof of SPE_EV_SPE2925 for Software: Practice And Experience ready for review
     1295
     1296Dear Dr. Peter Buhr,
     1297
     1298The proof of your Software: Practice And Experience article Advanced control-flow in Cforall is now available for review:
     1299
     1300Edit Article https://wiley.eproofing.in/Proof.aspx?token=ab7739d5678447fbbe5036f3bcba2445081500061
     1301
     1302To review your article, please complete the following steps, ideally within 48 hours*, so we can publish your article as quickly as possible.
     1303
     13041. Open your proof in the online proofing system using the button above.
     13052. Check the article for correctness and respond to all queries.For instructions on using the system, please see the "Help" menu in the upper right corner.
     13063. Submit your changes by clicking the "Submit" button in the proofing system.
     1307
     1308Helpful Tips
     1309
     1310*  Your manuscript has been formatted following the style requirements for the journal. Any requested changes that go against journal style will not be made.
     1311*  Your proof will include queries. These must be replied to using the system before the proof can be submitted.
     1312*  The only acceptable changes at this stage are corrections to grammatical errors or data accuracy, or to provide higher resolution figure files (if requested by the typesetter).
     1313*  Any changes to scientific content or authorship will require editorial review and approval.
     1314*  Once your changes are complete, submit the article after which no additional corrections can be requested.
     1315*  Most authors complete their corrections within 48 hours. Returning any corrections promptly will accelerate publication of your article.
     1316
     1317If you encounter any problems or have questions, please contact the production office at (SPEproofs@wiley.com). For the quickest response, include the journal name and your article ID (found in the subject line) in all correspondence.
     1318
     1319Best regards,
     1320Software: Practice And Experience Production Office
     1321
     1322* We appreciate that the COVID-19 pandemic may create conditions for you that make it difficult for you to review your proof within standard timeframes. If you have any problems keeping to this schedule, please reach out to me at (SPEproofs@wiley.com) to discuss alternatives.
     1323
     1324
     1325
    12901326From: "Pacaanas, Joel -" <jpacaanas@wiley.com>
    12911327To: "Peter A. Buhr" <pabuhr@uwaterloo.ca>
     
    13451381
    13461382Since the proof was reset, your added corrections before has also been removed. Please add them back.
    1347 
    13481383Please return your corrections at your earliest convenience.
    13491384
     
    13841419Best regards,
    13851420Joel Pacaanas
     1421
     1422
     1423
     1424Date: Wed, 2 Dec 2020 08:49:52 +0000
     1425From: <cs-author@wiley.com>
     1426To: <pabuhr@uwaterloo.ca>
     1427Subject: Published: Your article is now published in Early View!
     1428
     1429Dear Peter Buhr,
     1430
     1431Your article Advanced Control-flow and Concurrency in C A in Software: Practice and Experience has the following publication status: Published as Early View
     1432
     1433To access your article, please click the following link to register or log in:
     1434
     1435  https://authorservices.wiley.com/index.html#register
     1436
     1437You can also access your published article via this link: http://dx.doi.org/10.1002/spe.2925
     1438
     1439If you need any assistance, please click here https://hub.wiley.com/community/support/authorservices to view our Help section.
     1440
     1441Sincerely,                                                                                 
     1442Wiley Author Services
     1443
     1444
     1445Date: Wed, 2 Dec 2020 02:16:23 -0500
     1446From: <no-reply@copyright.com>
     1447To: <pabuhr@uwaterloo.ca>
     1448CC: <SPEproofs@wiley.com>
     1449Subject: Please submit your publication fee(s) SPE2925
     1450 
     1451John Wiley and Sons
     1452Please submit your selection and payment for publication fee(s).
     1453
     1454Dear Peter A. Buhr,
     1455
     1456Congratulations, your article in Software: Practice and Experience has published online:
     1457
     1458Manuscript DOI: 10.1002/spe.2925
     1459Manuscript ID: SPE2925
     1460Manuscript Title: Advanced control-flow in Cforall
     1461Published by: John Wiley and Sons
     1462
     1463Please carefully review your publication options. If you wish your colour
     1464figures to be printed in colour, you must select and pay for that option now
     1465using the RightsLink e-commerce solution from CCC.
     1466
     1467  Review my options & pay charges 
     1468  https://oa.copyright.com/apc-payment-ui/overview?id=f46ba36a-2565-4c8d-8865-693bb94d87e5&chargeset=CHARGES 
     1469
     1470To review and pay your charge(s), please click here
     1471<https://oa.copyright.com/apc-payment-ui/overview?id=f46ba36a-2565-4c8d-8865-693bb94d87e5&chargeset=CHARGES>. You
     1472can also forward this link to another party for processing.
     1473
     1474To complete a secure transaction, you will need a RightsLink account
     1475<https://oa.copyright.com/apc-payment-ui/registration?id=f46ba36a-2565-4c8d-8865-693bb94d87e5&chargeset=CHARGES>. If
     1476you do not have one already, you will be prompted to register as you are
     1477checking out your author charges. This is a very quick process; the majority of
     1478your registration form will be pre-populated automatically with information we
     1479have already supplied to RightsLink.
     1480
     1481If you have any questions about these charges, please contact CCC Customer
     1482Service <wileysupport@copyright.com> using the information below. Please do not
     1483reply directly to this email as this is an automated email notification sent
     1484from an unmonitored account.
     1485
     1486Sincerely,
     1487John Wiley and Sons
     1488       
     1489Tel.: +1-877-622-5543 / +1-978-646-2777
     1490wileysupport@copyright.com
     1491www.copyright.com
     1492       
     1493Copyright Clearance Center
     1494RightsLink
     1495       
     1496This message (including attachments) is confidential, unless marked
     1497otherwise. It is intended for the addressee(s) only. If you are not an intended
     1498recipient, please delete it without further distribution and reply to the
     1499sender that you have received the message in error.
     1500
     1501
     1502
     1503From: "Pacaanas, Joel -" <jpacaanas@wiley.com>
     1504To: "Peter A. Buhr" <pabuhr@uwaterloo.ca>
     1505Subject: RE: Please submit your publication fee(s) SPE2925
     1506Date: Thu, 3 Dec 2020 08:45:10 +0000
     1507
     1508Dear Dr Buhr,
     1509
     1510Thank you for your email and concern with regard to the RightsLink account. As
     1511you have mentioned that all figures will be printed as black and white, then I
     1512have selected it manually from the system to proceed further.
     1513
     1514Best regards,
     1515Joel
     1516
     1517Joel Q. Pacaanas
     1518Production Editor
     1519On behalf of Wiley
     1520Manila
     1521We partner with global experts to further innovative research.
     1522
     1523E-mail: jpacaanas@wiley.com
     1524Tel: +632 88558618
     1525Fax: +632 5325 0768
     1526
     1527-----Original Message-----
     1528From: Peter A. Buhr [mailto:pabuhr@uwaterloo.ca]
     1529Sent: Thursday, December 3, 2020 12:28 AM
     1530To: SPE Proofs <speproofs@wiley.com>
     1531Subject: Re: Please submit your publication fee(s) SPE2925
     1532
     1533I am trying to complete the forms to submit my publication fee.
     1534
     1535I clicked all the boxs to print in Black and White, so there is no fee.
     1536
     1537I then am asked to create RightsLink account, which I did.
     1538
     1539However, it requires that I click a box agreeing to:
     1540
     1541   I consent to have my contact information shared with my publisher and/or
     1542   funding organization, as needed, to facilitate APC payment(s), reporting and
     1543   customer care.
     1544
     1545I do not agree to this sharing and will not click this button.
     1546
     1547How would you like to proceed?
     1548
     1549
     1550
     1551From: "Pacaanas, Joel -" <jpacaanas@wiley.com>
     1552To: "Peter A. Buhr" <pabuhr@uwaterloo.ca>
     1553Subject: RE: Please submit your publication fee(s) SPE2925
     1554Date: Fri, 4 Dec 2020 07:55:59 +0000
     1555
     1556Dear Peter,
     1557
     1558Yes, you are now done with this selection.
     1559
     1560Thank you.
     1561
     1562Best regards,
     1563Joel
     1564
     1565Joel Q. Pacaanas
     1566Production Editor
     1567On behalf of Wiley
     1568Manila
     1569We partner with global experts to further innovative research.
     1570
     1571E-mail: jpacaanas@wiley.com
     1572Tel: +632 88558618
     1573Fax: +632 5325 0768
     1574
     1575-----Original Message-----
     1576From: Peter A. Buhr [mailto:pabuhr@uwaterloo.ca]
     1577Sent: Thursday, December 3, 2020 10:29 PM
     1578To: Pacaanas, Joel - <jpacaanas@wiley.com>
     1579Subject: Re: Please submit your publication fee(s) SPE2925
     1580
     1581    Thank you for your email and concern with regard to the RightsLink
     1582    account. As you have mentioned that all figures will be printed as black and
     1583    white, then I have selected it manually from the system to proceed further.
     1584
     1585Just be clear, am I done? Meaning I do not have to go back to that web-page again.
  • doc/theses/andrew_beach_MMath/.gitignore

    r342af53 r8e4aa05  
    33
    44# Final Files:
    5 thesis.pdf
     5*.pdf
    66
    77# The Makefile here is not generated.
  • doc/theses/andrew_beach_MMath/Makefile

    r342af53 r8e4aa05  
    11### Makefile for Andrew Beach's Masters Thesis
    22
    3 DOC=thesis.pdf
     3DOC=uw-ethesis.pdf
    44BUILD=out
    55TEXSRC=$(wildcard *.tex)
     
    77STYSRC=$(wildcard *.sty)
    88CLSSRC=$(wildcard *.cls)
    9 TEXLIB= .:${BUILD}:
     9TEXLIB= .:../../LaTeXmacros:${BUILD}:
    1010BIBLIB= .:../../bibliography
    1111
     
    2929        ${LATEX} ${BASE}
    3030        ${BIBTEX} ${BUILD}/${BASE}
     31        ${LATEX} ${BASE}
    3132        ${GLOSSARY} ${BUILD}/${BASE}
    3233        ${LATEX} ${BASE}
  • doc/theses/andrew_beach_MMath/existing.tex

    r342af53 r8e4aa05  
    1 \chapter{\CFA{} Existing Features}
    2 
    3 \section{Overloading and extern}
    4 Cforall has overloading, allowing multiple definitions of the same name to
    5 be defined.
    6 
    7 This also adds name mangling so that the assembly symbols are unique for
    8 different overloads. For compatability with names in C there is also a
    9 syntax to diable the name mangling. These unmangled names cannot be overloaded
    10 but act as the interface between C and \CFA code.
    11 
    12 The syntax for disabling mangling is:
    13 \begin{lstlisting}
    14 extern "C" {
    15     ...
    16 }
    17 \end{lstlisting}
    18 
    19 To re-enable mangling once it is disabled the syntax is:
    20 \begin{lstlisting}
    21 extern "Cforall" {
    22     ...
    23 }
    24 \end{lstlisting}
    25 
    26 Both should occur at the declaration level and effect all the declarations
    27 in \texttt{...}. Neither care about the state of mangling when they begin
    28 and will return to that state after the group is finished. So re-enabling
    29 is only used to nest areas of mangled and unmangled declarations.
    30 
    31 \section{References}
    32 \CFA adds references to C. These are auto-dereferencing pointers and use the
    33 same syntax as pointers except they use ampersand (\codeCFA{\&}) instead of
    34 the asterisk (\codeCFA{*}). They can also be constaint or mutable, if they
    35 are mutable they may be assigned to by using the address-of operator
    36 (\codeCFA\&) which converts them into a pointer.
     1\chapter{\CFA Existing Features}
     2
     3\CFA (C-for-all)~\cite{Cforall} is an open-source project extending ISO C with
     4modern safety and productivity features, while still ensuring backwards
     5compatibility with C and its programmers.  \CFA is designed to have an
     6orthogonal feature-set based closely on the C programming paradigm
     7(non-object-oriented) and these features can be added incrementally to an
     8existing C code-base allowing programmers to learn \CFA on an as-needed basis.
     9
     10Only those \CFA features pertinent to this thesis are discussed.  Many of the
     11\CFA syntactic and semantic features used in the thesis should be fairly
     12obvious to the reader.
     13
     14\section{Overloading and \lstinline{extern}}
     15\CFA has extensive overloading, allowing multiple definitions of the same name
     16to be defined.~\cite{Moss18}
     17\begin{cfa}
     18char i; int i; double i;                        $\C[3.75in]{// variable overload}$
     19int f(); double f();                            $\C{// return overload}$
     20void g( int ); void g( double );        $\C{// parameter overload}\CRT$
     21\end{cfa}
     22This feature requires name mangling so the assembly symbols are unique for
     23different overloads. For compatibility with names in C, there is also a syntax
     24to disable name mangling. These unmangled names cannot be overloaded but act as
     25the interface between C and \CFA code.  The syntax for disabling/enabling
     26mangling is:
     27\begin{cfa}
     28// name mangling
     29int i; // _X1ii_1
     30@extern "C"@ {  // no name mangling
     31        int j; // j
     32        @extern "Cforall"@ {  // name mangling
     33                int k; // _X1ki_1
     34        }
     35        // no name mangling
     36}
     37// name mangling
     38\end{cfa}
     39Both forms of @extern@ affect all the declarations within their nested lexical
     40scope and transition back to the previous mangling state when the lexical scope
     41ends.
     42
     43\section{Reference Type}
     44\CFA adds a rebindable reference type to C, but more expressive than the \Cpp
     45reference.  Multi-level references are allowed and act like auto-dereferenced
     46pointers using the ampersand (@&@) instead of the pointer asterisk (@*@). \CFA
     47references may also be mutable or non-mutable. If mutable, a reference variable
     48may be assigned to using the address-of operator (@&@), which converts the
     49reference to a pointer.
     50\begin{cfa}
     51int i, j;
     52int @&@ ri = i, @&&@ rri = ri;
     53rri = 3;  // auto-dereference assign to i
     54@&@ri = @&@j; // rebindable
     55ri = 5;   // assign to j
     56\end{cfa}
    3757
    3858\section{Constructors and Destructors}
    3959
    4060Both constructors and destructors are operators, which means they are just
    41 functions with special names. The special names are used to define them and
    42 may be used to call the functions expicately. The \CFA special names are
    43 constructed by taking the tokens in the operators and putting \texttt{?} where
    44 the arguments would go. So multiplication is \texttt{?*?} while dereference
    45 is \texttt{*?}. This also make it easy to tell the difference between
    46 pre-fix operations (such as \texttt{++?}) and post-fix operations
    47 (\texttt{?++}).
    48 
    49 The special name for contructors is \texttt{?\{\}}, which comes from the
    50 initialization syntax in C. The special name for destructors is
    51 \texttt{\^{}?\{\}}. % I don't like the \^{} symbol but $^\wedge$ isn't better.
    52 
    53 Any time a type T goes out of scope the destructor matching
    54 \codeCFA{void ^?\{\}(T \&);} is called. In theory this is also true of
    55 primitive types such as \codeCFA{int}, but in practice those are no-ops and
    56 are usually omitted for optimization.
     61functions with special operator names rather than type names in \Cpp. The
     62special operator names may be used to call the functions explicitly (not
     63allowed in \Cpp for constructors).
     64
     65In general, operator names in \CFA are constructed by bracketing an operator
     66token with @?@, which indicates where the arguments. For example, infixed
     67multiplication is @?*?@ while prefix dereference is @*?@. This syntax make it
     68easy to tell the difference between prefix operations (such as @++?@) and
     69post-fix operations (@?++@).
     70
     71The special name for a constructor is @?{}@, which comes from the
     72initialization syntax in C. The special name for a destructor is @^{}@, where
     73the @^@ has no special meaning.
     74% I don't like the \^{} symbol but $^\wedge$ isn't better.
     75\begin{cfa}
     76struct T { ... };
     77void ?@{}@(@T &@ this, ...) { ... }  // constructor
     78void ?@^{}@(@T &@ this, ...) { ... } // destructor
     79{
     80        T s = @{@ ... @}@;  // same constructor/initialization braces
     81} // destructor call automatically generated
     82\end{cfa}
     83The first parameter is a reference parameter to the type for the
     84constructor/destructor. Destructors may have multiple parameters.  The compiler
     85implicitly matches an overloaded constructor @void ^?{}(T &, ...);@ to an
     86object declaration with associated initialization, and generates a construction
     87call after the object is allocated. When an object goes out of scope, the
     88matching overloaded destructor @void ^?{}(T &);@ is called.  Without explicit
     89definition, \CFA creates a default and copy constructor, destructor and
     90assignment (like \Cpp). It is possible to define constructors/destructors for
     91basic and existing types.
    5792
    5893\section{Polymorphism}
    59 \CFA uses polymorphism to create functions and types that are defined over
    60 different types. \CFA polymorphic declarations serve the same role as \CPP
    61 templates or Java generics.
    62 
    63 Polymorphic declaractions start with a forall clause that goes before the
    64 standard (monomorphic) declaration. These declarations have the same syntax
    65 except that you may use the names introduced by the forall clause in them.
    66 
    67 Forall clauses are written \codeCFA{forall( ... )} where \codeCFA{...} becomes
    68 the list of polymorphic variables (local type names) and assertions, which
    69 repersent required operations on those types.
    70 
    71 \begin{lstlisting}
    72 forall(dtype T | { void do_once(T &); })
    73 void do_twice(T & value) {
    74     do_once(value);
    75     do_once(value);
    76 }
    77 \end{lstlisting}
    78 
    79 A polymorphic function can be used in the same way normal functions are.
    80 The polymorphics variables are filled in with concrete types and the
    81 assertions are checked. An assertion checked by seeing if that name of that
    82 type (with all the variables replaced with the concrete types) is defined at
    83 the the call site.
    84 
    85 As an example, even if no function named \codeCFA{do_once} is not defined
    86 near the definition of \codeCFA{do_twice} the following code will work.
    87 \begin{lstlisting}
     94\CFA uses parametric polymorphism to create functions and types that are
     95defined over multiple types. \CFA polymorphic declarations serve the same role
     96as \Cpp templates or Java generics. The ``parametric'' means the polymorphism is
     97accomplished by passing argument operations to associate \emph{parameters} at
     98the call site, and these parameters are used in the function to differentiate
     99among the types the function operates on.
     100
     101Polymorphic declarations start with a universal @forall@ clause that goes
     102before the standard (monomorphic) declaration. These declarations have the same
     103syntax except they may use the universal type names introduced by the @forall@
     104clause.  For example, the following is a polymorphic identity function that
     105works on any type @T@:
     106\begin{cfa}
     107@forall( T )@ @T@ identity( @T@ val ) { return val; }
     108int forty_two = identity( 42 ); // T bound to int, forty_two == 42
     109\end{cfa}
     110
     111To allow a polymorphic function to be separately compiled, the type @T@ must be
     112constrained by the operations used on @T@ in the function body. The @forall@
     113clauses is augmented with a list of polymorphic variables (local type names)
     114and assertions (constraints), which represent the required operations on those
     115types used in a function, \eg:
     116\begin{cfa}
     117forall( T @| { void do_once(T); }@) // assertion
     118void do_twice(T value) {
     119        do_once(value);
     120        do_once(value);
     121}
     122void do_once(int i) { ... }  // provide assertion
     123int i;
     124do_twice(i); // implicitly pass assertion do_once to do_twice
     125\end{cfa}
     126Any object with a type fulfilling the assertion may be passed as an argument to
     127a @do_twice@ call.
     128
     129A polymorphic function can be used in the same way as a normal function.  The
     130polymorphic variables are filled in with concrete types and the assertions are
     131checked. An assertion is checked by verifying each assertion operation (with
     132all the variables replaced with the concrete types from the arguments) is
     133defined at a call site.
     134
     135Note, a function named @do_once@ is not required in the scope of @do_twice@ to
     136compile it, unlike \Cpp template expansion. Furthermore, call-site inferencing
     137allows local replacement of the most specific parametric functions needs for a
     138call.
     139\begin{cfa}
     140void do_once(double y) { ... } // global
    88141int quadruple(int x) {
    89     void do_once(int & y) {
    90         y = y * 2;
    91     }
    92     do_twice(x);
    93     return x;
    94 }
    95 \end{lstlisting}
    96 This is not the recommended way to implement a quadruple function but it
    97 does work. The complier will deduce that \codeCFA{do_twice}'s T is an
    98 integer from the argument. It will then look for a definition matching the
    99 assertion which is the \codeCFA{do_once} defined within the function. That
    100 function will be passed in as a function pointer to \codeCFA{do_twice} and
    101 called within it.
    102 
    103 To avoid typing out long lists of assertions again and again there are also
    104 traits which collect assertions into convenent packages that can then be used
    105 in assertion lists instead of all of their components.
    106 \begin{lstlisting}
    107 trait done_once(dtype T) {
    108     void do_once(T &);
    109 }
    110 \end{lstlisting}
    111 
    112 After this the forall list in the previous example could instead be written
    113 with the trait instead of the assertion itself.
    114 \begin{lstlisting}
    115 forall(dtype T | done_once(T))
    116 \end{lstlisting}
    117 
    118 Traits can have arbitrary number of assertions in them and are usually used to
    119 create short hands for, and give descriptive names to, commond groupings of
    120 assertions.
    121 
    122 Polymorphic structures and unions may also be defined by putting a forall
    123 clause before the declaration. The type variables work the same way except
    124 are now used in field declaractions instead of parameters and local variables.
    125 
    126 \begin{lstlisting}
     142        void do_once(int y) { y = y * 2; } // local
     143        do_twice(x); // using local "do_once"
     144        return x;
     145}
     146\end{cfa}
     147Specifically, the complier deduces that @do_twice@'s T is an integer from the
     148argument @x@. It then looks for the most specific definition matching the
     149assertion, which is the nested integral @do_once@ defined within the
     150function. The matched assertion function is then passed as a function pointer
     151to @do_twice@ and called within it.
     152
     153To avoid typing long lists of assertions, constraints can be collect into
     154convenient packages called a @trait@, which can then be used in an assertion
     155instead of the individual constraints.
     156\begin{cfa}
     157trait done_once(T) {
     158        void do_once(T);
     159}
     160\end{cfa}
     161and the @forall@ list in the previous example is replaced with the trait.
     162\begin{cfa}
     163forall(dtype T | @done_once(T)@)
     164\end{cfa}
     165In general, a trait can contain an arbitrary number of assertions, both
     166functions and variables, and are usually used to create a shorthand for, and
     167give descriptive names to, common groupings of assertions describing a certain
     168functionality, like @sumable@, @listable@, \etc.
     169
     170Polymorphic structures and unions are defined by qualifying the aggregate type
     171with @forall@. The type variables work the same except they are used in field
     172declarations instead of parameters, returns, and local variable declarations.
     173\begin{cfa}
    127174forall(dtype T)
    128175struct node {
    129     node(T) * next;
    130     T * data;
    131 }
    132 \end{lstlisting}
    133 
    134 The \codeCFA{node(T)} is a use of a polymorphic structure. Polymorphic types
    135 must be provided their polymorphic parameters.
    136 
    137 There are many other features of polymorphism that have not given here but
    138 these are the ones used by the exception system.
     176        node(T) * next;  // generic linked node
     177        T * data;
     178}
     179\end{cfa}
     180The generic type @node(T)@ is an example of a polymorphic-type usage.  Like \Cpp
     181templates usage, a polymorphic-type usage must specify a type parameter.
     182
     183There are many other polymorphism features in \CFA but these are the ones used
     184by the exception system.
    139185
    140186\section{Concurrency}
    141 
    142 \CFA has a number of concurrency features, \codeCFA{thread}s,
    143 \codeCFA{monitor}s and \codeCFA{mutex} parameters, \codeCFA{coroutine}s and
    144 \codeCFA{generator}s. The two features that interact with the exception system
    145 are \codeCFA{thread}s and \codeCFA{coroutine}s; they and their supporting
    146 constructs will be described here.
    147 
    148 \subsection{Coroutines}
    149 Coroutines are routines that do not have to finish execution to hand control
    150 back to their caller, instead they may suspend their execution at any time
    151 and resume it later.
    152 Coroutines are not true concurrency but share some similarities and many of
    153 the same underpinnings and so are included as part of the \CFA threading
    154 library.
    155 
    156 In \CFA coroutines are created using the \codeCFA{coroutine} keyword which
    157 works just like \codeCFA{struct} except that the created structure will be
    158 modified by the compiler to satify the \codeCFA{is_coroutine} trait.
    159 
    160 These structures act as the interface between callers and the coroutine,
    161 the fields are used to pass information in and out. Here is a simple example
    162 where the single field is used to pass the next number in a sequence out.
    163 \begin{lstlisting}
     187\CFA has a number of concurrency features: @thread@, @monitor@, @mutex@
     188parameters, @coroutine@ and @generator@. The two features that interact with
     189the exception system are @thread@ and @coroutine@; they and their supporting
     190constructs are described here.
     191
     192\subsection{Coroutine}
     193A coroutine is a type with associated functions, where the functions are not
     194required to finish execution when control is handed back to the caller. Instead
     195they may suspend execution at any time and be resumed later at the point of
     196last suspension. (Generators are stackless and coroutines are stackful.) These
     197types are not concurrent but share some similarities along with common
     198underpinnings, so they are combined with the \CFA threading library. Further
     199discussion in this section only refers to the coroutine because generators are
     200similar.
     201
     202In \CFA, a coroutine is created using the @coroutine@ keyword, which is an
     203aggregate type like @struct,@ except the structure is implicitly modified by
     204the compiler to satisfy the @is_coroutine@ trait; hence, a coroutine is
     205restricted by the type system to types that provide this special trait.  The
     206coroutine structure acts as the interface between callers and the coroutine,
     207and its fields are used to pass information in and out of coroutine interface
     208functions.
     209
     210Here is a simple example where a single field is used to pass (communicate) the
     211next number in a sequence.
     212\begin{cfa}
    164213coroutine CountUp {
    165     unsigned int next;
    166 }
    167 \end{lstlisting}
    168 
    169 The routine part of the coroutine is a main function for the coroutine. It
    170 takes a reference to a coroutine object and returns nothing. In this function,
    171 and any functions called by this function, the suspend statement may be used
    172 to return execution to the coroutine's caller. When control returns to the
    173 function it continue from that same suspend statement instead of at the top
    174 of the function.
    175 \begin{lstlisting}
    176 void main(CountUp & this) {
    177     unsigned int next = 0;
    178     while (true) {
    179         this.next = next;
    180         suspend;
    181         next = next + 1;
    182     }
    183 }
    184 \end{lstlisting}
    185 
    186 Control is passed to the coroutine with the resume function. This includes the
    187 first time when the coroutine is starting up. The resume function takes a
    188 reference to the coroutine structure and returns the same reference. The
    189 return value is for easy access to communication variables. For example the
    190 next value from a count-up can be generated and collected in a single
    191 expression: \codeCFA{resume(count).next}.
     214        unsigned int next; // communication variable
     215}
     216CountUp countup;
     217\end{cfa}
     218Each coroutine has @main@ function, which takes a reference to a coroutine
     219object and returns @void@.
     220\begin{cfa}[numbers=left]
     221void main(@CountUp & this@) { // argument matches trait is_coroutine
     222        unsigned int up = 0;  // retained between calls
     223        while (true) {
     224                next = up; // make "up" available outside function
     225                @suspend;@$\label{suspend}$
     226                up += 1;
     227        }
     228}
     229\end{cfa}
     230In this function, or functions called by this function (helper functions), the
     231@suspend@ statement is used to return execution to the coroutine's caller
     232without terminating the coroutine.
     233
     234A coroutine is resumed by calling the @resume@ function, \eg @resume(countup)@.
     235The first resume calls the @main@ function at the top. Thereafter, resume calls
     236continue a coroutine in the last suspended function after the @suspend@
     237statement, in this case @main@ line~\ref{suspend}.  The @resume@ function takes
     238a reference to the coroutine structure and returns the same reference. The
     239return value allows easy access to communication variables defined in the
     240coroutine object. For example, the @next@ value for coroutine object @countup@
     241is both generated and collected in the single expression:
     242@resume(countup).next@.
    192243
    193244\subsection{Monitors and Mutex}
    194 
    195 True concurrency does not garrenty ordering. To get some of that ordering back
    196 \CFA uses monitors and mutex (mutual exclution) parameters. A monitor is
    197 another special declaration that contains a lock and is compatable with mutex
    198 parameters.
    199 
    200 Function parameters can have the \codeCFA{mutex} qualifiers on reference
    201 arguments, for example \codeCFA{void example(a_monitor & mutex arg);}. When the
    202 function is called it will acquire the lock on all of the mutex parameters.
    203 
    204 This means that all functions that mutex on a type are part of a critical
    205 section and only one will ever run at a time.
     245Concurrency does not guarantee ordering; without ordering results are
     246non-deterministic. To claw back ordering, \CFA uses monitors and @mutex@
     247(mutual exclusion) parameters. A monitor is another kind of aggregate, where
     248the compiler implicitly inserts a lock and instances are compatible with
     249@mutex@ parameters.
     250
     251A function that requires deterministic (ordered) execution, acquires mutual
     252exclusion on a monitor object by qualifying an object reference parameter with
     253@mutex@.
     254\begin{cfa}
     255void example(MonitorA & @mutex@ argA, MonitorB & @mutex@ argB);
     256\end{cfa}
     257When the function is called, it implicitly acquires the monitor lock for all of
     258the mutex parameters without deadlock.  This semantics means all functions with
     259the same mutex type(s) are part of a critical section for objects of that type
     260and only one runs at a time.
    206261
    207262\subsection{Threads}
    208 While coroutines allow new things to be done with a single execution path
    209 threads actually introduce new paths of execution that continue independently.
    210 Now for threads to work together their must be some communication between them
    211 and that means the timing of certain operations does have to be known. There
    212 or various means of syncronization and mutual exclution provided by \CFA but
    213 for exceptions only the basic two -- fork and join -- are needed.
    214 
    215 Threads are created like coroutines except the keyword is changed:
    216 \begin{lstlisting}
     263Functions, generators, and coroutines are sequential so there is only a single
     264(but potentially sophisticated) execution path in a program. Threads introduce
     265multiple execution paths that continue independently.
     266
     267For threads to work safely with objects requires mutual exclusion using
     268monitors and mutex parameters. For threads to work safely with other threads,
     269also requires mutual exclusion in the form of a communication rendezvous, which
     270also supports internal synchronization as for mutex objects. For exceptions
     271only the basic two basic operations are important: thread fork and join.
     272
     273Threads are created like coroutines with an associated @main@ function:
     274\begin{cfa}
    217275thread StringWorker {
    218     const char * input;
    219     int result;
     276        const char * input;
     277        int result;
    220278};
    221 
    222279void main(StringWorker & this) {
    223     const char * localCopy = this.input;
    224     // ... do some work, perhaps hashing the string ...
    225     this.result = result;
    226 }
    227 \end{lstlisting}
    228 The main function will start executing after the fork operation and continue
    229 executing until it is finished. If another thread joins with this one it will
    230 wait until main has completed execution. In other words everything the thread
    231 does is between fork and join.
    232 
    233 From the outside this is the creation and destruction of the thread object.
    234 Fork happens after the constructor is run and join happens before the
    235 destructor runs. Join also happens during the \codeCFA{join} function which
    236 can be used to join a thread earlier. If it is used the destructor does not
    237 join as that has already been completed.
     280        const char * localCopy = this.input;
     281        // ... do some work, perhaps hashing the string ...
     282        this.result = result;
     283}
     284{
     285        StringWorker stringworker; // fork thread running in "main"
     286} // implicitly join with thread $\(\Rightarrow\)$ wait for completion
     287\end{cfa}
     288The thread main is where a new thread starts execution after a fork operation
     289and then the thread continues executing until it is finished. If another thread
     290joins with an executing thread, it waits until the executing main completes
     291execution. In other words, everything a thread does is between a fork and join.
     292
     293From the outside, this behaviour is accomplished through creation and
     294destruction of a thread object.  Implicitly, fork happens after a thread
     295object's constructor is run and join happens before the destructor runs. Join
     296can also be specified explicitly using the @join@ function to wait for a
     297thread's completion independently from its deallocation (\ie destructor
     298call). If @join@ is called explicitly, the destructor does not implicitly join.
  • doc/theses/andrew_beach_MMath/features.tex

    r342af53 r8e4aa05  
    1 \chapter{Features}
    2 
    3 This chapter covers the design and user interface of the \CFA exception
    4 handling mechanism.
    5 
    6 \section{Virtual Casts}
    7 
    8 Virtual casts and virtual types are not truly part of the exception system but
    9 they did not exist in \CFA and are useful in exceptions. So a minimal version
    10 of they virtual system was designed and implemented.
    11 
    12 Virtual types are organizied in simple hierarchies. Each virtual type may have
    13 a parent and can have any number of children. A type's descendants are its
    14 children and its children's descendants. A type may not be its own descendant.
    15 
    16 Each virtual type has an associated virtual table type. A virtual table is a
    17 structure that has fields for all the virtual members of a type. A virtual
    18 type has all the virtual members of its parent and can add more. It may also
    19 update the values of the virtual members.
    20 
    21 Except for virtual casts, this is only used internally in the exception
    22 system. There is no general purpose interface for the other features. A
    23 a virtual cast has the following syntax:
    24 
    25 \begin{lstlisting}
     1\chapter{Exception Features}
     2
     3This chapter covers the design and user interface of the \CFA
     4exception-handling mechanism.
     5
     6\section{Virtuals}
     7Virtual types and casts are not part of the exception system nor are they
     8required for an exception system. But an object-oriented style hierarchy is a
     9great way of organizing exceptions so a minimal virtual system has been added
     10to \CFA.
     11
     12The pattern of a simple hierarchy was borrowed from object-oriented
     13programming was chosen for several reasons.
     14The first is that it allows new exceptions to be added in user code
     15and in libraries independently of each other. Another is it allows for
     16different levels of exception grouping (all exceptions, all IO exceptions or
     17a particular IO exception). Also it also provides a simple way of passing
     18data back and forth across the throw.
     19
     20Virtual types and casts are not required for a basic exception-system but are
     21useful for advanced exception features. However, \CFA is not object-oriented so
     22there is no obvious concept of virtuals. Hence, to create advanced exception
     23features for this work, I needed to design and implement a virtual-like
     24system for \CFA.
     25
     26% NOTE: Maybe we should but less of the rational here.
     27Object-oriented languages often organized exceptions into a simple hierarchy,
     28\eg Java.
     29\begin{center}
     30\setlength{\unitlength}{4000sp}%
     31\begin{picture}(1605,612)(2011,-1951)
     32\put(2100,-1411){\vector(1, 0){225}}
     33\put(3450,-1411){\vector(1, 0){225}}
     34\put(3550,-1411){\line(0,-1){225}}
     35\put(3550,-1636){\vector(1, 0){150}}
     36\put(3550,-1636){\line(0,-1){225}}
     37\put(3550,-1861){\vector(1, 0){150}}
     38\put(2025,-1490){\makebox(0,0)[rb]{\LstBasicStyle{exception}}}
     39\put(2400,-1460){\makebox(0,0)[lb]{\LstBasicStyle{arithmetic}}}
     40\put(3750,-1460){\makebox(0,0)[lb]{\LstBasicStyle{underflow}}}
     41\put(3750,-1690){\makebox(0,0)[lb]{\LstBasicStyle{overflow}}}
     42\put(3750,-1920){\makebox(0,0)[lb]{\LstBasicStyle{zerodivide}}}
     43\end{picture}%
     44\end{center}
     45The hierarchy provides the ability to handle an exception at different degrees
     46of specificity (left to right). Hence, it is possible to catch a more general
     47exception-type in higher-level code where the implementation details are
     48unknown, which reduces tight coupling to the lower-level implementation.
     49Otherwise, low-level code changes require higher-level code changes, \eg,
     50changing from raising @underflow@ to @overflow@ at the low level means changing
     51the matching catch at the high level versus catching the general @arithmetic@
     52exception. In detail, each virtual type may have a parent and can have any
     53number of children. A type's descendants are its children and its children's
     54descendants. A type may not be its own descendant.
     55
     56The exception hierarchy allows a handler (@catch@ clause) to match multiple
     57exceptions, \eg a base-type handler catches both base and derived
     58exception-types.
     59\begin{cfa}
     60try {
     61        ...
     62} catch(arithmetic &) {
     63        ... // handle arithmetic, underflow, overflow, zerodivide
     64}
     65\end{cfa}
     66Most exception mechanisms perform a linear search of the handlers and select
     67the first matching handler, so the order of handers is now important because
     68matching is many to one.
     69
     70Each virtual type needs an associated virtual table. A virtual table is a
     71structure with fields for all the virtual members of a type. A virtual type has
     72all the virtual members of its parent and can add more. It may also update the
     73values of the virtual members and often does.
     74
     75While much of the virtual infrastructure is created, it is currently only used
     76internally for exception handling. The only user-level feature is the virtual
     77cast, which is the same as the \Cpp \lstinline[language=C++]|dynamic_cast|.
     78\label{p:VirtualCast}
     79\begin{cfa}
    2680(virtual TYPE)EXPRESSION
    27 \end{lstlisting}
    28 
    29 This has the same precidence as a traditional C-cast and can be used in the
    30 same places. This will convert the result of EXPRESSION to the type TYPE. Both
    31 the type of EXPRESSION and TYPE must be pointers to virtual types.
    32 
    33 The cast is checked and will either return the original value or null, based
    34 on the result of the check. The check is does the object pointed at have a
    35 type that is a descendant of the target type. If it is the result is the
    36 pointer, otherwise the result is null.
    37 
    38 \section{Exceptions}
     81\end{cfa}
     82Note, the syntax and semantics matches a C-cast, rather than the function-like
     83\Cpp syntax for special casts. Both the type of @EXPRESSION@ and @TYPE@ must be
     84a pointer to a virtual type.
     85The cast dynamically checks if the @EXPRESSION@ type is the same or a subtype
     86of @TYPE@, and if true, returns a pointer to the
     87@EXPRESSION@ object, otherwise it returns @0p@ (null pointer).
     88
     89\section{Exception}
    3990% Leaving until later, hopefully it can talk about actual syntax instead
    4091% of my many strange macros. Syntax aside I will also have to talk about the
    4192% features all exceptions support.
    4293
    43 \section{Termination}
    44 
    45 Termination exception throws are likely the most framilar kind, as they are
    46 used in several popular programming languages. A termination will throw an
    47 exception, search the stack for a handler, unwind the stack to where the
    48 handler is defined, execute the handler and then continue execution after
    49 the handler. They are used when execution cannot continue here.
    50 
    51 Termination has two pieces of syntax it uses. The first is the throw:
    52 \begin{lstlisting}
     94Exceptions are defined by the trait system; there are a series of traits, and
     95if a type satisfies them, then it can be used as an exception. The following
     96is the base trait all exceptions need to match.
     97\begin{cfa}
     98trait is_exception(exceptT &, virtualT &) {
     99        virtualT const & get_exception_vtable(exceptT *);
     100};
     101\end{cfa}
     102The trait is defined over two types, the exception type and the virtual table
     103type. This should be one-to-one, each exception type has only one virtual
     104table type and vice versa. The only assertion in the trait is
     105@get_exception_vtable@, which takes a pointer of the exception type and
     106returns a reference to the virtual table type instance.
     107
     108The function @get_exception_vtable@ is actually a constant function.
     109Recardless of the value passed in (including the null pointer) it should
     110return a reference to the virtual table instance for that type.
     111The reason it is a function instead of a constant is that it make type
     112annotations easier to write as you can use the exception type instead of the
     113virtual table type; which usually has a mangled name.
     114% Also \CFA's trait system handles functions better than constants and doing
     115% it this way reduce the amount of boiler plate we need.
     116
     117% I did have a note about how it is the programmer's responsibility to make
     118% sure the function is implemented correctly. But this is true of every
     119% similar system I know of (except Agda's I guess) so I took it out.
     120
     121There are two more traits for exceptions @is_termination_exception@ and
     122@is_resumption_exception@. They are defined as follows:
     123
     124\begin{cfa}
     125trait is_termination_exception(
     126                exceptT &, virtualT & | is_exception(exceptT, virtualT)) {
     127        void defaultTerminationHandler(exceptT &);
     128};
     129
     130trait is_resumption_exception(
     131                exceptT &, virtualT & | is_exception(exceptT, virtualT)) {
     132        void defaultResumptionHandler(exceptT &);
     133};
     134\end{cfa}
     135
     136In other words they make sure that a given type and virtual type is an
     137exception and defines one of the two default handlers. These default handlers
     138are used in the main exception handling operations \see{Exception Handling}
     139and their use will be detailed there.
     140
     141However all three of these traits can be trickly to use directly.
     142There is a bit of repetition required but
     143the largest issue is that the virtual table type is mangled and not in a user
     144facing way. So there are three macros that can be used to wrap these traits
     145when you need to refer to the names:
     146@IS_EXCEPTION@, @IS_TERMINATION_EXCEPTION@ and @IS_RESUMPTION_EXCEPTION@.
     147
     148All take one or two arguments. The first argument is the name of the
     149exception type. Its unmangled and mangled form are passed to the trait.
     150The second (optional) argument is a parenthesized list of polymorphic
     151arguments. This argument should only with polymorphic exceptions and the
     152list will be passed to both types.
     153In the current set-up the base name and the polymorphic arguments have to
     154match so these macros can be used without losing flexability.
     155
     156For example consider a function that is polymorphic over types that have a
     157defined arithmetic exception:
     158\begin{cfa}
     159forall(Num | IS_EXCEPTION(Arithmetic, (Num)))
     160void some_math_function(Num & left, Num & right);
     161\end{cfa}
     162
     163\section{Exception Handling}
     164\CFA provides two kinds of exception handling, termination and resumption.
     165These twin operations are the core of the exception handling mechanism and
     166are the reason for the features of exceptions.
     167This section will cover the general patterns shared by the two operations and
     168then go on to cover the details each individual operation.
     169
     170Both operations follow the same set of steps to do their operation. They both
     171start with the user preforming a throw on an exception.
     172Then there is the search for a handler, if one is found than the exception
     173is caught and the handler is run. After that control returns to normal
     174execution.
     175
     176If the search fails a default handler is run and then control
     177returns to normal execution immediately. That is where the default handlers
     178@defaultTermiationHandler@ and @defaultResumptionHandler@ are used.
     179
     180\subsection{Termination}
     181\label{s:Termination}
     182
     183Termination handling is more familiar kind and used in most programming
     184languages with exception handling.
     185It is dynamic, non-local goto. If a throw is successful then the stack will
     186be unwound and control will (usually) continue in a different function on
     187the call stack. They are commonly used when an error has occured and recovery
     188is impossible in the current function.
     189
     190% (usually) Control can continue in the current function but then a different
     191% control flow construct should be used.
     192
     193A termination throw is started with the @throw@ statement:
     194\begin{cfa}
    53195throw EXPRESSION;
    54 \end{lstlisting}
    55 
    56 The expression must evaluate to a reference to a termination exception. A
    57 termination exception is any exception with a
    58 \codeCFA{void defaultTerminationHandler(T &);} (the default handler) defined
    59 on it. The handler is taken from the call sight with \CFA's trait system and
    60 passed into the exception system along with the exception itself.
    61 
    62 The exception passed into the system is then copied into managed memory.
    63 This is to ensure it remains in scope during unwinding. It is the user's
    64 responsibility to make sure the original exception is freed when it goes out
    65 of scope. Being allocated on the stack is sufficient for this.
    66 
    67 Then the exception system will search the stack starting from the throw and
    68 proceding towards the base of the stack, from callee to caller. As it goes
    69 it will check any termination handlers it finds:
    70 
    71 \begin{lstlisting}
    72 try {
    73     TRY_BLOCK
    74 } catch (EXCEPTION_TYPE * NAME) {
    75     HANDLER
    76 }
    77 \end{lstlisting}
    78 
    79 This shows a try statement with a single termination handler. The statements
    80 in TRY\_BLOCK will be executed when control reaches this statement. While
    81 those statements are being executed if a termination exception is thrown and
    82 it is not handled by a try statement further up the stack the EHM will check
    83 all of the terminations handlers attached to the try block, top to bottom.
    84 
    85 At each handler the EHM will check to see if the thrown exception is a
    86 descendant of EXCEPTION\_TYPE. If it is the pointer to the exception is
    87 bound to NAME and the statements in HANDLER are executed. If control reaches
    88 the end of the handler then it exits the block, the exception is freed and
    89 control continues after the try statement.
    90 
    91 The default handler is only used if no handler for the exception is found
    92 after the entire stack is searched. When that happens the default handler
    93 is called with a reference to the exception as its only argument. If the
    94 handler returns control continues from after the throw statement.
    95 
    96 \paragraph{Conditional Catches}
    97 
    98 Catch clauses may also be written as:
    99 \begin{lstlisting}
     196\end{cfa}
     197The expression must return a reference to a termination exception, where the
     198termination exception is any type that satifies @is_termination_exception@
     199at the call site.
     200Through \CFA's trait system the functions in the traits are passed into the
     201throw code. A new @defaultTerminationHandler@ can be defined in any scope to
     202change the throw's behavior (see below).
     203
     204The throw will copy the provided exception into managed memory. It is the
     205user's responcibility to ensure the original exception is cleaned up if the
     206stack is unwound (allocating it on the stack should be sufficient).
     207
     208Then the exception system searches the stack using the copied exception.
     209It starts starts from the throw and proceeds to the base of the stack,
     210from callee to caller.
     211At each stack frame, a check is made for resumption handlers defined by the
     212@catch@ clauses of a @try@ statement.
     213\begin{cfa}
     214try {
     215        GUARDED_BLOCK
     216} catch (EXCEPTION_TYPE$\(_1\)$ * NAME$\(_1\)$) {
     217        HANDLER_BLOCK$\(_1\)$
     218} catch (EXCEPTION_TYPE$\(_2\)$ * NAME$\(_2\)$) {
     219        HANDLER_BLOCK$\(_2\)$
     220}
     221\end{cfa}
     222When viewed on its own a try statement will simply exceute the statements in
     223@GUARDED_BLOCK@ and when those are finished the try statement finishes.
     224
     225However, while the guarded statements are being executed, including any
     226functions they invoke, all the handlers following the try block are now
     227or any functions invoked from those
     228statements, throws an exception, and the exception
     229is not handled by a try statement further up the stack, the termination
     230handlers are searched for a matching exception type from top to bottom.
     231
     232Exception matching checks the representation of the thrown exception-type is
     233the same or a descendant type of the exception types in the handler clauses. If
     234it is the same of a descendent of @EXCEPTION_TYPE@$_i$ then @NAME@$_i$ is
     235bound to a pointer to the exception and the statements in @HANDLER_BLOCK@$_i$
     236are executed. If control reaches the end of the handler, the exception is
     237freed and control continues after the try statement.
     238
     239If no handler is found during the search then the default handler is run.
     240Through \CFA's trait system the best match at the throw sight will be used.
     241This function is run and is passed the copied exception. After the default
     242handler is run control continues after the throw statement.
     243
     244There is a global @defaultTerminationHandler@ that cancels the current stack
     245with the copied exception. However it is generic over all exception types so
     246new default handlers can be defined for different exception types and so
     247different exception types can have different default handlers.
     248
     249\subsection{Resumption}
     250\label{s:Resumption}
     251
     252Resumption exception handling is a less common form than termination but is
     253just as old~\cite{Goodenough75} and is in some sense simpler.
     254It is a dynamic, non-local function call. If the throw is successful a
     255closure will be taken from up the stack and executed, after which the throwing
     256function will continue executing.
     257These are most often used when an error occured and if the error is repaired
     258then the function can continue.
     259
     260A resumption raise is started with the @throwResume@ statement:
     261\begin{cfa}
     262throwResume EXPRESSION;
     263\end{cfa}
     264The semantics of the @throwResume@ statement are like the @throw@, but the
     265expression has return a reference a type that satifies the trait
     266@is_resumption_exception@. The assertions from this trait are available to
     267the exception system while handling the exception.
     268
     269At runtime, no copies are made. As the stack is not unwound the exception and
     270any values on the stack will remain in scope while the resumption is handled.
     271
     272Then the exception system searches the stack using the provided exception.
     273It starts starts from the throw and proceeds to the base of the stack,
     274from callee to caller.
     275At each stack frame, a check is made for resumption handlers defined by the
     276@catchResume@ clauses of a @try@ statement.
     277\begin{cfa}
     278try {
     279        GUARDED_BLOCK
     280} catchResume (EXCEPTION_TYPE$\(_1\)$ * NAME$\(_1\)$) {
     281        HANDLER_BLOCK$\(_1\)$
     282} catchResume (EXCEPTION_TYPE$\(_2\)$ * NAME$\(_2\)$) {
     283        HANDLER_BLOCK$\(_2\)$
     284}
     285\end{cfa}
     286If the handlers are not involved in a search this will simply execute the
     287@GUARDED_BLOCK@ and then continue to the next statement.
     288Its purpose is to add handlers onto the stack.
     289(Note, termination and resumption handlers may be intermixed in a @try@
     290statement but the kind of throw must be the same as the handler for it to be
     291considered as a possible match.)
     292
     293If a search for a resumption handler reaches a try block it will check each
     294@catchResume@ clause, top-to-bottom.
     295At each handler if the thrown exception is or is a child type of
     296@EXCEPTION_TYPE@$_i$ then the a pointer to the exception is bound to
     297@NAME@$_i$ and then @HANDLER_BLOCK@$_i$ is executed. After the block is
     298finished control will return to the @throwResume@ statement.
     299
     300Like termination, if no resumption handler is found, the default handler
     301visible at the throw statement is called. It will use the best match at the
     302call sight according to \CFA's overloading rules. The default handler is
     303passed the exception given to the throw. When the default handler finishes
     304execution continues after the throw statement.
     305
     306There is a global @defaultResumptionHandler@ is polymorphic over all
     307termination exceptions and preforms a termination throw on the exception.
     308The @defaultTerminationHandler@ for that throw is matched at the original
     309throw statement (the resumption @throwResume@) and it can be customized by
     310introducing a new or better match as well.
     311
     312% \subsubsection?
     313
     314A key difference between resumption and termination is that resumption does
     315not unwind the stack. A side effect that is that when a handler is matched
     316and run it's try block (the guarded statements) and every try statement
     317searched before it are still on the stack. This can lead to the recursive
     318resumption problem.
     319
     320The recursive resumption problem is any situation where a resumption handler
     321ends up being called while it is running.
     322Consider a trivial case:
     323\begin{cfa}
     324try {
     325        throwResume (E &){};
     326} catchResume(E *) {
     327        throwResume (E &){};
     328}
     329\end{cfa}
     330When this code is executed the guarded @throwResume@ will throw, start a
     331search and match the handler in the @catchResume@ clause. This will be
     332call and placed on the stack on top of the try-block. The second throw then
     333throws and will seach the same try block and put call another instance of the
     334same handler leading to an infinite loop.
     335
     336This situation is trivial and easy to avoid, but much more complex cycles
     337can form with multiple handlers and different exception types.
     338
     339To prevent all of these cases we mask sections of the stack, or equvilantly
     340the try statements on the stack, so that the resumption seach skips over
     341them and continues with the next unmasked section of the stack.
     342
     343A section of the stack is marked when it is searched to see if it contains
     344a handler for an exception and unmarked when that exception has been handled
     345or the search was completed without finding a handler.
     346
     347% This might need a diagram. But it is an important part of the justification
     348% of the design of the traversal order.
     349\begin{verbatim}
     350       throwResume2 ----------.
     351            |                 |
     352 generated from handler       |
     353            |                 |
     354         handler              |
     355            |                 |
     356        throwResume1 -----.   :
     357            |             |   :
     358           try            |   : search skip
     359            |             |   :
     360        catchResume  <----'   :
     361            |                 |
     362\end{verbatim}
     363
     364The rules can be remembered as thinking about what would be searched in
     365termination. So when a throw happens in a handler; a termination handler
     366skips everything from the original throw to the original catch because that
     367part of the stack has been unwound, a resumption handler skips the same
     368section of stack because it has been masked.
     369A throw in a default handler will preform the same search as the original
     370throw because; for termination nothing has been unwound, for resumption
     371the mask will be the same.
     372
     373The symmetry with termination is why this pattern was picked. Other patterns,
     374such as marking just the handlers that caught, also work but lack the
     375symmetry whih means there is more to remember.
     376
     377\section{Conditional Catch}
     378Both termination and resumption handler clauses can be given an additional
     379condition to further control which exceptions they handle:
     380\begin{cfa}
    100381catch (EXCEPTION_TYPE * NAME ; CONDITION)
    101 \end{lstlisting}
    102 This has the same behaviour as a regular catch clause except that if the
    103 exception matches the given type the condition is also run. If the result is
    104 true only then is this considered a matching handler. If the result is false
    105 then the handler does not match and the search continues with the next clause
    106 in the try block.
    107 
    108 The condition considers all names in scope at the beginning of the try block
    109 to be in scope along with the name introduce in the catch clause itself.
    110 
    111 \paragraph{Re-Throwing}
    112 
    113 You can also rethrow the most recent termination exception with
    114 \codeCFA{throw;}. % This is terrible and you should never do it.
    115 This can be done in a handler or any function that could be called from a
    116 handler.
    117 
    118 This will start another termination throw reusing the exception, meaning it
    119 does not copy the exception or allocated any more memory for it. However the
    120 default handler is still at the original through and could refer to data that
    121 was on the unwound section of the stack. So instead a new default handler that
    122 does a program level abort is used.
    123 
    124 \section{Resumption}
    125 
    126 Resumption exceptions are less popular then termination but in many
    127 regards are simpler and easier to understand. A resumption throws an exception,
    128 searches for a handler on the stack, executes that handler on top of the stack
    129 and then continues execution from the throw. These are used when a problem
    130 needs to be fixed before execution continues.
    131 
    132 A resumption is thrown with a throw resume statement:
    133 \begin{lstlisting}
    134 throwResume EXPRESSION;
    135 \end{lstlisting}
    136 The result of EXPRESSION must be a resumption exception type. A resumption
    137 exception type is any type that satifies the assertion
    138 \codeCFA{void defaultResumptionHandler(T &);} (the default handler). When the
    139 statement is executed the expression is evaluated and the result is thrown.
    140 
    141 Handlers are declared using clauses in try statements:
    142 \begin{lstlisting}
    143 try {
    144     TRY_BLOCK
    145 } catchResume (EXCEPTION_TYPE * NAME) {
    146     HANDLER
    147 }
    148 \end{lstlisting}
    149 This is a simple example with the try block and a single resumption handler.
    150 Multiple resumption handlers can be put in a try statement and they can be
    151 mixed with termination handlers.
    152 
    153 When a resumption begins it will start searching the stack starting from
    154 the throw statement and working its way to the callers. In each try statement
    155 handlers will be tried top to bottom. Each handler is checked by seeing if
    156 the thrown exception is a descendant of EXCEPTION\_TYPE. If not the search
    157 continues. Otherwise NAME is bound to a pointer to the exception and the
    158 HANDLER statements are executed. After they are finished executing control
    159 continues from the throw statement.
    160 
    161 If no approprate handler is found then the default handler is called. The
    162 throw statement acts as a regular function call passing the exception to
    163 the default handler and after the handler finishes executing control continues
    164 from the throw statement.
    165 
    166 The exception system also tracks the position of a search on the stack. If
    167 another resumption exception is thrown while a resumption handler is running
    168 it will first check handlers pushed to the stack by the handler and any
    169 functions it called, then it will continue from the try statement that the
    170 handler is a part of; except for the default handler where it continues from
    171 the throw the default handler was passed to.
    172 
    173 This makes the search pattern for resumption reflect the one for termination,
    174 which is what most users expect.
    175 
    176 % This might need a diagram. But it is an important part of the justifaction
    177 % of the design of the traversal order.
    178 It also avoids the recursive resumption problem. If the entire stack is
    179 searched loops of resumption can form. Consider a handler that handles an
    180 exception of type A by resuming an exception of type B and on the same stack,
    181 later in the search path, is a second handler that handles B by resuming A.
    182 
    183 Assuming no other handlers on the stack handle A or B then in either traversal
    184 system an A resumed from the top of the stack will be handled by the first
    185 handler. A B resumed from the top or from the first handler it will be handled
    186 by the second hander. The only difference is when A is thrown from the second
    187 handler. The entire stack search will call the first handler again, creating a
    188 loop. Starting from the position in the stack though will break this loop.
    189 
    190 \paragraph{Conditional Catches}
    191 
    192 Resumption supports conditional catch clauses like termination does. They
    193 use the same syntax except the keyword is changed:
    194 \begin{lstlisting}
    195 catchResume (EXCEPTION_TYPE * NAME ; CONDITION) 
    196 \end{lstlisting}
    197 
    198 It also has the same behaviour, after the exception type has been matched
    199 with the EXCEPTION\_TYPE the CONDITION is evaluated with NAME in scope. If
    200 the result is true then the hander is run, otherwise the search continues
    201 just as if there had been a type mismatch.
    202 
    203 \paragraph{Re-Throwing}
    204 
    205 You may also re-throw resumptions with a \codeCFA{throwResume;} statement.
    206 This can only be done from inside of a \codeCFA{catchResume} block.
    207 
    208 Outside of any side effects of any code already run in the handler this will
    209 have the same effect as if the exception had not been caught in the first
    210 place.
     382\end{cfa}
     383First, the same semantics is used to match the exception type. Second, if the
     384exception matches, @CONDITION@ is executed. The condition expression may
     385reference all names in scope at the beginning of the try block and @NAME@
     386introduced in the handler clause. If the condition is true, then the handler
     387matches. Otherwise, the exception search continues as if the exception type
     388did not match.
     389\begin{cfa}
     390try {
     391        f1 = open( ... );
     392        f2 = open( ... );
     393        ...
     394} catch( IOFailure * f ; fd( f ) == f1 ) {
     395        // only handle IO failure for f1
     396}
     397\end{cfa}
     398Note, catching @IOFailure@, checking for @f1@ in the handler, and reraising the
     399exception if not @f1@ is different because the reraise does not examine any of
     400remaining handlers in the current try statement.
     401
     402\section{Rethrowing}
     403\colour{red}{From Andrew: I recomend we talk about why the language doesn't
     404have rethrows/reraises instead.}
     405
     406\label{s:Rethrowing}
     407Within the handler block or functions called from the handler block, it is
     408possible to reraise the most recently caught exception with @throw@ or
     409@throwResume@, respectively.
     410\begin{cfa}
     411try {
     412        ...
     413} catch( ... ) {
     414        ... throw;
     415} catchResume( ... ) {
     416        ... throwResume;
     417}
     418\end{cfa}
     419The only difference between a raise and a reraise is that reraise does not
     420create a new exception; instead it continues using the current exception, \ie
     421no allocation and copy. However the default handler is still set to the one
     422visible at the raise point, and hence, for termination could refer to data that
     423is part of an unwound stack frame. To prevent this problem, a new default
     424handler is generated that does a program-level abort.
    211425
    212426\section{Finally Clauses}
    213 
    214 A \codeCFA{finally} clause may be placed at the end of a try statement after
    215 all the handler clauses. In the simply case, with no handlers, it looks like
    216 this:
    217 
    218 \begin{lstlisting}
    219 try {
    220     TRY_BLOCK
    221 } finally {
    222     FINAL_STATEMENTS
    223 }
    224 \end{lstlisting}
    225 
    226 Any number of termination handlers and resumption handlers may proceed the
    227 finally clause.
    228 
    229 The FINAL\_STATEMENTS, the finally block, are executed whenever the try
    230 statement is removed from the stack. This includes: the TRY\_BLOCK finishes
    231 executing, a termination exception finishes executing and the stack unwinds.
    232 
    233 Execution of the finally block should finish by letting control run off
    234 the end of the block. This is because after the finally block is complete
    235 control will continue to where ever it would if the finally clause was not
    236 present.
    237 
    238 Because of this local control flow out of the finally block is forbidden.
    239 The compiler rejects uses of \codeCFA{break}, \codeCFA{continue},
    240 \codeCFA{fallthru} and \codeCFA{return} that would cause control to leave
    241 the finally block. Other ways to leave the finally block - such as a long
    242 jump or termination - are much harder to check, at best requiring additional
    243 runtime overhead, and so are merely discouraged.
     427Finally clauses are used to preform unconditional clean-up when leaving a
     428scope. They are placed at the end of a try statement:
     429\begin{cfa}
     430try {
     431        GUARDED_BLOCK
     432} ... // any number or kind of handler clauses
     433... finally {
     434        FINALLY_BLOCK
     435}
     436\end{cfa}
     437The @FINALLY_BLOCK@ is executed when the try statement is removed from the
     438stack, including when the @GUARDED_BLOCK@ finishes, any termination handler
     439finishes or during an unwind.
     440The only time the block is not executed is if the program is exited before
     441the stack is unwound.
     442
     443Execution of the finally block should always finish, meaning control runs off
     444the end of the block. This requirement ensures always continues as if the
     445finally clause is not present, \ie finally is for cleanup not changing control
     446flow. Because of this requirement, local control flow out of the finally block
     447is forbidden. The compiler precludes any @break@, @continue@, @fallthru@ or
     448@return@ that causes control to leave the finally block. Other ways to leave
     449the finally block, such as a long jump or termination are much harder to check,
     450and at best requiring additional run-time overhead, and so are mearly
     451discouraged.
     452
     453Not all languages with exceptions have finally clauses. Notably \Cpp does
     454without it as descructors serve a similar role. Although destructors and
     455finally clauses can be used in many of the same areas they have their own
     456use cases like top-level functions and lambda functions with closures.
     457Destructors take a bit more work to set up but are much easier to reuse while
     458finally clauses are good for once offs and can include local information.
    244459
    245460\section{Cancellation}
    246 
    247 Cancellation can be thought of as a stack-level abort or as an uncatchable
    248 termination. It unwinds the entirety of the current exception and if possible
    249 passes an exception to a different stack as a message.
    250 
    251 There is no special statement for starting a cancellation, instead you call
    252 the standard libary function \codeCFA{cancel\_stack} which takes an exception.
    253 Unlike in a throw this exception is not used in control flow but is just there
    254 to pass information about why the cancellation happened.
    255 
    256 The handler is decided entirely by which stack is being cancelled. There are
    257 three handlers that apply to three different groups of stacks:
    258 \begin{itemize}
    259 \item Main Stack:
    260 The main stack is the one on which the program main is called at the beginning
    261 of your program. It is also the only stack you have without the libcfathreads.
    262 
    263 Because of this there is no other stack ``above" (or possibly at all) for main
    264 to notify when a cancellation occurs. So after the stack is unwound we do a
    265 program level abort.
    266 
    267 \item Thread Stack:
    268 Thread stacks are those created \codeCFA{thread} or otherwise satify the
    269 \codeCFA{is\_thread} trait.
    270 
    271 Threads only have two structural points of communication that must happen,
    272 start and join. As the thread must be running to preform a cancellation it
    273 will be after start and before join, so join is one cancellation uses.
    274 
    275 After the stack is unwound the thread will halt as if had completed normally
    276 and wait for another thread to join with it. The other thread, when it joins,
    277 checks for a cancellation. If so it will throw the resumption exception
    278 \codeCFA{ThreadCancelled}.
    279 
    280 There is a difference here in how explicate joins (with the \codeCFA{join}
    281 function) and implicate joins (from a destructor call). Explicate joins will
    282 take the default handler (\codeCFA{defaultResumptionHandler}) from the context
    283 and use like a regular through does if the exception is not caught. The
    284 implicate join does a program abort instead.
    285 
    286 This is for safety. One of the big problems in exceptions is you cannot handle
    287 two terminations or cancellations on the same stack as either can destroy the
    288 context required for the other. This can happen with join but as the
    289 destructors will always be run when the stack is being unwound and one
    290 termination/cancellation is already active. Also since they are implicite they
    291 are easier to forget about.
    292 
    293 \item Coroutine Stack:
    294 Coroutine stacks are those created with \codeCFA{coroutine} or otherwise
    295 satify the \codeCFA{is\_coroutine} trait.
    296 
    297 A coroutine knows of two other coroutines, its starter and its last resumer.
    298 The last resumer is ``closer" so that is the one notified.
    299 
    300 After the stack is unwound control goes to the last resumer.
    301 Resume will resume throw a \codeCFA{CoroutineCancelled} exception, which is
    302 polymorphic over the coroutine type and has a pointer to the coroutine being
    303 cancelled and the cancelling exception. The resume function also has an
    304 assertion that the \codeCFA{defaultResumptionHandler} for the exception. So it
    305 will use the default handler like a regular throw.
    306 
    307 \end{itemize}
     461Cancellation is a stack-level abort, which can be thought of as as an
     462uncatchable termination. It unwinds the entirety of the current stack, and if
     463possible forwards the cancellation exception to a different stack.
     464
     465Cancellation is not an exception operation like termination or resumption.
     466There is no special statement for starting a cancellation; instead the standard
     467library function @cancel_stack@ is called passing an exception. Unlike a
     468throw, this exception is not used in matching only to pass information about
     469the cause of the cancellation.
     470(This also means matching cannot fail so there is no default handler either.)
     471
     472After @cancel_stack@ is called the exception is copied into the exception
     473handling mechanism's memory. Then the entirety of the current stack is
     474unwound. After that it depends one which stack is being cancelled.
     475\begin{description}
     476\item[Main Stack:]
     477The main stack is the one used by the program main at the start of execution,
     478and is the only stack in a sequential program. Even in a concurrent program
     479the main stack is only dependent on the environment that started the program.
     480Hence, when the main stack is cancelled there is nowhere else in the program
     481to notify. After the stack is unwound, there is a program-level abort.
     482
     483\item[Thread Stack:]
     484A thread stack is created for a @thread@ object or object that satisfies the
     485@is_thread@ trait. A thread only has two points of communication that must
     486happen: start and join. As the thread must be running to perform a
     487cancellation, it must occur after start and before join, so join is used
     488for communication here.
     489After the stack is unwound, the thread halts and waits for
     490another thread to join with it. The joining thread checks for a cancellation,
     491and if present, resumes exception @ThreadCancelled@.
     492
     493There is a subtle difference between the explicit join (@join@ function) and
     494implicit join (from a destructor call). The explicit join takes the default
     495handler (@defaultResumptionHandler@) from its calling context, which is used if
     496the exception is not caught. The implicit join does a program abort instead.
     497
     498This semantics is for safety. If an unwind is triggered while another unwind
     499is underway only one of them can proceed as they both want to ``consume'' the
     500stack. Letting both try to proceed leads to very undefined behaviour.
     501Both termination and cancellation involve unwinding and, since the default
     502@defaultResumptionHandler@ preforms a termination that could more easily
     503happen in an implicate join inside a destructor. So there is an error message
     504and an abort instead.
     505\todo{Perhaps have a more general disucssion of unwind collisions before
     506this point.}
     507
     508The recommended way to avoid the abort is to handle the intial resumption
     509from the implicate join. If required you may put an explicate join inside a
     510finally clause to disable the check and use the local
     511@defaultResumptionHandler@ instead.
     512
     513\item[Coroutine Stack:] A coroutine stack is created for a @coroutine@ object
     514or object that satisfies the @is_coroutine@ trait. A coroutine only knows of
     515two other coroutines, its starter and its last resumer. Of the two the last
     516resumer has the tightest coupling to the coroutine it activated and the most
     517up-to-date information.
     518
     519Hence, cancellation of the active coroutine is forwarded to the last resumer
     520after the stack is unwound. When the resumer restarts, it resumes exception
     521@CoroutineCancelled@, which is polymorphic over the coroutine type and has a
     522pointer to the cancelled coroutine.
     523
     524The resume function also has an assertion that the @defaultResumptionHandler@
     525for the exception. So it will use the default handler like a regular throw.
     526\end{description}
  • doc/theses/andrew_beach_MMath/future.tex

    r342af53 r8e4aa05  
    11\chapter{Future Work}
    22
     3\section{Language Improvements}
     4\CFA is a developing programming language. As such, there are partially or
     5unimplemented features of the language (including several broken components)
     6that I had to workaround while building an exception handling system largely in
     7the \CFA language (some C components).  The following are a few of these
     8issues, and once implemented/fixed, how this would affect the exception system.
     9\begin{itemize}
     10\item
     11The implementation of termination is not portable because it includes
     12hand-crafted assembly statements. These sections must be ported by hand to
     13support more hardware architectures, such as the ARM processor.
     14\item
     15Due to a type-system problem, the catch clause cannot bind the exception to a
     16reference instead of a pointer. Since \CFA has a very general reference
     17capability, programmers will want to use it. Once fixed, this capability should
     18result in little or no change in the exception system.
     19\item
     20Termination handlers cannot use local control-flow transfers, \eg by @break@,
     21@return@, \etc. The reason is that current code generation hoists a handler
     22into a nested function for convenience (versus assemble-code generation at the
     23@try@ statement). Hence, when the handler runs, its code is not in the lexical
     24scope of the @try@ statement, where the local control-flow transfers are
     25meaningful.
     26\item
     27There is no detection of colliding unwinds. It is possible for clean-up code
     28run during an unwind to trigger another unwind that escapes the clean-up code
     29itself; such as a termination exception caught further down the stack or a
     30cancellation. There do exist ways to handle this but currently they are not
     31even detected and the first unwind will simply be forgotten, often leaving
     32it in a bad state.
     33\item
     34Also the exception system did not have a lot of time to be tried and tested.
     35So just letting people use the exception system more will reveal new
     36quality of life upgrades that can be made with time.
     37\end{itemize}
     38
    339\section{Complete Virtual System}
    4 The virtual system should be completed. It was never supposed to be part of
    5 this project and so minimal work was done on it. A draft of what the complete
    6 system might look like was created but it was never finalized or implemented.
    7 A future project in \CFA would be to complete that work and to update the
    8 parts of the exception system that use the current version.
     40The virtual system should be completed. It was not supposed to be part of this
     41project, but was thrust upon it to do exception inheritance; hence, only
     42minimal work was done. A draft for a complete virtual system is available but
     43it is not finalized.  A future \CFA project is to complete that work and then
     44update the exception system that uses the current version.
    945
    10 For instance a full virtual system would probably allow for several
    11 improvements to the exception traits. Although they do currently work they
    12 could be made easier to use by making the virtual table type implitate in the
    13 trait (which would remove the need for those wrapper marcos) or allowing
    14 for assertions that give the layout of a virtual table for safety.
     46There are several improvements to the virtual system that would improve the
     47exception traits. The most important one is an assertion to check one virtual
     48type is a child of another. This check precisely captures many of the
     49correctness requirements.
    1550
    16 \section{Additional Throws}
    17 Several other kinds of throws, beyond the termination throw (\codeCFA{throw}),
    18 the resumption throw (\codeCFA{throwResume}) and the re-throws, were considered.
    19 None were as useful as the core throws but they would likely be worth
    20 revising.
     51The full virtual system might also include other improvement like associated
     52types to allow traits to refer to types not listed in their header. This
     53feature allows exception traits to not refer to the virtual-table type
     54explicitly, removing the need for the current interface macros.
    2155
    22 The first ones are throws for asynchronous exceptions, throwing exceptions
    23 from one stack to another. These act like signals allowing for communication
    24 between the stacks. This is usually used with resumption as it allows the
    25 target stack to continue execution normally after the exception has been
    26 handled.
     56\section{Additional Raises}
     57Several other kinds of exception raises were considered beyond termination
     58(@throw@), resumption (@throwResume@), and reraise.
    2759
    28 This would much more coordination between the concurrency system and the
    29 exception system to handle. Most of the interesting design decisions around
    30 applying asynchronous exceptions appear to be around masking (controlling
    31 which exceptions may be thrown at a stack). It would likely require more of
    32 the virtual system and would also effect how default handlers are set.
     60The first is a non-local/concurrent raise providing asynchronous exceptions,
     61\ie raising an exception on another stack. This semantics acts like signals
     62allowing for out-of-band communication among coroutines and threads. This kind
     63of raise is often restricted to resumption to allow the target stack to
     64continue execution normally after the exception has been handled. That is,
     65allowing one coroutine/thread to unwind the stack of another via termination is
     66bad software engineering.
    3367
    34 The other throws were designed to mimic bidirectional algebraic effects.
    35 Algebraic effects are used in some functional languages and allow a function
     68Non-local/concurrent requires more coordination between the concurrency system
     69and the exception system. Many of the interesting design decisions centre
     70around masking (controlling which exceptions may be thrown at a stack). It
     71would likely require more of the virtual system and would also effect how
     72default handlers are set.
     73
     74Other raises were considered to mimic bidirectional algebraic effects.
     75Algebraic effects are used in some functional languages allowing one function
    3676to have another function on the stack resolve an effect (which is defined with
    37 a function-like interface).
    38 These can be mimiced with resumptions and the the new throws were designed
    39 to try and mimic bidirectional algebraic effects, where control can go back
    40 and forth between the function effect caller and handler while the effect
    41 is underway.
     77a functional-like interface).  This semantics can be mimicked with resumptions
     78and new raises were discussed to mimic bidirectional algebraic-effects, where
     79control can go back and forth between the function-effect caller and handler
     80while the effect is underway.
    4281% resume-top & resume-reply
     82These raises would be like the resumption raise except using different search
     83patterns to find the handler.
    4384
    44 These throws would likely be just like the resumption throw except they would
    45 use different search patterns to find the handler to reply to.
     85\section{Zero-Cost Try}
     86\CFA does not have zero-cost try-statements because the compiler generates C
     87code rather than assembler code \see{\VPageref{p:zero-cost}}. When the compiler
     88does create its own assembly (or LLVM byte-code), then zero-cost try-statements
     89are possible. The downside of zero-cost try-statements is the LSDA complexity,
     90its size (program bloat), and the high cost of raising an exception.
    4691
    47 \section{Zero-Cost Exceptions}
    48 \CFA does not have zero-cost exceptions because it does not generate assembly
    49 but instead generates C code. See the implementation section. When the
    50 compiler does start to create its own assembly (or LLVM byte code) then
    51 zero-cost exceptions could be implemented.
     92Alternatively, some research could be done into the simpler alternative method
     93with a non-zero-cost try-statement but much lower cost exception raise. For
     94example, programs are starting to use exception in the normal control path, so
     95more exceptions are thrown. In these cases, the cost balance switches towards
     96low-cost raise. Unfortunately, while exceptions remain exceptional, the
     97libunwind model will probably remain the most effective option.
    5298
    53 Now in zero-cost exceptions the only part that is zero-cost are the try
    54 blocks. Some research could be done into the alternative methods for systems
    55 that expect a lot more exceptions to be thrown, allowing some overhead in
    56 entering and leaving try blocks to make throws faster. But while exceptions
    57 remain exceptional the libunwind model will probably remain the most effective
    58 option.
     99Zero-cost resumptions is still an open problem. First, because libunwind does
     100not support a successful-exiting stack-search without doing an unwind.
     101Workarounds are possible but awkward. Ideally an extension to libunwind could
     102be made, but that would either require separate maintenance or gain enough
     103support to have it folded into the standard.
    59104
    60 Zero-cost resumptions have more problems to solve. First because libunwind
    61 does not support a successful exiting stack search without doing an unwind.
    62 There are several ways to hack that functionality in. Ideally an extension to
    63 libunwind could be made, but that would either require seperate maintenance
    64 or gain enough support to have it folded into the standard.
     105Also new techniques to skip previously searched parts of the stack need to be
     106developed to handle the recursive resume problem and support advanced algebraic
     107effects.
    65108
    66 Also new techniques to skip previously searched parts of the stack will have
    67 to be developed.
     109\section{Signal Exceptions}
     110Goodenough~\cite{Goodenough75} suggests three types of exceptions: escape,
     111notify and signal.  Escape are termination exceptions, notify are resumption
     112exceptions, leaving signal unimplemented.
    68113
    69 \section{Support for More Platforms}
    70 Termination is not portable because it is implemented with inline assembly.
    71 Those sections will have to be rewritten to support different architectures
     114A signal exception allows either behaviour, \ie after an exception is handled,
     115the handler has the option of returning to the raise or after the @try@
     116statement. Currently, \CFA fixes the semantics of the handler return
     117syntactically by the @catch@ or @catchResume@ clause.
    72118
    73 \section{Quality-of-Life Improvements}
    74 Finally come various improvements to the usability of \CFA. Most of these
    75 would just require time. Time that would not lead to interesting research so
    76 it has been left aside for now. A few examples are included here but there
    77 are more:
     119Signal exception should be reexamined and possibly be supported in \CFA. A very
     120direct translation is to have a new raise and catch pair, and a new statement
     121(or statements) would indicate if the handler returns to the raise or continues
     122where it is; but there may be other options.
    78123
    79 \begin{itemize}
    80 \item Allowing exception handler to bind the exception to a reference instead
    81 of a pointer. This should actually result in no change in behaviour so there
    82 is no reason not to allow it. It is however a small improvement; giving a bit
    83 of flexibility to the user in what style they want to use.
    84 \item Enabling local control flow (by \codeCFA{break}, \codeCFA{return} and
    85 similar statements) out of a termination handler. The current set-up makes
    86 this very difficult but the catch function that runs the handler after it has
    87 been matched could be inlined into the function's body, which would make this
    88 much easier. (To do the same for try blocks would probably wait for zero-cost
    89 exceptions, which would allow the try block to be inlined as well.)
    90 \item Enabling local control flow out of a resumption handler. This would be
    91 a weighty operation, causing a stack unwind like a termination, so there might
    92 be a different statement or a statement modifier to make sure the user does
    93 this purposefully.
     124For instance, resumption could be extended to cover this use by allowing local
     125control flow out of it. This approach would require an unwind as part of the
     126transition as there are stack frames that have to be removed.  This approach
     127means there is no notify raise, but because \CFA does not have exception
     128signatures, a termination can be thrown from within any resumption handler so
     129there is already a way to do mimic this in existing \CFA.
    94130
    95 However this would require the more complex system as they cannot be inlined
    96 into the original function as they can be run at a different place on the
    97 stack. So instead the unwinding will have to carry with it information on
    98 which one of these points to continue at and possibly also the return value
    99 for the function if a \codeCFA{return} statement was used.
    100 \end{itemize}
     131% Maybe talk about the escape; and escape CONTROL_STMT; statements or how
     132% if we could choose if _Unwind_Resume proceeded to the clean-up stage this
     133% would be much easier to implement.
  • doc/theses/andrew_beach_MMath/implement.tex

    r342af53 r8e4aa05  
    22% Goes over how all the features are implemented.
    33
     4The implementation work for this thesis covers two components: the virtual
     5system and exceptions. Each component is discussed in detail.
     6
    47\section{Virtual System}
     8\label{s:VirtualSystem}
    59% Virtual table rules. Virtual tables, the pointer to them and the cast.
    6 The \CFA virtual system only has one public facing feature: virtual casts.
    7 However there is a lot of structure to support that and provide some other
    8 features for the standard library.
    9 
    10 All of this is accessed through a field inserted at the beginning of every
    11 virtual type. Currently it is called \codeC{virtual_table} but it is not
    12 ment to be accessed by the user. This field is a pointer to the type's
    13 virtual table instance. It is assigned once during the object's construction
    14 and left alone after that.
    15 
    16 \subsection{Virtual Table Construction}
    17 For each virtual type a virtual table is constructed. This is both a new type
    18 and an instance of that type. Other instances of the type could be created
    19 but the system doesn't use them. So this section will go over the creation of
    20 the type and the instance.
    21 
    22 Creating the single instance is actually very important. The address of the
    23 table acts as the unique identifier for the virtual type. Similarly the first
    24 field in every virtual table is the parent's id; a pointer to the parent
    25 virtual table instance.
    26 
    27 The remaining fields contain the type's virtual members. First come the ones
    28 present on the parent type, in the same order as they were the parent, and
    29 then any that this type introduces. The types of the ones inherited from the
    30 parent may have a slightly modified type, in that references to the
    31 dispatched type are replaced with the current virtual type. These are always
    32 taken by pointer or reference.
    33 
    34 The structure itself is created where the virtual type is created. The name
    35 of the type is created by mangling the name of the base type. The name of the
    36 instance is also generated by name mangling.
    37 
    38 The fields are initialized automatically.
     10While the \CFA virtual system currently has only one public feature, virtual
     11cast \see{\VPageref{p:VirtualCast}}, substantial structure is required to
     12support it, and provide features for exception handling and the standard
     13library.
     14
     15\subsection{Virtual Type}
     16Virtual types only have one change to their structure, the addition of a
     17pointer to the virtual table. This is always the first field so that
     18if it is cast to a supertype the field's location is still known.
     19
     20This field is set as part of all new generated constructors.
     21\todo{They only come as part exceptions and don't work.}
     22After the object is created the field is constant.
     23
     24However it can be read from, internally it is just a regular field called
     25@virtual_table@. Dereferencing it gives the virtual table and access to the
     26type's virtual members.
     27
     28\subsection{Virtual Table}
     29Every time a virtual type is defined the new virtual table type must also be
     30defined.
     31
     32The unique instance is important because the address of the virtual table
     33instance is used as the identifier for the virtual type. So a pointer to the
     34virtual table and the ID for the virtual type are interchangable.
     35\todo{Unique instances might be going so we will have to talk about the new
     36system instead.}
     37
     38The first step in putting it all together is to create the virtual table type.
     39The virtual table type is just a structure and can be described in terms of
     40its fields. The first field is always the parent type ID (or a pointer to
     41the parent virtual table) or 0 (the null pointer).
     42Next are other fields on the parent virtual table are repeated.
     43Finally are the fields used to store any new virtual members of the new
     44The virtual type
     45
     46The virtual system is accessed through a private constant field inserted at the
     47beginning of every virtual type, called the virtual-table pointer. This field
     48points at a type's virtual table and is assigned during the object's
     49construction. The address of a virtual table acts as the unique identifier for
     50the virtual type, and the first field of a virtual table is a pointer to the
     51parent virtual-table or @0p@. The remaining fields are duplicated from the
     52parent tables in this type's inheritance chain, followed by any fields this type
     53introduces. Parent fields are duplicated so they can be changed (all virtual
     54members are overridable), so that references to the dispatched type
     55are replaced with the current virtual type.
     56% These are always taken by pointer or reference.
     57
     58% Simple ascii diragram:
     59\begin{verbatim}
     60parent_pointer  \
     61parent_field0   |
     62...             | Same layout as parent.
     63parent_fieldN   /
     64child_field0
     65...
     66child_fieldN
     67\end{verbatim}
     68\todo{Refine the diagram}
     69
     70% For each virtual type, a virtual table is constructed. This is both a new type
     71% and an instance of that type. Other instances of the type could be created
     72% but the system doesn't use them. So this section will go over the creation of
     73% the type and the instance.
     74
     75A virtual table is created when the virtual type is created. The name of the
     76type is created by mangling the name of the base type. The name of the instance
     77is also generated by name mangling. The fields are initialized automatically.
    3978The parent field is initialized by getting the type of the parent field and
    4079using that to calculate the mangled name of the parent's virtual table type.
    4180There are two special fields that are included like normal fields but have
    42 special initialization rules: the \codeC{size} field is the type's size and is
    43 initialized with a sizeof expression, the \codeC{align} field is the type's
    44 alignment and uses an alignof expression. The remaining fields are resolved
    45 to a name matching the field's name and type using the normal visibility
    46 and overload resolution rules of the type system.
    47 
    48 These operations are split up into several groups depending on where they
    49 take place which can vary for monomorphic and polymorphic types. The first
    50 devision is between the declarations and the definitions. Declarations, such
    51 as a function signature or a structure's name, must always be visible but may
    52 be repeated so they go in headers. Definitions, such as function bodies and a
    53 structure's layout, don't have to be visible on use but must occur exactly
    54 once and go into source files.
    55 
     81special initialization rules: the @size@ field is the type's size and is
     82initialized with a @sizeof@ expression, the @align@ field is the type's
     83alignment and uses an @alignof@ expression. The remaining fields are resolved
     84to a name matching the field's name and type using the normal visibility and
     85overload resolution rules of the type system.
     86
     87These operations are split up into several groups depending on where they take
     88place which varies for monomorphic and polymorphic types. The first devision is
     89between the declarations and the definitions. Declarations, such as a function
     90signature or a aggregate's name, must always be visible but may be repeated in
     91the form of forward declarations in headers. Definitions, such as function
     92bodies and a aggregate's layout, can be separately compiled but must occur
     93exactly once in a source file.
     94
     95\begin{sloppypar}
    5696The declarations include the virtual type definition and forward declarations
    5797of the virtual table instance, constructor, message function and
    58 \codeCFA{get_exception_vtable}. The definition includes the storage and
    59 initialization of the virtual table instance and the bodies of the three
    60 functions.
     98@get_exception_vtable@. The definition includes the storage and initialization
     99of the virtual table instance and the bodies of the three functions.
     100\end{sloppypar}
    61101
    62102Monomorphic instances put all of these two groups in one place each.
    63 
    64 Polymorphic instances also split out the core declarations and definitions
    65 from the per-instance information. The virtual table type and most of the
    66 functions are polymorphic so they are all part of the core. The virtual table
    67 instance and the \codeCFA{get_exception_vtable} function.
    68 
    69 Coroutines and threads need instances of \codeCFA{CoroutineCancelled} and
    70 \codeCFA{ThreadCancelled} respectively to use all of their functionality.
    71 When a new data type is declared with \codeCFA{coroutine} or \codeCFA{thread}
    72 the forward declaration for the instance is created as well. The definition
    73 of the virtual table is created at the definition of the main function.
     103Polymorphic instances also split out the core declarations and definitions from
     104the per-instance information. The virtual table type and most of the functions
     105are polymorphic so they are all part of the core. The virtual table instance
     106and the @get_exception_vtable@ function.
     107
     108\begin{sloppypar}
     109Coroutines and threads need instances of @CoroutineCancelled@ and
     110@ThreadCancelled@ respectively to use all of their functionality. When a new
     111data type is declared with @coroutine@ or @thread@ the forward declaration for
     112the instance is created as well. The definition of the virtual table is created
     113at the definition of the main function.
     114\end{sloppypar}
    74115
    75116\subsection{Virtual Cast}
    76 Virtual casts are implemented as a function call that does the check and a
    77 old C-style cast to do the type conversion. The C-cast is just to make sure
    78 the generated code is correct so the rest of the section is about that
    79 function.
    80 
    81 The function is \codeC{__cfa__virtual_cast} and it is implemented in the
    82 standard library. It takes a pointer to the target type's virtual table and
    83 the object pointer being cast. The function is very simple, getting the
    84 object's virtual table pointer and then checking to see if it or any of
    85 its ancestors, by using the parent pointers, are the same as the target type
    86 virtual table pointer. It does this in a simple loop.
    87 
    88 For the generated code a forward decaration of the virtual works as follows.
    89 There is a forward declaration of \codeC{__cfa__virtual_cast} in every cfa
    90 file so it can just be used. The object argument is the expression being cast
    91 so that is just placed in the argument list.
    92 
    93 To build the target type parameter the compiler will create a mapping from
    94 concrete type-name -- so for polymorphic types the parameters are filled in
    95 -- to virtual table address. Every virtual table declaraction is added to the
    96 this table; repeats are ignored unless they have conflicting definitions.
    97 This does mean the declaractions have to be in scope, but they should usually
    98 be introduced as part of the type definition.
     117Virtual casts are implemented as a function call that does the subtype check
     118and a C coercion-cast to do the type conversion.
     119% The C-cast is just to make sure the generated code is correct so the rest of
     120% the section is about that function.
     121The function is
     122\begin{cfa}
     123void * __cfa__virtual_cast(
     124        struct __cfa__parent_vtable const * parent,
     125        struct __cfa__parent_vtable const * const * child );
     126\end{cfa}
     127and it is implemented in the standard library. The structure reperents the
     128head of a vtable which is the pointer to the parent virtual table. The
     129@parent@ points directly at the parent type virtual table while the @child@
     130points at the object of the (possibe) child type.
     131
     132In terms of the virtual cast expression, @parent@ comes from looking up the
     133type being cast to and @child@ is the result of the expression being cast.
     134Because the complier outputs C code, some type C type casts are also used.
     135The last bit of glue is an map that saves every virtual type the compiler
     136sees. This is used to check the type used in a virtual cast is a virtual
     137type and to get its virtual table.
     138(It also checks for conflicting definitions.)
     139
     140Inside the function it is a simple conditional. If the type repersented by
     141@parent@ is or is an ancestor of the type repersented by @*child@ (it
     142requires one more level of derefence to pass through the object) then @child@
     143is returned, otherwise the null pointer is returned.
     144
     145The check itself is preformed is a simple linear search. If the child
     146virtual table or any of its ancestors (which are retreved through the first
     147field of every virtual table) are the same as the parent virtual table then
     148the cast succeeds.
    99149
    100150\section{Exceptions}
     
    106156% resumption doesn't as well.
    107157
    108 Many modern languages work with an interal stack that function push and pop
    109 their local data to. Stack unwinding removes large sections of the stack,
    110 often across functions.
    111 
    112 At a very basic level this can be done with \codeC{setjmp} \& \codeC{longjmp}
    113 which simply move the top of the stack, discarding everything on the stack
    114 above a certain point. However this ignores all the clean-up code that should
    115 be run when certain sections of the stack are removed (for \CFA these are from
    116 destructors and finally clauses) and also requires that the point to which the
    117 stack is being unwound is known ahead of time. libunwind is used to address
    118 both of these problems.
    119 
    120 Libunwind, provided in \texttt{unwind.h} on most platorms, is a C library
    121 that provides \CPP style stack unwinding. Its operation is divided into two
    122 phases. The search phase -- phase 1 -- is used to scan the stack and decide
    123 where the unwinding will stop, this allows for a dynamic target. The clean-up
    124 phase -- phase 2 -- does the actual unwinding and also runs any clean-up code
    125 as it goes.
    126 
    127 To use the libunwind each function must have a personality function and an
    128 LSDA (Language Specific Data Area). Libunwind actually does very little, it
    129 simply moves down the stack from function to function. Most of the actions are
    130 implemented by the personality function which libunwind calls on every
    131 function. Since this is shared across many functions or even every function in
    132 a language it will need a bit more information. This is provided by the LSDA
    133 which has the unique information for each function.
    134 
    135 Theoretically the LSDA can contain anything but conventionally it is a table
    136 with entries reperenting areas of the function and what has to be done there
    137 during unwinding. These areas are described in terms of where the instruction
    138 pointer is. If the current value of the instruction pointer is between two
    139 values reperenting the beginning and end of a region then execution is
    140 currently being executed. These are used to mark out try blocks and the
    141 scopes of objects with destructors to run.
    142 
    143 GCC will generate an LSDA and attach its personality function with the
    144 \texttt{-fexceptions} flag. However this only handles the cleanup attribute.
    145 This attribute is used on a variable and specifies a function that should be
    146 run when the variable goes out of scope. The function is passed a pointer to
    147 the object as well so it can be used to mimic destructors. It however cannot
    148 be used to mimic try statements.
    149 
    150 \subsection{Implementing Personality Functions}
    151 Personality functions have a complex interface specified by libunwind.
    152 This section will cover some of the important parts of that interface.
    153 
    154 \begin{lstlisting}
    155 typedef _Unwind_Reason_Code (*_Unwind_Personality_Fn)(
    156     int version,
    157     _Unwind_Action action,
    158     _Unwind_Exception_Class exception_class,
    159     _Unwind_Exception * exception,
    160     struct _Unwind_Context * context);
     158% Many modern languages work with an interal stack that function push and pop
     159% their local data to. Stack unwinding removes large sections of the stack,
     160% often across functions.
     161
     162Stack unwinding is the process of removing stack frames (activations) from the
     163stack. On function entry and return, unwinding is handled directly by the code
     164embedded in the function. Usually, the stack-frame size is known statically
     165based on parameter and local variable declarations. For dynamically-sized
     166local variables, a runtime computation is necessary to know the frame
     167size. Finally, a function's frame-size may change during execution as local
     168variables (static or dynamic sized) go in and out of scope.
     169Allocating/deallocating stack space is usually an $O(1)$ operation achieved by
     170bumping the hardware stack-pointer up or down as needed.
     171
     172Unwinding across multiple stack frames is more complex because individual stack
     173management code associated with each frame is bypassed. That is, the location
     174of a function's frame-management code is largely unknown and dispersed
     175throughout the function, hence the current frame size managed by that code is
     176also unknown. Hence, code unwinding across frames does not have direct
     177knowledge about what is on the stack, and hence, how much of the stack needs to
     178be removed.
     179
     180% At a very basic level this can be done with @setjmp@ \& @longjmp@ which simply
     181% move the top of the stack, discarding everything on the stack above a certain
     182% point. However this ignores all the cleanup code that should be run when
     183% certain sections of the stack are removed (for \CFA these are from destructors
     184% and finally clauses) and also requires that the point to which the stack is
     185% being unwound is known ahead of time. libunwind is used to address both of
     186% these problems.
     187
     188The traditional unwinding mechanism for C is implemented by saving a snap-shot
     189of a function's state with @setjmp@ and restoring that snap-shot with
     190@longjmp@. This approach bypasses the need to know stack details by simply
     191reseting to a snap-shot of an arbitrary but existing function frame on the
     192stack. It is up to the programmer to ensure the snap-shot is valid when it is
     193reset, making this unwinding approach fragile with potential errors that are
     194difficult to debug because the stack becomes corrupted.
     195
     196However, many languages define cleanup actions that must be taken when objects
     197are deallocated from the stack or blocks end, such as running a variable's
     198destructor or a @try@ statement's @finally@ clause. Handling these mechanisms
     199requires walking the stack and checking each stack frame for these potential
     200actions.
     201
     202For exceptions, it must be possible to walk the stack frames in search of @try@
     203statements to match and execute a handler. For termination exceptions, it must
     204also be possible to unwind all stack frames from the throw to the matching
     205catch, and each of these frames must be checked for cleanup actions. Stack
     206walking is where most of the complexity and expense of exception handling
     207appears.
     208
     209One of the most popular tools for stack management is libunwind, a low-level
     210library that provides tools for stack walking, handler execution, and
     211unwinding. What follows is an overview of all the relevant features of
     212libunwind needed for this work, and how \CFA uses them to implement exception
     213handling.
     214
     215\subsection{libunwind Usage}
     216Libunwind, accessed through @unwind.h@ on most platforms, is a C library that
     217provides \CC-style stack-unwinding. Its operation is divided into two phases:
     218search and cleanup. The dynamic target search -- phase 1 -- is used to scan the
     219stack and decide where unwinding should stop (but no unwinding occurs). The
     220cleanup -- phase 2 -- does the unwinding and also runs any cleanup code.
     221
     222To use libunwind, each function must have a personality function and a Language
     223Specific Data Area (LSDA). The LSDA has the unique information for each
     224function to tell the personality function where a function is executing, its
     225current stack frame, and what handlers should be checked. Theoretically, the
     226LSDA can contain any information but conventionally it is a table with entries
     227representing regions of the function and what has to be done there during
     228unwinding. These regions are bracketed by the instruction pointer. If the
     229instruction pointer is within a region's start/end, then execution is currently
     230executing in that region. Regions are used to mark out the scopes of objects
     231with destructors and try blocks.
     232
     233% Libunwind actually does very little, it simply moves down the stack from
     234% function to function. Most of the actions are implemented by the personality
     235% function which libunwind calls on every function. Since this is shared across
     236% many functions or even every function in a language it will need a bit more
     237% information.
     238
     239The GCC compilation flag @-fexceptions@ causes the generation of an LSDA and
     240attaches its personality function. However, this
     241flag only handles the cleanup attribute:
     242\todo{Peter: What is attached? Andrew: It uses the .cfi\_personality directive
     243and that's all I know.}
     244\begin{cfa}
     245void clean_up( int * var ) { ... }
     246int avar __attribute__(( cleanup(clean_up) ));
     247\end{cfa}
     248which is used on a variable and specifies a function, in this case @clean_up@,
     249run when the variable goes out of scope.
     250The function is passed a pointer to the object being removed from the stack
     251so it can be used to mimic destructors.
     252However, this feature cannot be used to mimic @try@ statements as it cannot
     253control the unwinding.
     254
     255\subsection{Personality Functions}
     256Personality functions have a complex interface specified by libunwind. This
     257section covers some of the important parts of the interface.
     258
     259A personality function can preform different actions depending on how it is
     260called.
     261\begin{lstlisting}[language=C,{moredelim=**[is][\color{red}]{@}{@}}]
     262typedef _Unwind_Reason_Code (*@_Unwind_Personality_Fn@) (
     263        _Unwind_Action @action@,
     264        _Unwind_Exception_Class @exception_class@,
     265        _Unwind_Exception * @exception@,
     266        struct _Unwind_Context * @context@
     267);
    161268\end{lstlisting}
    162 
    163 The return value, the reason code, is an enumeration of possible messages
     269The @action@ argument is a bitmask of possible actions:
     270\begin{enumerate}
     271\item
     272@_UA_SEARCH_PHASE@ specifies a search phase and tells the personality function
     273to check for handlers. If there is a handler in a stack frame, as defined by
     274the language, the personality function returns @_URC_HANDLER_FOUND@; otherwise
     275it return @_URC_CONTINUE_UNWIND@.
     276
     277\item
     278@_UA_CLEANUP_PHASE@ specifies a cleanup phase, where the entire frame is
     279unwound and all cleanup code is run. The personality function does whatever
     280cleanup the language defines (such as running destructors/finalizers) and then
     281generally returns @_URC_CONTINUE_UNWIND@.
     282
     283\item
     284\begin{sloppypar}
     285@_UA_HANDLER_FRAME@ specifies a cleanup phase on a function frame that found a
     286handler. The personality function must prepare to return to normal code
     287execution and return @_URC_INSTALL_CONTEXT@.
     288\end{sloppypar}
     289
     290\item
     291@_UA_FORCE_UNWIND@ specifies a forced unwind call. Forced unwind only performs
     292the cleanup phase and uses a different means to decide when to stop
     293\see{\VRef{s:ForcedUnwind}}.
     294\end{enumerate}
     295
     296The @exception_class@ argument is a copy of the
     297\lstinline[language=C]|exception|'s @exception_class@ field.
     298
     299The \lstinline[language=C]|exception| argument is a pointer to the user
     300provided storage object. It has two public fields, the exception class, which
     301is actually just a number, identifying the exception handling mechanism that
     302created it, and the cleanup function. The cleanup function is called if
     303required by the exception.
     304
     305The @context@ argument is a pointer to an opaque type passed to helper
     306functions called inside the personality function.
     307
     308The return value, @_Unwind_Reason_Code@, is an enumeration of possible messages
    164309that can be passed several places in libunwind. It includes a number of
    165310messages for special cases (some of which should never be used by the
    166311personality function) and error codes but unless otherwise noted the
    167 personality function should always return \codeC{_URC_CONTINUE_UNWIND}.
    168 
    169 The \codeC{version} argument is the verson of the implementation that is
    170 calling the personality function. At this point it appears to always be 1 and
    171 it will likely stay that way until a new version of the API is updated.
    172 
    173 The \codeC{action} argument is set of flags that tell the personality
    174 function when it is being called and what it must do on this invocation.
    175 The flags are as follows:
    176 \begin{itemize}
    177 \item\codeC{_UA_SEARCH_PHASE}: This flag is set whenever the personality
    178 function is called during the search phase. The personality function should
    179 decide if unwinding will stop in this function or not. If it does then the
    180 personality function should return \codeC{_URC_HANDLER_FOUND}.
    181 \item\codeC{_UA_CLEANUP_PHASE}: This flag is set whenever the personality
    182 function is called during the cleanup phase. If no other flags are set this
    183 means the entire frame will be unwound and all cleanup code should be run.
    184 \item\codeC{_UA_HANDLER_FRAME}: This flag is set during the cleanup phase
    185 on the function frame that found the handler. The personality function must
    186 prepare to return to normal code execution and return
    187 \codeC{_URC_INSTALL_CONTEXT}.
    188 \item\codeC{_UA_FORCE_UNWIND}: This flag is set if the personality function
    189 is called through a forced unwind call. Forced unwind only performs the
    190 cleanup phase and uses a different means to decide when to stop. See its
    191 section below.
    192 \end{itemize}
    193 
    194 The \codeC{exception_class} argument is a copy of the \codeC{exception}'s
    195 \codeC{exception_class} field.
    196 
    197 The \codeC{exception} argument is a pointer to the user provided storage
    198 object. It has two public fields, the exception class which is actually just
    199 a number that identifies the exception handling mechanism that created it and
    200 the other is the clean-up function. The clean-up function is called if the
    201 exception needs to
    202 
    203 The \codeC{context} argument is a pointer to an opaque type. This is passed
    204 to the many helper functions that can be called inside the personality
    205 function.
     312personality function should always return @_URC_CONTINUE_UNWIND@.
    206313
    207314\subsection{Raise Exception}
    208 This could be considered the central function of libunwind. It preforms the
    209 two staged unwinding the library is built around and most of the rest of the
    210 interface of libunwind is here to support it. It's signature is as follows:
    211 
    212 \begin{lstlisting}
     315Raising an exception is the central function of libunwind and it performs a
     316two-staged unwinding.
     317\begin{cfa}
    213318_Unwind_Reason_Code _Unwind_RaiseException(_Unwind_Exception *);
     319\end{cfa}
     320First, the function begins the search phase, calling the personality function
     321of the most recent stack frame. It continues to call personality functions
     322traversing the stack from newest to oldest until a function finds a handler or
     323the end of the stack is reached. In the latter case, raise exception returns
     324@_URC_END_OF_STACK@.
     325
     326Second, when a handler is matched, raise exception continues onto the cleanup
     327phase.
     328Once again, it calls the personality functions of each stack frame from newest
     329to oldest. This pass stops at the stack frame containing the matching handler.
     330If that personality function has not install a handler, it is an error.
     331
     332If an error is encountered, raise exception returns either
     333@_URC_FATAL_PHASE1_ERROR@ or @_URC_FATAL_PHASE2_ERROR@ depending on when the
     334error occurred.
     335
     336\subsection{Forced Unwind}
     337\label{s:ForcedUnwind}
     338Forced Unwind is the other central function in libunwind.
     339\begin{cfa}
     340_Unwind_Reason_Code _Unwind_ForcedUnwind( _Unwind_Exception *,
     341        _Unwind_Stop_Fn, void *);
     342\end{cfa}
     343It also unwinds the stack but it does not use the search phase. Instead another
     344function, the stop function, is used to stop searching. The exception is the
     345same as the one passed to raise exception. The extra arguments are the stop
     346function and the stop parameter. The stop function has a similar interface as a
     347personality function, except it is also passed the stop parameter.
     348\begin{lstlisting}[language=C,{moredelim=**[is][\color{red}]{@}{@}}]
     349typedef _Unwind_Reason_Code (*@_Unwind_Stop_Fn@)(
     350        _Unwind_Action @action@,
     351        _Unwind_Exception_Class @exception_class@,
     352        _Unwind_Exception * @exception@,
     353        struct _Unwind_Context * @context@,
     354        void * @stop_parameter@);
    214355\end{lstlisting}
    215356
    216 When called the function begins the search phase, calling the personality
    217 function of the most recent stack frame. It will continue to call personality
    218 functions traversing the stack new-to-old until a function finds a handler or
    219 the end of the stack is reached. In the latter case raise exception will
    220 return with \codeC{_URC_END_OF_STACK}.
    221 
    222 Once a handler has been found raise exception continues onto the the cleanup
    223 phase. Once again it will call the personality functins of each stack frame
    224 from newest to oldest. This pass will stop at the stack frame that found the
    225 handler last time, if that personality function does not install the handler
    226 it is an error.
    227 
    228 If an error is encountered raise exception will return either
    229 \codeC{_URC_FATAL_PHASE1_ERROR} or \codeC{_URC_FATAL_PHASE2_ERROR} depending
    230 on when the error occured.
    231 
    232 \subsection{Forced Unwind}
    233 This is the second big function in libunwind. It also unwinds a stack but it
    234 does not use the search phase. Instead another function, the stop function,
    235 is used to decide when to stop.
    236 
    237 \begin{lstlisting}
    238 _Unwind_Reason_Code _Unwind_ForcedUnwind(
    239     _Unwind_Exception *, _Unwind_Stop_Fn, void *);
    240 \end{lstlisting}
    241 
    242 The exception is the same as the one passed to raise exception. The extra
    243 arguments are the stop function and the stop parameter. The stop function has
    244 a similar interface as a personality function, except it is also passed the
    245 stop parameter.
    246 
    247 \begin{lstlisting}
    248 typedef _Unwind_Reason_Code (*_Unwind_Stop_Fn)(
    249     int version,
    250     _Unwind_Action action,
    251     _Unwind_Exception_Class exception_class,
    252     _Unwind_Exception * exception,
    253     struct _Unwind_Context * context,
    254     void * stop_parameter);
    255 \end{lstlisting}
    256 
    257357The stop function is called at every stack frame before the personality
    258 function is called and then once more once after all frames of the stack have
    259 been unwound.
    260 
    261 Each time it is called the stop function should return \codeC{_URC_NO_REASON}
    262 or transfer control directly to other code outside of libunwind. The
    263 framework does not provide any assistance here.
    264 
    265 Its arguments are the same as the paired personality function.
    266 The actions \codeC{_UA_CLEANUP_PHASE} and \codeC{_UA_FORCE_UNWIND} are always
    267 set when it is called. By the official standard that is all but both GCC and
    268 Clang add an extra action on the last call at the end of the stack:
    269 \codeC{_UA_END_OF_STACK}.
     358function is called and then once more after all frames of the stack are
     359unwound.
     360
     361Each time it is called, the stop function should return @_URC_NO_REASON@ or
     362transfer control directly to other code outside of libunwind. The framework
     363does not provide any assistance here.
     364
     365\begin{sloppypar}
     366Its arguments are the same as the paired personality function. The actions
     367@_UA_CLEANUP_PHASE@ and @_UA_FORCE_UNWIND@ are always set when it is
     368called. Beyond the libunwind standard, both GCC and Clang add an extra action
     369on the last call at the end of the stack: @_UA_END_OF_STACK@.
     370\end{sloppypar}
    270371
    271372\section{Exception Context}
    272373% Should I have another independent section?
    273374% There are only two things in it, top_resume and current_exception. How it is
    274 % stored changes depending on wheither or not the thread-library is linked.
    275 
    276 The exception context is a piece of global storage used to maintain data
    277 across different exception operations and to communicate between different
    278 components.
    279 
    280 Each stack has its own exception context. In a purely sequental program, using
    281 only core Cforall, there is only one stack and the context is global. However
    282 if the library \texttt{libcfathread} is linked then there can be multiple
    283 stacks so they will each need their own.
    284 
    285 To handle this code always gets the exception context from the function
    286 \codeC{this_exception_context}. The main exception handling code is in
    287 \texttt{libcfa} and that library also defines the function as a weak symbol
    288 so it acts as a default. Meanwhile in \texttt{libcfathread} the function is
    289 defined as a strong symbol that replaces it when the libraries are linked
    290 together.
    291 
    292 The version of the function defined in \texttt{libcfa} is very simple. It
    293 returns a pointer to a global static variable. With only one stack this
    294 global instance is associated with the only stack.
    295 
    296 The version of the function defined in \texttt{libcfathread} has to handle
    297 more as there are multiple stacks. The exception context is included as
    298 part of the per-stack data stored as part of coroutines. In the cold data
    299 section, stored at the base of each stack, is the exception context for that
    300 stack. The \codeC{this_exception_context} uses the concurrency library to get
    301 the current coroutine and through it the cold data section and the exception
    302 context.
     375% stored changes depending on whether or not the thread-library is linked.
     376
     377The exception context is global storage used to maintain data across different
     378exception operations and to communicate among different components.
     379
     380Each stack must have its own exception context. In a sequential \CFA program,
     381there is only one stack with a single global exception-context. However, when
     382the library @libcfathread@ is linked, there are multiple stacks where each
     383needs its own exception context.
     384
     385General access to the exception context is provided by function
     386@this_exception_context@. For sequential execution, this function is defined as
     387a weak symbol in the \CFA system-library, @libcfa@. When a \CFA program is
     388concurrent, it links with @libcfathread@, where this function is defined with a
     389strong symbol replacing the sequential version.
     390
     391The sequential @this_exception_context@ returns a hard-coded pointer to the
     392global execption context.
     393The concurrent version adds the exception context to the data stored at the
     394base of each stack. When @this_exception_context@ is called it retrieves the
     395active stack and returns the address of the context saved there.
    303396
    304397\section{Termination}
     
    306399% catches. Talk about GCC nested functions.
    307400
    308 Termination exceptions use libunwind quite heavily because it matches the
    309 intended use from \CPP exceptions very closely. The main complication is that
    310 since the \CFA compiler works by translating to C code it cannot generate the
    311 assembly to form the LSDA for try blocks or destructors.
     401Termination exceptions use libunwind heavily because it matches the intended
     402use from \CC exceptions closely. The main complication for \CFA is that the
     403compiler generates C code, making it very difficult to generate the assembly to
     404form the LSDA for try blocks or destructors.
    312405
    313406\subsection{Memory Management}
    314 The first step of termination is to copy the exception into memory managed by
    315 the exception system. Currently the system just uses malloc, without reserved
    316 memory or and ``small allocation" optimizations. The exception handling
    317 mechanism manages memory for the exception as well as memory for libunwind
    318 and the system's own per-exception storage.
    319 
    320 Exceptions are stored in variable sized block. The first component is a fixed
    321 sized data structure that contains the information for libunwind and the
    322 exception system. The second component is a blob of memory that is big enough
    323 to store the exception. Macros with pointer arthritic and type cast are
    324 used to move between the components or go from the embedded
    325 \codeC{_Unwind_Exception} to the entire node.
    326 
    327 All of these nodes are strung together in a linked list. One linked list per
    328 stack, with the head stored in the exception context. Within each linked list
    329 the most recently thrown exception is at the head and the older exceptions
    330 are further down the list. This list format allows exceptions to be thrown
    331 while a different exception is being handled. Only the exception at the head
    332 of the list is currently being handled, the other will wait for the
    333 exceptions before them to be removed.
    334 
    335 The virtual members in the exception's virtual table. The size of the
    336 exception, the copy function and the free function are all in the virtual
    337 table so they are decided per-exception type. The size and copy function are
    338 used right away when the exception is copied in to managed memory. After the
    339 exception is handled the free function is used to clean up the exception and
    340 then the entire node is passed to free.
    341 
    342 \subsection{Try Statements \& Catch Clauses}
    343 The try statements with termination handlers have a pretty complex conversion
    344 to compensate for the lack of assembly generation. Libunwind requires an LSDA
    345 (Language Specific Data Area) and personality function for a function to
    346 unwind across it. The LSDA in particular is hard to generate at the level of
    347 C which is what the \CFA compiler outputs so a work-around is used.
    348 
    349 This work around is a function called \codeC{__cfaehm_try_terminate} in the
    350 standard library. The contents of a try block and the termination handlers
    351 are converted into functions. These are then passed to the try terminate
    352 function and it calls them. This puts the try statements in their own
    353 functions so that no function has to deal with both termination handlers and
    354 destructors.
    355 
    356 This function has some custom embedded assembly that defines its personality
    357 function and LSDA. This is hand coded in C which is why there is only one
    358 version of it, the compiler has no capability to generate it. The personality
    359 function is structured so that it may be expanded, but really it only handles
    360 this one function. Notably it does not handle any destructors so the function
    361 is constructed so that it does need to run it.
     407The first step of a termination raise is to copy the exception into memory
     408managed by the exception system. Currently, the system uses @malloc@, rather
     409than reserved memory or the stack top. The exception handling mechanism manages
     410memory for the exception as well as memory for libunwind and the system's own
     411per-exception storage.
     412
     413[Quick ASCII diagram to get started.]
     414\begin{verbatim}
     415Fixed Header  | _Unwind_Exception   <- pointer target
     416              |
     417              | Cforall storage
     418              |
     419Variable Body | the exception       <- fixed offset
     420              V ...
     421\end{verbatim}
     422
     423Exceptions are stored in variable-sized blocks.
     424The first component is a fixed sized data structure that contains the
     425information for libunwind and the exception system. The second component is an
     426area of memory big enough to store the exception. Macros with pointer arthritic
     427and type cast are used to move between the components or go from the embedded
     428@_Unwind_Exception@ to the entire node.
     429
     430All of these nodes are linked together in a list, one list per stack, with the
     431list head stored in the exception context. Within each linked list, the most
     432recently thrown exception is at the head followed by older thrown
     433exceptions. This format allows exceptions to be thrown, while a different
     434exception is being handled. The exception at the head of the list is currently
     435being handled, while other exceptions wait for the exceptions before them to be
     436removed.
     437
     438The virtual members in the exception's virtual table provide the size of the
     439exception, the copy function, and the free function, so they are specific to an
     440exception type. The size and copy function are used immediately to copy an
     441exception into managed memory. After the exception is handled the free function
     442is used to clean up the exception and then the entire node is passed to free
     443so the memory can be given back to the heap.
     444
     445\subsection{Try Statements and Catch Clauses}
     446The try statement with termination handlers is complex because it must
     447compensate for the lack of assembly-code generated from \CFA. Libunwind
     448requires an LSDA and personality function for control to unwind across a
     449function. The LSDA in particular is hard to mimic in generated C code.
     450
     451The workaround is a function called @__cfaehm_try_terminate@ in the standard
     452library. The contents of a try block and the termination handlers are converted
     453into functions. These are then passed to the try terminate function and it
     454calls them.
     455Because this function is known and fixed (and not an arbitrary function that
     456happens to contain a try statement) this means the LSDA can be generated ahead
     457of time.
     458
     459Both the LSDA and the personality function are set ahead of time using
     460embedded assembly. This is handcrafted using C @asm@ statements and contains
     461enough information for the single try statement the function repersents.
    362462
    363463The three functions passed to try terminate are:
    364 \begin{itemize}
    365 \item The try function: This function is the try block, all the code inside
    366 the try block is placed inside the try function. It takes no parameters and
    367 has no return value. This function is called during regular execution to run
    368 the try block.
    369 \item The match function: This function decides if this try statement should
    370 handle any given termination exception. It takes a pointer to the exception
    371 and returns 0 if the exception is not handled here. Otherwise the return value
    372 is the id of the handler that should handle the exception. It is called
    373 during the search phase.
    374 It is constructed from the conditional part of each handler. It runs each
    375 check in turn, first checking to see if the object
    376 \item The catch function: This function handles the exception. It takes a
    377 pointer to the exception and the handler's id and returns nothing. It is
    378 called after the clean-up phase.
    379 It is constructed by stitching together the bodies of each handler
    380 \end{itemize}
    381 All three are created with GCC nested functions. GCC nested functions can be
    382 used to create closures, functions that can refer to the state of other
    383 functions on the stack. This allows the functions to refer to the main
    384 function and all the variables in scope.
    385 
    386 These nested functions and all other functions besides
    387 \codeC{__cfaehm_try_terminate} in \CFA use the GCC personality function and
    388 the \texttt{-fexceptions} flag to generate the LSDA. This allows destructors
    389 to be implemented with the cleanup attribute.
     464\begin{description}
     465\item[try function:] This function is the try block, all the code inside the
     466try block is placed inside the try function. It takes no parameters and has no
     467return value. This function is called during regular execution to run the try
     468block.
     469
     470\item[match function:] This function is called during the search phase and
     471decides if a catch clause matches the termination exception. It is constructed
     472from the conditional part of each handler and runs each check, top to bottom,
     473in turn, first checking to see if the exception type matches and then if the
     474condition is true. It takes a pointer to the exception and returns 0 if the
     475exception is not handled here. Otherwise the return value is the id of the
     476handler that matches the exception.
     477
     478\item[handler function:] This function handles the exception. It takes a
     479pointer to the exception and the handler's id and returns nothing. It is called
     480after the cleanup phase. It is constructed by stitching together the bodies of
     481each handler and dispatches to the selected handler.
     482\end{description}
     483All three functions are created with GCC nested functions. GCC nested functions
     484can be used to create closures, functions that can refer to the state of other
     485functions on the stack. This approach allows the functions to refer to all the
     486variables in scope for the function containing the @try@ statement. These
     487nested functions and all other functions besides @__cfaehm_try_terminate@ in
     488\CFA use the GCC personality function and the @-fexceptions@ flag to generate
     489the LSDA. This allows destructors to be implemented with the cleanup attribute.
    390490
    391491\section{Resumption}
    392492% The stack-local data, the linked list of nodes.
    393493
    394 Resumption uses a list of nodes for its stack traversal. The head of the list
    395 is stored in the exception context. The nodes in the list just have a pointer
     494Resumption simple to implement because there is no stack unwinding. The
     495resumption raise uses a list of nodes for its stack traversal. The head of the
     496list is stored in the exception context. The nodes in the list have a pointer
    396497to the next node and a pointer to the handler function.
    397498
    398 The on a resumption throw the this list is traversed. At each node the
    399 handler function is called and is passed the exception by pointer. It returns
    400 true if the exception was handled and false otherwise.
    401 
    402 The handler function does both the matching and catching. It tries each
    403 the condition of \codeCFA{catchResume} in order, top-to-bottom and until it
    404 finds a handler that matches. If no handler matches then the function returns
    405 false. Otherwise the matching handler is run, if it completes successfully
    406 the function returns true. Rethrows, through the \codeCFA{throwResume;}
    407 statement, cause the function to return true.
    408 
    409 \subsection{Libunwind Compatibility}
    410 Resumption does not use libunwind for two simple reasons. The first is that
    411 it does not have to unwind anything so would never need to use the clean-up
    412 phase. Still the search phase could be used to make it free to enter or exit
    413 a try statement with resumption handlers in the same way termination handlers
    414 are for the same trade off in the cost of the throw. This is where the second
    415 reason comes in, there is no way to return from a search without installing
    416 a handler or raising an error.
    417 
    418 Although work arounds could be created none seemed to be worth it for the
    419 prototype. This implementation has no difference in behaviour and is much
    420 simpler.
     499A resumption raise traverses this list. At each node the handler function is
     500called, passing the exception by pointer. It returns true if the exception is
     501handled and false otherwise.
     502
     503The handler function does both the matching and handling. It computes the
     504condition of each @catchResume@ in top-to-bottom order, until it finds a
     505handler that matches. If no handler matches then the function returns
     506false. Otherwise the matching handler is run; if it completes successfully, the
     507function returns true. Rethrowing, through the @throwResume;@ statement,
     508causes the function to return true.
     509
     510% Recursive Resumption Stuff:
     511Search skipping \see{\VPageref{p:searchskip}}, which ignores parts of the stack
     512already examined, is accomplished by updating the front of the list as the
     513search continues. Before the handler at a node is called the head of the list
     514is updated to the next node of the current node. After the search is complete,
     515successful or not, the head of the list is reset.
     516
     517This mechanism means the current handler and every handler that has already
     518been checked are not on the list while a handler is run. If a resumption is
     519thrown during the handling of another resumption the active handlers and all
     520the other handler checked up to this point are not checked again.
     521
     522This structure also supports new handler added while the resumption is being
     523handled. These are added to the front of the list, pointing back along the
     524stack -- the first one points over all the checked handlers -- and the ordering
     525is maintained.
     526
     527\label{p:zero-cost}
     528Note, the resumption implementation has a cost for entering/exiting a @try@
     529statement with @catchResume@ clauses, whereas a @try@ statement with @catch@
     530clauses has zero-cost entry/exit. While resumption does not need the stack
     531unwinding and cleanup provided by libunwind, it could use the search phase to
     532providing zero-cost enter/exit using the LSDA. Unfortunately, there is no way
     533to return from a libunwind search without installing a handler or raising an
     534error. Although workarounds might be possible, they are beyond the scope of
     535this thesis. The current resumption implementation has simplicity in its
     536favour.
    421537% Seriously, just compare the size of the two chapters and then consider
    422538% that unwind is required knowledge for that chapter.
     
    424540\section{Finally}
    425541% Uses destructors and GCC nested functions.
    426 Finally clauses are a simple decomposition to some of the existing features.
    427 The code in the block is placed into a GCC nested function with a unique name,
    428 no arguments or return values. This nested function is then set as the
    429 clean-up function of an empty object that is declared at the beginning of a
    430 block placed around the contexts of the try statement.
     542Finally clauses is placed into a GCC nested-function with a unique name, and no
     543arguments or return values. This nested function is then set as the cleanup
     544function of an empty object that is declared at the beginning of a block placed
     545around the context of the associated @try@ statement.
    431546
    432547The rest is handled by GCC. The try block and all handlers are inside the
    433 block. When they are complete control exits the block and the empty object
    434 is cleaned up, which runs the function that contains the finally code.
     548block. At completion, control exits the block and the empty object is cleaned
     549up, which runs the function that contains the finally code.
    435550
    436551\section{Cancellation}
     
    438553
    439554Cancellation also uses libunwind to do its stack traversal and unwinding,
    440 however it uses a different primary function \codeC{_Unwind_ForcedUnwind}.
    441 Details of its interface can be found in the unwind section.
    442 
    443 The first step of cancellation is to find the stack was cancelled and which
    444 type of stack it is. Luckily the threads library stores the main thread
    445 pointer and the current thread pointer and every thread stores a pointer to
     555however it uses a different primary function @_Unwind_ForcedUnwind@. Details
     556of its interface can be found in the \VRef{s:ForcedUnwind}.
     557
     558The first step of cancellation is to find the cancelled stack and its type:
     559coroutine or thread. Fortunately, the thread library stores the main thread
     560pointer and the current thread pointer, and every thread stores a pointer to
    446561its main coroutine and the coroutine it is currently executing.
    447562
    448 So if the the current thread's main and current coroutine do not match, it is
    449 a coroutine cancellation. Otherwise if the main and current thread do not
    450 match, it is a thread cancellation. Otherwise it is a main thread
    451 cancellation.
    452 
    453 However if the threading library is not linked then execution must be on the
    454 main stack as that is the only one that exists. So the entire check is skipped
    455 using the linker and weak symbols. Instead the main thread cancellation is
    456 unconditionally preformed.
    457 
    458 Regardless of how they are choosen afterwords the stop function and the stop
    459 parameter are passed to the forced unwind functon. The general pattern of all
    460 three stop functions is the same, they continue unwinding until the end of
    461 stack when they do there primary work.
    462 
    463 Main stack cancellation it is very simple. The ``transfer" is just an abort,
    464 the program stops executing.
    465 
    466 The coroutine cancellation stores the exception on the coroutine and then
    467 does a coroutine context switch. The rest is handled inside resume. Every time
    468 control returns from a resumed thread there is a check to see if it is
    469 cancelled. If it is the exception is retrieved and the CoroutineCancelled
    470 exception is constructed and loaded. It is then thrown as a regular exception
    471 with the default handler coming from the context of the resumption call.
    472 
    473 The thread cancellation stores the exception on the thread's main stack and
    474 then returns to the scheduler. The rest is handled by the joiner. The wait
    475 for the joined thread to finish works the same but after that it checks
    476 to see if there was a cancellation. If there was the exception is retrieved
    477 and the ThreadCancelled exception is constructed. The default handler is
    478 passed in as a function pointer. If it is null (as it is for the
    479 auto-generated joins on destructor call) it a default is used that simply
    480 calls abort; which gives the required handling on implicate join.
     563So if the active thread's main and current coroutine are the same. If they
     564are then the current stack is a thread stack, otherwise it is a coroutine
     565stack. If it is a thread stack then an equality check with the stored main
     566thread pointer and current thread pointer is enough to tell if the current
     567thread is the main thread or not.
     568
     569However, if the threading library is not linked, the sequential execution is on
     570the main stack. Hence, the entire check is skipped because the weak-symbol
     571function is loaded. Therefore, a main thread cancellation is unconditionally
     572performed.
     573
     574Regardless of how the stack is chosen, the stop function and parameter are
     575passed to the forced-unwind function. The general pattern of all three stop
     576functions is the same: they continue unwinding until the end of stack when they
     577do there primary work.
     578
     579For main stack cancellation, the transfer is just a program abort.
     580
     581For coroutine cancellation, the exception is stored on the coroutine's stack,
     582and the coroutine context switches to its last resumer. The rest is handled on
     583the backside of the resume, which check if the resumed coroutine is
     584cancelled. If cancelled, the exception is retrieved from the resumed coroutine,
     585and a @CoroutineCancelled@ exception is constructed and loaded with the
     586cancelled exception. It is then resumed as a regular exception with the default
     587handler coming from the context of the resumption call.
     588
     589For thread cancellation, the exception is stored on the thread's main stack and
     590then context switched to the scheduler. The rest is handled by the thread
     591joiner. When the join is complete, the joiner checks if the joined thread is
     592cancelled. If cancelled, the exception is retrieved and the joined thread, and
     593a @ThreadCancelled@ exception is constructed and loaded with the cancelled
     594exception. The default handler is passed in as a function pointer. If it is
     595null (as it is for the auto-generated joins on destructor call), the default is
     596used, which is a program abort.
     597%; which gives the required handling on implicate join.
  • doc/theses/andrew_beach_MMath/thesis-frontpgs.tex

    r342af53 r8e4aa05  
    3636
    3737        A thesis \\
    38         presented to the University of Waterloo \\ 
     38        presented to the University of Waterloo \\
    3939        in fulfillment of the \\
    4040        thesis requirement for the degree of \\
     
    6464\cleardoublepage
    6565
    66  
     66
    6767%----------------------------------------------------------------------
    6868% EXAMINING COMMITTEE (Required for Ph.D. theses only)
     
    7171\begin{center}\textbf{Examining Committee Membership}\end{center}
    7272  \noindent
    73 The following served on the Examining Committee for this thesis. The decision of the Examining Committee is by majority vote.
    74   \bigskip
    75  
    76   \noindent
    77 \begin{tabbing}
    78 Internal-External Member: \=  \kill % using longest text to define tab length
    79 External Examiner: \>  Bruce Bruce \\
     73The following served on the Examining Committee for this thesis. The decision
     74of the Examining Committee is by majority vote.
     75  \bigskip
     76
     77  \noindent
     78\begin{tabbing}
     79Internal-External Member: \=  \kill % using longest text to define tab length
     80External Examiner: \>  Bruce Bruce \\
    8081\> Professor, Dept. of Philosophy of Zoology, University of Wallamaloo \\
    81 \end{tabbing} 
    82   \bigskip
    83  
     82\end{tabbing}
     83  \bigskip
     84
    8485  \noindent
    8586\begin{tabbing}
     
    9192\end{tabbing}
    9293  \bigskip
    93  
     94
    9495  \noindent
    9596  \begin{tabbing}
     
    99100\end{tabbing}
    100101  \bigskip
    101  
     102
    102103  \noindent
    103104\begin{tabbing}
     
    107108\end{tabbing}
    108109  \bigskip
    109  
     110
    110111  \noindent
    111112\begin{tabbing}
     
    123124  % December 13th, 2006.  It is designed for an electronic thesis.
    124125  \noindent
    125 I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners.
    126 
    127   \bigskip
    128  
     126I hereby declare that I am the sole author of this thesis. This is a true copy
     127of the thesis, including any required final revisions, as accepted by my
     128examiners.
     129
     130  \bigskip
     131
    129132  \noindent
    130133I understand that my thesis may be made electronically available to the public.
  • doc/theses/andrew_beach_MMath/thesis.tex

    r342af53 r8e4aa05  
    4545% FRONT MATERIAL
    4646%----------------------------------------------------------------------
    47 \input{thesis-frontpgs} 
     47\input{thesis-frontpgs}
    4848
    4949%----------------------------------------------------------------------
     
    6565A \gls{computer} could compute $\pi$ all day long. In fact, subsets of digits
    6666of $\pi$'s decimal approximation would make a good source for psuedo-random
    67 vectors, \gls{rvec} . 
     67vectors, \gls{rvec} .
    6868
    6969%----------------------------------------------------------------------
     
    9696
    9797\begin{itemize}
    98 \item A well-prepared PDF should be 
     98\item A well-prepared PDF should be
    9999  \begin{enumerate}
    100100    \item Of reasonable size, {\it i.e.} photos cropped and compressed.
    101     \item Scalable, to allow enlargment of text and drawings. 
    102   \end{enumerate} 
     101    \item Scalable, to allow enlargment of text and drawings.
     102  \end{enumerate}
    103103\item Photos must be bit maps, and so are not scaleable by definition. TIFF and
    104104BMP are uncompressed formats, while JPEG is compressed. Most photos can be
    105105compressed without losing their illustrative value.
    106 \item Drawings that you make should be scalable vector graphics, \emph{not} 
     106\item Drawings that you make should be scalable vector graphics, \emph{not}
    107107bit maps. Some scalable vector file formats are: EPS, SVG, PNG, WMF. These can
    108 all be converted into PNG or PDF, that pdflatex recognizes. Your drawing 
    109 package probably can export to one of these formats directly. Otherwise, a 
    110 common procedure is to print-to-file through a Postscript printer driver to 
    111 create a PS file, then convert that to EPS (encapsulated PS, which has a 
    112 bounding box to describe its exact size rather than a whole page). 
     108all be converted into PNG or PDF, that pdflatex recognizes. Your drawing
     109package probably can export to one of these formats directly. Otherwise, a
     110common procedure is to print-to-file through a Postscript printer driver to
     111create a PS file, then convert that to EPS (encapsulated PS, which has a
     112bounding box to describe its exact size rather than a whole page).
    113113Programs such as GSView (a Ghostscript GUI) can create both EPS and PDF from
    114114PS files. Appendix~\ref{AppendixA} shows how to generate properly sized Matlab
    115115plots and save them as PDF.
    116116\item It's important to crop your photos and draw your figures to the size that
    117 you want to appear in your thesis. Scaling photos with the 
    118 includegraphics command will cause loss of resolution. And scaling down 
     117you want to appear in your thesis. Scaling photos with the
     118includegraphics command will cause loss of resolution. And scaling down
    119119drawings may cause any text annotations to become too small.
    120120\end{itemize}
    121  
     121
    122122For more information on \LaTeX\, see the uWaterloo Skills for the
    123 Academic Workplace \href{https://uwaterloo.ca/information-systems-technology/services/electronic-thesis-preparation-and-submission-support/ethesis-guide/creating-pdf-version-your-thesis/creating-pdf-files-using-latex/latex-ethesis-and-large-documents}{course notes}. 
     123Academic Workplace \href{https://uwaterloo.ca/information-systems-technology/services/electronic-thesis-preparation-and-submission-support/ethesis-guide/creating-pdf-version-your-thesis/creating-pdf-files-using-latex/latex-ethesis-and-large-documents}{course notes}.
    124124\footnote{
    125125Note that while it is possible to include hyperlinks to external documents,
    126 it is not wise to do so, since anything you can't control may change over time. 
    127 It \emph{would} be appropriate and necessary to provide external links to 
    128 additional resources for a multimedia ``enhanced'' thesis. 
    129 But also note that if the \package{hyperref} package is not included, 
    130 as for the print-optimized option in this thesis template, any \cmmd{href} 
     126it is not wise to do so, since anything you can't control may change over time.
     127It \emph{would} be appropriate and necessary to provide external links to
     128additional resources for a multimedia ``enhanced'' thesis.
     129But also note that if the \package{hyperref} package is not included,
     130as for the print-optimized option in this thesis template, any \cmmd{href}
    131131commands in your logical document are no longer defined.
    132132A work-around employed by this thesis template is to define a dummy
    133 \cmmd{href} command (which does nothing) in the preamble of the document, 
    134 before the \package{hyperref} package is included. 
     133\cmmd{href} command (which does nothing) in the preamble of the document,
     134before the \package{hyperref} package is included.
    135135The dummy definition is then redifined by the
    136136\package{hyperref} package when it is included.
     
    138138
    139139The classic book by Leslie Lamport \cite{lamport.book}, author of \LaTeX , is
    140 worth a look too, and the many available add-on packages are described by 
     140worth a look too, and the many available add-on packages are described by
    141141Goossens \textit{et al} \cite{goossens.book}.
    142142
     
    180180Export Setup button in the figure Property Editor.
    181181
    182 \section{From the Command Line} 
     182\section{From the Command Line}
    183183All figure properties can also be manipulated from the command line. Here's an
    184 example: 
     184example:
    185185\begin{verbatim}
    186186x=[0:0.1:pi];
  • doc/theses/andrew_beach_MMath/unwinding.tex

    r342af53 r8e4aa05  
    11\chapter{Unwinding in \CFA}
    22
    3 Stack unwinding is the process of removing things from the stack. Within
    4 functions and on function return this is handled directly by the code in the
    5 function itself as it knows exactly what is on the stack just from the
    6 current location in the function. Unwinding across stack frames means that it
    7 is no longer knows exactly what is on the stack or even how much of the stack
    8 needs to be removed.
     3Stack unwinding is the process of removing stack frames (activations) from the
     4stack. On function entry and return, unwinding is handled directly by the code
     5embedded in the function. Usually, the stack-frame size is known statically
     6based on parameters and local variable declarations.  For dynamically-sized
     7local variables, a runtime computation is necessary to know the frame
     8size. Finally, a function's frame-size may change during execution as local
     9variables (static or dynamic sized) go in and out of scope.
     10Allocating/deallocating stack space is usually an $O(1)$ operation achieved by
     11bumping the hardware stack-pointer up or down as needed.
    912
    10 Even this is fairly simple if nothing needs to happen when the stack unwinds.
    11 Traditional C can unwind the stack by saving and restoring state (with
    12 \codeC{setjmp} \& \codeC{longjmp}). However many languages define actions that
    13 have to be taken when something is removed from the stack, such as running
    14 a variable's destructor or a \codeCFA{try} statement's \codeCFA{finally}
    15 clause. Handling this requires walking the stack going through each stack
    16 frame.
     13Unwinding across multiple stack frames is more complex because individual stack
     14management code associated with each frame is bypassed. That is, the location
     15of a function's frame code is largely unknown and dispersed throughout the
     16function, hence the current stack-frame size managed by that code is also
     17unknown. Hence, code unwinding across frames does not have direct knowledge
     18about what is on the stack, and hence, how much of the stack needs to be
     19removed.
    1720
    18 For exceptions, this means everything from the point the exception is raised
    19 to the point it is caught, while checking each frame for handlers during the
    20 stack walk to find out where it should be caught. This is where the most of
    21 the expense and complexity of exception handling comes from.
     21The traditional unwinding mechanism for C is implemented by saving a snap-shot
     22of a function's state with @setjmp@ and restoring that snap-shot with
     23@longjmp@. This approach bypasses the need to know stack details by simply
     24reseting to a snap-shot of an arbitrary but existing function frame on the
     25stack. It is up to the programmer to ensure the snap-shot is valid when it is
     26reset, making the code fragile with potential errors that are difficult to
     27debug because the stack becomes corrupted.
    2228
    23 To do all of this we use libunwind, a low level library that provides tools
    24 for stack walking and stack unwinding. What follows is an overview of all the
    25 relivant features of libunwind and then how \CFA uses them to implement its
    26 exception handling.
     29However, many languages define cleanup actions that have to be taken when
     30something is deallocated from the stack or blocks end, such as running a
     31variable's destructor or a @try@ statement's @finally@ clause. Handling these
     32mechanisms requires walking the stack and checking each stack frame for these
     33potential actions.
     34
     35For exceptions, it must be possible to walk the stack frames in search of try
     36statements with handlers to perform exception matching. For termination
     37exceptions, it must be possible to unwind all stack frames from the throw to
     38the matching catch, and each of these frames must be checked for cleanup
     39actions. Stack walking is where the most of the complexity and expense of
     40exception handling comes from.
     41
     42One of the most popular tools for stack management is libunwind, a low level
     43library that provides tools for stack walking and unwinding. What follows is an
     44overview of all the relevant features of libunwind and how \CFA uses them to
     45implement its exception handling.
    2746
    2847\section{libunwind Usage}
    29 
    30 \CFA uses two primary functions in libunwind to create most of its
    31 exceptional control-flow: \codeC{_Unwind_RaiseException} and
    32 \codeC{_Unwind_ForcedUnwind}.
    33 Their operation is divided into two phases: search and clean-up. The search
    34 phase -- phase 1 -- is used to scan the stack but not unwinding it. The
    35 clean-up phase -- phase 2 -- is used for unwinding.
     48\CFA uses two primary functions in libunwind to create most of its exceptional
     49control-flow: @_Unwind_RaiseException@ and @_Unwind_ForcedUnwind@.  Their
     50operation is divided into two phases: search and clean-up. The search phase --
     51phase 1 -- is used to scan the stack but not unwinding it. The clean-up phase
     52-- phase 2 -- is used for unwinding.
    3653
    3754The raise-exception function uses both phases. It starts by searching for a
     
    4461A personality function performs three tasks, although not all have to be
    4562present. The tasks performed are decided by the actions provided.
    46 \codeC{_Unwind_Action} is a bitmask of possible actions and an argument of
    47 this type is passed into the personality function.
     63@_Unwind_Action@ is a bitmask of possible actions and an argument of this type
     64is passed into the personality function.
    4865\begin{itemize}
    49 \item\codeC{_UA_SEARCH_PHASE} is passed in search phase and tells the
    50 personality function to check for handlers. If there is a handler in this
    51 stack frame, as defined by the language, the personality function should
    52 return \codeC{_URC_HANDLER_FOUND}. Otherwise it should return
    53 \codeC{_URC_CONTINUE_UNWIND}.
    54 \item\codeC{_UA_CLEANUP_PHASE} is passed in during the clean-up phase and
    55 means part or all of the stack frame is removed. The personality function
    56 should do whatever clean-up the language defines
    57 (such as running destructors/finalizers) and then generally returns
    58 \codeC{_URC_CONTINUE_UNWIND}.
    59 \item\codeC{_UA_HANDLER_FRAME} means the personality function must install
    60 a handler. It is also passed in during the clean-up phase and is in addition
    61 to the clean-up action. libunwind provides several helpers for the personality
    62 function here. Once it is done, the personality function must return
    63 \codeC{_URC_INSTALL_CONTEXT}.
     66\item
     67\begin{sloppypar}
     68@_UA_SEARCH_PHASE@ is passed in for the search phase and tells the personality
     69function to check for handlers. If there is a handler in a stack frame, as
     70defined by the language, the personality function returns @_URC_HANDLER_FOUND@;
     71otherwise it return @_URC_CONTINUE_UNWIND@.
     72\end{sloppypar}
     73\item
     74@_UA_CLEANUP_PHASE@ is passed in during the clean-up phase and means part or
     75all of the stack frame is removed. The personality function does whatever
     76clean-up the language defines (such as running destructors/finalizers) and then
     77generally returns @_URC_CONTINUE_UNWIND@.
     78\item
     79@_UA_HANDLER_FRAME@ means the personality function must install a handler. It
     80is also passed in during the clean-up phase and is in addition to the clean-up
     81action. libunwind provides several helpers for the personality function. Once
     82it is done, the personality function returns @_URC_INSTALL_CONTEXT@.
    6483\end{itemize}
    65 The personality function is given a number of other arguments. Some are for
    66 compatability and there is the \codeC{struct _Unwind_Context} pointer which
    67 passed to many helpers to get information about the current stack frame.
     84The personality function is given a number of other arguments. Some arguments
     85are for compatibility, and there is the @struct _Unwind_Context@ pointer which
     86is passed to many helpers to get information about the current stack frame.
    6887
    69 Forced-unwind only performs the clean-up phase. It takes three arguments:
    70 a pointer to the exception, a pointer to the stop function and a pointer to
    71 the stop parameter. It does most of the same things as phase two of
    72 raise-exception but with some extras.
    73 The first it passes in an extra action to the personality function on each
    74 stack frame, \codeC{_UA_FORCE_UNWIND}, which means a handler cannot be
     88For cancellation, forced-unwind only performs the clean-up phase. It takes
     89three arguments: a pointer to the exception, a pointer to the stop function and
     90a pointer to the stop parameter. It does most of the same actions as phase two
     91of raise-exception but passes in an extra action to the personality function on
     92each stack frame, @_UA_FORCE_UNWIND@, which means a handler cannot be
    7593installed.
    7694
    77 The big change is that forced-unwind calls the stop function. Each time it
    78 steps into a frame, before calling the personality function, it calls the
    79 stop function. The stop function receives all the same arguments as the
    80 personality function will and the stop parameter supplied to forced-unwind.
     95As well, forced-unwind calls the stop function each time it steps into a frame,
     96before calling the personality function. The stop function receives all the
     97same arguments as the personality function and the stop parameter supplied to
     98forced-unwind.
    8199
    82100The stop function is called one more time at the end of the stack after all
    83 stack frames have been removed. By the standard API this is marked by setting
     101stack frames have been removed. The standard API marks this frame by setting
    84102the stack pointer inside the context passed to the stop function. However both
    85 GCC and Clang add an extra action for this case \codeC{_UA_END_OF_STACK}.
     103GCC and Clang add an extra action for this case @_UA_END_OF_STACK@.
    86104
    87 Each time function the stop function is called it can do one or two things.
    88 When it is not the end of the stack it can return \codeC{_URC_NO_REASON} to
    89 continue unwinding.
     105Each time the stop function is called, it can do one or two things.  When it is
     106not the end of the stack it can return @_URC_NO_REASON@ to continue unwinding.
    90107% Is there a reason that NO_REASON is used instead of CONTINUE_UNWIND?
    91 Its only other option is to use its own means to transfer control elsewhere
    92 and never return to its caller. It may always do this and no additional tools
    93 are provided to do it.
     108The other option is to use some other means to transfer control elsewhere and
     109never return to its caller. libunwind provides no additional tools for
     110alternate transfers of control.
    94111
    95112\section{\CFA Implementation}
    96113
    97 To use libunwind, \CFA provides several wrappers, its own storage,
    98 personality functions, and a stop function.
     114To use libunwind, \CFA provides several wrappers, its own storage, personality
     115functions, and a stop function.
    99116
    100117The wrappers perform three tasks: set-up, clean-up and controlling the
     
    108125The core control code is called every time a throw -- after set-up -- or
    109126re-throw is run. It uses raise-exception to search for a handler and to run it
    110 if one is found. If no handler is found and raise-exception returns then
     127if one is found. If no handler is found and raise-exception returns, then
    111128forced-unwind is called to run all destructors on the stack before terminating
    112129the process.
    113130
    114 The stop function is very simple. It checks the end of stack flag to see if
    115 it is finished unwinding. If so, it calls \codeC{exit} to end the process,
    116 otherwise it returns with no-reason to continue unwinding.
     131The stop function is simple. It checks for the end of stack flag to see if
     132unwinding is finished. If so, it calls @exit@ to end the process, otherwise it
     133returns with no-reason to continue unwinding.
    117134% Yeah, this is going to have to change.
    118135
    119136The personality routine is more complex because it has to obtain information
    120 about the function by scanning the LSDA (Language Specific Data Area). This
     137about the function by scanning the Language Specific Data Area (LSDA). This
    121138step allows a single personality function to be used for multiple functions and
    122 let that personaliity function figure out exactly where in the function
    123 execution was, what is currently in the stack frame and what handlers should
    124 be checked.
     139lets that personality function figure out exactly where in the function
     140execution is, what is currently in the stack frame, and what handlers should be
     141checked.
    125142% Not that we do that yet.
    126143
    127 However, generating the LSDA is difficult. It requires knowledge about the
    128 location of the instruction pointer and stack layout, which varies with
    129 compiler and optimization levels. So for frames where there are only
    130 destructors, GCC's attribute cleanup with the \texttt{-fexception} flag is
    131 sufficient to handle unwinding.
     144It is also necessary to generate the LSDA, which is difficult. It requires
     145knowledge about the location of the instruction pointer and stack layout, which
     146varies with compiler and optimization levels. Fortunately, for frames where
     147there are only destructors, GCC's attribute cleanup with the @-fexception@ flag
     148is sufficient to handle unwinding.
    132149
    133 The only functions that require more than that are those that contain
    134 \codeCFA{try} statements. A \codeCFA{try} statement has a \codeCFA{try}
    135 clause, some number of \codeCFA{catch} clauses and \codeCFA{catchResume}
    136 clauses and may have a \codeCFA{finally} clause. Of these only \codeCFA{try}
    137 statements with \codeCFA{catch} clauses need to be transformed and only they
    138 and the \codeCFA{try} clause are involved.
     150The only functions that require more information are those containing @try@
     151statements. Specifically, only @try@ statements with @catch@ clauses need to be
     152transformed.  The @try@ statement is converted into a series of closures that
     153can access other parts of the function according to scoping rules but can be
     154passed around. The @catch@ clauses are converted into two functions: the match
     155function and the handler function.
    139156
    140 The \codeCFA{try} statement is converted into a series of closures which can
    141 access other parts of the function according to scoping rules but can be
    142 passed around. The \codeCFA{try} clause is converted into the try functions,
    143 almost entirely unchanged. The \codeCFA{catch} clauses are converted into two
    144 functions; the match function and the catch function.
     157Together the match function and the catch function form the code that runs when
     158an exception passes out of the guarded block for a try statement. The match
     159function is used during the search phase: it is passed an exception and checks
     160each handler to see if the raised exception matches the handler exception. It
     161returns an index that represents which handler matched or there is no
     162match. The catch function is used during the clean-up phase, it is passed an
     163exception and the index of a handler. It casts the exception to the exception
     164type declared in that handler and then runs the handler's body.
    145165
    146 Together the match function and the catch function form the code that runs
    147 when an exception passes out of a try block. The match function is used during
    148 the search phase, it is passed an exception and checks each handler to see if
    149 it will handle the exception. It returns an index that repersents which
    150 handler matched or that none of them did. The catch function is used during
    151 the clean-up phase, it is passed an exception and the index of a handler. It
    152 casts the exception to the exception type declared in that handler and then
    153 runs the handler's body.
    154 
    155 These three functions are passed to \codeC{try_terminate}. This is an
     166These three functions are passed to @try_terminate@, which is an
    156167% Maybe I shouldn't quote that, it isn't its actual name.
    157 internal hand-written function that has its own personality function and
    158 custom assembly LSD does the exception handling in \CFA. During normal
    159 execution all this function does is call the try function and then return.
    160 It is only when exceptions are thrown that anything interesting happens.
     168internal hand-written function that has its own personality function and custom
     169assembly LSDA for doing the exception handling in \CFA. During normal
     170execution, this function calls the try function and then return.  It is only
     171when exceptions are thrown that anything interesting happens.
    161172
    162173During the search phase the personality function gets the pointer to the match
    163 function and calls it. If the match function returns a handler index the
     174function and calls it. If the match function returns a handler index, the
    164175personality function saves it and reports that the handler has been found,
    165 otherwise unwinding continues.
    166 During the clean-up phase the personality function only does anything if the
    167 handler was found in this frame. If it was then the personality function
    168 installs the handler, which is setting the instruction pointer in
    169 \codeC{try_terminate} to an otherwise unused section that calls the catch
    170 function, passing it the current exception and handler index.
    171 \codeC{try_terminate} returns as soon as the catch function returns.
     176otherwise unwinding continues.  During the clean-up phase, the personality
     177function only performs an action, when a handler is found in a frame. For each
     178found frame, the personality function installs the handler, which sets the
     179instruction pointer in @try_terminate@ to an otherwise unused section that
     180calls the catch function, passing it the current exception and handler index.
     181@try_terminate@ returns as soon as the catch function returns.  At this point
     182control has returned to normal control flow.
    172183
    173 At this point control has returned to normal control flow.
     184\PAB{Maybe a diagram would be helpful?}
  • doc/theses/fangren_yu_COOP_F20/Report.tex

    r342af53 r8e4aa05  
    1717\usepackage[usenames]{color}
    1818\input{common}                                          % common CFA document macros
    19 \usepackage[dvips,plainpages=false,pdfpagelabels,pdfpagemode=UseNone,colorlinks=true,pagebackref=true,linkcolor=blue,citecolor=blue,urlcolor=blue,pagebackref=true,breaklinks=true]{hyperref}
     19\usepackage[dvips,plainpages=false,pdfpagelabels,pdfpagemode=UseNone,colorlinks=true,linkcolor=blue,citecolor=blue,urlcolor=blue,pagebackref=true,breaklinks=true]{hyperref}
    2020\usepackage{breakurl}
    2121\urlstyle{sf}
     
    7676\renewcommand{\subsectionmark}[1]{\markboth{\thesubsection\quad #1}{\thesubsection\quad #1}}
    7777\pagenumbering{roman}
    78 \linenumbers                                            % comment out to turn off line numbering
     78%\linenumbers                                            % comment out to turn off line numbering
    7979
    8080\maketitle
    8181\pdfbookmark[1]{Contents}{section}
    82 \tableofcontents
    83 
    84 \clearpage
     82
    8583\thispagestyle{plain}
    8684\pagenumbering{arabic}
    8785
    8886\begin{abstract}
     87\CFA is an evolutionary, non-object-oriented extension of the C programming language, featuring a parametric type-system, and is currently under active development. The reference compiler for the \CFA language, @cfa-cc@, has some of its major components dated back to the early 2000s, which are based on inefficient data structures and algorithms. This report introduces improvements targeting the expression resolution algorithm, suggested by a recent prototype experiment on a simplified model, which are implemented in @cfa-cc@ to support the full \CFA language. These optimizations speed up the compiler by a factor of 20 across the existing \CFA codebase, bringing the compilation time of a mid-sized \CFA source file down to the 10-second level. A few problem cases derived from realistic code examples are analyzed in detail, with proposed solutions. This work is a critical step in the \CFA project development to achieve its eventual goal of being used alongside C for large software systems.
    8988\end{abstract}
    9089
     90\clearpage
     91\section*{Acknowledgements}
     92\begin{sloppypar}
     93I would like to thank everyone in the \CFA team for their contribution towards this project. Programming language design and development is a tough subject and requires a lot of teamwork. Without the collaborative efforts from the team, this project could not have been a success. Specifically, I would like to thank Andrew Beach for introducing me to the \CFA codebase, Thierry Delisle for maintaining the test and build automation framework, Michael Brooks for providing example programs of various experimental language and type system features, and most importantly, Professor Martin Karsten for recommending me to the \CFA team, and my supervisor, Professor Peter Buhr for encouraging me to explore deeply into intricate compiler algorithms. Finally, I gratefully acknowledge the help from Aaron Moss, former graduate from the team and the author of the precedent thesis work, to participate in the \CFA team's virtual conferences and email correspondence, and provide many critical arguments and suggestions. 2020 had been an unusually challenging year for everyone and we managed to keep a steady pace.
     94\end{sloppypar}
     95
     96\clearpage
     97\tableofcontents
     98
     99\clearpage
    91100\section{Introduction}
    92101
    93 \section{Completed work}
     102\CFA language, developed by the Programming Language Group at the University of Waterloo, has a long history, with the initial language design in 1992 by Glen Ditchfield~\cite{Ditchfield92} and the first proof-of-concept compiler built in 2003 by Richard Bilson~\cite{Bilson03}. Many new features have been added to the language over time, but the core of \CFA's type-system --- parametric functions introduced by the @forall@ clause (hence the name of the language) providing parametric overloading --- remains mostly unchanged.
     103
     104The current \CFA reference compiler, @cfa-cc@, is designed using the visitor pattern~\cite{vistorpattern} over an abstract syntax tree (AST), where multiple passes over the AST modify it for subsequent passes. @cfa-cc@ still includes many parts taken directly from the original Bilson implementation, which served as the starting point for this enhancement work to the type system. Unfortunately, the prior implementation did not provide the efficiency required for the language to be practical: a \CFA source file of approximately 1000 lines of code can take multiple minutes to compile. The cause of the problem is that the old compiler used inefficient data structures and algorithms for expression resolution, which involved significant copying and redundant work.
     105
     106This report presents a series of optimizations to the performance-critical parts of the resolver, with a major rework of the compiler data-structures using a functional-programming approach to reduce memory complexity. The improvements were suggested by running the compiler builds with a performance profiler against the \CFA standard-library source-code and a test suite to find the most underperforming components in the compiler algorithm.
     107
     108The \CFA team endorses a pragmatic philosophy that focuses on practical implications of language design and implementation rather than theoretical limits. In particular, the compiler is designed to be expressive with respect to code reuse while maintaining type safety, but compromise theoretical soundness in extreme corner cases. However, when these corner cases do appear in actual usage, they need to be thoroughly investigated. A case-by-case analysis is presented for several of these corner cases, some of which point to certain weaknesses in the language design with solutions proposed based on experimental results.
     109
     110\section{AST restructuring}
    94111
    95112\subsection{Memory model with sharing}
    96113
    97 A major rework of the abstract syntax tree (AST) data structure in the compiler is completed as the first step of the project. The majority of work were documented in the reference manual of the compiler~\cite{cfa-cc}. To summarize:
    98 \begin{itemize}
    99 \item
    100 AST nodes (and therefore subtrees) can be shared without copying when reused.
    101 \item
    102 Modifications apply the functional programming principle, making copies for local changes without affecting the original data shared by other owners. In-place mutations are permitted as a special case when sharing does not happen. The logic is implemented by reference counting.
    103 \item
    104 Memory allocation and freeing are performed automatically using smart pointers.
    105 \end{itemize}
    106 The resolver algorithm designed for overload resolution naturally introduces a significant amount of reused intermediate representations, especially in the following two places:
    107 \begin{itemize}
    108 \item
    109 Function overload candidates are computed by combining the argument candidates bottom-up, with many of them being a common term. For example, if $n$ overloads of a function @f@ all take an integer for the first parameter but different types for the second (@f( int, int )@, @f( int, double )@, etc.) the first term is reused $n$ times for each of the generated candidate expressions. This effect is particularly bad for deep expression trees.
    110 \item
    111 In the unification algorithm and candidate elimination step, actual types are obtained by substituting the type parameters by their bindings. Let $n$ be the complexity (\ie number of nodes in representation) of the original type, $m$ be the complexity of bound type for parameters, and $k$ be the number of occurrences of type parameters in the original type. If everything needs to be deep-copied, the substitution step takes $O(n+mk)$ time and memory, while using shared nodes it is reduced to $O(n)$ time and $O(k)$ memory.
    112 \end{itemize}
    113 One of the worst examples for the old compiler is a long chain of I/O operations
    114 \begin{cfa}
    115 sout | 1 | 2 | 3 | 4 | ...
    116 \end{cfa}
    117 The pipe operator is overloaded by \CFA I/O library for every primitive type in C language, as well as I/O manipulators defined by the library. In total there are around 50 overloads for the output stream operation. On resolving the $n$-th pipe operator in the sequence, the first term, which is the result of sub-expression containing $n-1$ pipe operators, is reused to resolve every overload. Therefore at least $O(n^2)$ copies of expression nodes are made during resolution, not even counting type unification cost; combined with two large factors from number of overloads of pipe operators, and that the ``output stream type'' in \CFA is a trait with 27 assertions (which adds to complexity of the pipe operator's type) this makes compiling a long output sequence extremely slow. In new AST representation only $O(n)$ copies are required and type of pipe operator is not copied at all.
    118 
    119 Reduction in space complexity is especially important, as preliminary profiling result on the old compiler build shows that over half of time spent in expression resolution are on memory allocations.
    120  
     114A major rework of the AST data-structure in the compiler was completed as the first step of the project. The majority of this work is documented in my prior report documenting the compiler reference-manual~\cite{cfa-cc}. To summarize:
     115\begin{itemize}
     116\item
     117AST nodes (and therefore subtrees) can be shared without copying.
     118\item
     119Modifications are performed using functional-programming principles, making copies for local changes without affecting the original data shared by other owners. In-place mutations are permitted as a special case when there is no sharing. The logic is implemented by reference counting.
     120\item
     121Memory allocation and freeing are performed automatically using smart pointers~\cite{smartpointers}.
     122\end{itemize}
     123
     124The resolver algorithm, designed for overload resolution, allows a significant amount of code reused, and hence copying, for the intermediate representations, especially in the following two places:
     125\begin{itemize}
     126\item
     127Function overload candidates are computed by combining the argument candidates bottom-up, with many being a common term. For example, if $n$ overloads of a function @f@ all take an integer for the first parameter but different types for the second, \eg @f( int, int )@, @f( int, double )@, etc., the first term is copied $n$ times for each of the generated candidate expressions. This copying is particularly bad for deep expression trees.
     128\item
     129In the unification algorithm and candidate elimination step, actual types are obtained by substituting the type parameters by their bindings. Let $n$ be the complexity (\ie number of nodes in representation) of the original type, $m$ be the complexity of the bound type for parameters, and $k$ be the number of occurrences of type parameters in the original type. If every substitution needs to be deep-copied, these copy step takes $O(n+mk)$ time and memory, while using shared nodes it is reduced to $O(n)$ time and $O(k)$ memory.
     130\end{itemize}
     131One of the worst examples for the old compiler is a long chain of I/O operations:
     132\begin{cfa}
     133sout | 1 | 2 | 3 | 4 | ...;   // print integer constants
     134\end{cfa}
     135The pipe operator is overloaded by the \CFA I/O library for every primitive type in the C language, as well as I/O manipulators defined by the library. In total, there are around 50 overloads for the output stream operation. On resolving the $n$-th pipe operator in the sequence, the first term, which is the result of sub-expression containing $n-1$ pipe operators, is reused to resolve every overload. Therefore at least $O(n^2)$ copies of expression nodes are made during resolution, not even counting type unification cost; combined with the two large factors from number of overloads of pipe operators, and that the ``output stream type'' in \CFA is a trait with 27 assertions (which adds to complexity of the pipe operator's type) this makes compiling a long output sequence extremely slow. In the new AST representation, only $O(n)$ copies are required and the type of the pipe operator is not copied at all.
     136Reduction in space complexity is especially important, as preliminary profiling results on the old compiler build showed over half of the time spent in expression resolution is on memory allocations.
     137
     138Since the compiler codebase is large and the new memory model mostly benefits expression resolution, some of the old data structures are still kept, and a conversion pass happens before and after the general resolve phase. Rewriting every compiler module will take longer, and whether the new model is correct was unknown when this project started, therefore only the resolver is currently implemented with the new data structure.
     139
    121140
    122141\subsection{Merged resolver calls}
    123142
    124 The pre-resolve phase of compilation, inadequately called ``validate'' in the compiler source code, does more than just simple syntax validation, as it also normalizes input program. Some of them, however, requires type information on expressions and therefore needs to call the resolver before the general resolve phase. There are three notable places where the resolver is invoked:
    125 \begin{itemize}
    126 \item
    127 Attempt to generate default constructor, copy constructor and destructor for user-defined @struct@ types
    128 \item
    129 Resolve @with@ statements (the same as in Python, which introduces fields of a structure directly in scope)
     143The pre-resolve phase of compilation, inappropriately called ``validate'' in the compiler source code, has a number of passes that do more than simple syntax and semantic validation; some passes also normalizes the input program. A few of these passes require type information for expressions, and therefore, need to call the resolver before the general resolve phase. There are three notable places where the resolver is invoked:
     144\begin{itemize}
     145\item
     146Generate default constructor, copy constructor and destructor for user-defined @struct@ types.
     147\item
     148Resolve @with@ statements (the same as in Pascal~\cite{pascal}), which introduces fields of a structure directly into a scope.
    130149\item
    131150Resolve @typeof@ expressions (cf. @decltype@ in \CC); note that this step may depend on symbols introduced by @with@ statements.
    132151\end{itemize}
    133 Since the compiler codebase is large and the new memory model mostly only benefits expression resolution, the old data structure is still kept, and a conversion pass happens before and after resolve phase. Rewriting every compiler module will take a long time, and whether the new model is correct is still unknown when started, therefore only the resolver is implemented with the new data structure.
    134 
    135 Since the constructor calls were one of the most expensive to resolve (reason will be shown in the next section), pre-resolve phase were taking more time after resolver moves to the more efficient new implementation. To better facilitate the new resolver, every step that requires type information are reintegrated as part of resolver.
    136 
    137 A by-product of this work is that the reversed dependence of @with@ statement and @typeof@ can now be handled. Previously, the compiler is unable to handle cases such as
     152
     153Since the constructor calls are one of the most expensive to resolve (reason given in~\VRef{s:SpecialFunctionLookup}), this pre-resolve phase was taking a large amount of time even after the resolver was changed to the more efficient new implementation. The problem is that multiple resolutions repeat a significant amount of work. Therefore, to better facilitate the new resolver, every step that requires type information should be integrated as part of the general resolver phase.
     154
     155A by-product of this work is that reversed dependence between @with@ statement and @typeof@ can now be handled. Previously, the compiler was unable to handle cases such as:
    138156\begin{cfa}
    139157struct S { int x; };
    140158S foo();
    141159typeof( foo() ) s; // type is S
    142 with (s) { 
     160with (s) {
    143161        x; // refers to s.x
    144162}
    145163\end{cfa}
    146 since type of @s@ is still unresolved when handling @with@ expressions. Instead, the new (and correct) approach is to evaluate @typeof@ expressions when the declaration is first seen, and it suffices because of the declaration-before-use rule.
     164since the type of @s@ is unresolved when handling @with@ expressions because the @with@ pass follows the @typeof@ pass (interchanging passes only interchanges the problem). Instead, the new (and correct) approach is to evaluate @typeof@ expressions when the declaration is first seen during resolution, and it suffices because of the declaration-before-use rule.
    147165
    148166
    149167\subsection{Special function lookup}
    150 
    151 Reducing the number of functions looked up for overload resolution is an effective way to gain performance when there are many overloads but most of them are trivially wrong. In practice, most functions have few (if any) overloads but there are notable exceptions. Most importantly, constructor @?{}@, destructor @^?{}@, and assignment @?=?@ are generated for every user-defined type, and in a large source file there can be hundreds of them. Furthermore, many calls to them are generated for initializing variables and passing arguments. This fact makes them the most overloaded and most called functions.
    152 
    153 In an object-oriented programming language, object has methods declared with their types, so a call such as @obj.f()@ only needs to perform lookup in the method table corresponding to type of @obj@. \CFA on the other hand, does not have methods, and all types are open (\ie new operations can be defined on them), so a similar approach will not work in general. However, the ``big 3'' operators have a unique property enforced by the language rules, such that the first parameter must have a reference type. Since \CFA does not have class inheritance, reference type must always match exactly. Therefore, argument-dependent lookup can be implemented for these operators, by using a dedicated symbol table.
    154 
    155 The lookup key used for the special functions is the mangled type name of the first parameter, which acts as the @this@ parameter in an object-oriented language. To handle generic types, the type parameters are stripped off, and only the base type is matched. Note that a constructor (destructor, assignment operator) taking arbitrary @this@ argument, for example @forall( dtype T ) void ?{}( T & );@ is not allowed, and it guarantees that if the @this@ type is known, all possible overloads can be found by searching with the given type. In case that the @this@ argument itself is overloaded, it is resolved first and all possible result types are used for lookup.
    156 
    157 Note that for the generated expressions, the particular variable for @this@ argument is fully known, without overloads, so the majority of constructor call resolutions only need to check for one given object type. Explicit constructor calls and assignment statements sometimes may require lookup for multiple types. In the extremely rare case that type of @this@ argument is yet unbound, everything will have to be checked, just like without the argument-dependent lookup algorithm; fortunately, this case almost never happens in practice. An example is found in the library function @new@:
     168\label{s:SpecialFunctionLookup}
     169
     170Reducing the number of function looked ups for overload resolution is an effective way to gain performance when there are many overloads but most of them are trivially wrong. In practice, most functions have few (if any) overloads but there are notable exceptions. Most importantly, constructor @?{}@, destructor @^?{}@, and assignment @?=?@ are generated for every user-defined type (@struct@ and @union@ in C), and in a large source file there can be hundreds of them. Furthermore, many calls are generated for initializing variables, passing arguments and copying values. This fact makes them the most overloaded and most called functions.
     171
     172In an object-oriented programming language, the object-method types are scoped within a class, so a call such as @obj.f()@ only needs to perform lookup in the method table corresponding to the type of @obj@. \CFA on the other hand, does not have methods, and all types are open, \ie new operations can be defined on them without inheritance; at best a \CFA type can be constrained by a translation unit. However, the ``big 3'' operators have a unique property enforced by the language rules: the first parameter must be a reference to its associated type, which acts as the @this@ parameter in an object-oriented language. Since \CFA does not have class inheritance, the reference type must always match exactly. Therefore, argument-dependent lookup can be implemented for these operators by using a dedicated, fast symbol-table.
     173
     174The lookup key for the special functions is the mangled type name of the first parameter. To handle generic types, the type parameters are stripped off, and only the base type is matched. Note a constructor (destructor, assignment operator) may not take an arbitrary @this@ argument, \eg @forall( dtype T ) void ?{}( T & )@, thus guaranteeing that if the @this@ type is known, all possible overloads can be found by searching with this given type. In the case where the @this@ argument itself is overloaded, it is resolved first and all possible result types are used for lookup.
     175
     176Note that for a generated expression, the particular variable for the @this@ argument is fully known, without overloads, so the majority of constructor-call resolutions only need to check for one given object type. Explicit constructor calls and assignment statements sometimes require lookup for multiple types. In the extremely rare case that the @this@-argument type is unbound, all necessary types are guaranteed to be checked, as for the previous lookup without the argument-dependent lookup; fortunately, this complex case almost never happens in practice. An example is found in the library function @new@:
    158177\begin{cfa}
    159178forall( dtype T | sized( T ), ttype TT | { void ?{}( T &, TT ); } )
    160179T * new( TT p ) { return &(*malloc()){ p }; }
    161180\end{cfa}
    162 as @malloc@ may return a pointer to any type, depending on context. 
    163 
    164 Interestingly, this particular line of code actually caused another complicated issue, where the unusually massive work of checking every constructor in presence makes the case even worse. Section~\ref{s:TtypeResolutionInfiniteRecursion} presents a detailed analysis for the problem.
    165 
    166 The ``callable'' operator @?()@ (cf. @operator()@ in \CC) could also be included in the special operator list, as it is usually only on user-defined types, and the restriction that first argument must be a reference seems reasonable in this case.
     181as @malloc@ may return a pointer to any type, depending on context.
     182
     183Interestingly, this particular declaration actually causes another complicated issue, making the complex checking of every constructor even worse. \VRef[Section]{s:TtypeResolutionInfiniteRecursion} presents a detailed analysis of this problem.
     184
     185The ``callable'' operator @?()@ (cf. @operator()@ in \CC) can also be included in this special operator list, as it is usually only on user-defined types, and the restriction that the first argument must be a reference seems reasonable in this case.
    167186
    168187
    169188\subsection{Improvement of function type representation}
    170189
    171 Since substituting type parameters with their bound types is one fundamental operation in many parts of resolver algorithm (particularly unification and environment binding), making as few copies of type nodes as possible helps reducing memory complexity. Even with the new memory management model, allocation is still a significant factor of resolver performance. Conceptually, operations on type nodes of AST should be performed in functional programming style, treating the data structure as immutable and only copy when necessary. The in-place mutation is a mere optimization that does not change logic of operations.
    172 The model was broken on function types by an inappropriate design. Function types require some special treatment due to the existence of assertions. In particular, it must be able to distinguish two different kinds of type parameter usage:
     190Since substituting type parameters with their bound types is one fundamental operation in many parts of resolver algorithm (particularly unification and environment binding), making as few copies of type nodes as possible helps reducing memory complexity. Even with the new memory management model, allocation is still a significant factor of resolver performance. Conceptually, operations on type nodes of the AST should be performed in functional-programming style, treating the data structure as immutable and only copying when necessary. The in-place mutation is a mere optimization that does not change the logic for operations.
     191
     192However, the model was broken for function types by an inappropriate design. Function types require special treatment due to the existence of assertions that constrain the types it supports. Specifically, it must be possible to distinguish two different kinds of type parameter usage:
    173193\begin{cfa}
    174194forall( dtype T ) void foo( T * t ) {
    175         forall( dtype U ) void bar( T * t, U * u ) { ... }
    176 }
    177 \end{cfa}
    178 Here, only @U@ is a free parameter in declaration of @bar@, as it appears in the function's own forall clause; while @T@ is not free.
    179 
    180 Moreover, the resolution algorithm also has to distinguish type bindings of multiple calls to the same function, for example with
     195        forall( dtype U ) void bar( @T@ * t, @U@ * u ) { ... }
     196}
     197\end{cfa}
     198Here, only @U@ is a free parameter in the nested declaration of function @bar@, as @T@ must be bound at the call site when resolving @bar@.
     199
     200Moreover, the resolution algorithm also has to distinguish type bindings of multiple calls to the same function, \eg:
    181201\begin{cfa}
    182202forall( dtype T ) int foo( T x );
    183 foo( foo( 1.0 ) );
    184 \end{cfa}
    185 The inner call has binding (T: double) while the outer call has binding (T: int). Therefore a unique representation of free parameters in each expression is required. This was previously done by creating a copy of the parameter declarations inside function type, and fixing references afterwards. However, fixing references is an inherently deep operation that does not work well with functional programming model, as it must be evaluated eagerly on the entire syntax tree representing the function type.
    186 
    187 The revised approach generates a unique ID value for each function call expression instance and represents an occurrence of free parameter type with a pair of generated ID and the original parameter declaration, so that references do not need to be fixed, and a shallow copy of function type is possible.
    188 
    189 Note that after the change, all declaration nodes in syntax tree representation maps one-to-one with the actual declarations in the program, and therefore are guaranteed to be unique. Such property can potentially enable more optimizations, and some related ideas are presented after Section~\ref{s:SharedSub-ExpressionCaseUniqueExpressions}.
     203int i = foo( foo( 1.0 ) );
     204\end{cfa}
     205The inner call has binding (T: double) while the outer call has binding (T: int). Therefore a unique representation for the free parameters is required in each expression. This type binding was previously done by creating a copy of the parameter declarations inside the function type and fixing references afterwards. However, fixing references is an inherently deep operation that does not work well with the functional-programming style, as it forces eager evaluation on the entire syntax tree representing the function type.
     206
     207The revised approach generates a unique ID value for each function call expression instance and represents an occurrence of a free-parameter type with a pair of generated ID and original parameter declaration, so references are unique and a shallow copy of the function type is possible.
     208
     209Note that after the change, all declaration nodes in the syntax-tree representation now map one-to-one with the actual declarations in the program, and therefore are guaranteed to be unique. This property can potentially enable more optimizations, and some related ideas are presented at the end of \VRef{s:SharedSub-ExpressionCaseUniqueExpressions}.
    190210
    191211
    192212\subsection{Improvement of pruning steps}
    193213
    194 A minor improvement for candidate elimination is to skip the step on the function overloads themselves and only perform on results of function application. As function calls are usually by name, the name resolution rule dictates that every function candidate necessarily has a different type; indirect function calls are rare, and when they do appear, they usually will not have many possible interpretations, and those rarely matches exactly in argument type. Since function types have a much more complex representation than data types (with multiple parameters and assertions), checking equality on them also takes longer.
    195 
    196 A brief test of this approach shows that the number of function overloads considered in expression resolution increases by a negligible amount of less than 1 percent, while type comparisons in candidate elimination are cut by more than half. Improvement is consistent over all \CFA source files in the test suite.
     214A minor improvement for candidate elimination is to skip the step on the function overloads and only check the results of function application. As function calls are usually by name (versus pointers to functions), the name resolution rule dictates that every function candidate necessarily has a different type; indirect function calls are rare, and when they do appear, there are even fewer cases with multiple interpretations, and these rarely match exactly in argument type. Since function types have a much more complex representation (with multiple parameters and assertions) than data types, checking equality on them also takes longer.
     215
     216A brief test of this approach shows that the number of function overloads considered in expression resolution increases by an amount of less than 1 percent, while type comparisons in candidate elimination are reduced by more than half. This improvement is consistent over all \CFA source files in the test suite.
    197217
    198218
     
    200220\label{s:SharedSub-ExpressionCaseUniqueExpressions}
    201221
    202 Unique expression denotes an expression that must be evaluated only once, to prevent unwanted side effects. It is currently only a compiler artifact, generated on tuple member expression of the form
     222Unique expression denotes an expression evaluated only once to prevent unwanted side effects. It is currently only a compiler artifact, generated for tuple-member expression of the form:
    203223\begin{cfa}
    204224struct S { int a; int b; };
     
    206226s.[a, b]; // tuple member expression, type is [int, int]
    207227\end{cfa}
    208 If the aggregate expression contains function calls, it cannot be evaluated multiple times:
     228If the aggregate expression is function call, it cannot be evaluated multiple times:
    209229\begin{cfa}
    210230S makeS();
    211 makeS().[a, b]; // this should only make one S
     231makeS().[a, b]; // this should only generate a unique S
    212232\end{cfa}
    213233Before code generation, the above expression is internally represented as
     
    226246\end{cfa}
    227247at code generation, where @_unique_var@ and @_unique_var_evaluated@ are generated variables whose scope covers all appearances of the same expression.
    228 
    229 Note that although the unique expression is only used for tuple expansion now, it is a generally useful construction, and can be seen in other languages, such as Scala's @lazy val@~\cite{Scala}; therefore it could be worthwhile to introduce the unique expression to a broader context in \CFA and even make it directly available to programmers.
    230 
    231 In the compiler's visitor pattern, however, this creates a problem where multiple paths to a logically unique expression exist, so it may be modified more than once and become ill-formed; some specific intervention is required to ensure that unique expressions are only visited once. Furthermore, a unique expression appearing in more than one places will be copied on mutation so its representation is no longer unique. Some hacks are required to keep it in sync, and the methods are different when mutating the unique expression instance itself or its underlying expression.
    232 
    233 Example when mutating the underlying expression (visit-once guard)
     248The conditional check ensures a single call to @makeS()@ even though there are logically multiple calls because of the tuple field expansion.
     249
     250Note that although the unique expression is only used for tuple expansion now, it is a generally useful construction, and is seen in other programming languages, such as Scala's @lazy val@~\cite{Scala}; therefore it may be worthwhile to introduce the unique expression to a broader context in \CFA and even make it directly available to programmers.
     251
     252In the compiler's visitor pattern, however, this creates a problem where multiple paths to a logically unique expression exist, so it may be modified more than once and become ill-formed; some specific intervention is required to ensure unique expressions are only visited once. Furthermore, a unique expression appearing in more than one places is copied on mutation so its representation is no longer unique.
     253
     254Currently, special cases are required to keep everything synchronized, and the methods are different when mutating the unique expression instance itself or its underlying expression:
     255\begin{itemize}
     256\item
     257When mutating the underlying expression (visit-once guard)
    234258\begin{cfa}
    235259void InsertImplicitCalls::previsit( const ast::UniqueExpr * unqExpr ) {
    236         if ( visitedIds.count( unqExpr->id ) ) visit_children = false;
     260        @if ( visitedIds.count( unqExpr->id ) ) visit_children = false;@
    237261        else visitedIds.insert( unqExpr->id );
    238262}
    239263\end{cfa}
    240 Example when mutating the unique instance itself, which actually creates copies
     264\item
     265When mutating the unique instance itself, which actually creates copies
    241266\begin{cfa}
    242267auto mutExpr = mutate( unqExpr ); // internally calls copy when shared
    243 if ( ! unqMap.count( unqExpr->id ) ) {
     268@if ( ! unqMap.count( unqExpr->id ) ) {@
    244269        ...
    245270} else {
     
    248273}
    249274\end{cfa}
    250 Such workaround seems difficult to be fit into a common visitor template. This suggests the memory model may need different kinds of nodes to accurately represent the syntax tree.
    251 
    252 Together with the fact that declaration nodes are always unique, it is possible that AST nodes can be classified by three different types:
    253 \begin{itemize}
    254 \item
    255 \textbf{Strictly unique} with only one owner (declarations);
    256 \item
    257 \textbf{Logically unique} with (possibly) many owners but should not be copied (unique expression example presented here);
    258 \item
    259 \textbf{Shared} by functional programming model, which assume immutable data structure and are copied on mutation.
     275\end{itemize}
     276Such workarounds are difficult to fit into the common visitor pattern, which suggests the memory model may need different kinds of nodes to accurately represent this feature in the AST.
     277
     278Given that declaration nodes are unique, it is possible for AST nodes to be divided into three different types:
     279\begin{itemize}
     280\item
     281\textbf{Singleton} with only one owner (declarations);
     282\item
     283\textbf{No-copy} with multiple owners but cannot be copied (unique expression example presented here);
     284\item
     285\textbf{Copy} by functional-programming style, which assumes immutable data structures that are copied on mutation.
    260286\end{itemize}
    261287The boilerplate code can potentially handle these three cases differently.
     
    264290\section{Analysis of resolver algorithm complexity}
    265291
    266 The focus of this chapter is to identify and analyze some realistic cases that cause resolver algorithm to have an exponential run time. As previous work has shown [3], the overload resolution problem in \CFA has worst-case exponential complexity; however, only few specific patterns can trigger the exponential complexity in practice. Implementing heuristic-based optimization for those selected cases is helpful to alleviate the problem.
     292The focus of this section is to identify and analyze some realistic cases that cause the resolver algorithm to have an exponential runtime. As previous work has shown~\cite[\S~4.2.1]{Moss19}, the overload resolution problem in \CFA has worst-case exponential complexity; however, only few specific patterns can trigger the exponential complexity in practice. Implementing heuristic-based optimization for those selected cases is helpful to alleviate the problem.
    267293
    268294
     
    270296\label{s:UnboundReturnType}
    271297
    272 The interaction of return type overloading and polymorphic functions creates this problem of function calls with unbound return type, and is further complicated by the presence of assertions.
     298The interaction of return-type overloading and polymorphic functions creates function calls with unbounded return-type, and is further complicated by the presence of assertions.
    273299The prime example of a function with unbound return type is the type-safe version of C @malloc@:
    274300\begin{cfa}
    275 // size deduced from type, so no need to provide the size argument
    276 forall( dtype T | sized( T ) ) T * malloc( void );
    277 \end{cfa}
    278 Unbound return type can be problematic in resolver algorithm complexity because a single match of function call with unbound return type may create multiple candidates. In the worst case, consider a function declared to return any @otype@:
     301forall( dtype T | sized( T ) )
     302T * malloc( void ) { return (T *)malloc( sizeof(T) ); } // call C malloc
     303int * i = malloc();  // type deduced from left-hand size $\(\Rightarrow\)$ no size argument or return cast
     304\end{cfa}
     305An unbound return-type is problematic in resolver complexity because a single match of a function call with an unbound return type may create multiple candidates. In the worst case, consider a function declared that returns any @otype@ (defined \VPageref{otype}):
    279306\begin{cfa}
    280307forall( otype T ) T anyObj( void );
    281308\end{cfa}
    282 As the resolver attempts to satisfy the otype constraint on @T@, a single call to @anyObj()@ without the result type known creates at least as many candidates as the number of complete types currently in scope; with generic types it becomes even worse, for example, assuming a declaration of generic pair is available at that point:
     309As the resolver attempts to satisfy the otype constraint on @T@, a call to @anyObj()@ in an expression, without the result type known, creates at least as many candidates as the number of complete types currently in scope; with generic types it becomes even worse, \eg assuming a declaration of a generic @pair@ is available at that point:
    283310\begin{cfa}
    284311forall( otype T, otype U ) struct pair { T first; U second; };
    285312\end{cfa}
    286 Then an @anyObj()@ call can result in arbitrarily complex types, such as @pair( pair( int,int ), pair( int,int ) )@, and the depth can grow indefinitely until the specified parameter depth limit, thus creating exponentially many candidates. However, the expected types allowed by parent expressions are practically very few, so most of those interpretations are invalid; if the result type is never bound up to top level, by the semantic rules it is ambiguous if there are more than one valid bindings, and resolution can fail fast. It is therefore reasonable to delay resolving assertions on an unbound parameter in return type; however, with the current cost model, such behavior may further cause irregularities in candidate selection, such that the presence of assertions can change the preferred candidate, even when order of expression costs are supposed to stay the same. Detailed analysis of this issue will be presented later, in the correctness part.
     313Then an @anyObj()@ call can result in arbitrarily complex types, such as @pair( pair( int, int ), pair( int, int ) )@, and the depth can grow indefinitely until a specified parameter-depth limit, thus creating exponentially many candidates. However, the expected types allowed by parent expressions are practically very few, so most of those interpretations are invalid; if the result type is never bound up to the top level, by the semantic rules it is ambiguous if there is more than one valid binding and resolution fails quickly. It is therefore reasonable to delay resolving assertions on an unbound parameter in a return type; however, with the current cost model, such behavior may further cause irregularities in candidate selection, such that the presence of assertions can change the preferred candidate, even when order of expression costs are supposed to stay the same. A detailed analysis of this issue is presented in \VRef{s:AnalysisTypeSystemCorrectness}.
    287314
    288315
     
    290317\label{s:TtypeResolutionInfiniteRecursion}
    291318
    292 @ttype@ (``tuple type'') is a relatively new addition to the language that attempts to provide type-safe variadic argument semantics. Unlike regular @dtype@ parameters, @ttype@ is only valid in function parameter list, and may only appear once as the type of last parameter. At the call site, a @ttype@ parameter is bound to the tuple type of all remaining function call arguments.
     319@ttype@ (``tuple type'') is a relatively new addition to the language that attempts to provide type-safe variadic argument semantics. Unlike regular @dtype@ parameters, @ttype@ is only valid in a function parameter-list, and may only appear once as the last parameter type. At the call site, a @ttype@ parameter is bound to the tuple type of all remaining function-call arguments.
    293320
    294321There are two kinds of idiomatic @ttype@ usage: one is to provide flexible argument forwarding, similar to the variadic template in \CC (\lstinline[language=C++]|template<typename... args>|), as shown below in the implementation of @unique_ptr@
     
    298325        T * data;
    299326};
    300 forall( dtype T | sized( T ), ttype Args | { void ?{}( T &, Args ); })
    301 void ?{}( unique_ptr( T ) & this, Args args ) {
    302         this.data = new( args );
    303 }
    304 \end{cfa}
    305 the other is to implement structural recursion in the first-rest manner:
    306 \begin{cfa}
    307 forall( otype T, ttype Params | { void process( T ); void func( Params ); })
     327forall( dtype T | sized( T ), @ttype Args@ | { void ?{}( T &, Args ); })
     328void ?{}( unique_ptr( T ) & this, Args @args@ ) {
     329        this.data = new( @args@ );  // forward constructor arguments to dynamic allocator
     330}
     331\end{cfa}
     332The other usage is to implement structural recursion in the first-rest pattern:
     333\begin{cfa}
     334forall( otype T, @ttype Params@ | { void process( T ); void func( Params ); })
    308335void func( T arg1, Params p ) {
    309336        process( arg1 );
    310         func( p );
    311 }
    312 \end{cfa}
    313 For the second use case, it is important that the number of parameters in the recursive call go down, since the call site must deduce all assertion candidates, and that is only possible if by just looking at argument types (and not their values), the recursion is known to be completed in a finite number of steps.
    314 
    315 In recent experiments, however, some flaw in the type binding rules can lead to the first kind of @ttype@ use case produce an invalid candidate that the resolver enters an infinite loop.
    316 
    317 This bug was discovered in an attempt to raise assertion recursive depth limit and one of the library program takes exponentially longer time to compile. The cause of the problem is identified to be the following set of functions.
    318 File @memory.cfa@ contains
    319 \begin{cfa}
    320 #include "memory.hfa"
    321 #include "stdlib.hfa"
    322 \end{cfa}
    323 where file @memory.hfa@ contains the @unique_ptr@ declaration above, and two other similar functions with @ttype@ parameter:
    324 \begin{cfa}
    325 forall( dtype T | sized( T ), ttype Args | { void ?{}( T &, Args ); }) {
     337        func( @p@ );  // recursive call until base case of one argument
     338}
     339\end{cfa}
     340For the second use case, it is imperative the number of parameters in the recursive call goes down, since the call site must deduce all assertion candidates, and that is only possible if by observation of the argument types (and not their values), the recursion is known to be completed in a finite number of steps.
     341
     342In recent experiments, however, a flaw in the type-binding rules can lead to the first kind of @ttype@ use case producing an invalid candidate and the resolver enters an infinite loop.
     343This bug was discovered in an attempt to raise the assertion recursive-depth limit and one of the library programs took exponentially longer to compile. The cause of the problem is the following set of functions:
     344\begin{cfa}
     345// unique_ptr  declaration from above
     346
     347forall( dtype T | sized( T ), ttype Args | { void ?{}( T &, Args ); } ) { // distribute forall clause
    326348        void ?{}( counter_data( T ) & this, Args args );
    327349        void ?{}( counter_ptr( T ) & this, Args args );
    328350        void ?{}( unique_ptr( T ) & this, Args args );
    329351}
    330 \end{cfa}
    331 File @stdlib.hfa@ contains
    332 \begin{cfa}
     352
    333353forall( dtype T | sized( T ), ttype TT | { void ?{}( T &, TT ); } )
    334 T * new( TT p ) { return &(*malloc()){ p }; }
    335 \end{cfa}
    336 
    337 In the expression @(*malloc()){p}@, the type of object being constructed is yet unknown, since the return type information is not immediately provided. That caused every constructor to be searched, and while normally a bound @ttype@ cannot be unified with any free parameter, it is possible with another free @ttype@. Therefore in addition to the correct option provided by assertion, 3 wrong options are examined, each of which again requires the same assertion, for an unknown base type T and @ttype@ arguments, and that becomes an infinite loop, until the specified recursion limit and resolution is forced to fail. Moreover, during the recursion steps, number of candidates grows exponentially, since there are always 3 options at each step.
    338 
    339 Unfortunately, @ttype@ to @ttype@ binding is necessary, to allow calling the function provided by assertion indirectly.
    340 \begin{cfa}
    341 forall( dtype T | sized( T ), ttype Args | { void ?{}( T &, Args ); })
    342 void ?{}( unique_ptr( T ) & this, Args args ) { this.data = (T * )new( args ); }
    343 \end{cfa}
    344 Here the constructor assertion is used for the @new( args )@ call.
     354T * new( TT p ) { return @&(*malloc()){ p };@ }
     355\end{cfa}
     356In the expression @(*malloc()){p}@, the type of the object being constructed is unknown, since the return-type information is not immediately available. That causes every constructor to be searched, and while normally a bound @ttype@ cannot be unified with any free parameter, it is possible with another free @ttype@. Therefore, in addition to the correct option provided by the assertion, 3 wrong options are examined, each of which again requires the same assertion, for an unknown base-type @T@ and @ttype@ argument, which becomes an infinite loop until the specified recursion limit and resolution is fails. Moreover, during the recursion steps, the number of candidates grows exponentially, since there are always 3 options at each step.
     357
     358Unfortunately, @ttype@ to @ttype@ binding is necessary, to allow indirectly calling a function provided in an assertion.
     359\begin{cfa}
     360forall( dtype T | sized( T ), ttype Args | { @void ?{}( T &, Args );@ })
     361void ?{}( unique_ptr( T ) & this, Args args ) { this.data = (T *)@new( args )@; } // constructor call
     362\end{cfa}
     363Here the constructor assertion is used by the @new( args )@ call to indirectly call the constructor on the allocated storage.
    345364Therefore, it is hard, perhaps impossible, to solve this problem by tweaking the type binding rules. An assertion caching algorithm can help improve this case by detecting cycles in recursion.
    346365
    347 Meanwhile, without the caching algorithm implemented, some changes in the \CFA source code are enough to eliminate this problem, at least in the current codebase. Note that the issue only happens with an overloaded variadic function, which rarely appears in practice, since the idiomatic use cases are for argument forwarding and self-recursion. The only overloaded @ttype@ function so far discovered in all of \CFA standard library code is the constructor, and by utilizing the argument-dependent lookup process described in Section~\ref{s:UnboundReturnType}, adding a cast before constructor call gets rid of the issue.
    348 \begin{cfa}
    349 T * new( TT p ) { return &(*(T * )malloc()){ p }; }
     366Meanwhile, without a caching algorithm implemented, some changes in the \CFA source code are enough to eliminate this problem, at least in the current codebase. Note that the issue only happens with an overloaded variadic function, which rarely appears in practice, since the idiomatic use cases are for argument forwarding and self-recursion. The only overloaded @ttype@ function so far discovered in all of \CFA standard library is the constructor, and by utilizing the argument-dependent lookup process described in \VRef{s:UnboundReturnType}, adding a cast before the constructor call removes the issue.
     367\begin{cfa}
     368T * new( TT p ) { return &(*@(T * )@malloc()){ p }; }
    350369\end{cfa}
    351370
     
    353372\subsection{Reused assertions in nested generic type}
    354373
    355 The following test of deeply nested dynamic generic type reveals that locally caching reused assertions is necessary, rather than just a resolver optimization, because recomputing assertions can result in bloated generated code size:
     374The following test of deeply nested, dynamic generic type reveals that locally caching reused assertions is necessary, rather than just a resolver optimization, because recomputing assertions can result in bloated generated code size:
    356375\begin{cfa}
    357376struct nil {};
     
    361380int main() {
    362381        #if   N==0
    363         nil x;   
     382        nil @x@;
    364383        #elif N==1
    365         cons( size_t, nil ) x;
     384        cons( size_t, nil ) @x@;
    366385        #elif N==2
    367         cons( size_t, cons( size_t, nil ) ) x;
     386        cons( size_t, cons( size_t, nil ) ) @x@;
    368387        #elif N==3
    369         cons( size_t, cons( size_t, cons( size_t, nil ) ) ) x;
     388        cons( size_t, cons( size_t, cons( size_t, nil ) ) ) @x@;
    370389        // similarly for N=4,5,6
    371390        #endif
    372391}
    373392\end{cfa}
    374 At the declaration of @x@, it is implicitly initialized by generated constructor call, whose signature is given by
     393At the declaration of @x@, it is implicitly initialized by generated constructor call, with signature:
    375394\begin{cfa}
    376395forall( otype L, otype R ) void ?{}( cons( L, R ) & );
    377396\end{cfa}
    378 Note that the @otype@ constraint contains 4 assertions:
     397where the @otype@ constraint contains the 4 assertions:\label{otype}
    379398\begin{cfa}
    380399void ?{}( L & ); // default constructor
     
    383402L & ?=?( L &, L & ); // assignment
    384403\end{cfa}
    385 Now since the right hand side of outermost cons is again a cons, recursive assertions are required. When the compiler cannot cache and reuse already resolved assertions, it becomes a problem, as each of those 4 pending assertions again asks for 4 more assertions one level below. Without any caching, number of resolved assertions grows exponentially, while that is obviously unnecessary since there are only $n+1$ different types involved. Even worse, this causes exponentially many wrapper functions generated later at the codegen step, and results in huge compiled binary.
    386 
    387 \begin{table}[h]
     404
     405\begin{table}[htb]
     406\centering
    388407\caption{Compilation results of nested cons test}
     408\label{t:NestedConsTest}
    389409\begin{tabular}{|r|r|r|}
    390410\hline
     
    402422\end{table}
    403423
    404 As the local functions are implemented by emitting executable code on the stack~\cite{gcc-nested-func}, it eventually means that compiled code also has exponential run time. This problem has evident practical implications, as nested collection types are frequently used in real production code.
    405 
     424Now since the right hand side of outermost cons is again a cons, recursive assertions are required. \VRef[Table]{t:NestedConsTest} shows when the compiler does not cache and reuse already resolved assertions, it becomes a problem, as each of these 4 pending assertions again asks for 4 more assertions one level below. Without caching, the number of resolved assertions grows exponentially, which is unnecessary since there are only $n+1$ different types involved. Even worse, this problem causes exponentially many wrapper functions to be generated at the backend, resulting in a huge binary. As the local functions are implemented by emitting executable code on the stack~\cite{gcc-nested-func}, it means that compiled code also has exponential run time. This problem has practical implications, as nested collection types are frequently used in real production code.
    406425
    407426\section{Analysis of type system correctness}
     427\label{s:AnalysisTypeSystemCorrectness}
    408428
    409429In Moss' thesis~\cite[\S~4.1.2,~p.~45]{Moss19}, the author presents the following example:
     
    412432\begin{cfa}
    413433void f( int );
    414 double g$_1$( int );
    415 int g$_2$( long );
     434double g$\(_1\)$( int );
     435int g$\(_2\)$( long );
    416436f( g( 42 ) );
    417437\end{cfa}
     
    422442From the set of candidates whose parameter and argument types have been unified and whose assertions have been satisfied, those whose sub-expression interpretations have the smallest total cost of conversion are selected ... The total cost of conversion for each of these candidates is then calculated based on the implicit conversions and polymorphism involved in adapting the types of the sub-expression interpretations to the formal parameter types.
    423443\end{quote}
    424 With this model, the algorithm picks @g1@ in resolving the @f( g( 42 ) )@ call, which seems to be undesirable.
    425 
    426 There are further evidence that shows the Bilson model is fundamentally incorrect, following the discussion of unbound return type in Section~\ref{s:UnboundReturnType}. By the conversion cost specification, a binding from a polymorphic type parameter to a concrete type incurs a polymorphic cost of 1. It remains unspecified \emph{when} the type parameters should become bound. When the parameterized types appear in the function parameters, they can be deduced from the argument type, and there is no ambiguity. In the unbound return case, however, the binding may happen at any stage in expression resolution, therefore it is impossible to define a unique local conversion cost. Note that type binding happens exactly once per parameter in resolving the entire expression, so the global binding cost is unambiguously 1.
    427 
    428 As per the current compiler implementation, it does have a notable inconsistency in handling such case. For any unbound parameter that does \emph{not} come with an associated assertion, it remains unbound to the parent expression; for those that does however, they are immediately bound in the assertion resolution step, and concrete result types are used in the parent expressions.
    429 
     444With this model, the algorithm picks @g1@ in resolving the @f( g( 42 ) )@ call, which is undesirable.
     445
     446There is further evidence that shows the Bilson model is fundamentally incorrect, following the discussion of unbound return type in \VRef{s:UnboundReturnType}. By the conversion-cost specification, a binding from a polymorphic type-parameter to a concrete type incurs a polymorphic cost of 1. It remains unspecified \emph{when} the type parameters should become bound. When the parameterized types appear in function parameters, they can be deduced from the argument type, and there is no ambiguity. In the unbound return case, however, the binding may happen at any stage in expression resolution, therefore it is impossible to define a unique local conversion cost. Note that type binding happens exactly once per parameter in resolving the entire expression, so the global binding cost is unambiguously 1.
     447
     448In the current compiler implementation, there is a notable inconsistency in handling this case. For any unbound parameter that does \emph{not} come with an associated assertion, it remains unbound to the parent expression; for those that do, however, they are immediately bound in the assertion resolution step, and concrete result types are used in the parent expressions.
    430449Consider the following example:
    431450\begin{cfa}
     
    433452void h( int * );
    434453\end{cfa}
    435 The expression @h( f() )@ eventually has a total cost of 1 from binding (T: int), but in the eager resolution model, the cost of 1 may occur either at call to @f@ or at call to @h@, and with the assertion resolution triggering a binding, the local cost of @f()@ is (0 poly, 0 spec) with no assertions, but (1 poly, -1 spec) with an assertion:
    436 \begin{cfa}
    437 forall( dtype T | { void g( T * ); } ) T * f( void );
     454The expression @h( f() )@ eventually has a total cost of 1 from binding (T: int), but in the eager-resolution model, the cost of 1 may occur either at the call to @f@ or at call to @h@, and with the assertion resolution triggering a binding, the local cost of @f()@ is (0 poly, 0 spec) with no assertions, but (1 poly, -1 spec) with an assertion:
     455\begin{cfa}
     456forall( dtype T | @{ void g( T * ); }@ ) T * f( void );
    438457void g( int * );
    439458void h( int * );
    440459\end{cfa}
    441 and that contradicts the principle that adding assertions should make expression cost lower. Furthermore, the time at which type binding and assertion resolution happens is an implementation detail of the compiler, but not a part of language definition. That means two compliant \CFA compilers, one performing immediate assertion resolution at each step, and one delaying assertion resolution on unbound types, can produce different expression costs and therefore different candidate selection, making the language rule itself partially undefined and therefore unsound. By the above reasoning, the updated cost model using global sum of costs should be accepted as the standard. It also allows the compiler to freely choose when to resolve assertions, as the sum of total costs is independent of that choice; more optimizations regarding assertion resolution can also be implemented.
     460and that contradicts the principle that adding assertions should make expression cost lower. Furthermore, the time at which type binding and assertion resolution happens is an implementation detail of the compiler, not part of the language definition. That means two compliant \CFA compilers, one performing immediate assertion resolution at each step, and one delaying assertion resolution on unbound types, can produce different expression costs and therefore different candidate selection, making the language rule itself partially undefined, and therefore, unsound. By the above reasoning, the updated cost model using global sum of costs should be accepted as the standard. It also allows the compiler to freely choose when to resolve assertions, as the sum of total costs is independent of that choice; more optimizations regarding assertion resolution can also be implemented.
    442461
    443462
    444463\section{Timing results}
    445464
    446 For the timing results presented here, the \CFA compiler is built with gcc 9.3.0, and tested on a server machine running Ubuntu 20.04, 64GB RAM and 32-core 2.2 GHz CPU, results reported by the time command, and using only 8 cores in parallel such that the time is close to the case with 100% CPU utilization on a single thread.
    447 
    448 On the most recent build, the \CFA standard library (~1.3 MB of source code) compiles in 4 minutes 47 seconds total processor time (single thread equivalent), with the slowest file taking 13 seconds. The test suite (178 test cases, ~2.2MB of source code) completes within 25 minutes total processor time,\footnote{Including a few runtime tests; total time spent in compilation is approximately 21 minutes.} with the slowest file taking 23 seconds. In contrast, the library build on old compiler takes 85 minutes total, 5 minutes for the slowest file. Full test suite takes too long with old compiler build and is therefore not run, but the slowest test cases take approximately 5 minutes. Overall, the most recent build compared to old build in April 2020, before the project started, is consistently faster by a factor of 20.
    449 
    450 Additionally, 6 selected \CFA source files with distinct features from library and test suite are used to test compiler performance after each of the optimizations are implemented. Test files are from the most recent build and run through C preprocessor to eliminate the factor of header file changes. The selected tests are:
    451 \begin{itemize}
    452 \item
    453 @lib/fstream@ (112 KB)\footnote{File sizes are after preprocessing, with no line information (\lstinline|gcc -E -P|).}: implementation of I/O library
     465For the timing results presented here, the \CFA compiler is built with gcc 9.3.0, and tested on a server machine running Ubuntu 20.04, 64GB RAM and 32-core 2.2 GHz CPU.
     466Timing is reported by the @time@ command and an experiment is run using 8 cores, where each core is at 100\% CPU utilization.
     467
     468On the most recent build, the \CFA standard library ($\approx$1.3 MB of source code) compiles in 4 minutes 47 seconds total processor time (single thread equivalent), with the slowest file taking 13 seconds. The test suite (178 test cases, $\approx$2.2MB of source code) completes within 25 minutes total processor time,
     469% PAB: I do not understand this footnote.
     470%\footnote{Including a few runtime tests; total time spent in compilation is approximately 21 minutes.}
     471with the slowest file taking 23 seconds. In contrast, the library build with the old compiler takes 85 minutes total, 5 minutes for the slowest file. The full test-suite takes too long with old compiler build and is therefore not run, but the slowest test cases take approximately 5 minutes. Overall, the most recent build compared to an old build is consistently faster by a factor of 20.
     472
     473Additionally, 6 selected \CFA source files with distinct features from the library and test suite are used to illustrate the compiler performance change after each of the implemented optimizations. Test files are from the most recent build and run through the C preprocessor to expand header file, perform macro expansions, but no line number information (@gcc -E -P@).
     474\VRef[Table]{t:SelectedFileByCompilerBuild} shows the selected tests:
     475\begin{itemize}
     476\item
     477@lib/fstream@ (112 KB)
    454478\item
    455479@lib/mutex@ (166 KB): implementation of concurrency primitive
     
    459483@lib/stdlib@ (64 KB): type-safe wrapper to @void *@-based C standard library functions
    460484\item
    461 @test/ISO2@ (55 KB): application of I/O library
     485@test/io2@ (55 KB): application of I/O library
    462486\item
    463487@test/thread@ (188 KB): application of threading library
    464488\end{itemize}
    465 
    466 The \CFA compiler builds are picked from git commit history that passed the test suite, and implement the optimizations incrementally:
    467 \begin{itemize}
    468 \item
    469 \#0 is the first working build of new AST data structure
     489versus \CFA compiler builds picked from the git commit history that implement the optimizations incrementally:
     490\begin{itemize}
     491\item
     492old resolver
     493\item
     494\#0 is the first working build of the new AST data structure
    470495\item
    471496\#1 implements special symbol table and argument-dependent lookup
    472497\item
    473 \#2 implements late assertion satisfaction
    474 \item
    475 \#3 implements revised function type representation
    476 \item
    477 \#4 skips pruning on expressions with function type (most recent build)
    478 \end{itemize}
    479 The old resolver with no memory sharing and none of the optimizations above is also tested.
    480 \begin{table}
     498\#2 implements late assertion-satisfaction
     499\item
     500\#3 implements revised function-type representation
     501\item
     502\#4 skips pruning on expressions for function types (most recent build)
     503\end{itemize}
     504Reading left to right for a test shows the benefit of each optimization on the cost of compilation.
     505
     506\begin{table}[htb]
     507\centering
    481508\caption{Compile time of selected files by compiler build, in seconds}
     509\label{t:SelectedFileByCompilerBuild}
    482510\begin{tabular}{|l|r|r|r|r|r|r|}
    483511\hline
     
    502530\end{table}
    503531
    504 
    505532\section{Conclusion}
    506533
    507 Over the course of 8 months of active research and development in \CFA type system and compiler algorithm, performance of the reference \CFA compiler, cfa-cc, has been greatly improved, allowing mid-sized \CFA programs to be compiled and built reasonably fast. As there are also ongoing efforts in the team on building a standard library, evaluating the runtime performance, and attempting to incorporate \CFA with existing software written in C, this project is especially meaningful for practical purposes.
    508 
    509 Analysis conducted in the project were based significantly on heuristics and practical evidence, as the theoretical bounds and average cases for the expression resolution problem differ. This approach was difficult at start to follow, with an unacceptably slow compiler, since running the program through debugger and validation tools (\eg @gdb@, @valgrind@) adds another order of magnitude to run time, which was already in minutes. However, near the end of the project, many significant improvements have already been made and new optimizations can be tested immediately. The positive feedback in development cycle benefits the \CFA team as a whole, more than just for the compiler optimizations.
    510 
    511 Some potential issues of the language that may happen frequently in practice have been identified. Due to the time constraint and complex nature of these problems, a handful of them remain unsolved, but some constructive proposals are made. Notably, introducing a local assertion cache in the resolver is a common solution for a few remaining problems, so that should be the focus of work soon.
    512 
    513 The \CFA team are planning on a public alpha release of the language as the compiler performance becomes promising, and other parts of the system, such as a standard library, are also being enhanced. Ideally, the remaining problems should be resolved before release, and the solutions will also be integral to drafting a formal specification.
     534Over the course of 8 months of active research and development of the \CFA type system and compiler algorithms, performance of the reference \CFA compiler, cfa-cc, has been greatly improved. Now, mid-sized \CFA programs are compiled reasonably fast. Currently, there are ongoing efforts by the \CFA team to augment the standard library and evaluate its runtime performance, and incorporate \CFA with existing software written in C; therefore this project is especially meaningful for these practical purposes.
     535
     536Accomplishing this work was difficult. Analysis conducted in the project is based significantly on heuristics and practical evidence, as the theoretical bounds and average cases for the expression resolution problem differ. As well, the slowness of the initial compiler made attempts to understand why and where problems exist extremely difficult because both debugging and validation tools (\eg @gdb@, @valgrind@, @pref@) further slowed down compilation time. However, by the end of the project, I had found and fixed several significant problems and new optimizations are easier to introduce and test. The reduction in the development cycle benefits the \CFA team as a whole.
     537
     538Some potential issues of the language, which happen frequently in practice, have been identified. Due to the time constraint and complex nature of these problems, a handful of them remain unsolved, but some constructive proposals are made. Notably, introducing a local assertion cache in the resolver is a reasonable solution for a few remaining problems, so that should be the focus of future work.
     539
     540The \CFA team are planning on a public alpha release of the language as the compiler performance, given my recent improvements, is now useable. Other parts of the system, such as the standard library, have made significant gains due to the speed up in the development cycle. Ideally, the remaining problems should be resolved before release, and the solutions will also be integral to drafting a formal specification.
    514541
    515542\addcontentsline{toc}{section}{\refname}
  • doc/theses/fangren_yu_COOP_S20/Report.tex

    r342af53 r8e4aa05  
    1717\usepackage[usenames]{color}
    1818\input{common}                                          % common CFA document macros
    19 \usepackage[dvips,plainpages=false,pdfpagelabels,pdfpagemode=UseNone,colorlinks=true,pagebackref=true,linkcolor=blue,citecolor=blue,urlcolor=blue,pagebackref=true,breaklinks=true]{hyperref}
     19\usepackage[dvips,plainpages=false,pdfpagelabels,pdfpagemode=UseNone,colorlinks=true,linkcolor=blue,citecolor=blue,urlcolor=blue,pagebackref=true,breaklinks=true]{hyperref}
    2020\usepackage{breakurl}
    2121\urlstyle{sf}
  • doc/theses/thierry_delisle_PhD/thesis/Makefile

    r342af53 r8e4aa05  
    88BibTeX = BIBINPUTS=${TeXLIB} && export BIBINPUTS && bibtex
    99
    10 MAKEFLAGS = --no-print-directory --silent
     10MAKEFLAGS = --no-print-directory # --silent
    1111VPATH = ${Build} ${Figures}
    1212
     
    3232        emptytree \
    3333        fairness \
     34        io_uring \
     35        pivot_ring \
    3436        system \
    3537}
     
    4345## Define the documents that need to be made.
    4446all: thesis.pdf
    45 thesis.pdf: ${TEXTS} ${FIGURES} ${PICTURES} glossary.tex local.bib
     47thesis.pdf: ${TEXTS} ${FIGURES} ${PICTURES} thesis.tex glossary.tex local.bib
    4648
    4749DOCUMENT = thesis.pdf
     
    4951
    5052# Directives #
     53
     54.NOTPARALLEL:                                           # cannot make in parallel
    5155
    5256.PHONY : all clean                                      # not file names
     
    8185        ${LaTeX} $<
    8286
    83 build/fairness.svg : fig/fairness.py | ${Build}
    84         python3 $< $@
    85 
    8687## Define the default recipes.
    8788
     
    105106        sed -i 's/$@/${Build}\/$@/g' ${Build}/$@_t
    106107
     108build/fairness.svg : fig/fairness.py | ${Build}
     109        python3 $< $@
     110
    107111## pstex with inverted colors
    108112%.dark.pstex : fig/%.fig Makefile | ${Build}
  • doc/theses/thierry_delisle_PhD/thesis/local.bib

    r342af53 r8e4aa05  
    512512}
    513513
     514@manual{MAN:bsd/kqueue,
     515  title = {KQUEUE(2) - FreeBSD System Calls Manual},
     516  url   = {https://www.freebsd.org/cgi/man.cgi?query=kqueue},
     517  year  = {2020},
     518  month = {may}
     519}
     520
    514521% Apple's MAC OS X
    515522@manual{MAN:apple/scheduler,
     
    577584
    578585% --------------------------------------------------
     586% Man Pages
     587@manual{MAN:open,
     588  key        = "open",
     589  title      = "open(2) Linux User's Manual",
     590  year       = "2020",
     591  month      = "February",
     592}
     593
     594@manual{MAN:accept,
     595  key        = "accept",
     596  title      = "accept(2) Linux User's Manual",
     597  year       = "2019",
     598  month      = "March",
     599}
     600
     601@manual{MAN:select,
     602  key        = "select",
     603  title      = "select(2) Linux User's Manual",
     604  year       = "2019",
     605  month      = "March",
     606}
     607
     608@manual{MAN:poll,
     609  key        = "poll",
     610  title      = "poll(2) Linux User's Manual",
     611  year       = "2019",
     612  month      = "July",
     613}
     614
     615@manual{MAN:epoll,
     616  key        = "epoll",
     617  title      = "epoll(7) Linux User's Manual",
     618  year       = "2019",
     619  month      = "March",
     620}
     621
     622@manual{MAN:aio,
     623  key        = "aio",
     624  title      = "aio(7) Linux User's Manual",
     625  year       = "2019",
     626  month      = "March",
     627}
     628
     629@misc{MAN:io_uring,
     630  title   = {Efficient IO with io\_uring},
     631  author  = {Axboe, Jens},
     632  year    = "2019",
     633  month   = "March",
     634  version = {0,4},
     635  howpublished = {\url{https://kernel.dk/io_uring.pdf}}
     636}
     637
     638% --------------------------------------------------
    579639% Wikipedia Entries
    580640@misc{wiki:taskparallel,
     
    617677  note = "[Online; accessed 2-January-2021]"
    618678}
     679
     680@misc{wiki:future,
     681  author = "{Wikipedia contributors}",
     682  title = "Futures and promises --- {W}ikipedia{,} The Free Encyclopedia",
     683  year = "2020",
     684  url = "https://en.wikipedia.org/wiki/Futures_and_promises",
     685  note = "[Online; accessed 9-February-2021]"
     686}
  • doc/theses/thierry_delisle_PhD/thesis/text/core.tex

    r342af53 r8e4aa05  
    4949
    5050\section{Design}
    51 In general, a na\"{i}ve \glsxtrshort{fifo} ready-queue does not scale with increased parallelism from \glspl{hthrd}, resulting in decreased performance. The problem is adding/removing \glspl{thrd} is a single point of contention. As shown in the evaluation sections, most production schedulers do scale when adding \glspl{hthrd}. The common solution to the single point of contention is to shard the ready-queue so each \gls{hthrd} can access the ready-queue without contention, increasing performance though lack of contention.
     51In general, a na\"{i}ve \glsxtrshort{fifo} ready-queue does not scale with increased parallelism from \glspl{hthrd}, resulting in decreased performance. The problem is adding/removing \glspl{thrd} is a single point of contention. As shown in the evaluation sections, most production schedulers do scale when adding \glspl{hthrd}. The common solution to the single point of contention is to shard the ready-queue so each \gls{hthrd} can access the ready-queue without contention, increasing performance.
    5252
    5353\subsection{Sharding} \label{sec:sharding}
    54 An interesting approach to sharding a queue is presented in \cit{Trevors paper}. This algorithm presents a queue with a relaxed \glsxtrshort{fifo} guarantee using an array of strictly \glsxtrshort{fifo} sublists as shown in Figure~\ref{fig:base}. Each \emph{cell} of the array has a timestamp for the last operation and a pointer to a linked-list with a lock and each node in the list is marked with a timestamp indicating when it is added to the list. A push operation is done by picking a random cell, acquiring the list lock, and pushing to the list. If the cell is locked, the operation is simply retried on another random cell until a lock is acquired. A pop operation is done in a similar fashion except two random cells are picked. If both cells are unlocked with non-empty lists, the operation pops the node with the oldest cell timestamp. If one of the cells is unlocked and non-empty, the operation pops from that cell. If both cells are either locked or empty, the operation picks two new random cells and tries again.
     54An interesting approach to sharding a queue is presented in \cit{Trevors paper}. This algorithm presents a queue with a relaxed \glsxtrshort{fifo} guarantee using an array of strictly \glsxtrshort{fifo} sublists as shown in Figure~\ref{fig:base}. Each \emph{cell} of the array has a timestamp for the last operation and a pointer to a linked-list with a lock. Each node in the list is marked with a timestamp indicating when it is added to the list. A push operation is done by picking a random cell, acquiring the list lock, and pushing to the list. If the cell is locked, the operation is simply retried on another random cell until a lock is acquired. A pop operation is done in a similar fashion except two random cells are picked. If both cells are unlocked with non-empty lists, the operation pops the node with the oldest timestamp. If one of the cells is unlocked and non-empty, the operation pops from that cell. If both cells are either locked or empty, the operation picks two new random cells and tries again.
    5555
    5656\begin{figure}
     
    100100\paragraph{Local Information} Figure~\ref{fig:emptytls} shows an approach using dense information, similar to the bitmap, but each \gls{hthrd} keeps its own independent copy. While this approach can offer good scalability \emph{and} low latency, the liveliness and discovery of the information can become a problem. This case is made worst in systems with few processors where even blind random picks can find \glspl{thrd} in a few tries.
    101101
    102 I built a prototype of these approaches and none of these techniques offer satisfying performance when few threads are present. All of these approach hit the same 2 problems. First, randomly picking sub-queues is very fast but means any improvement to the hit rate can easily be countered by a slow-down in look-up speed when there are empty lists. Second, the array is already as sharded to avoid contention bottlenecks, so any denser data structure tends to become a bottleneck. In all cases, these factors meant the best cases scenario, \ie many threads, would get worst throughput, and the worst-case scenario, few threads, would get a better hit rate, but an equivalent poor throughput. As a result I tried an entirely different approach.
     102I built a prototype of these approaches and none of these techniques offer satisfying performance when few threads are present. All of these approach hit the same 2 problems. First, randomly picking sub-queues is very fast. That speed means any improvement to the hit rate can easily be countered by a slow-down in look-up speed, whether or not there are empty lists. Second, the array is already sharded to avoid contention bottlenecks, so any denser data structure tends to become a bottleneck. In all cases, these factors meant the best cases scenario, \ie many threads, would get worst throughput, and the worst-case scenario, few threads, would get a better hit rate, but an equivalent poor throughput. As a result I tried an entirely different approach.
    103103
    104104\subsection{Dynamic Entropy}\cit{https://xkcd.com/2318/}
    105 In the worst-case scenario there are only few \glspl{thrd} ready to run, or more precisely given $P$ \glspl{proc}\footnote{For simplicity, this assumes there is a one-to-one match between \glspl{proc} and \glspl{hthrd}.}, $T$ \glspl{thrd} and $\epsilon$ a very small number, than the worst case scenario can be represented by $\epsilon \ll P$, than $T = P + \epsilon$. It is important to note in this case that fairness is effectively irrelevant. Indeed, this case is close to \emph{actually matching} the model of the ``Ideal multi-tasking CPU'' on page \pageref{q:LinuxCFS}. In this context, it is possible to use a purely internal-locality based approach and still meet the fairness requirements. This approach simply has each \gls{proc} running a single \gls{thrd} repeatedly. Or from the shared ready-queue viewpoint, each \gls{proc} pushes to a given sub-queue and then popes from the \emph{same} subqueue. In cases where $T \gg P$, the scheduler should also achieves similar performance without affecting the fairness guarantees.
     105In the worst-case scenario there are only few \glspl{thrd} ready to run, or more precisely given $P$ \glspl{proc}\footnote{For simplicity, this assumes there is a one-to-one match between \glspl{proc} and \glspl{hthrd}.}, $T$ \glspl{thrd} and $\epsilon$ a very small number, than the worst case scenario can be represented by $T = P + \epsilon$, with $\epsilon \ll P$. It is important to note in this case that fairness is effectively irrelevant. Indeed, this case is close to \emph{actually matching} the model of the ``Ideal multi-tasking CPU'' on page \pageref{q:LinuxCFS}. In this context, it is possible to use a purely internal-locality based approach and still meet the fairness requirements. This approach simply has each \gls{proc} running a single \gls{thrd} repeatedly. Or from the shared ready-queue viewpoint, each \gls{proc} pushes to a given sub-queue and then pops from the \emph{same} subqueue. The challenge is for the the scheduler to achieve good performance in both the $T = P + \epsilon$ case and the $T \gg P$ case, without affecting the fairness guarantees in the later.
    106106
    107 To handle this case, I use a pseudo random-number generator, \glsxtrshort{prng} in a novel way. When the scheduler uses a \glsxtrshort{prng} instance per \gls{proc} exclusively, the random-number seed effectively starts an encoding that produces a list of all accessed subqueues, from latest to oldest. The novel approach is to be able to ``replay'' the \glsxtrshort{prng} backwards and there exist \glsxtrshort{prng}s that are fast, compact \emph{and} can be run forward and backwards. Linear congruential generators~\cite{wiki:lcg} are an example of \glsxtrshort{prng}s that match these requirements.
     107To handle this case, I use a \glsxtrshort{prng}\todo{Fix missing long form} in a novel way. There exist \glsxtrshort{prng}s that are fast, compact and can be run forward \emph{and} backwards.  Linear congruential generators~\cite{wiki:lcg} are an example of \glsxtrshort{prng}s of such \glsxtrshort{prng}s. The novel approach is to use the ability to run backwards to ``replay'' the \glsxtrshort{prng}. The scheduler uses an exclusive \glsxtrshort{prng} instance per \gls{proc}, the random-number seed effectively starts an encoding that produces a list of all accessed subqueues, from latest to oldest. Replaying the \glsxtrshort{prng} to identify cells accessed recently and which probably have data still cached.
    108108
    109109The algorithm works as follows:
  • doc/theses/thierry_delisle_PhD/thesis/text/intro.tex

    r342af53 r8e4aa05  
    77While previous work on the concurrent package of \CFA focused on features and interfaces, this thesis focuses on performance, introducing \glsxtrshort{api} changes only when required by performance considerations. More specifically, this thesis concentrates on scheduling and \glsxtrshort{io}. Prior to this work, the \CFA runtime used a strictly \glsxtrshort{fifo} \gls{rQ}.
    88
    9 This work exclusively concentrates on Linux as it's operating system since the existing \CFA runtime and compiler does not already support other operating systems. Furthermore, as \CFA is yet to be released, supporting version of Linux older that the latest version is not a goal of this work.
     9This work exclusively concentrates on Linux as it's operating system since the existing \CFA runtime and compiler does not already support other operating systems. Furthermore, as \CFA is yet to be released, supporting version of Linux older than the latest version is not a goal of this work.
  • doc/theses/thierry_delisle_PhD/thesis/text/io.tex

    r342af53 r8e4aa05  
    1 \chapter{User Level \glsxtrshort{io}}
    2 As mentionned in Section~\ref{prev:io}, User-Level \glsxtrshort{io} requires multiplexing the \glsxtrshort{io} operations of many \glspl{thrd} onto fewer \glspl{proc} using asynchronous \glsxtrshort{io} operations. Various operating systems offer various forms of asynchronous operations and as mentioned in Chapter~\ref{intro}, this work is exclusively focuesd on Linux.
     1\chapter{User Level \io}
     2As mentioned in Section~\ref{prev:io}, User-Level \io requires multiplexing the \io operations of many \glspl{thrd} onto fewer \glspl{proc} using asynchronous \io operations. Different operating systems offer various forms of asynchronous operations and as mentioned in Chapter~\ref{intro}, this work is exclusively focused on the Linux operating-system.
    33
    4 \section{Existing options}
    5 Since \glsxtrshort{io} operations are generally handled by the
     4\section{Kernel Interface}
     5Since this work fundamentally depends on operating-system support, the first step of any design is to discuss the available interfaces and pick one (or more) as the foundations of the non-blocking \io subsystem.
    66
    7 \subsection{\lstinline|epoll|, \lstinline|poll| and \lstinline|select|}
     7\subsection{\lstinline{O_NONBLOCK}}
     8In Linux, files can be opened with the flag @O_NONBLOCK@~\cite{MAN:open} (or @SO_NONBLOCK@~\cite{MAN:accept}, the equivalent for sockets) to use the file descriptors in ``nonblocking mode''. In this mode, ``Neither the @open()@ nor any subsequent \io operations on the [opened file descriptor] will cause the calling
     9process to wait''~\cite{MAN:open}. This feature can be used as the foundation for the non-blocking \io subsystem. However, for the subsystem to know when an \io operation completes, @O_NONBLOCK@ must be use in conjunction with a system call that monitors when a file descriptor becomes ready, \ie, the next \io operation on it does not cause the process to wait\footnote{In this context, ready means \emph{some} operation can be performed without blocking. It does not mean an operation returning \lstinline{EAGAIN} succeeds on the next try. For example, a ready read may only return a subset of bytes and the read must be issues again for the remaining bytes, at which point it may return \lstinline{EAGAIN}.}.
     10This mechanism is also crucial in determining when all \glspl{thrd} are blocked and the application \glspl{kthrd} can now block.
    811
    9 \subsection{Linux's AIO}
     12There are three options to monitor file descriptors in Linux\footnote{For simplicity, this section omits \lstinline{pselect} and \lstinline{ppoll}. The difference between these system calls and \lstinline{select} and \lstinline{poll}, respectively, is not relevant for this discussion.}, @select@~\cite{MAN:select}, @poll@~\cite{MAN:poll} and @epoll@~\cite{MAN:epoll}. All three of these options offer a system call that blocks a \gls{kthrd} until at least one of many file descriptors becomes ready. The group of file descriptors being waited is called the \newterm{interest set}.
    1013
     14\paragraph{\lstinline{select}} is the oldest of these options, it takes as an input a contiguous array of bits, where each bits represent a file descriptor of interest. On return, it modifies the set in place to identify which of the file descriptors changed status. This destructive change means that calling select in a loop requires re-initializing the array each time and the number of file descriptors supported has a hard limit. Another limit of @select@ is that once the call is started, the interest set can no longer be modified. Monitoring a new file descriptor generally requires aborting any in progress call to @select@\footnote{Starting a new call to \lstinline{select} is possible but requires a distinct kernel thread, and as a result is not an acceptable multiplexing solution when the interest set is large and highly dynamic unless the number of parallel calls to \lstinline{select} can be strictly bounded.}.
    1115
     16\paragraph{\lstinline{poll}} is an improvement over select, which removes the hard limit on the number of file descriptors and the need to re-initialize the input on every call. It works using an array of structures as an input rather than an array of bits, thus allowing a more compact input for small interest sets. Like @select@, @poll@ suffers from the limitation that the interest set cannot be changed while the call is blocked.
     17
     18\paragraph{\lstinline{epoll}} further improves these two functions by allowing the interest set to be dynamically added to and removed from while a \gls{kthrd} is blocked on an @epoll@ call. This dynamic capability is accomplished by creating an \emph{epoll instance} with a persistent interest set, which is used across multiple calls. This capability significantly reduces synchronization overhead on the part of the caller (in this case the \io subsystem), since the interest set can be modified when adding or removing file descriptors without having to synchronize with other \glspl{kthrd} potentially calling @epoll@.
     19
     20However, all three of these system calls have limitations. The @man@ page for @O_NONBLOCK@ mentions that ``[@O_NONBLOCK@] has no effect for regular files and block devices'', which means none of these three system calls are viable multiplexing strategies for these types of \io operations. Furthermore, @epoll@ has been shown to have problems with pipes and ttys~\cit{Peter's examples in some fashion}. Finally, none of these are useful solutions for multiplexing \io operations that do not have a corresponding file descriptor and can be awkward for operations using multiple file descriptors.
     21
     22\subsection{POSIX asynchronous I/O (AIO)}
     23An alternative to @O_NONBLOCK@ is the AIO interface. Its interface lets programmers enqueue operations to be performed asynchronously by the kernel. Completions of these operations can be communicated in various ways: either by spawning a new \gls{kthrd}, sending a Linux signal, or by polling for completion of one or more operation. For this work, spawning a new \gls{kthrd} is counter-productive but a related solution is discussed in Section~\ref{io:morethreads}. Using interrupts handlers can also lead to fairly complicated interactions between subsystems. Leaving polling for completion, which is similar to the previous system calls. While AIO only supports read and write operations to file descriptors, it does not have the same limitation as @O_NONBLOCK@, \ie, the file descriptors can be regular files and blocked devices. It also supports batching multiple operations in a single system call.
     24
     25AIO offers two different approach to polling: @aio_error@ can be used as a spinning form of polling, returning @EINPROGRESS@ until the operation is completed, and @aio_suspend@ can be used similarly to @select@, @poll@ or @epoll@, to wait until one or more requests have completed. For the purpose of \io multiplexing, @aio_suspend@ is the best interface. However, even if AIO requests can be submitted concurrently, @aio_suspend@ suffers from the same limitation as @select@ and @poll@, \ie, the interest set cannot be dynamically changed while a call to @aio_suspend@ is in progress. AIO also suffers from the limitation of specifying which requests have completed, \ie programmers have to poll each request in the interest set using @aio_error@ to identify the completed requests. This limitation means that, like @select@ and @poll@ but not @epoll@, the time needed to examine polling results increases based on the total number of requests monitored, not the number of completed requests.
     26Finally, AIO does not seem to be a popular interface, which I believe is due in part to this poor polling interface. Linus Torvalds talks about this interface as follows:
    1227
    1328\begin{displayquote}
    14         AIO is a horrible ad-hoc design, with the main excuse being "other,
     29        AIO is a horrible ad-hoc design, with the main excuse being ``other,
    1530        less gifted people, made that design, and we are implementing it for
    1631        compatibility because database people - who seldom have any shred of
    17         taste - actually use it".
     32        taste - actually use it''.
    1833
    1934        But AIO was always really really ugly.
     
    2439\end{displayquote}
    2540
    26 Interestingly, in this e-mail answer, Linus goes on to describe
     41Interestingly, in this e-mail, Linus goes on to describe
    2742``a true \textit{asynchronous system call} interface''
    2843that does
     
    3045in
    3146``some kind of arbitrary \textit{queue up asynchronous system call} model''.
    32 This description is actually quite close to the interface of the interface described in the next section.
     47This description is actually quite close to the interface described in the next section.
    3348
    34 \subsection{\texttt{io\_uring}}
    35 A very recent addition to Linux, @io_uring@\cit{io\_uring} is a framework that aims to solve many of the problems listed with the above mentioned solutions.
     49\subsection{\lstinline{io_uring}}
     50A very recent addition to Linux, @io_uring@~\cite{MAN:io_uring}, is a framework that aims to solve many of the problems listed in the above interfaces. Like AIO, it represents \io operations as entries added to a queue. But like @epoll@, new requests can be submitted while a blocking call waiting for requests to complete is already in progress. The @io_uring@ interface uses two ring buffers (referred to simply as rings) at its core: a submit ring to which programmers push \io requests and a completion ring from which programmers poll for completion.
     51
     52One of the big advantages over the prior interfaces is that @io_uring@ also supports a much wider range of operations. In addition to supporting reads and writes to any file descriptor like AIO, it supports other operations like @open@, @close@, @fsync@, @accept@, @connect@, @send@, @recv@, @splice@, \etc.
     53
     54On top of these, @io_uring@ adds many extras like avoiding copies between the kernel and user-space using shared memory, allowing different mechanisms to communicate with device drivers, and supporting chains of requests, \ie, requests that automatically trigger followup requests on completion.
    3655
    3756\subsection{Extra Kernel Threads}\label{io:morethreads}
    38 Finally, if the operating system does not offer any satisfying forms of asynchronous \glsxtrshort{io} operations, a solution is to fake it by creating a pool of \glspl{kthrd} and delegating operations to them in order to avoid blocking \glspl{proc}.
     57Finally, if the operating system does not offer a satisfactory form of asynchronous \io operations, an ad-hoc solution is to create a pool of \glspl{kthrd} and delegate operations to it to avoid blocking \glspl{proc}, which is a compromise for multiplexing. In the worst case, where all \glspl{thrd} are consistently blocking on \io, it devolves into 1-to-1 threading. However, regardless of the frequency of \io operations, it achieves the fundamental goal of not blocking \glspl{proc} when \glspl{thrd} are ready to run. This approach is used by languages like Go\cit{Go} and frameworks like libuv\cit{libuv}, since it has the advantage that it can easily be used across multiple operating systems. This advantage is especially relevant for languages like Go, which offer a homogeneous \glsxtrshort{api} across all platforms. As opposed to C, which has a very limited standard api for \io, \eg, the C standard library has no networking.
    3958
    4059\subsection{Discussion}
     60These options effectively fall into two broad camps: waiting for \io to be ready versus waiting for \io to complete. All operating systems that support asynchronous \io must offer an interface along one of these lines, but the details vary drastically. For example, Free BSD offers @kqueue@~\cite{MAN:bsd/kqueue}, which behaves similarly to @epoll@, but with some small quality of use improvements, while Windows (Win32)~\cit{https://docs.microsoft.com/en-us/windows/win32/fileio/synchronous-and-asynchronous-i-o} offers ``overlapped I/O'', which handles submissions similarly to @O_NONBLOCK@ with extra flags on the synchronous system call, but waits for completion events, similarly to @io_uring@.
    4161
     62For this project, I selected @io_uring@, in large parts because to its generality. While @epoll@ has been shown to be a good solution for socket \io (\cite{DBLP:journals/pomacs/KarstenB20}), @io_uring@'s transparent support for files, pipes, and more complex operations, like @splice@ and @tee@, make it a better choice as the foundation for a general \io subsystem.
    4263
    4364\section{Event-Engine}
     65An event engine's responsibility is to use the kernel interface to multiplex many \io operations onto few \glspl{kthrd}. In concrete terms, this means \glspl{thrd} enter the engine through an interface, the event engines then starts the operation and parks the calling \glspl{thrd}, returning control to the \gls{proc}. The parked \glspl{thrd} are then rescheduled by the event engine once the desired operation has completed.
     66
     67\subsection{\lstinline{io_uring} in depth}
     68Before going into details on the design of my event engine, more details on @io_uring@ usage are presented, each important in the design of the engine.
     69Figure~\ref{fig:iouring} shows an overview of an @io_uring@ instance.
     70Two ring buffers are used to communicate with the kernel: one for submissions~(left) and one for completions~(right).
     71The submission ring contains entries, \newterm{Submit Queue Entries} (SQE), produced (appended) by the application when an operation starts and then consumed by the kernel.
     72The completion ring contains entries, \newterm{Completion Queue Entries} (CQE), produced (appended) by the kernel when an operation completes and then consumed by the application.
     73The submission ring contains indexes into the SQE array (denoted \emph{S}) containing entries describing the I/O operation to start;
     74the completion ring contains entries for the completed I/O operation.
     75Multiple @io_uring@ instances can be created, in which case they each have a copy of the data structures in the figure.
     76
     77\begin{figure}
     78        \centering
     79        \input{io_uring.pstex_t}
     80        \caption{Overview of \lstinline{io_uring}}
     81%       \caption[Overview of \lstinline{io_uring}]{Overview of \lstinline{io_uring} \smallskip\newline Two ring buffer are used to communicate with the kernel, one for completions~(right) and one for submissions~(left). The completion ring contains entries, \newterm{CQE}s: Completion Queue Entries, that are produced by the kernel when an operation completes and then consumed by the application. On the other hand, the application produces \newterm{SQE}s: Submit Queue Entries, which it appends to the submission ring for the kernel to consume. Unlike the completion ring, the submission ring does not contain the entries directly, it indexes into the SQE array (denoted \emph{S}) instead.}
     82        \label{fig:iouring}
     83\end{figure}
     84
     85New \io operations are submitted to the kernel following 4 steps, which use the components shown in the figure.
     86\begin{enumerate}
     87\item
     88An SQE is allocated from the pre-allocated array (denoted \emph{S} in Figure~\ref{fig:iouring}). This array is created at the same time as the @io_uring@ instance, is in kernel-locked memory visible by both the kernel and the application, and has a fixed size determined at creation. How these entries are allocated is not important for the functioning of @io_uring@, the only requirement is that no entry is reused before the kernel has consumed it.
     89\item
     90The SQE is filled according to the desired operation. This step is straight forward, the only detail worth mentioning is that SQEs have a @user_data@ field that must be filled in order to match submission and completion entries.
     91\item
     92The SQE is submitted to the submission ring by appending the index of the SQE to the ring following regular ring buffer steps: \lstinline{buffer[head] = item; head++}. Since the head is visible to the kernel, some memory barriers may be required to prevent the compiler from reordering these operations. Since the submission ring is a regular ring buffer, more than one SQE can be added at once and the head is updated only after all entries are updated.
     93\item
     94The kernel is notified of the change to the ring using the system call @io_uring_enter@. The number of elements appended to the submission ring is passed as a parameter and the number of elements consumed is returned. The @io_uring@ instance can be constructed so this step is not required, but this requires elevated privilege.% and an early version of @io_uring@ had additional restrictions.
     95\end{enumerate}
     96
     97\begin{sloppypar}
     98The completion side is simpler: applications call @io_uring_enter@ with the flag @IORING_ENTER_GETEVENTS@ to wait on a desired number of operations to complete. The same call can be used to both submit SQEs and wait for operations to complete. When operations do complete, the kernel appends a CQE to the completion ring and advances the head of the ring. Each CQE contains the result of the operation as well as a copy of the @user_data@ field of the SQE that triggered the operation. It is not necessary to call @io_uring_enter@ to get new events because the kernel can directly modify the completion ring. The system call is only needed if the application wants to block waiting for operations to complete.
     99\end{sloppypar}
     100
     101The @io_uring_enter@ system call is protected by a lock inside the kernel. This protection means that concurrent call to @io_uring_enter@ using the same instance are possible, but there is no performance gained from parallel calls to @io_uring_enter@. It is possible to do the first three submission steps in parallel, however, doing so requires careful synchronization.
     102
     103@io_uring@ also introduces constraints on the number of simultaneous operations that can be ``in flight''. Obviously, SQEs are allocated from a fixed-size array, meaning that there is a hard limit to how many SQEs can be submitted at once. In addition, the @io_uring_enter@ system call can fail because ``The  kernel [...] ran out of resources to handle [a request]'' or ``The application is attempting to overcommit the number of requests it can  have  pending.''. This restriction means \io request bursts may have to be subdivided and submitted in chunks at a later time.
     104
     105\subsection{Multiplexing \io: Submission}
     106The submission side is the most complicated aspect of @io_uring@ and its design largely dictates the completion side.
     107
     108While it is possible to do the first steps of submission in parallel, the duration of the system call scales with number of entries submitted. The consequence is that the amount of parallelism used to prepare submissions for the next system call is limited. Beyond this limit, the length of the system call is the throughput limiting factor. I concluded from early experiments that preparing submissions seems to take about as long as the system call itself, which means that with a single @io_uring@ instance, there is no benefit in terms of \io throughput to having more than two \glspl{hthrd}. Therefore the design of the submission engine must manage multiple instances of @io_uring@ running in parallel, effectively sharding @io_uring@ instances. Similarly to scheduling, this sharding can be done privately, \ie, one instance per \glspl{proc}, or in decoupled pools, \ie, a pool of \glspl{proc} use a pool of @io_uring@ instances without one-to-one coupling between any given instance and any given \gls{proc}.
     109
     110\subsubsection{Pool of Instances}
     111One approach is to have multiple shared instances. \Glspl{thrd} attempting \io operations pick one of the available instances and submits operations to that instance. Since the completion will be sent to the same instance, all instances with pending operations must be polled continuously\footnote{As will be described in Chapter~\ref{practice}, this does not translate into constant CPU usage.}. Since there is no coupling between \glspl{proc} and @io_uring@ instances in this approach, \glspl{thrd} running on more than one \gls{proc} can attempt to submit to the same instance concurrently. Since @io_uring@ effectively sets the amount of sharding needed to avoid contention on its internal locks, performance in this approach is based on two aspects: the synchronization needed to submit does not induce more contention than @io_uring@ already does and the scheme to route \io requests to specific @io_uring@ instances does not introduce contention. This second aspect has an oversized importance because it comes into play before the sharding of instances, and as such, all \glspl{hthrd} can contend on the routing algorithm.
     112
     113Allocation in this scheme can be handled fairly easily. Free SQEs, \ie, SQEs that aren't currently being used to represent a request, can be written to safely and have a field called @user_data@ which the kernel only reads to copy to CQEs. Allocation also requires no ordering guarantee as all free SQEs are interchangeable. This requires a simple concurrent bag. The only added complexity is that the number of SQEs is fixed, which means allocation can fail. This failure needs to be pushed up to the routing algorithm, \glspl{thrd} attempting \io operations must not be directed to @io_uring@ instances without any available SQEs. Ideally, the routing algorithm would block operations up-front if none of the instances have available SQEs.
     114
     115Once an SQE is allocated, \glspl{thrd} can fill them normally, they simply need to keep track of the SQE index and which instance it belongs to.
     116
     117Once an SQE is filled in, what needs to happen is that the SQE must be added to the submission ring buffer, an operation that is not thread-safe on itself, and the kernel must be notified using the @io_uring_enter@ system call. The submission ring buffer is the same size as the pre-allocated SQE buffer, therefore pushing to the ring buffer cannot fail\footnote{This is because it is invalid to have the same \lstinline{sqe} multiple times in the ring buffer.}. However, as mentioned, the system call itself can fail with the expectation that it will be retried once some of the already submitted operations complete. Since multiple SQEs can be submitted to the kernel at once, it is important to strike a balance between batching and latency. Operations that are ready to be submitted should be batched together in few system calls, but at the same time, operations should not be left pending for long period of times before being submitted. This can be handled by either designating one of the submitting \glspl{thrd} as the being responsible for the system call for the current batch of SQEs or by having some other party regularly submitting all ready SQEs, \eg, the poller \gls{thrd} mentioned later in this section.
     118
     119In the case of designating a \gls{thrd}, ideally, when multiple \glspl{thrd} attempt to submit operations to the same @io_uring@ instance, all requests would be batched together and one of the \glspl{thrd} would do the system call on behalf of the others, referred to as the \newterm{submitter}. In practice however, it is important that the \io requests are not left pending indefinitely and as such, it may be required to have a current submitter and a next submitter. Indeed, as long as there is a ``next'' submitter, \glspl{thrd} submitting new \io requests can move on, knowing that some future system call will include their request. Once the system call is done, the submitter must also free SQEs so that the allocator can reused them.
     120
     121Finally, the completion side is much simpler since the @io_uring@ system call enforces a natural synchronization point. Polling simply needs to regularly do the system call, go through the produced CQEs and communicate the result back to the originating \glspl{thrd}. Since CQEs only own a signed 32 bit result, in addition to the copy of the @user_data@ field, all that is needed to communicate the result is a simple future~\cite{wiki:future}. If the submission side does not designate submitters, polling can also submit all SQEs as it is polling events.  A simple approach to polling is to allocate a \gls{thrd} per @io_uring@ instance and simply let the poller \glspl{thrd} poll their respective instances when scheduled. This design is especially convenient for reasons explained in Chapter~\ref{practice}.
     122
     123With this pool of instances approach, the big advantage is that it is fairly flexible. It does not impose restrictions on what \glspl{thrd} submitting \io operations can and cannot do between allocations and submissions. It also can gracefully handle running out of resources, SQEs or the kernel returning @EBUSY@. The down side to this is that many of the steps used for submitting need complex synchronization to work properly. The routing and allocation algorithm needs to keep track of which ring instances have available SQEs, block incoming requests if no instance is available, prevent barging if \glspl{thrd} are already queued up waiting for SQEs and handle SQEs being freed. The submission side needs to safely append SQEs to the ring buffer, make sure no SQE is dropped or left pending forever, notify the allocation side when SQEs can be reused and handle the kernel returning @EBUSY@. Sharding the @io_uring@ instances should alleviate much of the contention caused by this, but all this synchronization may still have non-zero cost.
     124
     125\subsubsection{Private Instances}
     126Another approach is to simply create one ring instance per \gls{proc}. This alleviate the need for synchronization on the submissions, requiring only that \glspl{thrd} are not interrupted in between two submission steps. This is effectively the same requirement as using @thread_local@ variables. Since SQEs that are allocated must be submitted to the same ring, on the same \gls{proc}, this effectively forces the application to submit SQEs in allocation order\footnote{The actual requirement is that \glspl{thrd} cannot context switch between allocation and submission. This requirement means that from the subsystem's point of view, the allocation and submission are sequential. To remove this requirement, a \gls{thrd} would need the ability to ``yield to a specific \gls{proc}'', \ie, park with the promise that it will be run next on a specific \gls{proc}, the \gls{proc} attached to the correct ring. This is not a current or planned feature of \CFA.}, greatly simplifying both allocation and submission. In this design, allocation and submission form a ring partitioned ring buffer as shown in Figure~\ref{fig:pring}. Once added to the ring buffer, the attached \gls{proc} has a significant amount of flexibility with regards to when to do the system call. Possible options are: when the \gls{proc} runs out of \glspl{thrd} to run, after running a given number of threads \glspl{thrd}, etc.
     127
     128\begin{figure}
     129        \centering
     130        \input{pivot_ring.pstex_t}
     131        \caption[Partitioned ring buffer]{Partitioned ring buffer \smallskip\newline Allocated sqes are appending to the first partition. When submitting, the partition is simply advanced to include all the sqes that should be submitted. The kernel considers the partition as the head of the ring.}
     132        \label{fig:pring}
     133\end{figure}
     134
     135This approach has the advantage that it does not require much of the synchronization needed in the shared approach. This comes at the cost that \glspl{thrd} submitting \io operations have less flexibility, they cannot park or yield, and several exceptional cases are handled poorly. Instances running out of SQEs cannot run \glspl{thrd} wanting to do \io operations, in such a case the \gls{thrd} needs to be moved to a different \gls{proc}, the only current way of achieving this would be to @yield()@ hoping to be scheduled on a different \gls{proc}, which is not guaranteed. Another problematic case is that \glspl{thrd} that do not park for long periods of time will delay the submission of any SQE not already submitted. This issue is similar to fairness issues which schedulers that use work-stealing mentioned in the previous chapter.
     136
    44137
    45138
    46139\section{Interface}
     140Finally, the last important part of the \io subsystem is it's interface. There are multiple approaches that can be offered to programmers, each with advantages and disadvantages. The new \io subsystem can replace the C runtime's API or extend it. And in the later case the interface can go from very similar to vastly different. The following sections discuss some useful options using @read@ as an example. The standard Linux interface for C is :
     141
     142@ssize_t read(int fd, void *buf, size_t count);@.
     143
     144\subsection{Replacement}
     145Replacing the C \glsxtrshort{api}
     146
     147\subsection{Synchronous Extension}
     148
     149\subsection{Asynchronous Extension}
     150
     151\subsection{Interface directly to \lstinline{io_uring}}
  • doc/theses/thierry_delisle_PhD/thesis/text/runtime.tex

    r342af53 r8e4aa05  
    1111
    1212\section{Clusters}
    13 \CFA allows the option to group user-level threading, in the form of clusters. Both \glspl{thrd} and \glspl{proc} belong to a specific cluster. \Glspl{thrd} are only be scheduled onto \glspl{proc} in the same cluster and scheduling is done independently of other clusters. Figure~\ref{fig:system} shows an overview of the \CFA runtime, which allows programmers to tightly control parallelism. It also opens the door to handling effects like NUMA, by pining clusters to a specific NUMA node\footnote{This is not currently implemented in \CFA, but the only hurdle left is creating a generic interface for cpu masks.}.
     13\CFA allows the option to group user-level threading, in the form of clusters. Both \glspl{thrd} and \glspl{proc} belong to a specific cluster. \Glspl{thrd} are only scheduled onto \glspl{proc} in the same cluster and scheduling is done independently of other clusters. Figure~\ref{fig:system} shows an overview of the \CFA runtime, which allows programmers to tightly control parallelism. It also opens the door to handling effects like NUMA, by pining clusters to a specific NUMA node\footnote{This is not currently implemented in \CFA, but the only hurdle left is creating a generic interface for cpu masks.}.
    1414
    1515\begin{figure}
     
    2525
    2626\section{\glsxtrshort{io}}\label{prev:io}
    27 Prior to this work, the \CFA runtime did not add any particular support for \glsxtrshort{io} operations. %\CFA being built on C, this means that,
    28 While all I/O operations available in C are available in \CFA, \glsxtrshort{io} operations are designed for the POSIX threading model~\cite{pthreads}. Using these 1:1 threading operations in an M:N threading model means I/O operations block \glspl{proc} instead of \glspl{thrd}. While this can work in certain cases, it limits the number of concurrent operations to the number of \glspl{proc} rather than \glspl{thrd}. It also means deadlock can occur because all \glspl{proc} are blocked even if at least one \gls{thrd} is ready to run. A simple example of this type of deadlock would be as follows:
     27Prior to this work, the \CFA runtime did not add any particular support for \glsxtrshort{io} operations. While all \glsxtrshort{io} operations available in C are available in \CFA, \glsxtrshort{io} operations are designed for the POSIX threading model~\cite{pthreads}. Using these 1:1 threading operations in an M:N threading model means \glsxtrshort{io} operations block \glspl{proc} instead of \glspl{thrd}. While this can work in certain cases, it limits the number of concurrent operations to the number of \glspl{proc} rather than \glspl{thrd}. It also means deadlock can occur because all \glspl{proc} are blocked even if at least one \gls{thrd} is ready to run. A simple example of this type of deadlock would be as follows:
     28
    2929\begin{quote}
    3030Given a simple network program with 2 \glspl{thrd} and a single \gls{proc}, one \gls{thrd} sends network requests to a server and the other \gls{thrd} waits for a response from the server. If the second \gls{thrd} races ahead, it may wait for responses to requests that have not been sent yet. In theory, this should not be a problem, even if the second \gls{thrd} waits, because the first \gls{thrd} is still ready to run and should be able to get CPU time to send the request. With M:N threading, while the first \gls{thrd} is ready, the lone \gls{proc} \emph{cannot} run the first \gls{thrd} if it is blocked in the \glsxtrshort{io} operation of the second \gls{thrd}. If this happen, the system is in a synchronization deadlock\footnote{In this example, the deadlocked could be resolved if the server sends unprompted messages to the client. However, this solution is not general and may not be appropriate even in this simple case.}.
    3131\end{quote}
    32 Therefore, one of the objective of this work is to introduce \emph{User-Level \glsxtrshort{io}}, like \glslink{uthrding}{User-Level \emph{Threading}} blocks \glspl{thrd} rather than \glspl{proc} when doing \glsxtrshort{io} operations, which entails multiplexing the \glsxtrshort{io} operations of many \glspl{thrd} onto fewer \glspl{proc}. This multiplexing requires that a single \gls{proc} be able to execute multiple I/O operations in parallel. This requirement cannot be done with operations that block \glspl{proc}, \ie \glspl{kthrd}, since the first operation would prevent starting new operations for its blocking duration. Executing I/O operations in parallel requires \emph{asynchronous} \glsxtrshort{io}, sometimes referred to as \emph{non-blocking}, since the \gls{kthrd} does not block.
    3332
    34 \section{Interoperating with C}
     33Therefore, one of the objective of this work is to introduce \emph{User-Level \glsxtrshort{io}}, like \glslink{uthrding}{User-Level \emph{Threading}} blocks \glspl{thrd} rather than \glspl{proc} when doing \glsxtrshort{io} operations, which entails multiplexing the \glsxtrshort{io} operations of many \glspl{thrd} onto fewer \glspl{proc}. This multiplexing requires that a single \gls{proc} be able to execute multiple \glsxtrshort{io} operations in parallel. This requirement cannot be done with operations that block \glspl{proc}, \ie \glspl{kthrd}, since the first operation would prevent starting new operations for its blocking duration. Executing \glsxtrshort{io} operations in parallel requires \emph{asynchronous} \glsxtrshort{io}, sometimes referred to as \emph{non-blocking}, since the \gls{kthrd} does not block.
     34
     35\section{Interoperating with \texttt{C}}
    3536While \glsxtrshort{io} operations are the classical example of operations that block \glspl{kthrd}, the non-blocking challenge extends to all blocking system-calls. The POSIX standard states~\cite[\S~2.9.1]{POSIX17}:
    3637\begin{quote}
     
    4445\begin{enumerate}
    4546        \item Precisely identifying blocking C calls is difficult.
    46         \item Introducing new code can have a significant impact on general performance.
     47        \item Introducing control points code can have a significant impact on general performance.
    4748\end{enumerate}
    48 Because of these consequences, this work does not attempt to ``sandbox'' calls to C. Therefore, it is possible for an unidentified library calls to block a \gls{kthrd} leading to deadlocks in \CFA's M:N threading model, which would not occur in a traditional 1:1 threading model. Currently, all M:N thread systems interacting with UNIX without sandboxing suffer from this problem but manage to work very well in the majority of applications. Therefore, a complete solution to this problem is outside the scope of this thesis.
     49Because of these consequences, this work does not attempt to ``sandbox'' calls to C. Therefore, it is possible calls from an unidentified library will block a \gls{kthrd} leading to deadlocks in \CFA's M:N threading model, which would not occur in a traditional 1:1 threading model. Currently, all M:N thread systems interacting with UNIX without sandboxing suffer from this problem but manage to work very well in the majority of applications. Therefore, a complete solution to this problem is outside the scope of this thesis.
  • doc/theses/thierry_delisle_PhD/thesis/thesis.tex

    r342af53 r8e4aa05  
    1 % uWaterloo Thesis Template for LaTeX
    2 % Last Updated June 14, 2017 by Stephen Carr, IST Client Services
    3 % FOR ASSISTANCE, please send mail to rt-IST-CSmathsci@ist.uwaterloo.ca
    4 
    5 % Effective October 2006, the University of Waterloo
    6 % requires electronic thesis submission. See the uWaterloo thesis regulations at
     1%======================================================================
     2% University of Waterloo Thesis Template for LaTeX
     3% Last Updated November, 2020
     4% by Stephen Carr, IST Client Services,
     5% University of Waterloo, 200 University Ave. W., Waterloo, Ontario, Canada
     6% FOR ASSISTANCE, please send mail to request@uwaterloo.ca
     7
     8% DISCLAIMER
     9% To the best of our knowledge, this template satisfies the current uWaterloo thesis requirements.
     10% However, it is your responsibility to assure that you have met all requirements of the University and your particular department.
     11
     12% Many thanks for the feedback from many graduates who assisted the development of this template.
     13% Also note that there are explanatory comments and tips throughout this template.
     14%======================================================================
     15% Some important notes on using this template and making it your own...
     16
     17% The University of Waterloo has required electronic thesis submission since October 2006.
     18% See the uWaterloo thesis regulations at
    719% https://uwaterloo.ca/graduate-studies/thesis.
    8 
    9 % DON'T FORGET TO ADD YOUR OWN NAME AND TITLE in the "hyperref" package
    10 % configuration below. THIS INFORMATION GETS EMBEDDED IN THE PDF FINAL PDF DOCUMENT.
    11 % You can view the information if you view Properties of the PDF document.
    12 
    13 % Many faculties/departments also require one or more printed
    14 % copies. This template attempts to satisfy both types of output.
    15 % It is based on the standard "book" document class which provides all necessary
    16 % sectioning structures and allows multi-part theses.
    17 
    18 % DISCLAIMER
    19 % To the best of our knowledge, this template satisfies the current uWaterloo requirements.
    20 % However, it is your responsibility to assure that you have met all
    21 % requirements of the University and your particular department.
    22 % Many thanks for the feedback from many graduates that assisted the development of this template.
    23 
    24 % -----------------------------------------------------------------------
    25 
    26 % By default, output is produced that is geared toward generating a PDF
    27 % version optimized for viewing on an electronic display, including
    28 % hyperlinks within the PDF.
    29 
     20% This thesis template is geared towards generating a PDF version optimized for viewing on an electronic display, including hyperlinks within the PDF.
     21
     22% DON'T FORGET TO ADD YOUR OWN NAME AND TITLE in the "hyperref" package configuration below.
     23% THIS INFORMATION GETS EMBEDDED IN THE PDF FINAL PDF DOCUMENT.
     24% You can view the information if you view properties of the PDF document.
     25
     26% Many faculties/departments also require one or more printed copies.
     27% This template attempts to satisfy both types of output.
     28% See additional notes below.
     29% It is based on the standard "book" document class which provides all necessary sectioning structures and allows multi-part theses.
     30
     31% If you are using this template in Overleaf (cloud-based collaboration service), then it is automatically processed and previewed for you as you edit.
     32
     33% For people who prefer to install their own LaTeX distributions on their own computers, and process the source files manually, the following notes provide the sequence of tasks:
     34 
    3035% E.g. to process a thesis called "mythesis.tex" based on this template, run:
    3136
    3237% pdflatex mythesis     -- first pass of the pdflatex processor
    3338% bibtex mythesis       -- generates bibliography from .bib data file(s)
    34 % makeindex         -- should be run only if an index is used
     39% makeindex         -- should be run only if an index is used 
    3540% pdflatex mythesis     -- fixes numbering in cross-references, bibliographic references, glossaries, index, etc.
    36 % pdflatex mythesis     -- fixes numbering in cross-references, bibliographic references, glossaries, index, etc.
    37 
    38 % If you use the recommended LaTeX editor, Texmaker, you would open the mythesis.tex
    39 % file, then click the PDFLaTeX button. Then run BibTeX (under the Tools menu).
    40 % Then click the PDFLaTeX button two more times. If you have an index as well,
    41 % you'll need to run MakeIndex from the Tools menu as well, before running pdflatex
     41% pdflatex mythesis     -- it takes a couple of passes to completely process all cross-references
     42
     43% If you use the recommended LaTeX editor, Texmaker, you would open the mythesis.tex file, then click the PDFLaTeX button. Then run BibTeX (under the Tools menu).
     44% Then click the PDFLaTeX button two more times.
     45% If you have an index as well,you'll need to run MakeIndex from the Tools menu as well, before running pdflatex
    4246% the last two times.
    4347
    44 % N.B. The "pdftex" program allows graphics in the following formats to be
    45 % included with the "\includegraphics" command: PNG, PDF, JPEG, TIFF
    46 % Tip 1: Generate your figures and photos in the size you want them to appear
    47 % in your thesis, rather than scaling them with \includegraphics options.
    48 % Tip 2: Any drawings you do should be in scalable vector graphic formats:
    49 % SVG, PNG, WMF, EPS and then converted to PNG or PDF, so they are scalable in
    50 % the final PDF as well.
    51 % Tip 3: Photographs should be cropped and compressed so as not to be too large.
    52 
    53 % To create a PDF output that is optimized for double-sided printing:
    54 %
    55 % 1) comment-out the \documentclass statement in the preamble below, and
    56 % un-comment the second \documentclass line.
    57 %
    58 % 2) change the value assigned below to the boolean variable
    59 % "PrintVersion" from "false" to "true".
    60 
    61 % --------------------- Start of Document Preamble -----------------------
    62 
    63 % Specify the document class, default style attributes, and page dimensions
     48% N.B. The "pdftex" program allows graphics in the following formats to be included with the "\includegraphics" command: PNG, PDF, JPEG, TIFF
     49% Tip: Generate your figures and photos in the size you want them to appear in your thesis, rather than scaling them with \includegraphics options.
     50% Tip: Any drawings you do should be in scalable vector graphic formats: SVG, PNG, WMF, EPS and then converted to PNG or PDF, so they are scalable in the final PDF as well.
     51% Tip: Photographs should be cropped and compressed so as not to be too large.
     52
     53% To create a PDF output that is optimized for double-sided printing:
     54% 1) comment-out the \documentclass statement in the preamble below, and un-comment the second \documentclass line.
     55% 2) change the value assigned below to the boolean variable "PrintVersion" from " false" to "true".
     56
     57%======================================================================
     58%   D O C U M E N T   P R E A M B L E
     59% Specify the document class, default style attributes, and page dimensions, etc.
    6460% For hyperlinked PDF, suitable for viewing on a computer, use this:
    6561\documentclass[letterpaper,12pt,titlepage,oneside,final]{book}
    6662
    67 % For PDF, suitable for double-sided printing, change the PrintVersion variable below
    68 % to "true" and use this \documentclass line instead of the one above:
     63% For PDF, suitable for double-sided printing, change the PrintVersion variable below to "true" and use this \documentclass line instead of the one above:
    6964%\documentclass[letterpaper,12pt,titlepage,openright,twoside,final]{book}
    7065
    71 \newcommand{\href}[1]{#1} % does nothing, but defines the command so the
    72     % print-optimized version will ignore \href tags (redefined by hyperref pkg).
     66% Some LaTeX commands I define for my own nomenclature.
     67% If you have to, it's easier to make changes to nomenclature once here than in a million places throughout your thesis!
     68\newcommand{\package}[1]{\textbf{#1}} % package names in bold text
     69\newcommand{\cmmd}[1]{\textbackslash\texttt{#1}} % command name in tt font
     70\newcommand{\href}[1]{#1} % does nothing, but defines the command so the print-optimized version will ignore \href tags (redefined by hyperref pkg).
     71%\newcommand{\texorpdfstring}[2]{#1} % does nothing, but defines the command
     72% Anything defined here may be redefined by packages added below...
    7373
    7474% This package allows if-then-else control structures.
     
    7676\newboolean{PrintVersion}
    7777\setboolean{PrintVersion}{false}
    78 % CHANGE THIS VALUE TO "true" as necessary, to improve printed results for hard copies
    79 % by overriding some options of the hyperref package below.
     78% CHANGE THIS VALUE TO "true" as necessary, to improve printed results for hard copies by overriding some options of the hyperref package, called below.
    8079
    8180%\usepackage{nomencl} % For a nomenclature (optional; available from ctan.org)
    8281\usepackage{amsmath,amssymb,amstext} % Lots of math symbols and environments
     82\usepackage{xcolor}
    8383\usepackage{graphicx} % For including graphics
    8484
    8585% Hyperlinks make it very easy to navigate an electronic document.
    86 % In addition, this is where you should specify the thesis title
    87 % and author as they appear in the properties of the PDF document.
     86% In addition, this is where you should specify the thesis title and author as they appear in the properties of the PDF document.
    8887% Use the "hyperref" package
    8988% N.B. HYPERREF MUST BE THE LAST PACKAGE LOADED; ADD ADDITIONAL PKGS ABOVE
    9089\usepackage[pagebackref=false]{hyperref} % with basic options
    91                 % N.B. pagebackref=true provides links back from the References to the body text. This can cause trouble for printing.
     90%\usepackage[pdftex,pagebackref=true]{hyperref}
     91% N.B. pagebackref=true provides links back from the References to the body text. This can cause trouble for printing.
    9292\hypersetup{
    9393        plainpages=false,       % needed if Roman numbers in frontpages
    94         unicode=false,          % non-Latin characters in Acrobats bookmarks
    95         pdftoolbar=true,        % show Acrobats toolbar?
    96         pdfmenubar=true,        % show Acrobats menu?
     94        unicode=false,          % non-Latin characters in Acrobat's bookmarks
     95        pdftoolbar=true,        % show Acrobat's toolbar?
     96        pdfmenubar=true,        % show Acrobat's menu?
    9797        pdffitwindow=false,     % window fit to page when opened
    9898        pdfstartview={FitH},    % fits the width of the page to the window
     
    110110\ifthenelse{\boolean{PrintVersion}}{   % for improved print quality, change some hyperref options
    111111\hypersetup{    % override some previously defined hyperref options
    112         citecolor=black,
    113         filecolor=black,
    114         linkcolor=black,
     112        citecolor=black,%
     113        filecolor=black,%
     114        linkcolor=black,%
    115115        urlcolor=black
    116116}}{} % end of ifthenelse (no else)
     
    120120% although it's supposed to be in both the TeX Live and MikTeX distributions. There are also documentation and
    121121% installation instructions there.
    122 \renewcommand*{\glstextformat}[1]{\textsf{#1}}
     122\makeatletter
     123\newcommand*{\glsplainhyperlink}[2]{%
     124  \colorlet{currenttext}{.}% store current text color
     125  \colorlet{currentlink}{\@linkcolor}% store current link color
     126  \hypersetup{linkcolor=currenttext}% set link color
     127  \hyperlink{#1}{#2}%
     128  \hypersetup{linkcolor=currentlink}% reset to default
     129}
     130\let\@glslink\glsplainhyperlink
     131\makeatother
    123132
    124133\usepackage{csquotes}
     
    126135
    127136% Setting up the page margins...
    128 \setlength{\textheight}{9in}\setlength{\topmargin}{-0.45in}\setlength{\headsep}{0.25in}
     137\setlength{\textheight}{9in}
     138\setlength{\topmargin}{-0.45in}
     139\setlength{\headsep}{0.25in}
    129140% uWaterloo thesis requirements specify a minimum of 1 inch (72pt) margin at the
    130 % top, bottom, and outside page edges and a 1.125 in. (81pt) gutter
    131 % margin (on binding side). While this is not an issue for electronic
    132 % viewing, a PDF may be printed, and so we have the same page layout for
    133 % both printed and electronic versions, we leave the gutter margin in.
     141% top, bottom, and outside page edges and a 1.125 in. (81pt) gutter margin (on binding side).
     142% While this is not an issue for electronic viewing, a PDF may be printed, and so we have the same page layout for both printed and electronic versions, we leave the gutter margin in.
    134143% Set margins to minimum permitted by uWaterloo thesis regulations:
    135144\setlength{\marginparwidth}{0pt} % width of margin notes
     
    140149\setlength{\evensidemargin}{0.125in} % Adds 1/8 in. to binding side of all
    141150% even-numbered pages when the "twoside" printing option is selected
    142 \setlength{\oddsidemargin}{0.125in} % Adds 1/8 in. to the left of all pages
    143 % when "oneside" printing is selected, and to the left of all odd-numbered
    144 % pages when "twoside" printing is selected
    145 \setlength{\textwidth}{6.375in} % assuming US letter paper (8.5 in. x 11 in.) and
    146 % side margins as above
     151\setlength{\oddsidemargin}{0.125in} % Adds 1/8 in. to the left of all pages when "oneside" printing is selected, and to the left of all odd-numbered pages when "twoside" printing is selected
     152\setlength{\textwidth}{6.375in} % assuming US letter paper (8.5 in. x 11 in.) and side margins as above
    147153\raggedbottom
    148154
    149 % The following statement specifies the amount of space between
    150 % paragraphs. Other reasonable specifications are \bigskipamount and \smallskipamount.
     155% The following statement specifies the amount of space between paragraphs. Other reasonable specifications are \bigskipamount and \smallskipamount.
    151156\setlength{\parskip}{\medskipamount}
    152157
    153 % The following statement controls the line spacing.  The default
    154 % spacing corresponds to good typographic conventions and only slight
    155 % changes (e.g., perhaps "1.2"), if any, should be made.
     158% The following statement controls the line spacing.
     159% The default spacing corresponds to good typographic conventions and only slight changes (e.g., perhaps "1.2"), if any, should be made.
    156160\renewcommand{\baselinestretch}{1} % this is the default line space setting
    157161
    158 % By default, each chapter will start on a recto (right-hand side)
    159 % page.  We also force each section of the front pages to start on
    160 % a recto page by inserting \cleardoublepage commands.
    161 % In many cases, this will require that the verso page be
    162 % blank and, while it should be counted, a page number should not be
    163 % printed.  The following statements ensure a page number is not
    164 % printed on an otherwise blank verso page.
     162% By default, each chapter will start on a recto (right-hand side) page.
     163% We also force each section of the front pages to start on a recto page by inserting \cleardoublepage commands.
     164% In many cases, this will require that the verso (left-hand) page be blank, and while it should be counted, a page number should not be printed.
     165% The following statements ensure a page number is not printed on an otherwise blank verso page.
    165166\let\origdoublepage\cleardoublepage
    166167\newcommand{\clearemptydoublepage}{%
     
    194195\input{common}
    195196\CFAStyle                                               % CFA code-style for all languages
    196 \lstset{basicstyle=\linespread{0.9}\tt}
     197\lstset{language=CFA,basicstyle=\linespread{0.9}\tt}    % CFA default language
    197198
    198199% glossary of terms to use
     
    200201\makeindex
    201202
    202 %======================================================================
    203 %   L O G I C A L    D O C U M E N T -- the content of your thesis
     203\newcommand\io{\glsxtrshort{io}\xspace}%
     204
     205%======================================================================
     206%   L O G I C A L    D O C U M E N T
     207% The logical document contains the main content of your thesis.
     208% Being a large document, it is a good idea to divide your thesis into several files, each one containing one chapter or other significant chunk of content, so you can easily shuffle things around later if desired.
    204209%======================================================================
    205210\begin{document}
    206211
    207 % For a large document, it is a good idea to divide your thesis
    208 % into several files, each one containing one chapter.
    209 % To illustrate this idea, the "front pages" (i.e., title page,
    210 % declaration, borrowers' page, abstract, acknowledgements,
    211 % dedication, table of contents, list of tables, list of figures,
    212 % nomenclature) are contained within the file "uw-ethesis-frontpgs.tex" which is
    213 % included into the document by the following statement.
    214212%----------------------------------------------------------------------
    215213% FRONT MATERIAL
     214% title page,declaration, borrowers' page, abstract, acknowledgements,
     215% dedication, table of contents, list of tables, list of figures, nomenclature, etc.
    216216%----------------------------------------------------------------------
    217217\input{text/front.tex}
    218218
    219 
    220219%----------------------------------------------------------------------
    221220% MAIN BODY
    222 %----------------------------------------------------------------------
    223 % Because this is a short document, and to reduce the number of files
    224 % needed for this template, the chapters are not separate
    225 % documents as suggested above, but you get the idea. If they were
    226 % separate documents, they would each start with the \chapter command, i.e,
    227 % do not contain \documentclass or \begin{document} and \end{document} commands.
     221% We suggest using a separate file for each chapter of your thesis.
     222% Start each chapter file with the \chapter command.
     223% Only use \documentclass or \begin{document} and \end{document} commands in this master document.
     224% Tip: Putting each sentence on a new line is a way to simplify later editing.
     225%----------------------------------------------------------------------
     226
    228227\part{Introduction}
    229228\input{text/intro.tex}
     
    232231\part{Design}
    233232\input{text/core.tex}
     233\input{text/io.tex}
    234234\input{text/practice.tex}
    235 \input{text/io.tex}
    236235\part{Evaluation}
    237236\label{Evaluation}
     
    243242%----------------------------------------------------------------------
    244243% END MATERIAL
    245 %----------------------------------------------------------------------
    246 
    247 % B I B L I O G R A P H Y
    248 % -----------------------
    249 
    250 % The following statement selects the style to use for references.  It controls the sort order of the entries in the bibliography and also the formatting for the in-text labels.
     244% Bibliography, Appendices, Index, etc.
     245%----------------------------------------------------------------------
     246
     247% Bibliography
     248
     249% The following statement selects the style to use for references.
     250% It controls the sort order of the entries in the bibliography and also the formatting for the in-text labels.
    251251\bibliographystyle{plain}
    252252% This specifies the location of the file containing the bibliographic information.
    253 % It assumes you're using BibTeX (if not, why not?).
    254 \cleardoublepage % This is needed if the book class is used, to place the anchor in the correct page,
    255                  % because the bibliography will start on its own page.
    256                  % Use \clearpage instead if the document class uses the "oneside" argument
     253% It assumes you're using BibTeX to manage your references (if not, why not?).
     254\cleardoublepage % This is needed if the "book" document class is used, to place the anchor in the correct page, because the bibliography will start on its own page.
     255% Use \clearpage instead if the document class uses the "oneside" argument
    257256\phantomsection  % With hyperref package, enables hyperlinking from the table of contents to bibliography
    258257% The following statement causes the title "References" to be used for the bibliography section:
     
    263262
    264263\bibliography{local,pl}
    265 % Tip 5: You can create multiple .bib files to organize your references.
     264% Tip: You can create multiple .bib files to organize your references.
    266265% Just list them all in the \bibliogaphy command, separated by commas (no spaces).
    267266
    268 % % The following statement causes the specified references to be added to the bibliography% even if they were not
    269 % % cited in the text. The asterisk is a wildcard that causes all entries in the bibliographic database to be included (optional).
     267% The following statement causes the specified references to be added to the bibliography even if they were not cited in the text.
     268% The asterisk is a wildcard that causes all entries in the bibliographic database to be included (optional).
    270269% \nocite{*}
     270%----------------------------------------------------------------------
     271
     272% Appendices
    271273
    272274% The \appendix statement indicates the beginning of the appendices.
    273275\appendix
    274 % Add a title page before the appendices and a line in the Table of Contents
     276% Add an un-numbered title page before the appendices and a line in the Table of Contents
    275277\chapter*{APPENDICES}
    276278\addcontentsline{toc}{chapter}{APPENDICES}
     279% Appendices are just more chapters, with different labeling (letters instead of numbers).
    277280%======================================================================
    278281\chapter[PDF Plots From Matlab]{Matlab Code for Making a PDF Plot}
     
    312315%\input{thesis.ind}                             % index
    313316
    314 \phantomsection
    315 
    316 \end{document}
     317\phantomsection         % allows hyperref to link to the correct page
     318
     319%----------------------------------------------------------------------
     320\end{document} % end of logical document
  • doc/user/figures/Cdecl.fig

    r342af53 r8e4aa05  
    19192 2 0 1 0 7 50 -1 -1 0.000 0 0 -1 0 0 5
    2020         2850 1200 3600 1200 3600 1350 2850 1350 2850 1200
    21 4 1 0 50 -1 4 10 0.0000 2 120 90 2925 1325 0\001
    22 4 1 0 50 -1 4 10 0.0000 2 120 90 3075 1325 1\001
    23 4 1 0 50 -1 4 10 0.0000 2 120 90 3225 1325 2\001
    24 4 1 0 50 -1 4 10 0.0000 2 120 90 3375 1325 3\001
    25 4 1 0 50 -1 4 10 0.0000 2 120 90 3525 1325 4\001
     214 1 0 50 -1 4 11 0.0000 2 120 90 2925 1325 0\001
     224 1 0 50 -1 4 11 0.0000 2 120 90 3075 1325 1\001
     234 1 0 50 -1 4 11 0.0000 2 120 90 3225 1325 2\001
     244 1 0 50 -1 4 11 0.0000 2 120 90 3375 1325 3\001
     254 1 0 50 -1 4 11 0.0000 2 120 90 3525 1325 4\001
    2626-6
    27272 2 0 1 0 7 50 -1 -1 0.000 0 0 -1 0 0 5
     
    5555        1 1 1.00 45.00 60.00
    5656         2550 1275 2850 1275
    57 4 1 0 50 -1 4 10 0.0000 2 120 90 1350 1650 0\001
    58 4 1 0 50 -1 4 10 0.0000 2 120 90 1500 1650 1\001
    59 4 1 0 50 -1 4 10 0.0000 2 120 90 1650 1650 2\001
    60 4 1 0 50 -1 4 10 0.0000 2 120 90 1800 1650 3\001
    61 4 1 0 50 -1 4 10 0.0000 2 120 90 1950 1650 4\001
    62 4 1 0 50 -1 4 10 0.0000 2 90 90 1200 1325 x\001
    63 4 1 0 50 -1 4 10 0.0000 2 90 90 2400 1325 x\001
     574 1 0 50 -1 4 11 0.0000 2 120 90 1350 1650 0\001
     584 1 0 50 -1 4 11 0.0000 2 120 90 1500 1650 1\001
     594 1 0 50 -1 4 11 0.0000 2 120 90 1650 1650 2\001
     604 1 0 50 -1 4 11 0.0000 2 120 90 1800 1650 3\001
     614 1 0 50 -1 4 11 0.0000 2 120 90 1950 1650 4\001
     624 1 0 50 -1 4 11 0.0000 2 90 90 1200 1325 x\001
     634 1 0 50 -1 4 11 0.0000 2 90 90 2400 1325 x\001
  • doc/user/user.tex

    r342af53 r8e4aa05  
    1111%% Created On       : Wed Apr  6 14:53:29 2016
    1212%% Last Modified By : Peter A. Buhr
    13 %% Last Modified On : Mon Oct  5 08:57:29 2020
    14 %% Update Count     : 3998
     13%% Last Modified On : Mon Feb 15 13:48:53 2021
     14%% Update Count     : 4452
    1515%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
    1616
     
    3737\usepackage{mathptmx}                                   % better math font with "times"
    3838\usepackage[usenames]{color}
     39\usepackage[dvips,plainpages=false,pdfpagelabels,pdfpagemode=UseNone,colorlinks=true,pagebackref=true,linkcolor=blue,citecolor=blue,urlcolor=blue,pagebackref=true,breaklinks=true]{hyperref}
     40\usepackage{breakurl}
     41
     42\renewcommand\footnoterule{\kern -3pt\rule{0.3\linewidth}{0.15pt}\kern 2pt}
     43
     44\usepackage[pagewise]{lineno}
     45\renewcommand{\linenumberfont}{\scriptsize\sffamily}
     46\usepackage[firstpage]{draftwatermark}
     47\SetWatermarkLightness{0.9}
     48
     49% Default underscore is too low and wide. Cannot use lstlisting "literate" as replacing underscore
     50% removes it as a variable-name character so keywords in variables are highlighted. MUST APPEAR
     51% AFTER HYPERREF.
     52\renewcommand{\textunderscore}{\leavevmode\makebox[1.2ex][c]{\rule{1ex}{0.075ex}}}
     53
     54\setlength{\topmargin}{-0.45in}                                                 % move running title into header
     55\setlength{\headsep}{0.25in}
     56
     57%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
     58
    3959\newcommand{\CFALatin}{}
    4060% inline code ©...© (copyright symbol) emacs: C-q M-)
     
    4666% math escape $...$ (dollar symbol)
    4767\input{common}                                          % common CFA document macros
    48 \usepackage[dvips,plainpages=false,pdfpagelabels,pdfpagemode=UseNone,colorlinks=true,pagebackref=true,linkcolor=blue,citecolor=blue,urlcolor=blue,pagebackref=true,breaklinks=true]{hyperref}
    49 \usepackage{breakurl}
    50 
    51 \renewcommand\footnoterule{\kern -3pt\rule{0.3\linewidth}{0.15pt}\kern 2pt}
    52 
    53 \usepackage[pagewise]{lineno}
    54 \renewcommand{\linenumberfont}{\scriptsize\sffamily}
    55 \usepackage[firstpage]{draftwatermark}
    56 \SetWatermarkLightness{0.9}
    57 
    58 % Default underscore is too low and wide. Cannot use lstlisting "literate" as replacing underscore
    59 % removes it as a variable-name character so keywords in variables are highlighted. MUST APPEAR
    60 % AFTER HYPERREF.
    61 \renewcommand{\textunderscore}{\leavevmode\makebox[1.2ex][c]{\rule{1ex}{0.075ex}}}
    62 
    63 \setlength{\topmargin}{-0.45in}                                                 % move running title into header
    64 \setlength{\headsep}{0.25in}
    65 
    66 %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
    67 
    6868\CFAStyle                                                                                               % use default CFA format-style
     69\lstset{language=CFA}                                                                   % CFA default lnaguage
    6970\lstnewenvironment{C++}[1][]                            % use C++ style
    70 {\lstset{language=C++,moredelim=**[is][\protect\color{red}]{®}{®},#1}}
     71{\lstset{language=C++,moredelim=**[is][\protect\color{red}]{@}{@},#1}}
    7172{}
    7273
     
    8182\newcommand{\Emph}[2][red]{{\color{#1}\textbf{\emph{#2}}}}
    8283\newcommand{\R}[1]{\Textbf{#1}}
     84\newcommand{\RC}[1]{\Textbf{\LstBasicStyle{#1}}}
    8385\newcommand{\B}[1]{{\Textbf[blue]{#1}}}
    8486\newcommand{\G}[1]{{\Textbf[OliveGreen]{#1}}}
     
    103105
    104106\author{
    105 \huge \CFA Team \medskip \\
    106 \Large Andrew Beach, Richard Bilson, Peter A. Buhr, Thierry Delisle, \smallskip \\
    107 \Large Glen Ditchfield, Rodolfo G. Esteves, Aaron Moss, Rob Schluntz
     107\huge \CFA Team (past and present) \medskip \\
     108\Large Andrew Beach, Richard Bilson, Michael Brooks, Peter A. Buhr, Thierry Delisle, \smallskip \\
     109\Large Glen Ditchfield, Rodolfo G. Esteves, Aaron Moss, Colby Parsons, Rob Schluntz, \smallskip \\
     110\Large Fangren Yu, Mubeen Zulfiqar
    108111}% author
    109112
     
    126129\vspace*{\fill}
    127130\noindent
    128 \copyright\,2016 \CFA Project \\ \\
     131\copyright\,2016, 2018, 2021 \CFA Project \\ \\
    129132\noindent
    130133This work is licensed under the Creative Commons Attribution 4.0 International License.
     
    144147\section{Introduction}
    145148
    146 \CFA{}\index{cforall@\CFA}\footnote{Pronounced ``\Index*{C-for-all}'', and written \CFA, CFA, or \CFL.} is a modern general-purpose programming-language, designed as an evolutionary step forward for the C programming language.
     149\CFA{}\index{cforall@\CFA}\footnote{Pronounced ``\Index*{C-for-all}'', and written \CFA, CFA, or \CFL.} is a modern general-purpose concurrent programming-language, designed as an evolutionary step forward for the C programming language.
    147150The syntax of \CFA builds from C and should look immediately familiar to C/\Index*[C++]{\CC{}} programmers.
    148151% Any language feature that is not described here can be assumed to be using the standard \Celeven syntax.
    149 \CFA adds many modern programming-language features that directly lead to increased \emph{\Index{safety}} and \emph{\Index{productivity}}, while maintaining interoperability with existing C programs and achieving similar performance.
     152\CFA adds many modern features that directly lead to increased \emph{\Index{safety}} and \emph{\Index{productivity}}, while maintaining interoperability with existing C programs and achieving similar performance.
    150153Like C, \CFA is a statically typed, procedural (non-\Index{object-oriented}) language with a low-overhead runtime, meaning there is no global \Index{garbage-collection}, but \Index{regional garbage-collection}\index{garbage-collection!regional} is possible.
    151154The primary new features include polymorphic routines and types, exceptions, concurrency, and modules.
     
    157160instead, a programmer evolves a legacy program into \CFA by incrementally incorporating \CFA features.
    158161As well, new programs can be written in \CFA using a combination of C and \CFA features.
     162In many ways, \CFA is to C as \Index{Scala}~\cite{Scala} is to Java, providing a vehicle for new typing and control-flow capabilities on top of a highly popular programming language allowing immediate dissemination.
    159163
    160164\Index*[C++]{\CC{}}~\cite{c++:v1} had a similar goal 30 years ago, allowing object-oriented programming to be incrementally added to C.
     
    165169For example, the following programs compare the C, \CFA, and \CC I/O mechanisms, where the programs output the same result.
    166170\begin{center}
    167 \begin{tabular}{@{}l@{\hspace{1.5em}}l@{\hspace{1.5em}}l@{}}
    168 \multicolumn{1}{c@{\hspace{1.5em}}}{\textbf{C}} & \multicolumn{1}{c}{\textbf{\CFA}}     & \multicolumn{1}{c}{\textbf{\CC}}      \\
    169 \begin{cfa}
    170 #include <stdio.h>§\indexc{stdio.h}§
     171\begin{tabular}{@{}l@{\hspace{1em}}l@{\hspace{1em}}l@{}}
     172\multicolumn{1}{c@{\hspace{1em}}}{\textbf{C}}   & \multicolumn{1}{c}{\textbf{\CFA}}     & \multicolumn{1}{c}{\textbf{\CC}}      \\
     173\begin{cfa}
     174#include <stdio.h>$\indexc{stdio.h}$
    171175
    172176int main( void ) {
    173177        int x = 0, y = 1, z = 2;
    174         ®printf( "%d %d %d\n", x, y, z );®
     178        @printf( "%d %d %d\n", x, y, z );@
    175179}
    176180\end{cfa}
    177181&
    178182\begin{cfa}
    179 #include <fstream>§\indexc{fstream}§
     183#include <fstream>$\indexc{fstream}$
    180184
    181185int main( void ) {
    182186        int x = 0, y = 1, z = 2;
    183         ®sout | x | y | z;®§\indexc{sout}§
     187        @sout | x | y | z;@$\indexc{sout}$
    184188}
    185189\end{cfa}
    186190&
    187191\begin{cfa}
    188 #include <iostream>§\indexc{iostream}§
     192#include <iostream>$\indexc{iostream}$
    189193using namespace std;
    190194int main() {
    191195        int x = 0, y = 1, z = 2;
    192         ®cout<<x<<" "<<y<<" "<<z<<endl;®
     196        @cout<<x<<" "<<y<<" "<<z<<endl;@
    193197}
    194198\end{cfa}
    195199\end{tabular}
    196200\end{center}
    197 While the \CFA I/O looks similar to the \Index*[C++]{\CC{}} output style, there are important differences, such as automatic spacing between variables as in \Index*{Python} (see~\VRef{s:IOLibrary}).
     201While \CFA I/O \see{\VRef{s:StreamIOLibrary}} looks similar to \Index*[C++]{\CC{}}, there are important differences, such as automatic spacing between variables and an implicit newline at the end of the expression list, similar to \Index*{Python}~\cite{Python}.
    198202
    199203
     
    210214\section{Why fix C?}
    211215
    212 The C programming language is a foundational technology for modern computing with millions of lines of code implementing everything from hobby projects to commercial operating-systems.
     216The C programming language is a foundational technology for modern computing with billions of lines of code implementing everything from hobby projects to commercial operating-systems.
    213217This installation base and the programmers producing it represent a massive software-engineering investment spanning decades and likely to continue for decades more.
    214218Even with all its problems, C continues to be popular because it allows writing software at virtually any level in a computer system without restriction.
    215 For system programming, where direct access to hardware, storage management, and real-time issues are a requirement, C is usually the only language of choice.
    216 The TIOBE index~\cite{TIOBE} for February 2020 ranks the top six most \emph{popular} programming languages as \Index*{Java} 17.4\%, C 16.8\%, Python 9.3\%, \Index*[C++]{\CC{}} 6.2\%, \Csharp 5.9\%, Visual Basic 5.9\% = 61.5\%, where the next 50 languages are less than 2\% each, with a long tail.
     219For system programming, where direct access to hardware, storage management, and real-time issues are a requirement, C is the only language of choice.
     220The TIOBE index~\cite{TIOBE} for February 2021 ranks the top six most \emph{popular} programming languages as C 17.4\%, \Index*{Java} 12\%, Python 12\%, \Index*[C++]{\CC{}} 7.6\%, \Csharp 4\%, Visual Basic 3.8\% = 56.8\%, where the next 50 languages are less than 2\% each, with a long tail.
    217221The top 4 rankings over the past 35 years are:
    218222\begin{center}
    219223\setlength{\tabcolsep}{10pt}
    220224\begin{tabular}{@{}rcccccccc@{}}
    221                 & 2020  & 2015  & 2010  & 2005  & 2000  & 1995  & 1990  & 1985  \\ \hline
    222 Java    & 1             & 2             & 1             & 2             & 3             & -             & -             & -             \\
    223 \R{C}   & \R{2} & \R{1} & \R{2} & \R{1} & \R{1} & \R{2} & \R{1} & \R{1} \\
    224 Python  & 3             & 7             & 6             & 6             & 22    & 21    & -             & -             \\
    225 \CC             & 4             & 4             & 4             & 3             & 2             & 1             & 2             & 12    \\
     225                & 2021  & 2016  & 2011  & 2006  & 2001  & 1996  & 1991  & 1986  \\ \hline
     226\R{C}   & \R{1} & \R{2} & \R{2} & \R{1} & \R{1} & \R{1} & \R{1} & \R{1} \\
     227Java    & 2             & 1             & 1             & 2             & 3             & 28    & -             & -             \\
     228Python  & 3             & 5             & 6             & 7             & 23    & 13    & -             & -             \\
     229\CC             & 4             & 3             & 3             & 3             & 2             & 2             & 2             & 8             \\
    226230\end{tabular}
    227231\end{center}
     
    232236As stated, the goal of the \CFA project is to engineer modern language-features into C in an evolutionary rather than revolutionary way.
    233237\CC~\cite{C++14,C++} is an example of a similar project;
    234 however, it largely extended the C language, and did not address most of C's existing problems.\footnote{%
     238however, it largely extended the C language, and did not address many of C's existing problems.\footnote{%
    235239Two important existing problems addressed were changing the type of character literals from ©int© to ©char© and enumerator from ©int© to the type of its enumerators.}
    236240\Index*{Fortran}~\cite{Fortran08}, \Index*{Ada}~\cite{Ada12}, and \Index*{Cobol}~\cite{Cobol14} are examples of programming languages that took an evolutionary approach, where modern language-features (\eg objects, concurrency) are added and problems fixed within the framework of the existing language.
     
    241245
    242246The result of this project is a language that is largely backwards compatible with \Index*[C11]{\Celeven{}}~\cite{C11}, but fixes many of the well known C problems while adding modern language-features.
    243 To achieve these goals required a significant engineering exercise, where we had to ``think inside the existing C box''.
    244 Without these significant extension to C, it is unable to cope with the needs of modern programming problems and programmers;
    245 as a result, it will fade into disuse.
    246 Considering the large body of existing C code and programmers, there is significant impetus to ensure C is transformed into a modern programming language.
     247To achieve these goals required a significant engineering exercise, \ie ``thinking \emph{inside} the C box''.
     248Considering the large body of existing C code and programmers, there is significant impetus to ensure C is transformed into a modern language.
    247249While \Index*[C11]{\Celeven{}} made a few simple extensions to the language, nothing was added to address existing problems in the language or to augment the language with modern language-features.
    248250While some may argue that modern language-features may make C complex and inefficient, it is clear a language without modern capabilities is insufficient for the advanced programming problems existing today.
     
    251253\section{History}
    252254
    253 The \CFA project started with \Index*{Dave Till}\index{Till, Dave}'s \Index*{K-W C}~\cite{Buhr94a,Till89}, which extended C with new declaration syntax, multiple return values from routines, and advanced assignment capabilities using the notion of tuples.
    254 (See~\cite{Werther96} for similar work in \Index*[C++]{\CC{}}.)
     255The \CFA project started with \Index*{Dave Till}\index{Till, Dave}'s \Index*{K-W C}~\cite{Buhr94a,Till89}, which extended C with new declaration syntax, multiple return values from routines, and advanced assignment capabilities using the notion of tuples \see{\cite{Werther96} for similar work in \Index*[C++]{\CC{}}}.
    255256The first \CFA implementation of these extensions was by \Index*{Rodolfo Esteves}\index{Esteves, Rodolfo}~\cite{Esteves04}.
    256257
    257258The signature feature of \CFA is \emph{\Index{overload}able} \Index{parametric-polymorphic} functions~\cite{forceone:impl,Cormack90,Duggan96} with functions generalized using a ©forall© clause (giving the language its name):
    258259\begin{cfa}
    259 ®forall( otype T )® T identity( T val ) { return val; }
    260 int forty_two = identity( 42 ); §\C{// T is bound to int, forty\_two == 42}§
     260@forall( otype T )@ T identity( T val ) { return val; }
     261int forty_two = identity( 42 ); $\C{// T is bound to int, forty\_two == 42}$
    261262\end{cfa}
    262263% extending the C type system with parametric polymorphism and overloading, as opposed to the \Index*[C++]{\CC{}} approach of object-oriented extensions.
    263264\CFA{}\hspace{1pt}'s polymorphism was originally formalized by \Index*{Glen Ditchfield}\index{Ditchfield, Glen}~\cite{Ditchfield92}, and first implemented by \Index*{Richard Bilson}\index{Bilson, Richard}~\cite{Bilson03}.
    264265However, at that time, there was little interesting in extending C, so work did not continue.
    265 As the saying goes, ``\Index*{What goes around, comes around.}'', and there is now renewed interest in the C programming language because of legacy code-bases, so the \CFA project has been restarted.
     266As the saying goes, ``\Index*{What goes around, comes around.}'', and there is now renewed interest in the C programming language because of the legacy code-base, so the \CFA project was restarted in 2015.
    266267
    267268
     
    273274This feature allows \CFA programmers to take advantage of the existing panoply of C libraries to access thousands of external software features.
    274275Language developers often state that adequate \Index{library support} takes more work than designing and implementing the language itself.
    275 Fortunately, \CFA, like \Index*[C++]{\CC{}}, starts with immediate access to all exiting C libraries, and in many cases, can easily wrap library routines with simpler and safer interfaces, at very low cost.
     276Fortunately, \CFA, like \Index*[C++]{\CC{}}, starts with immediate access to all exiting C libraries, and in many cases, can easily wrap library routines with simpler and safer interfaces, at zero or very low cost.
    276277Hence, \CFA begins by leveraging the large repository of C libraries, and than allows programmers to incrementally augment their C programs with modern \Index{backward-compatible} features.
    277278
     
    286287
    287288double key = 5.0, vals[10] = { /* 10 sorted floating values */ };
    288 double * val = (double *)bsearch( &key, vals, 10, sizeof(vals[0]), comp ); §\C{// search sorted array}§
     289double * val = (double *)bsearch( &key, vals, 10, sizeof(vals[0]), comp ); $\C{// search sorted array}$
    289290\end{cfa}
    290291which can be augmented simply with a polymorphic, type-safe, \CFA-overloaded wrappers:
     
    295296
    296297forall( otype T | { int ?<?( T, T ); } ) unsigned int bsearch( T key, const T * arr, size_t size ) {
    297         T * result = bsearch( key, arr, size ); §\C{// call first version}§
    298         return result ? result - arr : size; } §\C{// pointer subtraction includes sizeof(T)}§
    299 
    300 double * val = bsearch( 5.0, vals, 10 ); §\C{// selection based on return type}§
     298        T * result = bsearch( key, arr, size ); $\C{// call first version}$
     299        return result ? result - arr : size; } $\C{// pointer subtraction includes sizeof(T)}$
     300
     301double * val = bsearch( 5.0, vals, 10 ); $\C{// selection based on return type}$
    301302int posn = bsearch( 5.0, vals, 10 );
    302303\end{cfa}
     
    310311\begin{cfa}
    311312forall( dtype T | sized(T) ) T * malloc( void ) { return (T *)malloc( sizeof(T) ); }
    312 int * ip = malloc(); §\C{// select type and size from left-hand side}§
     313int * ip = malloc(); $\C{// select type and size from left-hand side}$
    313314double * dp = malloc();
    314315struct S {...} * sp = malloc();
     
    319320However, it is necessary to differentiate between C and \CFA code because of name \Index{overload}ing, as for \CC.
    320321For example, the C math-library provides the following routines for computing the absolute value of the basic types: ©abs©, ©labs©, ©llabs©, ©fabs©, ©fabsf©, ©fabsl©, ©cabsf©, ©cabs©, and ©cabsl©.
    321 Whereas, \CFA wraps each of these routines into ones with the overloaded name ©abs©:
    322 \begin{cfa}
    323 char ®abs®( char );
    324 extern "C" { int ®abs®( int ); } §\C{// use default C routine for int}§
    325 long int ®abs®( long int );
    326 long long int ®abs®( long long int );
    327 float ®abs®( float );
    328 double ®abs®( double );
    329 long double ®abs®( long double );
    330 float _Complex ®abs®( float _Complex );
    331 double _Complex ®abs®( double _Complex );
    332 long double _Complex ®abs®( long double _Complex );
    333 \end{cfa}
    334 The problem is the name clash between the library routine ©abs© and the \CFA names ©abs©.
    335 Hence, names appearing in an ©extern "C"© block have \newterm*{C linkage}.
    336 Then overloading polymorphism uses a mechanism called \newterm{name mangling}\index{mangling!name} to create unique names that are different from C names, which are not mangled.
    337 Hence, there is the same need, as in \CC, to know if a name is a C or \CFA name, so it can be correctly formed.
    338 There is no way around this problem, other than C's approach of creating unique names for each pairing of operation and types.
    339 
    340 This example strongly illustrates a core idea in \CFA: \emph{the \Index{power of a name}}.
     322Whereas, \CFA wraps each of these routines into one overloaded name ©abs©:
     323\begin{cfa}
     324char @abs@( char );
     325extern "C" { int @abs@( int ); } $\C{// use default C routine for int}$
     326long int @abs@( long int );
     327long long int @abs@( long long int );
     328float @abs@( float );
     329double @abs@( double );
     330long double @abs@( long double );
     331float _Complex @abs@( float _Complex );
     332double _Complex @abs@( double _Complex );
     333long double _Complex @abs@( long double _Complex );
     334\end{cfa}
     335The problem is \Index{name clash} between the C name ©abs© and the \CFA names ©abs©, resulting in two name linkages\index{C linkage}: ©extern "C"© and ©extern "Cforall"© (default).
     336Overloaded names must use \newterm{name mangling}\index{mangling!name} to create unique names that are different from unmangled C names.
     337Hence, there is the same need as in \CC to know if a name is a C or \CFA name, so it can be correctly formed.
     338The only way around this problem is C's approach of creating unique names for each pairing of operation and type.
     339
     340This example illustrates a core idea in \CFA: \emph{the \Index{power of a name}}.
    341341The name ``©abs©'' evokes the notion of absolute value, and many mathematical types provide the notion of absolute value.
    342342Hence, knowing the name ©abs© is sufficient to apply it to any type where it is applicable.
     
    344344
    345345
    346 \section[Compiling a CFA Program]{Compiling a \CFA Program}
     346\section{\CFA Compilation}
    347347
    348348The command ©cfa© is used to compile a \CFA program and is based on the \Index{GNU} \Indexc{gcc} command, \eg:
    349349\begin{cfa}
    350 cfa§\indexc{cfa}\index{compilation!cfa@©cfa©}§ [ gcc-options ] [ C/§\CFA{}§ source-files ] [ assembler/loader files ]
    351 \end{cfa}
    352 \CFA programs having the following ©gcc© flags turned on:
    353 \begin{description}
     350cfa$\indexc{cfa}\index{compilation!cfa@©cfa©}$ [ gcc/$\CFA{}$-options ] [ C/$\CFA{}$ source-files ] [ assembler/loader files ]
     351\end{cfa}
     352There is no ordering among options (flags) and files, unless an option has an argument, which must appear immediately after the option possibly with or without a space separating option and argument.
     353
     354\CFA has the following ©gcc© flags turned on:
     355\begin{description}[topsep=0pt]
    354356\item
    355357\Indexc{-std=gnu11}\index{compilation option!-std=gnu11@{©-std=gnu11©}}
     
    359361Use the traditional GNU semantics for inline routines in C11 mode, which allows inline routines in header files.
    360362\end{description}
    361 The following new \CFA options are available:
    362 \begin{description}
     363
     364\CFA has the following new options:
     365\begin{description}[topsep=0pt]
    363366\item
    364367\Indexc{-CFA}\index{compilation option!-CFA@©-CFA©}
    365 Only the C preprocessor and the \CFA translator steps are performed and the transformed program is written to standard output, which makes it possible to examine the code generated by the \CFA translator.
     368Only the C preprocessor (flag ©-E©) and the \CFA translator steps are performed and the transformed program is written to standard output, which makes it possible to examine the code generated by the \CFA translator.
    366369The generated code starts with the standard \CFA \Index{prelude}.
     370
     371\item
     372\Indexc{-XCFA}\index{compilation option!-XCFA@©-XCFA©}
     373Pass next flag as-is to the ©cfa-cpp© translator (see details below).
    367374
    368375\item
    369376\Indexc{-debug}\index{compilation option!-debug@©-debug©}
    370377The program is linked with the debugging version of the runtime system.
    371 The debug version performs runtime checks to help during the debugging phase of a \CFA program, but can substantially slow program execution.
     378The debug version performs runtime checks to aid the debugging phase of a \CFA program, but can substantially slow program execution.
    372379The runtime checks should only be removed after the program is completely debugged.
    373380\textbf{This option is the default.}
     
    399406\item
    400407\Indexc{-no-include-stdhdr}\index{compilation option!-no-include-stdhdr@©-no-include-stdhdr©}
    401 Do not supply ©extern "C"© wrappers for \Celeven standard include files (see~\VRef{s:StandardHeaders}).
     408Do not supply ©extern "C"© wrappers for \Celeven standard include files \see{\VRef{s:StandardHeaders}}.
    402409\textbf{This option is \emph{not} the default.}
    403410\end{comment}
     
    430437\begin{cfa}
    431438#ifndef __CFORALL__
    432 #include <stdio.h>§\indexc{stdio.h}§ §\C{// C header file}§
     439#include <stdio.h>$\indexc{stdio.h}$ $\C{// C header file}$
    433440#else
    434 #include <fstream>§\indexc{fstream}§ §\C{// \CFA header file}§
     441#include <fstream>$\indexc{fstream}$ $\C{// \CFA header file}$
    435442#endif
    436443\end{cfa}
     
    438445
    439446The \CFA translator has multiple steps.
    440 The following flags control how the tranlator works, the stages run, and printing within a stage.
     447The following flags control how the translator works, the stages run, and printing within a stage.
    441448The majority of these flags are used by \CFA developers, but some are occasionally useful to programmers.
     449Each option must be escaped with \Indexc{-XCFA}\index{translator option!-XCFA@{©-XCFA©}} to direct it to the compiler step, similar to the ©-Xlinker© flag for the linker, \eg:
     450\begin{lstlisting}[language=sh]
     451cfa $test$.cfa -CFA -XCFA -p # print translated code without printing the standard prelude
     452cfa $test$.cfa -XCFA -P -XCFA parse -XCFA -n # show program parse without prelude
     453\end{lstlisting}
    442454\begin{description}[topsep=5pt,itemsep=0pt,parsep=0pt]
    443455\item
    444 \Indexc{-h}\index{translator option!-h@{©-h©}}, \Indexc{--help}\index{translator option!--help@{©--help©}} \, print help message
    445 \item
    446 \Indexc{-l}\index{translator option!-l@{©-l©}}, \Indexc{--libcfa}\index{translator option!--libcfa@{©--libcfa©}} \, generate libcfa.c
     456\Indexc{-c}\index{translator option!-c@{©-c©}}, \Indexc{--colors}\index{translator option!--colors@{©--colors©}} \, diagnostic color: ©never©, ©always©, \lstinline[deletekeywords=auto]{auto}
     457\item
     458\Indexc{-g}\index{translator option!-g@{©-g©}}, \Indexc{--gdb}\index{translator option!--gdb@{©--gdb©}} \, wait for gdb to attach
     459\item
     460\Indexc{-h}\index{translator option!-h@{©-h©}}, \Indexc{--help}\index{translator option!--help@{©--help©}} \, print translator help message
     461\item
     462\Indexc{-l}\index{translator option!-l@{©-l©}}, \Indexc{--libcfa}\index{translator option!--libcfa@{©--libcfa©}} \, generate ©libcfa.c©
    447463\item
    448464\Indexc{-L}\index{translator option!-L@{©-L©}}, \Indexc{--linemarks}\index{translator option!--linemarks@{©--linemarks©}} \, generate line marks
     
    454470\Indexc{-n}\index{translator option!-n@{©-n©}}, \Indexc{--no-prelude}\index{translator option!--no-prelude@{©--no-prelude©}} \, do not read prelude
    455471\item
    456 \Indexc{-p}\index{translator option!-p@{©-p©}}, \Indexc{--prototypes}\index{translator option!--prototypes@{©--prototypes©}} \, generate prototypes for prelude functions
     472\Indexc{-p}\index{translator option!-p@{©-p©}}, \Indexc{--prototypes}\index{translator option!--prototypes@{©--prototypes©}} \, do not generate prelude prototypes $\Rightarrow$ prelude not printed
     473\item
     474\Indexc{-d}\index{translator option!-d@{©-d©}}, \Indexc{--deterministic-out}\index{translator option!--deterministic-out@{©--deterministic-out©}} \, only print deterministic output
    457475\item
    458476\Indexc{-P}\index{translator option!-P@{©-P©}}, \Indexc{--print}\index{translator option!--print@{©--print©}} \, one of:
    459477\begin{description}[topsep=0pt,itemsep=0pt,parsep=0pt]
    460478\item
     479\Indexc{ascodegen}\index{translator option!-P@{©-P©}!©ascodegen©}\index{translator option!--print@{©-print©}!©ascodegen©} \, as codegen rather than AST
     480\item
     481\Indexc{asterr}\index{translator option!-P@{©-P©}!©asterr©}\index{translator option!--print@{©-print©}!©asterr©} \, AST on error
     482\item
     483\Indexc{declstats}\index{translator option!-P@{©-P©}!©declstats©}\index{translator option!--print@{©-print©}!©declstats©} \, code property statistics
     484\item
     485\Indexc{parse}\index{translator option!-P@{©-P©}!©parse©}\index{translator option!--print@{©-print©}!©parse©} \, yacc (parsing) debug information
     486\item
     487\Indexc{pretty}\index{translator option!-P@{©-P©}!©pretty©}\index{translator option!--print@{©-print©}!©pretty©} \, prettyprint for ©ascodegen© flag
     488\item
     489\Indexc{rproto}\index{translator option!-P@{©-P©}!©rproto©}\index{translator option!--print@{©-print©}!©rproto©} \, resolver-proto instance
     490\item
     491\Indexc{rsteps}\index{translator option!-P@{©-P©}!©rsteps©}\index{translator option!--print@{©-print©}!©rsteps©} \, resolver steps
     492\item
     493\Indexc{tree}\index{translator option!-P@{©-P©}!©tree©}\index{translator option!--print@{©-print©}!©tree©} \, parse tree
     494\item
     495\Indexc{ast}\index{translator option!-P@{©-P©}!©ast©}\index{translator option!--print@{©-print©}!©ast©} \, AST after parsing
     496\item
     497\Indexc{symevt}\index{translator option!-P@{©-P©}!©symevt©}\index{translator option!--print@{©-print©}!©symevt©} \, symbol table events
     498\item
    461499\Indexc{altexpr}\index{translator option!-P@{©-P©}!©altexpr©}\index{translator option!--print@{©-print©}!©altexpr©} \, alternatives for expressions
    462500\item
    463 \Indexc{ascodegen}\index{translator option!-P@{©-P©}!©ascodegen©}\index{translator option!--print@{©-print©}!©ascodegen©} \, as codegen rather than AST
    464 \item
    465 \Indexc{ast}\index{translator option!-P@{©-P©}!©ast©}\index{translator option!--print@{©-print©}!©ast©} \, AST after parsing
    466 \item
    467501\Indexc{astdecl}\index{translator option!-P@{©-P©}!©astdecl©}\index{translator option!--print@{©-print©}!©astdecl©} \, AST after declaration validation pass
    468502\item
    469 \Indexc{asterr}\index{translator option!-P@{©-P©}!©asterr©}\index{translator option!--print@{©-print©}!©asterr©} \, AST on error
     503\Indexc{resolver}\index{translator option!-P@{©-P©}!©resolver©}\index{translator option!--print@{©-print©}!©resolver©} \, before resolver step
    470504\item
    471505\Indexc{astexpr}\index{translator option!-P@{©-P©}!©astexpr©}\index{translator option!--print@{©-print©}!©altexpr©} \, AST after expression analysis
    472506\item
     507\Indexc{ctordtor}\index{translator option!-P@{©-P©}!©ctordtor©}\index{translator option!--print@{©-print©}!©ctordtor©} \, after ctor/dtor are replaced
     508\item
     509\Indexc{tuple}\index{translator option!-P@{©-P©}!©tuple©}\index{translator option!--print@{©-print©}!©tuple©} \, after tuple expansion
     510\item
    473511\Indexc{astgen}\index{translator option!-P@{©-P©}!©astgen©}\index{translator option!--print@{©-print©}!©astgen©} \, AST after instantiate generics
    474512\item
    475513\Indexc{box}\index{translator option!-P@{©-P©}!©box©}\index{translator option!--print@{©-print©}!©box©} \, before box step
    476514\item
    477 \Indexc{ctordtor}\index{translator option!-P@{©-P©}!©ctordtor©}\index{translator option!--print@{©-print©}!©ctordtor©} \, after ctor/dtor are replaced
    478 \item
    479515\Indexc{codegen}\index{translator option!-P@{©-P©}!©codegen©}\index{translator option!--print@{©-print©}!©codegen©} \, before code generation
    480 \item
    481 \Indexc{declstats}\index{translator option!-P@{©-P©}!©declstats©}\index{translator option!--print@{©-print©}!©declstats©} \, code property statistics
    482 \item
    483 \Indexc{parse}\index{translator option!-P@{©-P©}!©parse©}\index{translator option!--print@{©-print©}!©parse©} \, yacc (parsing) debug information
    484 \item
    485 \Indexc{pretty}\index{translator option!-P@{©-P©}!©pretty©}\index{translator option!--print@{©-print©}!©pretty©} \, prettyprint for ascodegen flag
    486 \item
    487 \Indexc{resolver}\index{translator option!-P@{©-P©}!©resolver©}\index{translator option!--print@{©-print©}!©resolver©} \, before resolver step
    488 \item
    489 \Indexc{rproto}\index{translator option!-P@{©-P©}!©rproto©}\index{translator option!--print@{©-print©}!©rproto©} \, resolver-proto instance
    490 \item
    491 \Indexc{rsteps}\index{translator option!-P@{©-P©}!©rsteps©}\index{translator option!--print@{©-print©}!©rsteps©} \, resolver steps
    492 \item
    493 \Indexc{symevt}\index{translator option!-P@{©-P©}!©symevt©}\index{translator option!--print@{©-print©}!©symevt©} \, symbol table events
    494 \item
    495 \Indexc{tree}\index{translator option!-P@{©-P©}!©tree©}\index{translator option!--print@{©-print©}!©tree©} \, parse tree
    496 \item
    497 \Indexc{tuple}\index{translator option!-P@{©-P©}!©tuple©}\index{translator option!--print@{©-print©}!©tuple©} \, after tuple expansion
    498516\end{description}
    499517\item
    500518\Indexc{--prelude-dir} <directory> \, prelude directory for debug/nodebug
    501519\item
    502 \Indexc{-S}\index{translator option!-S@{©-S©}!©counters,heap,time,all,none©}, \Indexc{--statistics}\index{translator option!--statistics@{©--statistics©}!©counters,heap,time,all,none©} <option-list> \, enable profiling information:
    503 \begin{description}[topsep=0pt,itemsep=0pt,parsep=0pt]
    504 \item
    505 \Indexc{counters,heap,time,all,none}
    506 \end{description}
     520\Indexc{-S}\index{translator option!-S@{©-S©}!©counters,heap,time,all,none©}, \Indexc{--statistics}\index{translator option!--statistics@{©--statistics©}!©counters,heap,time,all,none©} <option-list> \, enable profiling information: ©counters©, ©heap©, ©time©, ©all©, ©none©
    507521\item
    508522\Indexc{-t}\index{translator option!-t@{©-t©}}, \Indexc{--tree}\index{translator option!--tree@{©--tree©}} build in tree
     
    513527\label{s:BackquoteIdentifiers}
    514528
    515 \CFA introduces several new keywords (see \VRef{s:CFAKeywords}) that can clash with existing C variable-names in legacy code.
     529\CFA introduces several new keywords \see{\VRef{s:CFAKeywords}} that can clash with existing C variable-names in legacy code.
    516530Keyword clashes are accommodated by syntactic transformations using the \CFA backquote escape-mechanism:
    517531\begin{cfa}
    518 int ®``®otype = 3; §\C{// make keyword an identifier}§
    519 double ®``®forall = 3.5;
     532int @``@otype = 3; $\C{// make keyword an identifier}$
     533double @``@forall = 3.5;
    520534\end{cfa}
    521535
    522536Existing C programs with keyword clashes can be converted by enclosing keyword identifiers in backquotes, and eventually the identifier name can be changed to a non-keyword name.
    523 \VRef[Figure]{f:HeaderFileInterposition} shows how clashes in existing C header-files (see~\VRef{s:StandardHeaders}) can be handled using preprocessor \newterm{interposition}: ©#include_next© and ©-I filename©.
     537\VRef[Figure]{f:HeaderFileInterposition} shows how clashes in existing C header-files \see{\VRef{s:StandardHeaders}} can be handled using preprocessor \newterm{interposition}: ©#include_next© and ©-I filename©.
    524538Several common C header-files with keyword clashes are fixed in the standard \CFA header-library, so there is a seamless programming-experience.
    525539
     
    527541\begin{cfa}
    528542// include file uses the CFA keyword "with".
    529 #if ! defined( with ) §\C{// nesting ?}§
    530 #define with ®``®with §\C{// make keyword an identifier}§
     543#if ! defined( with )                                                   $\C{// nesting ?}$
     544#define with @``@with                                                   $\C{// make keyword an identifier}$
    531545#define __CFA_BFD_H__
    532546#endif
    533 §{\color{red}\#\textbf{include\_next} <bfdlink.h>}§ §\C{// must have internal check for multiple expansion}§
    534 #if defined( with ) && defined( __CFA_BFD_H__ ) §\C{// reset only if set}§
     547$\R{\#include\_next} <bfdlink.h>$                               $\C{// must have internal check for multiple expansion}$
     548#if defined( with ) && defined( __CFA_BFD_H__ ) $\C{// reset only if set}$
    535549#undef with
    536550#undef __CFA_BFD_H__
     
    544558\section{Constant Underscores}
    545559
    546 Numeric constants are extended to allow \Index{underscore}s\index{constant!underscore}, \eg:
    547 \begin{cfa}
    548 2®_®147®_®483®_®648; §\C{// decimal constant}§
    549 56®_®ul; §\C{// decimal unsigned long constant}§
    550 0®_®377; §\C{// octal constant}§
    551 0x®_®ff®_®ff; §\C{// hexadecimal constant}§
    552 0x®_®ef3d®_®aa5c; §\C{// hexadecimal constant}§
    553 3.141®_®592®_®654; §\C{// floating constant}§
    554 10®_®e®_®+1®_®00; §\C{// floating constant}§
    555 0x®_®ff®_®ff®_®p®_®3; §\C{// hexadecimal floating}§
    556 0x®_®1.ffff®_®ffff®_®p®_®128®_®l; §\C{// hexadecimal floating long constant}§
    557 L®_®§"\texttt{\textbackslash{x}}§®_®§\texttt{ff}§®_®§\texttt{ee}"§; §\C{// wide character constant}§
     560Numeric constants are extended to allow \Index{underscore}s\index{constant!underscore} as a separator, \eg:
     561\begin{cfa}
     5622@_@147@_@483@_@648; $\C{// decimal constant}$
     56356@_@ul; $\C{// decimal unsigned long constant}$
     5640@_@377; $\C{// octal constant}$
     5650x@_@ff@_@ff; $\C{// hexadecimal constant}$
     5660x@_@ef3d@_@aa5c; $\C{// hexadecimal constant}$
     5673.141@_@592@_@654; $\C{// floating constant}$
     56810@_@e@_@+1@_@00; $\C{// floating constant}$
     5690x@_@ff@_@ff@_@p@_@3; $\C{// hexadecimal floating}$
     5700x@_@1.ffff@_@ffff@_@p@_@128@_@l; $\C{// hexadecimal floating long constant}$
     571L@_@$"\texttt{\textbackslash{x}}$@_@$\texttt{ff}$@_@$\texttt{ee}"$; $\C{// wide character constant}$
    558572\end{cfa}
    559573The rules for placement of underscores are:
     
    574588It is significantly easier to read and enter long constants when they are broken up into smaller groupings (many cultures use comma and/or period among digits for the same purpose).
    575589This extension is backwards compatible, matches with the use of underscore in variable names, and appears in \Index*{Ada} and \Index*{Java} 8.
     590\CC uses the single quote (©'©) as a separator, restricted within a sequence of digits, \eg ©0xaa©©'©©ff©, ©3.141©©'©©592E1©©'©©1©.
    576591
    577592
    578593\section{Exponentiation Operator}
    579594
    580 C, \CC, and Java (and many other programming languages) have no exponentiation operator\index{exponentiation!operator}\index{operator!exponentiation}, \ie $x^y$, and instead use a routine, like \Indexc{pow(x,y)}, to perform the exponentiation operation.
    581 \CFA extends the basic operators with the exponentiation operator ©?®\®?©\index{?\\?@©?®\®?©} and ©?\=?©\index{?\\=?@©®\®=?©}, as in, ©x ®\® y© and ©x ®\®= y©, which means $x^y$ and $x \leftarrow x^y$.
    582 The priority of the exponentiation operator is between the cast and multiplicative operators, so that ©w * (int)x \ (int)y * z© is parenthesized as ©((w * (((int)x) \ ((int)y))) * z)©.
     595C, \CC, and Java (and other programming languages) have no exponentiation operator\index{exponentiation!operator}\index{operator!exponentiation}, \ie $x^y$, and instead use a routine, like \Indexc{pow(x,y)}, to perform the exponentiation operation.
     596\CFA extends the basic operators with the exponentiation operator ©?©\R{©\\©}©?©\index{?\\?@©?@\@?©} and ©?©\R{©\\©}©=?©\index{?\\=?@©@\@=?©}, as in, ©x ©\R{©\\©}© y© and ©x ©\R{©\\©}©= y©, which means $x^y$ and $x \leftarrow x^y$.
     597The priority of the exponentiation operator is between the cast and multiplicative operators, so that ©w * (int)x \ (int)y * z© is parenthesized as ©(w * (((int)x) \ ((int)y))) * z©.
    583598
    584599There are exponentiation operators for integral and floating types, including the builtin \Index{complex} types.
     
    587602Floating exponentiation\index{exponentiation!floating} is performed using \Index{logarithm}s\index{exponentiation!logarithm}, so the exponent cannot be negative.
    588603\begin{cfa}
    589 sout | 1 ®\® 0 | 1 ®\® 1 | 2 ®\® 8 | -4 ®\® 3 | 5 ®\® 3 | 5 ®\® 32 | 5L ®\® 32 | 5L ®\® 64 | -4 ®\® -3 | -4.0 ®\® -3 | 4.0 ®\® 2.1
    590            | (1.0f+2.0fi) ®\® (3.0f+2.0fi);
    591 1 1 256 -64 125 ®0® 3273344365508751233 ®0® ®0® -0.015625 18.3791736799526 0.264715-1.1922i
     604sout | 1 @\@ 0 | 1 @\@ 1 | 2 @\@ 8 | -4 @\@ 3 | 5 @\@ 3 | 5 @\@ 32 | 5L @\@ 32 | 5L @\@ 64 | -4 @\@ -3 | -4.0 @\@ -3 | 4.0 @\@ 2.1
     605           | (1.0f+2.0fi) @\@ (3.0f+2.0fi);
     6061 1 256 -64 125 @0@ 3273344365508751233 @0@ @0@ -0.015625 18.3791736799526 0.264715-1.1922i
    592607\end{cfa}
    593608Note, ©5 \ 32© and ©5L \ 64© overflow, and ©-4 \ -3© is a fraction but stored in an integer so all three computations generate an integral zero.
    594 Parenthesis are necessary for complex constants or the expression is parsed as ©1.0f+®(®2.0fi \ 3.0f®)®+2.0fi©.
     609Because exponentiation has higher priority than ©+©, parenthesis are necessary for exponentiation of \Index{complex constant}s or the expression is parsed as ©1.0f+©\R{©(©}©2.0fi \ 3.0f©\R{©)©}©+2.0fi©, requiring \R{©(©}©1.0f+2.0fi©\R{©)©}© \ ©\R{©(©}©3.0f+2.0fi©\R{©)©}.
     610
    595611The exponentiation operator is available for all the basic types, but for user-defined types, only the integral-computation version is available.
    596612\begin{cfa}
    597 forall( otype OT | { void ?{}( OT & this, one_t ); OT ?*?( OT, OT ); } )
    598 OT ?®\®?( OT ep, unsigned int y );
    599 forall( otype OT | { void ?{}( OT & this, one_t ); OT ?*?( OT, OT ); } )
    600 OT ?®\®?( OT ep, unsigned long int y );
     613forall( otype T | { void ?{}( T & this, one_t ); T ?*?( T, T ); } )
     614T ?@\@?( T ep, unsigned int y );
     615forall( otype T | { void ?{}( T & this, one_t ); T ?*?( T, T ); } )
     616T ?@\@?( T ep, unsigned long int y );
    601617\end{cfa}
    602618The user type ©T© must define multiplication, one (©1©), and ©*©.
     
    609625
    610626%\subsection{\texorpdfstring{\protect\lstinline@if@/\protect\lstinline@while@ Statement}{if Statement}}
    611 \subsection{\texorpdfstring{\LstKeywordStyle{if}/\LstKeywordStyle{while} Statement}{if/while Statement}}
    612 
    613 The ©if©/©while© expression allows declarations, similar to ©for© declaration expression.
    614 (Does not make sense for ©do©-©while©.)
    615 \begin{cfa}
    616 if ( ®int x = f()® ) ... §\C{// x != 0}§
    617 if ( ®int x = f(), y = g()® ) ... §\C{// x != 0 \&\& y != 0}§
    618 if ( ®int x = f(), y = g(); x < y® ) ... §\C{// relational expression}§
    619 if ( ®struct S { int i; } x = { f() }; x.i < 4® ) §\C{// relational expression}§
    620 
    621 while ( ®int x = f()® ) ... §\C{// x != 0}§
    622 while ( ®int x = f(), y = g()® ) ... §\C{// x != 0 \&\& y != 0}§
    623 while ( ®int x = f(), y = g(); x < y® ) ... §\C{// relational expression}§
    624 while ( ®struct S { int i; } x = { f() }; x.i < 4® ) ... §\C{// relational expression}§
    625 \end{cfa}
    626 Unless a relational expression is specified, each variable is compared not equal to 0, which is the standard semantics for the ©if©/©while© expression, and the results are combined using the logical ©&&© operator.\footnote{\CC only provides a single declaration always compared not equal to 0.}
    627 The scope of the declaration(s) is local to the @if@ statement but exist within both the ``then'' and ``else'' clauses.
     627\subsection{\texorpdfstring{\LstKeywordStyle{if} / \LstKeywordStyle{while} Statement}{if / while Statement}}
     628
     629The ©if©/©while© expression allows declarations, similar to ©for© declaration expression.\footnote{
     630Declarations in the ©do©-©while© condition are not useful because they appear after the loop body.}
     631\begin{cfa}
     632if ( @int x = f()@ ) ... $\C{// x != 0}$
     633if ( @int x = f(), y = g()@ ) ... $\C{// x != 0 \&\& y != 0}$
     634if ( @int x = f(), y = g(); x < y@ ) ... $\C{// relational expression}$
     635if ( @struct S { int i; } x = { f() }; x.i < 4@ ) $\C{// relational expression}$
     636
     637while ( @int x = f()@ ) ... $\C{// x != 0}$
     638while ( @int x = f(), y = g()@ ) ... $\C{// x != 0 \&\& y != 0}$
     639while ( @int x = f(), y = g(); x < y@ ) ... $\C{// relational expression}$
     640while ( @struct S { int i; } x = { f() }; x.i < 4@ ) ... $\C{// relational expression}$
     641\end{cfa}
     642Unless a relational expression is specified, each variable is compared not equal to 0, which is the standard semantics for the ©if©/©while© expression, and the results are combined using the logical ©&&© operator.
     643The scope of the declaration(s) is local to the ©if© statement but exist within both the \emph{then} and \emph{else} clauses.
     644\CC only provides a single declaration always compared ©!=© to 0.
    628645
    629646
    630647%\section{\texorpdfstring{\protect\lstinline@case@ Clause}{case Clause}}
    631648\subsection{\texorpdfstring{\LstKeywordStyle{case} Clause}{case Clause}}
     649\label{s:caseClause}
    632650
    633651C restricts the ©case© clause of a ©switch© statement to a single value.
     
    640658\begin{cfa}
    641659switch ( i ) {
    642   case ®1, 3, 5®:
     660  case @1, 3, 5@:
    643661        ...
    644   case ®2, 4, 6®:
     662  case @2, 4, 6@:
    645663        ...
    646664}
     
    670688\begin{cfa}
    671689switch ( i ) {
    672   case ®1~5:® §\C{// 1, 2, 3, 4, 5}§
     690  case @1~5:@ $\C{// 1, 2, 3, 4, 5}$
    673691        ...
    674   case ®10~15:® §\C{// 10, 11, 12, 13, 14, 15}§
     692  case @10~15:@ $\C{// 10, 11, 12, 13, 14, 15}$
    675693        ...
    676694}
     
    678696Lists of subranges are also allowed.
    679697\begin{cfa}
    680 case ®1~5, 12~21, 35~42®:
     698case @1~5, 12~21, 35~42@:
    681699\end{cfa}
    682700
     
    722740if ( argc == 3 ) {
    723741        // open output file
    724         ®// open input file
    725 ®} else if ( argc == 2 ) {
    726         ®// open input file (duplicate)
    727 
    728 ®} else {
     742        @// open input file
     743@} else if ( argc == 2 ) {
     744        @// open input file (duplicate)
     745
     746@} else {
    729747        // usage message
    730748}
     
    733751\end{cquote}
    734752In this example, case 2 is always done if case 3 is done.
    735 This control flow is difficult to simulate with if statements or a ©switch© statement without fall-through as code must be duplicated or placed in a separate routine.
     753This control flow is difficult to simulate with ©if© statements or a ©switch© statement without fall-through as code must be duplicated or placed in a separate routine.
    736754C also uses fall-through to handle multiple case-values resulting in the same action:
    737755\begin{cfa}
    738756switch ( i ) {
    739   ®case 1: case 3: case 5:®     // odd values
     757  @case 1: case 3: case 5:@     // odd values
    740758        // odd action
    741759        break;
    742   ®case 2: case 4: case 6:®     // even values
     760  @case 2: case 4: case 6:@     // even values
    743761        // even action
    744762        break;
    745763}
    746764\end{cfa}
    747 However, this situation is handled in other languages without fall-through by allowing a list of case values.
    748 While fall-through itself is not a problem, the problem occurs when fall-through is the default, as this semantics is unintuitive to many programmers and is different from virtually all other programming languages with a ©switch© statement.
     765This situation better handled without fall-through by allowing a list of case values \see{\VRef{s:caseClause}}.
     766While fall-through itself is not a problem, the problem occurs when fall-through is the default, as this semantics is unintuitive to many programmers and is different from most programming languages with a ©switch© statement.
    749767Hence, default fall-through semantics results in a large number of programming errors as programmers often \emph{forget} the ©break© statement at the end of a ©case© clause, resulting in inadvertent fall-through.
    750768
     
    756774        if ( j < k ) {
    757775                ...
    758           ®case 1:®             // transfer into "if" statement
     776          @case 1:@             // transfer into "if" statement
    759777                ...
    760778        } // if
     
    762780        while ( j < 5 ) {
    763781                ...
    764           ®case 3:®             // transfer into "while" statement
     782          @case 3:@             // transfer into "while" statement
    765783                ...
    766784        } // while
    767785} // switch
    768786\end{cfa}
    769 The problem with this usage is branching into control structures, which is known to cause both comprehension and technical difficulties.
    770 The comprehension problem occurs from the inability to determine how control reaches a particular point due to the number of branches leading to it.
     787This usage branches into control structures, which is known to cause both comprehension and technical difficulties.
     788The comprehension problem results from the inability to determine how control reaches a particular point due to the number of branches leading to it.
    771789The technical problem results from the inability to ensure declaration and initialization of variables when blocks are not entered at the beginning.
    772 There are no positive arguments for this kind of control flow, and therefore, there is a strong impetus to eliminate it.
     790There are few arguments for this kind of control flow, and therefore, there is a strong impetus to eliminate it.
    773791Nevertheless, C does have an idiom where this capability is used, known as ``\Index*{Duff's device}''~\cite{Duff83}:
    774792\begin{cfa}
     
    794812\item
    795813It is possible to place the ©default© clause anywhere in the list of labelled clauses for a ©switch© statement, rather than only at the end.
    796 Virtually all programming languages with a ©switch© statement require the ©default© clause to appear last in the case-clause list.
     814Most programming languages with a ©switch© statement require the ©default© clause to appear last in the case-clause list.
    797815The logic for this semantics is that after checking all the ©case© clauses without success, the ©default© clause is selected;
    798816hence, physically placing the ©default© clause at the end of the ©case© clause list matches with this semantics.
     
    803821\begin{cfa}
    804822switch ( x ) {
    805         ®int y = 1;® §\C{// unreachable initialization}§
    806         ®x = 7;® §\C{// unreachable code without label/branch}§
     823        @int y = 1;@ $\C{// unreachable initialization}$
     824        @x = 7;@ $\C{// unreachable code without label/branch}$
    807825  case 0: ...
    808826        ...
    809         ®int z = 0;® §\C{// unreachable initialization, cannot appear after case}§
     827        @int z = 0;@ $\C{// unreachable initialization, cannot appear after case}$
    810828        z = 2;
    811829  case 1:
    812         ®x = z;® §\C{// without fall through, z is uninitialized}§
     830        @x = z;@ $\C{// without fall through, z is uninitialized}$
    813831}
    814832\end{cfa}
    815833While the declaration of the local variable ©y© is useful with a scope across all ©case© clauses, the initialization for such a variable is defined to never be executed because control always transfers over it.
    816 Furthermore, any statements before the first ©case© clause can only be executed if labelled and transferred to using a ©goto©, either from outside or inside of the ©switch©, both of which are problematic.
    817 As well, the declaration of ©z© cannot occur after the ©case© because a label can only be attached to a statement, and without a fall through to case 3, ©z© is uninitialized.
    818 The key observation is that the ©switch© statement branches into control structure, \ie there are multiple entry points into its statement body.
     834Furthermore, any statements before the first ©case© clause can only be executed if labelled and transferred to using a ©goto©, either from outside or inside of the ©switch©, where both are problematic.
     835As well, the declaration of ©z© cannot occur after the ©case© because a label can only be attached to a statement, and without a fall-through to case 3, ©z© is uninitialized.
     836The key observation is that the ©switch© statement branches into a control structure, \ie there are multiple entry points into its statement body.
    819837\end{enumerate}
    820838
     
    842860Therefore, to preserve backwards compatibility, it is necessary to introduce a new kind of ©switch© statement, called ©choose©, with no implicit fall-through semantics and an explicit fall-through if the last statement of a case-clause ends with the new keyword ©fallthrough©/©fallthru©, \eg:
    843861\begin{cfa}
    844 ®choose® ( i ) {
     862@choose@ ( i ) {
    845863  case 1:  case 2:  case 3:
    846864        ...
    847         ®// implicit end of switch (break)
    848   ®case 5:
     865        @// implicit end of switch (break)
     866  @case 5:
    849867        ...
    850         ®fallthru®; §\C{// explicit fall through}§
     868        @fallthru@; $\C{// explicit fall through}$
    851869  case 7:
    852870        ...
    853         ®break® §\C{// explicit end of switch (redundant)}§
     871        @break@ $\C{// explicit end of switch (redundant)}$
    854872  default:
    855873        j = 3;
    856874}
    857875\end{cfa}
    858 Like the ©switch© statement, the ©choose© statement retains the fall-through semantics for a list of ©case© clauses;
     876Like the ©switch© statement, the ©choose© statement retains the fall-through semantics for a list of ©case© clauses.
    859877An implicit ©break© is applied only at the end of the \emph{statements} following a ©case© clause.
    860878An explicit ©fallthru© is retained because it is a C-idiom most C programmers expect, and its absence might discourage programmers from using the ©choose© statement.
     
    872890\begin{cfa}
    873891switch ( x ) {
    874         ®int i = 0;® §\C{// allowed only at start}§
     892        @int i = 0;@ $\C{// allowed only at start}$
    875893  case 0:
    876894        ...
    877         ®int j = 0;® §\C{// disallowed}§
     895        @int j = 0;@ $\C{// disallowed}$
    878896  case 1:
    879897        {
    880                 ®int k = 0;® §\C{// allowed at different nesting levels}§
     898                @int k = 0;@ $\C{// allowed at different nesting levels}$
    881899                ...
    882           ®case 2:® §\C{// disallow case in nested statements}§
     900          @case 2:@ $\C{// disallow case in nested statements}$
    883901        }
    884902  ...
     
    897915  case 3:
    898916        if ( ... ) {
    899                 ... ®fallthru;® // goto case 4
     917                ... @fallthru;@ // goto case 4
    900918        } else {
    901919                ...
     
    912930choose ( ... ) {
    913931  case 3:
    914         ... ®fallthrough common;®
     932        ... @fallthrough common;@
    915933  case 4:
    916         ... ®fallthrough common;®
    917 
    918   ®common:® // below fallthrough
     934        ... @fallthrough common;@
     935
     936  @common:@ // below fallthrough
    919937                          // at case-clause level
    920938        ...     // common code for cases 3/4
     
    932950                for ( ... ) {
    933951                        // multi-level transfer
    934                         ... ®fallthru common;®
     952                        ... @fallthru common;@
    935953                }
    936954                ...
    937955        }
    938956        ...
    939   ®common:® // below fallthrough
     957  @common:@ // below fallthrough
    940958                          // at case-clause level
    941959\end{cfa}
     
    948966
    949967\begin{figure}
    950 \begin{tabular}{@{}l|l@{}}
    951 \multicolumn{1}{c|}{loop control} & \multicolumn{1}{c}{output} \\
     968\begin{tabular}{@{}l@{\hspace{25pt}}|l@{}}
     969\multicolumn{1}{@{}c@{\hspace{25pt}}|}{loop control} & \multicolumn{1}{c@{}}{output} \\
    952970\hline
    953 \begin{cfa}[xleftmargin=0pt]
    954 while ®()® { sout | "empty"; break; }
    955 do { sout | "empty"; break; } while ®()®;
    956 for ®()® { sout | "empty"; break; }
    957 for ( ®0® ) { sout | "A"; } sout | "zero";
    958 for ( ®1® ) { sout | "A"; }
    959 for ( ®10® ) { sout | "A"; }
    960 for ( ®= 10® ) { sout | "A"; }
    961 for ( ®1 ~= 10 ~ 2® ) { sout | "B"; }
    962 for ( ®10 -~= 1 ~ 2® ) { sout | "C"; }
    963 for ( ®0.5 ~ 5.5® ) { sout | "D"; }
    964 for ( ®5.5 -~ 0.5® ) { sout | "E"; }
    965 for ( ®i; 10® ) { sout | i; }
    966 for ( ®i; = 10® ) { sout | i; }
    967 for ( ®i; 1 ~= 10 ~ 2® ) { sout | i; }
    968 for ( ®i; 10 -~= 1 ~ 2® ) { sout | i; }
    969 for ( ®i; 0.5 ~ 5.5® ) { sout | i; }
    970 for ( ®i; 5.5 -~ 0.5® ) { sout | i; }
    971 for ( ®ui; 2u ~= 10u ~ 2u® ) { sout | ui; }
    972 for ( ®ui; 10u -~= 2u ~ 2u® ) { sout | ui; }
     971\begin{cfa}
     972while @($\,$)@ { sout | "empty"; break; }
     973do { sout | "empty"; break; } while @($\,$)@;
     974for @($\,$)@ { sout | "empty"; break; }
     975for ( @0@ ) { sout | "A"; } sout | "zero";
     976for ( @1@ ) { sout | "A"; }
     977for ( @10@ ) { sout | "A"; }
     978for ( @= 10@ ) { sout | "A"; }
     979for ( @1 ~= 10 ~ 2@ ) { sout | "B"; }
     980for ( @10 -~= 1 ~ 2@ ) { sout | "C"; }
     981for ( @0.5 ~ 5.5@ ) { sout | "D"; }
     982for ( @5.5 -~ 0.5@ ) { sout | "E"; }
     983for ( @i; 10@ ) { sout | i; }
     984for ( @i; = 10@ ) { sout | i; }
     985for ( @i; 1 ~= 10 ~ 2@ ) { sout | i; }
     986for ( @i; 10 -~= 1 ~ 2@ ) { sout | i; }
     987for ( @i; 0.5 ~ 5.5@ ) { sout | i; }
     988for ( @i; 5.5 -~ 0.5@ ) { sout | i; }
     989for ( @ui; 2u ~= 10u ~ 2u@ ) { sout | ui; }
     990for ( @ui; 10u -~= 2u ~ 2u@ ) { sout | ui; }
    973991enum { N = 10 };
    974 for ( ®N® ) { sout | "N"; }
    975 for ( ®i; N® ) { sout | i; }
    976 for ( ®i; N -~ 0® ) { sout | i; }
     992for ( @N@ ) { sout | "N"; }
     993for ( @i; N@ ) { sout | i; }
     994for ( @i; N -~ 0@ ) { sout | i; }
    977995const int start = 3, comp = 10, inc = 2;
    978 for ( ®i; start ~ comp ~ inc + 1® ) { sout | i; }
    979 for ( i; 1 ~ ®@® ) { if ( i > 10 ) break; sout | i; }
    980 for ( i; 10 -~ ®@® ) { if ( i < 0 ) break; sout | i; }
    981 for ( i; 2 ~ ®@® ~ 2 ) { if ( i > 10 ) break; sout | i; }
    982 for ( i; 2.1 ~ ®@® ~ ®@® ) { if ( i > 10.5 ) break; sout | i; i += 1.7; }
    983 for ( i; 10 -~ ®@® ~ 2 ) { if ( i < 0 ) break; sout | i; }
    984 for ( i; 12.1 ~ ®@® ~ ®@® ) { if ( i < 2.5 ) break; sout | i; i -= 1.7; }
    985 for ( i; 5 ®:® j; -5 ~ @ ) { sout | i | j; }
    986 for ( i; 5 ®:® j; -5 -~ @ ) { sout | i | j; }
    987 for ( i; 5 ®:® j; -5 ~ @ ~ 2 ) { sout | i | j; }
    988 for ( i; 5 ®:® j; -5 -~ @ ~ 2 ) { sout | i | j; }
    989 for ( i; 5 ®:® j; -5 ~ @ ) { sout | i | j; }
    990 for ( i; 5 ®:® j; -5 -~ @ ) { sout | i | j; }
    991 for ( i; 5 ®:® j; -5 ~ @ ~ 2 ) { sout | i | j; }
    992 for ( i; 5 ®:® j; -5 -~ @ ~ 2 ) { sout | i | j; }
    993 for ( i; 5 ®:® j; -5 -~ @ ~ 2 ®:® k; 1.5 ~ @ ) { sout | i | j | k; }
    994 for ( i; 5 ®:® j; -5 -~ @ ~ 2 ®:® k; 1.5 ~ @ ) { sout | i | j | k; }
    995 for ( i; 5 ®:® k; 1.5 ~ @ ®:® j; -5 -~ @ ~ 2 ) { sout | i | j | k; }
     996for ( @i; start ~ comp ~ inc + 1@ ) { sout | i; }
     997for ( i; 1 ~ $\R{@}$ ) { if ( i > 10 ) break; sout | i; }
     998for ( i; 10 -~ $\R{@}$ ) { if ( i < 0 ) break; sout | i; }
     999for ( i; 2 ~ $\R{@}$ ~ 2 ) { if ( i > 10 ) break; sout | i; }
     1000for ( i; 2.1 ~ $\R{@}$ ~ $\R{@}$ ) { if ( i > 10.5 ) break; sout | i; i += 1.7; }
     1001for ( i; 10 -~ $\R{@}$ ~ 2 ) { if ( i < 0 ) break; sout | i; }
     1002for ( i; 12.1 ~ $\R{@}$ ~ $\R{@}$ ) { if ( i < 2.5 ) break; sout | i; i -= 1.7; }
     1003for ( i; 5 @:@ j; -5 ~ $@$ ) { sout | i | j; }
     1004for ( i; 5 @:@ j; -5 -~ $@$ ) { sout | i | j; }
     1005for ( i; 5 @:@ j; -5 ~ $@$ ~ 2 ) { sout | i | j; }
     1006for ( i; 5 @:@ j; -5 -~ $@$ ~ 2 ) { sout | i | j; }
     1007for ( i; 5 @:@ j; -5 ~ $@$ ) { sout | i | j; }
     1008for ( i; 5 @:@ j; -5 -~ $@$ ) { sout | i | j; }
     1009for ( i; 5 @:@ j; -5 ~ $@$ ~ 2 ) { sout | i | j; }
     1010for ( i; 5 @:@ j; -5 -~ $@$ ~ 2 ) { sout | i | j; }
     1011for ( i; 5 @:@ j; -5 -~ $@$ ~ 2 @:@ k; 1.5 ~ $@$ ) { sout | i | j | k; }
     1012for ( i; 5 @:@ j; -5 -~ $@$ ~ 2 @:@ k; 1.5 ~ $@$ ) { sout | i | j | k; }
     1013for ( i; 5 @:@ k; 1.5 ~ $@$ @:@ j; -5 -~ $@$ ~ 2 ) { sout | i | j | k; }
    9961014\end{cfa}
    9971015&
     
    10561074\subsection{Loop Control}
    10571075
    1058 The ©for©/©while©/©do-while© loop-control allows empty or simplified ranges (see Figure~\ref{f:LoopControlExamples}).
    1059 \begin{itemize}
     1076Looping a fixed number of times, possibly with a loop index, occurs frequently.
     1077\CFA condenses simply looping to facilitate coding speed and safety.
     1078The ©for©/©while©/©do-while© loop-control is augmented as follows \see{examples in \VRef[Figure]{f:LoopControlExamples}}:
     1079\begin{itemize}[itemsep=0pt]
     1080\item
     1081©0© is the implicit start value;
     1082\item
     1083©1© is the implicit increment value.
     1084\item
     1085The up-to range uses operator ©+=© for increment;
     1086\item
     1087The down-to range uses operator ©-=© for decrement.
    10601088\item
    10611089The loop index is polymorphic in the type of the comparison value N (when the start value is implicit) or the start value M.
     1090\begin{cfa}
     1091for ( i; @5@ )                                  $\C[2.5in]{// typeof(5) i; 5 is comparison value}$
     1092for ( i; @1.5@~5.5~0.5 )                $\C{// typeof(1.5) i; 1.5 is start value}$
     1093\end{cfa}
    10621094\item
    10631095An empty conditional implies comparison value of ©1© (true).
    1064 \item
    1065 A comparison N is implicit up-to exclusive range [0,N©®)®©.
    1066 \item
    1067 A comparison ©=© N is implicit up-to inclusive range [0,N©®]®©.
    1068 \item
    1069 The up-to range M ©~©\index{~@©~©} N means exclusive range [M,N©®)®©.
    1070 \item
    1071 The up-to range M ©~=©\index{~=@©~=©} N means inclusive range [M,N©®]®©.
    1072 \item
    1073 The down-to range M ©-~©\index{-~@©-~©} N means exclusive range [N,M©®)®©.
    1074 \item
    1075 The down-to range M ©-~=©\index{-~=@©-~=©} N means inclusive range [N,M©®]®©.
    1076 \item
    1077 ©0© is the implicit start value;
    1078 \item
    1079 ©1© is the implicit increment value.
    1080 \item
    1081 The up-to range uses operator ©+=© for increment;
    1082 \item
    1083 The down-to range uses operator ©-=© for decrement.
     1096\begin{cfa}
     1097while ( $\R{/*empty*/}$ )               $\C{// while ( true )}$
     1098for ( $\R{/*empty*/}$ )                 $\C{// for ( ; true; )}$
     1099do ... while ( $\R{/*empty*/}$ ) $\C{// do ... while ( true )}$
     1100\end{cfa}
     1101\item
     1102A comparison N is implicit up-to exclusive range [0,N\R{)}.
     1103\begin{cfa}
     1104for ( @5@ )                                             $\C{// for ( typeof(5) i; i < 5; i += 1 )}$
     1105\end{cfa}
     1106\item
     1107A comparison ©=© N is implicit up-to inclusive range [0,N\R{]}.
     1108\begin{cfa}
     1109for ( @=@5 )                                    $\C{// for ( typeof(5) i; i <= 5; i += 1 )}$
     1110\end{cfa}
     1111\item
     1112The up-to range M ©~©\index{~@©~©} N means exclusive range [M,N\R{)}.
     1113\begin{cfa}
     1114for ( 1@~@5 )                                   $\C{// for ( typeof(1) i = 1; i < 5; i += 1 )}$
     1115\end{cfa}
     1116\item
     1117The up-to range M ©~=©\index{~=@©~=©} N means inclusive range [M,N\R{]}.
     1118\begin{cfa}
     1119for ( 1@~=@5 )                                  $\C{// for ( typeof(1) i = 1; i <= 5; i += 1 )}$
     1120\end{cfa}
     1121\item
     1122The down-to range M ©-~©\index{-~@©-~©} N means exclusive range [N,M\R{)}.
     1123\begin{cfa}
     1124for ( 1@-~@5 )                                  $\C{// for ( typeof(1) i = 5; i > 0; i -= 1 )}$
     1125\end{cfa}
     1126\item
     1127The down-to range M ©-~=©\index{-~=@©-~=©} N means inclusive range [N,M\R{]}.
     1128\begin{cfa}
     1129for ( 1@-~=@5 )                                 $\C{// for ( typeof(1) i = 5; i >= 0; i -= 1 )}$
     1130\end{cfa}
    10841131\item
    10851132©@© means put nothing in this field.
     1133\begin{cfa}
     1134for ( 1~$\R{@}$~2 )                             $\C{// for ( typeof(1) i = 1; /*empty*/; i += 2 )}$
     1135\end{cfa}
    10861136\item
    10871137©:© means start another index.
     1138\begin{cfa}
     1139for ( i; 5 @:@ j; 2~12~3 )              $\C{// for ( typeof(i) i = 1, j = 2; i < 5 \&\& j < 12; i += 1, j += 3 )}\CRT$
     1140\end{cfa}
    10881141\end{itemize}
    10891142
     
    10921145\subsection{\texorpdfstring{Labelled \LstKeywordStyle{continue} / \LstKeywordStyle{break} Statement}{Labelled continue / break Statement}}
    10931146
    1094 While C provides ©continue© and ©break© statements for altering control flow, both are restricted to one level of nesting for a particular control structure.
    1095 Unfortunately, this restriction forces programmers to use \Indexc{goto} to achieve the equivalent control-flow for more than one level of nesting.
     1147C ©continue© and ©break© statements, for altering control flow, are restricted to one level of nesting for a particular control structure.
     1148This restriction forces programmers to use \Indexc{goto} to achieve the equivalent control-flow for more than one level of nesting.
    10961149To prevent having to switch to the ©goto©, \CFA extends the \Indexc{continue}\index{continue@©continue©!labelled}\index{labelled!continue@©continue©} and \Indexc{break}\index{break@©break©!labelled}\index{labelled!break@©break©} with a target label to support static multi-level exit\index{multi-level exit}\index{static multi-level exit}~\cite{Buhr85}, as in Java.
    10971150For both ©continue© and ©break©, the target label must be directly associated with a ©for©, ©while© or ©do© statement;
    10981151for ©break©, the target label can also be associated with a ©switch©, ©if© or compound (©{}©) statement.
    1099 \VRef[Figure]{f:MultiLevelExit} shows ©continue© and ©break© indicating the specific control structure, and the corresponding C program using only ©goto© and labels.
     1152\VRef[Figure]{f:MultiLevelExit} shows a comparison between labelled ©continue© and ©break© and the corresponding C equivalent using ©goto© and labels.
    11001153The innermost loop has 8 exit points, which cause continuation or termination of one or more of the 7 \Index{nested control-structure}s.
    11011154
     
    11041157\begin{lrbox}{\myboxA}
    11051158\begin{cfa}[tabsize=3]
    1106 ®Compound:® {
    1107         ®Try:® try {
    1108                 ®For:® for ( ... ) {
    1109                         ®While:® while ( ... ) {
    1110                                 ®Do:® do {
    1111                                         ®If:® if ( ... ) {
    1112                                                 ®Switch:® switch ( ... ) {
     1159@Compound:@ {
     1160        @Try:@ try {
     1161                @For:@ for ( ... ) {
     1162                        @While:@ while ( ... ) {
     1163                                @Do:@ do {
     1164                                        @If:@ if ( ... ) {
     1165                                                @Switch:@ switch ( ... ) {
    11131166                                                        case 3:
    1114                                                                 ®break Compound®;
    1115                                                                 ®break Try®;
    1116                                                                 ®break For®;      /* or */  ®continue For®;
    1117                                                                 ®break While®;  /* or */  ®continue While®;
    1118                                                                 ®break Do®;      /* or */  ®continue Do®;
    1119                                                                 ®break If®;
    1120                                                                 ®break Switch®;
     1167                                                                @break Compound@;
     1168                                                                @break Try@;
     1169                                                                @break For@;      /* or */  @continue For@;
     1170                                                                @break While@;  /* or */  @continue While@;
     1171                                                                @break Do@;      /* or */  @continue Do@;
     1172                                                                @break If@;
     1173                                                                @break Switch@;
    11211174                                                        } // switch
    11221175                                                } else {
    1123                                                         ... ®break If®; ...     // terminate if
     1176                                                        ... @break If@; ...     // terminate if
    11241177                                                } // if
    11251178                                } while ( ... ); // do
    11261179                        } // while
    11271180                } // for
    1128         } ®finally® { // always executed
     1181        } @finally@ { // always executed
    11291182        } // try
    11301183} // compound
     
    11361189{
    11371190
    1138                 ®ForC:® for ( ... ) {
    1139                         ®WhileC:® while ( ... ) {
    1140                                 ®DoC:® do {
     1191                @ForC:@ for ( ... ) {
     1192                        @WhileC:@ while ( ... ) {
     1193                                @DoC:@ do {
    11411194                                        if ( ... ) {
    11421195                                                switch ( ... ) {
    11431196                                                        case 3:
    1144                                                                 ®goto Compound®;
    1145                                                                 ®goto Try®;
    1146                                                                 ®goto ForB®;      /* or */  ®goto ForC®;
    1147                                                                 ®goto WhileB®;  /* or */  ®goto WhileC®;
    1148                                                                 ®goto DoB®;      /* or */  ®goto DoC®;
    1149                                                                 ®goto If®;
    1150                                                                 ®goto Switch®;
    1151                                                         } ®Switch:® ;
     1197                                                                @goto Compound@;
     1198                                                                @goto Try@;
     1199                                                                @goto ForB@;      /* or */  @goto ForC@;
     1200                                                                @goto WhileB@;  /* or */  @goto WhileC@;
     1201                                                                @goto DoB@;      /* or */  @goto DoC@;
     1202                                                                @goto If@;
     1203                                                                @goto Switch@;
     1204                                                        } @Switch:@ ;
    11521205                                                } else {
    1153                                                         ... ®goto If®; ...      // terminate if
    1154                                                 } ®If:®;
    1155                                 } while ( ... ); ®DoB:® ;
    1156                         } ®WhileB:® ;
    1157                 } ®ForB:® ;
    1158 
    1159 
    1160 } ®Compound:® ;
     1206                                                        ... @goto If@; ...      // terminate if
     1207                                                } @If:@;
     1208                                } while ( ... ); @DoB:@ ;
     1209                        } @WhileB:@ ;
     1210                } @ForB:@ ;
     1211
     1212
     1213} @Compound:@ ;
    11611214\end{cfa}
    11621215\end{lrbox}
    11631216
    11641217\subfloat[\CFA]{\label{f:CFibonacci}\usebox\myboxA}
    1165 \hspace{2pt}
     1218\hspace{3pt}
    11661219\vrule
    1167 \hspace{2pt}
     1220\hspace{3pt}
    11681221\subfloat[C]{\label{f:CFAFibonacciGen}\usebox\myboxB}
    11691222\caption{Multi-level Exit}
     
    11801233This restriction prevents missing declarations and/or initializations at the start of a control structure resulting in undefined behaviour.
    11811234\end{itemize}
    1182 The advantage of the labelled ©continue©/©break© is allowing static multi-level exits without having to use the ©goto© statement, and tying control flow to the target control structure rather than an arbitrary point in a program.
     1235The advantage of the labelled ©continue©/©break© is allowing static multi-level exits without having to use the ©goto© statement, and tying control flow to the target control structure rather than an arbitrary point in a program via a label.
    11831236Furthermore, the location of the label at the \emph{beginning} of the target control structure informs the reader (\Index{eye candy}) that complex control-flow is occurring in the body of the control structure.
    11841237With ©goto©, the label is at the end of the control structure, which fails to convey this important clue early enough to the reader.
     
    11871240
    11881241
    1189 %\section{\texorpdfstring{\protect\lstinline@with@ Statement}{with Statement}}
    1190 \section{\texorpdfstring{\LstKeywordStyle{with} Statement}{with Statement}}
     1242%\subsection{\texorpdfstring{\protect\lstinline@with@ Statement}{with Statement}}
     1243\subsection{\texorpdfstring{\LstKeywordStyle{with} Statement}{with Statement}}
    11911244\label{s:WithStatement}
    11921245
    1193 Grouping heterogeneous data into \newterm{aggregate}s (structure/union) is a common programming practice, and an aggregate can be further organized into more complex structures, such as arrays and containers:
    1194 \begin{cfa}
    1195 struct S { §\C{// aggregate}§
    1196         char c; §\C{// fields}§
    1197         int i;
    1198         double d;
     1246Grouping heterogeneous data into an \newterm{aggregate} (structure/union) is a common programming practice, and aggregates may be nested:
     1247\begin{cfa}
     1248struct Person {                                                         $\C{// aggregate}$
     1249        struct Name { char first[20], last[20]; } name $\C{// nesting}$
     1250        struct Address { ... } address                  $\C{// nesting}$
     1251        int sex;
    11991252};
    1200 S s, as[10];
    1201 \end{cfa}
    1202 However, functions manipulating aggregates must repeat the aggregate name to access its containing fields:
    1203 \begin{cfa}
    1204 void f( S s ) {
    1205         ®s.®c; ®s.®i; ®s.®d; §\C{// access containing fields}§
    1206 }
    1207 \end{cfa}
    1208 which extends to multiple levels of qualification for nested aggregates.
    1209 A similar situation occurs in object-oriented programming, \eg \CC:
     1253\end{cfa}
     1254Functions manipulating aggregates must repeat the aggregate name to access its containing fields.
     1255\begin{cfa}
     1256Person p
     1257@p.@name; @p.@address; @p.@sex; $\C{// access containing fields}$
     1258\end{cfa}
     1259which extends to multiple levels of qualification for nested aggregates and multiple aggregates.
     1260\begin{cfa}
     1261struct Ticket { ... } t;
     1262@p.name@.first; @p.address@.street;             $\C{// access nested fields}$
     1263@t.@departure; @t.@cost;                                $\C{// access multiple aggregate}$
     1264\end{cfa}
     1265Repeated aggregate qualification is tedious and makes code difficult to read.
     1266Therefore, reducing aggregate qualification is a useful language design goal.
     1267
     1268C allows unnamed nested aggregates that open their scope into the containing aggregate.
     1269This feature is used to group fields for attributes and/or with ©union© aggregates.
     1270\begin{cfa}
     1271struct S {
     1272        struct { int g,  h; } __attribute__(( aligned(64) ));
     1273        int tag;
     1274        union {
     1275                struct { char c1,  c2; } __attribute__(( aligned(128) ));
     1276                struct { int i1,  i2; };
     1277                struct { double d1,  d2; };
     1278        };
     1279};
     1280s.g; s.h; s.tag; s.c1; s.c2; s.i1; s.i2; s.d1; s.d2;
     1281\end{cfa}
     1282
     1283Object-oriented languages reduce qualification for class variables within member functions, \eg \CC:
    12101284\begin{C++}
    12111285struct S {
    1212         char c; §\C{// fields}§
    1213         int i;
    1214         double d;
    1215         void f() { §\C{// implicit ``this'' aggregate}§
    1216                 ®this->®c; ®this->®i; ®this->®d; §\C{// access containing fields}§
     1286        char @c@;   int @i@;   double @d@;
     1287        void f( /* S * this */ ) {                              $\C{// implicit ``this'' parameter}$
     1288                @c@;   @i@;   @d@;                                      $\C{// this->c; this->i; this->d;}$
    12171289        }
    12181290}
    12191291\end{C++}
    1220 Object-oriented nesting of member functions in a \lstinline[language=C++]@class/struct@ allows eliding \lstinline[language=C++]@this->@ because of lexical scoping.
    1221 However, for other aggregate parameters, qualification is necessary:
    1222 \begin{cfa}
    1223 struct T { double m, n; };
    1224 int S::f( T & t ) { §\C{// multiple aggregate parameters}§
    1225         c; i; d; §\C{\color{red}// this--{\textgreater}.c, this--{\textgreater}.i, this--{\textgreater}.d}§
    1226         ®t.®m; ®t.®n; §\C{// must qualify}§
    1227 }
    1228 \end{cfa}
    1229 
    1230 To simplify the programmer experience, \CFA provides a ©with© statement (see Pascal~\cite[\S~4.F]{Pascal}) to elide aggregate qualification to fields by opening a scope containing the field identifiers.
    1231 Hence, the qualified fields become variables with the side-effect that it is easier to optimizing field references in a block.
    1232 \begin{cfa}
    1233 void f( S & this ) ®with ( this )® { §\C{// with statement}§
    1234         c; i; d; §\C{\color{red}// this.c, this.i, this.d}§
     1292In general, qualification is elided for the variables and functions in the lexical scopes visible from a member function.
     1293However, qualification is necessary for name shadowing and explicit aggregate parameters.
     1294\begin{cfa}
     1295struct T {
     1296        char @m@;   int @i@;   double @n@;              $\C{// derived class variables}$
     1297};
     1298struct S : public T {
     1299        char @c@;   int @i@;   double @d@;              $\C{// class variables}$
     1300        void g( double @d@, T & t ) {
     1301                d;   @t@.m;   @t@.i;   @t@.n;           $\C{// function parameter}$
     1302                c;   i;   @this->@d;   @S::@d;          $\C{// class S variables}$
     1303                m;   @T::@i;   n;                                       $\C{// class T variables}$
     1304        }
     1305};
     1306\end{cfa}
     1307Note the three different forms of qualification syntax in \CC, ©.©, ©->©, ©::©, which is confusing.
     1308
     1309Since \CFA in not object-oriented, it has no implicit parameter with its implicit qualification.
     1310Instead \CFA introduces a general mechanism using the ©with© statement \see{Pascal~\cite[\S~4.F]{Pascal}} to explicitly elide aggregate qualification by opening a scope containing the field identifiers.
     1311Hence, the qualified fields become variables with the side-effect that it is simpler to write, easier to read, and optimize field references in a block.
     1312\begin{cfa}
     1313void f( S & this ) @with ( this )@ {            $\C{// with statement}$
     1314        @c@;   @i@;   @d@;                                              $\C{// this.c, this.i, this.d}$
    12351315}
    12361316\end{cfa}
    12371317with the generality of opening multiple aggregate-parameters:
    12381318\begin{cfa}
    1239 void f( S & s, T & t ) ®with ( s, t )® { §\C{// multiple aggregate parameters}§
    1240         c; i; d; §\C{\color{red}// s.c, s.i, s.d}§
    1241         m; n; §\C{\color{red}// t.m, t.n}§
    1242 }
    1243 \end{cfa}
    1244 
    1245 In detail, the ©with© statement has the form:
    1246 \begin{cfa}
    1247 §\emph{with-statement}§:
    1248         'with' '(' §\emph{expression-list}§ ')' §\emph{compound-statement}§
    1249 \end{cfa}
    1250 and may appear as the body of a function or nested within a function body.
    1251 Each expression in the expression-list provides a type and object.
    1252 The type must be an aggregate type.
     1319void g( S & s, T & t ) @with ( s, t )@ {        $\C{// multiple aggregate parameters}$
     1320        c;   @s.@i;   d;                                                $\C{// s.c, s.i, s.d}$
     1321        m;   @t.@i;   n;                                                $\C{// t.m, t.i, t.n}$
     1322}
     1323\end{cfa}
     1324where qualification is only necessary to disambiguate the shadowed variable ©i©.
     1325
     1326In detail, the ©with© statement may appear as the body of a function or nested within a function body.
     1327The ©with© clause takes a list of expressions, where each expression provides an aggregate type and object.
    12531328(Enumerations are already opened.)
    1254 The object is the implicit qualifier for the open structure-fields.
    1255 
     1329To open a pointer type, the pointer must be dereferenced to obtain a reference to the aggregate type.
     1330\begin{cfa}
     1331S * sp;
     1332with ( *sp ) { ... }
     1333\end{cfa}
     1334The expression object is the implicit qualifier for the open structure-fields.
     1335\CFA's ability to overload variables \see{\VRef{s:VariableOverload}} and use the left-side of assignment in type resolution means most fields with the same name but different types are automatically disambiguated, eliminating qualification.
    12561336All expressions in the expression list are open in parallel within the compound statement.
    12571337This semantic is different from Pascal, which nests the openings from left to right.
    12581338The difference between parallel and nesting occurs for fields with the same name and type:
    12591339\begin{cfa}
    1260 struct S { int ®i®; int j; double m; } s, w;
    1261 struct T { int ®i®; int k; int m; } t, w;
    1262 with ( s, t ) {
    1263         j + k; §\C{// unambiguous, s.j + t.k}§
    1264         m = 5.0; §\C{// unambiguous, t.m = 5.0}§
    1265         m = 1; §\C{// unambiguous, s.m = 1}§
    1266         int a = m; §\C{// unambiguous, a = s.i }§
    1267         double b = m; §\C{// unambiguous, b = t.m}§
    1268         int c = s.i + t.i; §\C{// unambiguous, qualification}§
    1269         (double)m; §\C{// unambiguous, cast}§
    1270 }
    1271 \end{cfa}
    1272 For parallel semantics, both ©s.i© and ©t.i© are visible, so ©i© is ambiguous without qualification;
    1273 for nested semantics, ©t.i© hides ©s.i©, so ©i© implies ©t.i©.
    1274 \CFA's ability to overload variables means fields with the same name but different types are automatically disambiguated, eliminating most qualification when opening multiple aggregates.
    1275 Qualification or a cast is used to disambiguate.
    1276 
    1277 There is an interesting problem between parameters and the function-body ©with©, \eg:
    1278 \begin{cfa}
    1279 void ?{}( S & s, int i ) with ( s ) { §\C{// constructor}§
    1280         ®s.i = i;®  j = 3;  m = 5.5; §\C{// initialize fields}§
     1340struct Q { int @i@; int k; int @m@; } q, w;
     1341struct R { int @i@; int j; double @m@; } r, w;
     1342with ( r, q ) {
     1343        j + k;                                                                  $\C{// unambiguous, r.j + q.k}$
     1344        m = 5.0;                                                                $\C{// unambiguous, q.m = 5.0}$
     1345        m = 1;                                                                  $\C{// unambiguous, r.m = 1}$
     1346        int a = m;                                                              $\C{// unambiguous, a = r.i }$
     1347        double b = m;                                                   $\C{// unambiguous, b = q.m}$
     1348        int c = r.i + q.i;                                              $\C{// disambiguate with qualification}$
     1349        (double)m;                                                              $\C{// disambiguate with cast}$
     1350}
     1351\end{cfa}
     1352For parallel semantics, both ©r.i© and ©q.i© are visible, so ©i© is ambiguous without qualification;
     1353for nested semantics, ©q.i© hides ©r.i©, so ©i© implies ©q.i©.
     1354Pascal nested-semantics is possible by nesting ©with© statements.
     1355\begin{cfa}
     1356with ( r ) {
     1357        i;                                                                              $\C{// unambiguous, r.i}$
     1358        with ( q ) {
     1359                i;                                                                      $\C{// unambiguous, q.i}$
     1360        }
     1361}
     1362\end{cfa}
     1363A cast or qualification can be used to disambiguate variables within a ©with© \emph{statement}.
     1364A cast can be used to disambiguate among overload variables in a ©with© \emph{expression}:
     1365\begin{cfa}
     1366with ( w ) { ... }                                                      $\C{// ambiguous, same name and no context}$
     1367with ( (Q)w ) { ... }                                           $\C{// unambiguous, cast}$
     1368\end{cfa}
     1369Because there is no left-side in the ©with© expression to implicitly disambiguate between the ©w© variables, it is necessary to explicitly disambiguate by casting ©w© to type ©Q© or ©R©.
     1370
     1371Finally, there is an interesting problem between parameters and the function-body ©with©, \eg:
     1372\begin{cfa}
     1373void ?{}( S & s, int i ) with ( s ) { $\C{// constructor}$
     1374        @s.i = i;@  j = 3;  m = 5.5; $\C{// initialize fields}$
    12811375}
    12821376\end{cfa}
     
    12911385and implicitly opened \emph{after} a function-body open, to give them higher priority:
    12921386\begin{cfa}
    1293 void ?{}( S & s, int ®i® ) with ( s ) ®with( §\emph{\color{red}params}§ )® {
    1294         s.i = ®i®; j = 3; m = 5.5;
    1295 }
    1296 \end{cfa}
    1297 Finally, a cast may be used to disambiguate among overload variables in a ©with© expression:
    1298 \begin{cfa}
    1299 with ( w ) { ... } §\C{// ambiguous, same name and no context}§
    1300 with ( (S)w ) { ... } §\C{// unambiguous, cast}§
    1301 \end{cfa}
    1302 and ©with© expressions may be complex expressions with type reference (see Section~\ref{s:References}) to aggregate:
    1303 % \begin{cfa}
    1304 % struct S { int i, j; } sv;
    1305 % with ( sv ) { §\C{// implicit reference}§
    1306 %       S & sr = sv;
    1307 %       with ( sr ) { §\C{// explicit reference}§
    1308 %               S * sp = &sv;
    1309 %               with ( *sp ) { §\C{// computed reference}§
    1310 %                       i = 3; j = 4; §\C{\color{red}// sp--{\textgreater}i, sp--{\textgreater}j}§
    1311 %               }
    1312 %               i = 2; j = 3; §\C{\color{red}// sr.i, sr.j}§
    1313 %       }
    1314 %       i = 1; j = 2; §\C{\color{red}// sv.i, sv.j}§
    1315 % }
    1316 % \end{cfa}
    1317 
    1318 In \Index{object-oriented} programming, there is an implicit first parameter, often names \textbf{©self©} or \textbf{©this©}, which is elided.
    1319 \begin{C++}
    1320 class C {
    1321         int i, j;
    1322         int mem() { §\C{\color{red}// implicit "this" parameter}§
    1323                 i = 1; §\C{\color{red}// this->i}§
    1324                 j = 2; §\C{\color{red}// this->j}§
    1325         }
    1326 }
    1327 \end{C++}
    1328 Since \CFA is non-object-oriented, the equivalent object-oriented program looks like:
    1329 \begin{cfa}
    1330 struct S { int i, j; };
    1331 int mem( S & ®this® ) { §\C{// explicit "this" parameter}§
    1332         ®this.®i = 1; §\C{// "this" is not elided}§
    1333         ®this.®j = 2;
    1334 }
    1335 \end{cfa}
    1336 but it is cumbersome having to write ``©this.©'' many times in a member.
    1337 
    1338 \CFA provides a ©with© clause/statement (see Pascal~\cite[\S~4.F]{Pascal}) to elided the "©this.©" by opening a scope containing field identifiers, changing the qualified fields into variables and giving an opportunity for optimizing qualified references.
    1339 \begin{cfa}
    1340 int mem( S & this ) ®with( this )® { §\C{// with clause}§
    1341         i = 1; §\C{\color{red}// this.i}§
    1342         j = 2; §\C{\color{red}// this.j}§
    1343 }
    1344 \end{cfa}
    1345 which extends to multiple routine parameters:
    1346 \begin{cfa}
    1347 struct T { double m, n; };
    1348 int mem2( S & this1, T & this2 ) ®with( this1, this2 )® {
    1349         i = 1; j = 2;
    1350         m = 1.0; n = 2.0;
    1351 }
    1352 \end{cfa}
    1353 
    1354 The statement form is used within a block:
    1355 \begin{cfa}
    1356 int foo() {
    1357         struct S1 { ... } s1;
    1358         struct S2 { ... } s2;
    1359         ®with( s1 )® { §\C{// with statement}§
    1360                 // access fields of s1 without qualification
    1361                 ®with s2® { §\C{// nesting}§
    1362                         // access fields of s1 and s2 without qualification
    1363                 }
    1364         }
    1365         ®with s1, s2® {
    1366                 // access unambiguous fields of s1 and s2 without qualification
    1367         }
    1368 }
    1369 \end{cfa}
    1370 
    1371 When opening multiple structures, fields with the same name and type are ambiguous and must be fully qualified.
    1372 For fields with the same name but different type, context/cast can be used to disambiguate.
    1373 \begin{cfa}
    1374 struct S { int i; int j; double m; } a, c;
    1375 struct T { int i; int k; int m } b, c;
    1376 with( a, b )
    1377 {
    1378 }
    1379 \end{cfa}
    1380 
    1381 \begin{comment}
    1382 The components in the "with" clause
    1383 
    1384   with a, b, c { ... }
    1385 
    1386 serve 2 purposes: each component provides a type and object. The type must be a
    1387 structure type. Enumerations are already opened, and I think a union is opened
    1388 to some extent, too. (Or is that just unnamed unions?) The object is the target
    1389 that the naked structure-fields apply to. The components are open in "parallel"
    1390 at the scope of the "with" clause/statement, so opening "a" does not affect
    1391 opening "b", etc. This semantic is different from Pascal, which nests the
    1392 openings.
    1393 
    1394 Having said the above, it seems reasonable to allow a "with" component to be an
    1395 expression. The type is the static expression-type and the object is the result
    1396 of the expression. Again, the type must be an aggregate. Expressions require
    1397 parenthesis around the components.
    1398 
    1399   with( a, b, c ) { ... }
    1400 
    1401 Does this now make sense?
    1402 
    1403 Having written more CFA code, it is becoming clear to me that I *really* want
    1404 the "with" to be implemented because I hate having to type all those object
    1405 names for fields. It's a great way to drive people away from the language.
    1406 \end{comment}
     1387void ?{}( S & s, int @i@ ) with ( s ) @with( $\emph{\R{params}}$ )@ { // syntax not allowed, illustration only
     1388        s.i = @i@; j = 3; m = 5.5;
     1389}
     1390\end{cfa}
     1391This implicit semantic matches with programmer expectation.
     1392
    14071393
    14081394
     
    14141400Non-local transfer can cause stack unwinding, \ie non-local routine termination, depending on the kind of raise.
    14151401\begin{cfa}
    1416 exception_t E {}; §\C{// exception type}§
     1402exception_t E {}; $\C{// exception type}$
    14171403void f(...) {
    1418         ... throw E{}; ... §\C{// termination}§
    1419         ... throwResume E{}; ... §\C{// resumption}§
     1404        ... throw E{}; ... $\C{// termination}$
     1405        ... throwResume E{}; ... $\C{// resumption}$
    14201406}
    14211407try {
    14221408        f(...);
    1423 } catch( E e ; §boolean-predicate§ ) {          §\C{// termination handler}§
     1409} catch( E e ; $boolean-predicate$ ) {          $\C{// termination handler}$
    14241410        // recover and continue
    1425 } catchResume( E e ; §boolean-predicate§ ) { §\C{// resumption handler}§
     1411} catchResume( E e ; $boolean-predicate$ ) { $\C{// resumption handler}$
    14261412        // repair and return
    14271413} finally {
     
    14301416\end{cfa}
    14311417The kind of raise and handler match: ©throw© with ©catch© and ©throwResume© with ©catchResume©.
    1432 Then the exception type must match along with any additonal predicate must be true.
     1418Then the exception type must match along with any additional predicate must be true.
    14331419The ©catch© and ©catchResume© handlers may appear in any oder.
    14341420However, the ©finally© clause must appear at the end of the ©try© statement.
     
    14831469For example, a routine returning a \Index{pointer} to an array of integers is defined and used in the following way:
    14841470\begin{cfa}
    1485 int ®(*®f®())[®5®]® {...}; §\C{// definition}§
    1486  ... ®(*®f®())[®3®]® += 1; §\C{// usage}§
     1471int @(*@f@())[@5@]@ {...}; $\C{// definition}$
     1472 ... @(*@f@())[@3@]@ += 1; $\C{// usage}$
    14871473\end{cfa}
    14881474Essentially, the return type is wrapped around the routine name in successive layers (like an \Index{onion}).
     
    14991485\begin{tabular}{@{}l@{\hspace{3em}}l@{}}
    15001486\multicolumn{1}{c@{\hspace{3em}}}{\textbf{\CFA}}        & \multicolumn{1}{c}{\textbf{C}}        \\
    1501 \begin{cfa}
    1502 ß[5] *ß ®int® x1;
    1503 ß* [5]ß ®int® x2;
    1504 ß[* [5] int]ß f®( int p )®;
     1487\begin{cfa}[moredelim={**[is][\color{blue}]{\#}{\#}}]
     1488#[5] *# @int@ x1;
     1489#* [5]# @int@ x2;
     1490#[* [5] int]# f@( int p )@;
    15051491\end{cfa}
    15061492&
    1507 \begin{cfa}
    1508 ®int® ß*ß x1 ß[5]ß;
    1509 ®int® ß(*ßx2ß)[5]ß;
    1510 ßint (*ßf®( int p )®ß)[5]ß;
     1493\begin{cfa}[moredelim={**[is][\color{blue}]{\#}{\#}}]
     1494@int@ #*# x1 #[5]#;
     1495@int@ #(*#x2#)[5]#;
     1496#int (*#f@( int p )@#)[5]#;
    15111497\end{cfa}
    15121498\end{tabular}
     
    15201506\multicolumn{1}{c@{\hspace{3em}}}{\textbf{\CFA}}        & \multicolumn{1}{c}{\textbf{C}}        \\
    15211507\begin{cfa}
    1522 ®*® int x, y;
     1508@*@ int x, y;
    15231509\end{cfa}
    15241510&
    15251511\begin{cfa}
    1526 int ®*®x, ®*®y;
     1512int @*@x, @*@y;
    15271513\end{cfa}
    15281514\end{tabular}
     
    15331519\multicolumn{1}{c@{\hspace{3em}}}{\textbf{\CFA}}        & \multicolumn{1}{c}{\textbf{C}}        \\
    15341520\begin{cfa}
    1535 ®*® int x;
     1521@*@ int x;
    15361522int y;
    15371523\end{cfa}
    15381524&
    15391525\begin{cfa}
    1540 int ®*®x, y;
     1526int @*@x, y;
    15411527
    15421528\end{cfa}
     
    16471633
    16481634\section{Pointer / Reference}
     1635\label{s:PointerReference}
    16491636
    16501637C provides a \newterm{pointer type};
     
    16731660&
    16741661\begin{cfa}
    1675 int * ®const® x = (int *)100
     1662int * @const@ x = (int *)100
    16761663*x = 3;                 // implicit dereference
    1677 int * ®const® y = (int *)104;
     1664int * @const@ y = (int *)104;
    16781665*y = *x;                        // implicit dereference
    16791666\end{cfa}
     
    17131700\begin{tabular}{@{}l@{\hspace{2em}}l@{}}
    17141701\begin{cfa}
    1715 int x, y, ®*® p1, ®*® p2, ®**® p3;
    1716 p1 = ®&®x;     // p1 points to x
     1702int x, y, @*@ p1, @*@ p2, @**@ p3;
     1703p1 = @&@x;     // p1 points to x
    17171704p2 = p1;     // p2 points to x
    1718 p1 = ®&®y;     // p1 points to y
     1705p1 = @&@y;     // p1 points to y
    17191706p3 = &p2;  // p3 points to p2
    17201707\end{cfa}
     
    17281715For example, \Index*{Algol68}~\cite{Algol68} infers pointer dereferencing to select the best meaning for each pointer usage
    17291716\begin{cfa}
    1730 p2 = p1 + x; §\C{// compiler infers *p2 = *p1 + x;}§
     1717p2 = p1 + x; $\C{// compiler infers *p2 = *p1 + x;}$
    17311718\end{cfa}
    17321719Algol68 infers the following dereferencing ©*p2 = *p1 + x©, because adding the arbitrary integer value in ©x© to the address of ©p1© and storing the resulting address into ©p2© is an unlikely operation.
     
    17361723In C, objects of pointer type always manipulate the pointer object's address:
    17371724\begin{cfa}
    1738 p1 = p2; §\C{// p1 = p2\ \ rather than\ \ *p1 = *p2}§
    1739 p2 = p1 + x; §\C{// p2 = p1 + x\ \ rather than\ \ *p2 = *p1 + x}§
     1725p1 = p2; $\C{// p1 = p2\ \ rather than\ \ *p1 = *p2}$
     1726p2 = p1 + x; $\C{// p2 = p1 + x\ \ rather than\ \ *p2 = *p1 + x}$
    17401727\end{cfa}
    17411728even though the assignment to ©p2© is likely incorrect, and the programmer probably meant:
    17421729\begin{cfa}
    1743 p1 = p2; §\C{// pointer address assignment}§
    1744 ®*®p2 = ®*®p1 + x; §\C{// pointed-to value assignment / operation}§
     1730p1 = p2; $\C{// pointer address assignment}$
     1731@*@p2 = @*@p1 + x; $\C{// pointed-to value assignment / operation}$
    17451732\end{cfa}
    17461733The C semantics work well for situations where manipulation of addresses is the primary meaning and data is rarely accessed, such as storage management (©malloc©/©free©).
     
    17581745To support this common case, a reference type is introduced in \CFA, denoted by ©&©, which is the opposite dereference semantics to a pointer type, making the value at the pointed-to location the implicit semantics for dereferencing (similar but not the same as \CC \Index{reference type}s).
    17591746\begin{cfa}
    1760 int x, y, ®&® r1, ®&® r2, ®&&® r3;
    1761 ®&®r1 = &x; §\C{// r1 points to x}§
    1762 ®&®r2 = &r1; §\C{// r2 points to x}§
    1763 ®&®r1 = &y; §\C{// r1 points to y}§
    1764 ®&&®r3 = ®&®&r2; §\C{// r3 points to r2}§
    1765 r2 = ((r1 + r2) * (r3 - r1)) / (r3 - 15); §\C{// implicit dereferencing}§
     1747int x, y, @&@ r1, @&@ r2, @&&@ r3;
     1748@&@r1 = &x; $\C{// r1 points to x}$
     1749@&@r2 = &r1; $\C{// r2 points to x}$
     1750@&@r1 = &y; $\C{// r1 points to y}$
     1751@&&@r3 = @&@&r2; $\C{// r3 points to r2}$
     1752r2 = ((r1 + r2) * (r3 - r1)) / (r3 - 15); $\C{// implicit dereferencing}$
    17661753\end{cfa}
    17671754Except for auto-dereferencing by the compiler, this reference example is the same as the previous pointer example.
     
    17691756One way to conceptualize a reference is via a rewrite rule, where the compiler inserts a dereference operator before the reference variable for each reference qualifier in a declaration, so the previous example becomes:
    17701757\begin{cfa}
    1771 ®*®r2 = ((®*®r1 + ®*®r2) ®*® (®**®r3 - ®*®r1)) / (®**®r3 - 15);
     1758@*@r2 = ((@*@r1 + @*@r2) @*@ (@**@r3 - @*@r1)) / (@**@r3 - 15);
    17721759\end{cfa}
    17731760When a reference operation appears beside a dereference operation, \eg ©&*©, they cancel out.
     
    17781765For a \CFA reference type, the cancellation on the left-hand side of assignment leaves the reference as an address (\Index{lvalue}):
    17791766\begin{cfa}
    1780 (&®*®)r1 = &x; §\C{// (\&*) cancel giving address in r1 not variable pointed-to by r1}§
     1767(&@*@)r1 = &x; $\C{// (\&*) cancel giving address in r1 not variable pointed-to by r1}$
    17811768\end{cfa}
    17821769Similarly, the address of a reference can be obtained for assignment or computation (\Index{rvalue}):
    17831770\begin{cfa}
    1784 (&(&®*®)®*®)r3 = &(&®*®)r2; §\C{// (\&*) cancel giving address in r2, (\&(\&*)*) cancel giving address in r3}§
     1771(&(&@*@)@*@)r3 = &(&@*@)r2; $\C{// (\&*) cancel giving address in r2, (\&(\&*)*) cancel giving address in r3}$
    17851772\end{cfa}
    17861773Cancellation\index{cancellation!pointer/reference}\index{pointer!cancellation} works to arbitrary depth.
     
    17901777int x, *p1 = &x, **p2 = &p1, ***p3 = &p2,
    17911778                 &r1 = x,    &&r2 = r1,   &&&r3 = r2;
    1792 ***p3 = 3; §\C{// change x}§
    1793 r3 = 3; §\C{// change x, ***r3}§
    1794 **p3 = ...; §\C{// change p1}§
    1795 &r3 = ...; §\C{// change r1, (\&*)**r3, 1 cancellation}§
    1796 *p3 = ...; §\C{// change p2}§
    1797 &&r3 = ...; §\C{// change r2, (\&(\&*)*)*r3, 2 cancellations}§
    1798 &&&r3 = p3; §\C{// change r3 to p3, (\&(\&(\&*)*)*)r3, 3 cancellations}§
     1779***p3 = 3; $\C{// change x}$
     1780r3 = 3; $\C{// change x, ***r3}$
     1781**p3 = ...; $\C{// change p1}$
     1782&r3 = ...; $\C{// change r1, (\&*)**r3, 1 cancellation}$
     1783*p3 = ...; $\C{// change p2}$
     1784&&r3 = ...; $\C{// change r2, (\&(\&*)*)*r3, 2 cancellations}$
     1785&&&r3 = p3; $\C{// change r3 to p3, (\&(\&(\&*)*)*)r3, 3 cancellations}$
    17991786\end{cfa}
    18001787Furthermore, both types are equally performant, as the same amount of dereferencing occurs for both types.
     
    18031790As for a pointer type, a reference type may have qualifiers:
    18041791\begin{cfa}
    1805 const int cx = 5; §\C{// cannot change cx;}§
    1806 const int & cr = cx; §\C{// cannot change what cr points to}§
    1807 ®&®cr = &cx; §\C{// can change cr}§
    1808 cr = 7; §\C{// error, cannot change cx}§
    1809 int & const rc = x; §\C{// must be initialized}§
    1810 ®&®rc = &x; §\C{// error, cannot change rc}§
    1811 const int & const crc = cx; §\C{// must be initialized}§
    1812 crc = 7; §\C{// error, cannot change cx}§
    1813 ®&®crc = &cx; §\C{// error, cannot change crc}§
     1792const int cx = 5; $\C{// cannot change cx;}$
     1793const int & cr = cx; $\C{// cannot change what cr points to}$
     1794@&@cr = &cx; $\C{// can change cr}$
     1795cr = 7; $\C{// error, cannot change cx}$
     1796int & const rc = x; $\C{// must be initialized}$
     1797@&@rc = &x; $\C{// error, cannot change rc}$
     1798const int & const crc = cx; $\C{// must be initialized}$
     1799crc = 7; $\C{// error, cannot change cx}$
     1800@&@crc = &cx; $\C{// error, cannot change crc}$
    18141801\end{cfa}
    18151802Hence, for type ©& const©, there is no pointer assignment, so ©&rc = &x© is disallowed, and \emph{the address value cannot be the null pointer unless an arbitrary pointer is coerced\index{coercion} into the reference}:
    18161803\begin{cfa}
    1817 int & const cr = *0; §\C{// where 0 is the int * zero}§
     1804int & const cr = *0; $\C{// where 0 is the int * zero}$
    18181805\end{cfa}
    18191806Note, constant reference-types do not prevent \Index{addressing errors} because of explicit storage-management:
     
    18221809cr = 5;
    18231810free( &cr );
    1824 cr = 7; §\C{// unsound pointer dereference}§
     1811cr = 7; $\C{// unsound pointer dereference}$
    18251812\end{cfa}
    18261813
    18271814The position of the ©const© qualifier \emph{after} the pointer/reference qualifier causes confuse for C programmers.
    18281815The ©const© qualifier cannot be moved before the pointer/reference qualifier for C style-declarations;
    1829 \CFA-style declarations (see \VRef{s:AlternativeDeclarations}) attempt to address this issue:
     1816\CFA-style declarations \see{\VRef{s:AlternativeDeclarations}} attempt to address this issue:
    18301817\begin{cquote}
    18311818\begin{tabular}{@{}l@{\hspace{3em}}l@{}}
    18321819\multicolumn{1}{c@{\hspace{3em}}}{\textbf{\CFA}}        & \multicolumn{1}{c}{\textbf{C}}        \\
    18331820\begin{cfa}
    1834 ®const® * ®const® * const int ccp;
    1835 ®const® & ®const® & const int ccr;
     1821@const@ * @const@ * const int ccp;
     1822@const@ & @const@ & const int ccr;
    18361823\end{cfa}
    18371824&
    18381825\begin{cfa}
    1839 const int * ®const® * ®const® ccp;
     1826const int * @const@ * @const@ ccp;
    18401827
    18411828\end{cfa}
     
    18461833Finally, like pointers, references are usable and composable with other type operators and generators.
    18471834\begin{cfa}
    1848 int w, x, y, z, & ar[3] = { x, y, z }; §\C{// initialize array of references}§
    1849 &ar[1] = &w; §\C{// change reference array element}§
    1850 typeof( ar[1] ) p; §\C{// (gcc) is int, \ie the type of referenced object}§
    1851 typeof( &ar[1] ) q; §\C{// (gcc) is int \&, \ie the type of reference}§
    1852 sizeof( ar[1] ) == sizeof( int ); §\C{// is true, \ie the size of referenced object}§
    1853 sizeof( &ar[1] ) == sizeof( int *) §\C{// is true, \ie the size of a reference}§
     1835int w, x, y, z, & ar[3] = { x, y, z }; $\C{// initialize array of references}$
     1836&ar[1] = &w; $\C{// change reference array element}$
     1837typeof( ar[1] ) p; $\C{// (gcc) is int, \ie the type of referenced object}$
     1838typeof( &ar[1] ) q; $\C{// (gcc) is int \&, \ie the type of reference}$
     1839sizeof( ar[1] ) == sizeof( int ); $\C{// is true, \ie the size of referenced object}$
     1840sizeof( &ar[1] ) == sizeof( int *) $\C{// is true, \ie the size of a reference}$
    18541841\end{cfa}
    18551842
    18561843In contrast to \CFA reference types, \Index*[C++]{\CC{}}'s reference types are all ©const© references, preventing changes to the reference address, so only value assignment is possible, which eliminates half of the \Index{address duality}.
    18571844Also, \CC does not allow \Index{array}s\index{array!reference} of reference\footnote{
    1858 The reason for disallowing arrays of reference is unknown, but possibly comes from references being ethereal (like a textual macro), and hence, replaceable by the referant object.}
     1845The reason for disallowing arrays of reference is unknown, but possibly comes from references being ethereal (like a textual macro), and hence, replaceable by the referent object.}
    18591846\Index*{Java}'s reference types to objects (all Java objects are on the heap) are like C pointers, which always manipulate the address, and there is no (bit-wise) object assignment, so objects are explicitly cloned by shallow or deep copying, which eliminates half of the address duality.
    18601847
     
    18681855Therefore, for pointer/reference initialization, the initializing value must be an address not a value.
    18691856\begin{cfa}
    1870 int * p = &x; §\C{// assign address of x}§
    1871 ®int * p = x;® §\C{// assign value of x}§
    1872 int & r = x; §\C{// must have address of x}§
     1857int * p = &x; $\C{// assign address of x}$
     1858@int * p = x;@ $\C{// assign value of x}$
     1859int & r = x; $\C{// must have address of x}$
    18731860\end{cfa}
    18741861Like the previous example with C pointer-arithmetic, it is unlikely assigning the value of ©x© into a pointer is meaningful (again, a warning is usually given).
     
    18791866Similarly, when a reference type is used for a parameter/return type, the call-site argument does not require a reference operator for the same reason.
    18801867\begin{cfa}
    1881 int & f( int & r ); §\C{// reference parameter and return}§
    1882 z = f( x ) + f( y ); §\C{// reference operator added, temporaries needed for call results}§
     1868int & f( int & r ); $\C{// reference parameter and return}$
     1869z = f( x ) + f( y ); $\C{// reference operator added, temporaries needed for call results}$
    18831870\end{cfa}
    18841871Within routine ©f©, it is possible to change the argument by changing the corresponding parameter, and parameter ©r© can be locally reassigned within ©f©.
     
    18931880When a pointer/reference parameter has a ©const© value (immutable), it is possible to pass literals and expressions.
    18941881\begin{cfa}
    1895 void f( ®const® int & cr );
    1896 void g( ®const® int * cp );
    1897 f( 3 );                   g( ®&®3 );
    1898 f( x + y );             g( ®&®(x + y) );
     1882void f( @const@ int & cr );
     1883void g( @const@ int * cp );
     1884f( 3 );                   g( @&@3 );
     1885f( x + y );             g( @&@(x + y) );
    18991886\end{cfa}
    19001887Here, the compiler passes the address to the literal 3 or the temporary for the expression ©x + y©, knowing the argument cannot be changed through the parameter.
     
    19071894void f( int & r );
    19081895void g( int * p );
    1909 f( 3 );                   g( ®&®3 ); §\C{// compiler implicit generates temporaries}§
    1910 f( x + y );             g( ®&®(x + y) ); §\C{// compiler implicit generates temporaries}§
     1896f( 3 );                   g( @&@3 ); $\C{// compiler implicit generates temporaries}$
     1897f( x + y );             g( @&@(x + y) ); $\C{// compiler implicit generates temporaries}$
    19111898\end{cfa}
    19121899Essentially, there is an implicit \Index{rvalue} to \Index{lvalue} conversion in this case.\footnote{
     
    19191906\begin{cfa}
    19201907void f( int i );
    1921 void (* fp)( int ); §\C{// routine pointer}§
    1922 fp = f; §\C{// reference initialization}§
    1923 fp = &f; §\C{// pointer initialization}§
    1924 fp = *f; §\C{// reference initialization}§
    1925 fp(3); §\C{// reference invocation}§
    1926 (*fp)(3); §\C{// pointer invocation}§
     1908void (* fp)( int ); $\C{// routine pointer}$
     1909fp = f; $\C{// reference initialization}$
     1910fp = &f; $\C{// pointer initialization}$
     1911fp = *f; $\C{// reference initialization}$
     1912fp(3); $\C{// reference invocation}$
     1913(*fp)(3); $\C{// pointer invocation}$
    19271914\end{cfa}
    19281915While C's treatment of routine objects has similarity to inferring a reference type in initialization contexts, the examples are assignment not initialization, and all possible forms of assignment are possible (©f©, ©&f©, ©*f©) without regard for type.
    19291916Instead, a routine object should be referenced by a ©const© reference:
    19301917\begin{cfa}
    1931 ®const® void (®&® fr)( int ) = f; §\C{// routine reference}§
    1932 fr = ... §\C{// error, cannot change code}§
    1933 &fr = ...; §\C{// changing routine reference}§
    1934 fr( 3 ); §\C{// reference call to f}§
    1935 (*fr)(3); §\C{// error, incorrect type}§
     1918@const@ void (@&@ fr)( int ) = f; $\C{// routine reference}$
     1919fr = ... $\C{// error, cannot change code}$
     1920&fr = ...; $\C{// changing routine reference}$
     1921fr( 3 ); $\C{// reference call to f}$
     1922(*fr)(3); $\C{// error, incorrect type}$
    19361923\end{cfa}
    19371924because the value of the routine object is a routine literal, \ie the routine code is normally immutable during execution.\footnote{
     
    19461933\begin{itemize}
    19471934\item
    1948 if ©R© is an \Index{rvalue} of type ©T &©$_1\cdots$ ©&©$_r$, where $r \ge 1$ references (©&© symbols), than ©&R© has type ©T ®*®&©$_{\color{red}2}\cdots$ ©&©$_{\color{red}r}$, \ie ©T© pointer with $r-1$ references (©&© symbols).
    1949 
    1950 \item
    1951 if ©L© is an \Index{lvalue} of type ©T &©$_1\cdots$ ©&©$_l$, where $l \ge 0$ references (©&© symbols), than ©&L© has type ©T ®*®&©$_{\color{red}1}\cdots$ ©&©$_{\color{red}l}$, \ie ©T© pointer with $l$ references (©&© symbols).
     1935if ©R© is an \Index{rvalue} of type ©T &©$_1\cdots$ ©&©$_r$, where $r \ge 1$ references (©&© symbols), than ©&R© has type ©T ©\R{©*©}©&©\R{$_2$}$\cdots$ ©&©\R{$_r$}, \ie ©T© pointer with $r-1$ references (©&© symbols).
     1936
     1937\item
     1938if ©L© is an \Index{lvalue} of type ©T &©$_1\cdots$ ©&©$_l$, where $l \ge 0$ references (©&© symbols), than ©&L© has type ©T ©\R{©*©}©&©\R{$_1$}$\cdots$ ©&©\R{$_l$}, \ie ©T© pointer with $l$ references (©&© symbols).
    19521939\end{itemize}
    19531940The following example shows the first rule applied to different \Index{rvalue} contexts:
     
    19551942int x, * px, ** ppx, *** pppx, **** ppppx;
    19561943int & rx = x, && rrx = rx, &&& rrrx = rrx ;
    1957 x = rrrx; §\C[2.0in]{// rrrx is an lvalue with type int \&\&\& (equivalent to x)}§
    1958 px = &rrrx; §\C{// starting from rrrx, \&rrrx is an rvalue with type int *\&\&\& (\&x)}§
    1959 ppx = &&rrrx; §\C{// starting from \&rrrx, \&\&rrrx is an rvalue with type int **\&\& (\&rx)}§
    1960 pppx = &&&rrrx; §\C{// starting from \&\&rrrx, \&\&\&rrrx is an rvalue with type int ***\& (\&rrx)}§
    1961 ppppx = &&&&rrrx; §\C{// starting from \&\&\&rrrx, \&\&\&\&rrrx is an rvalue with type int **** (\&rrrx)}§
     1944x = rrrx; $\C[2.0in]{// rrrx is an lvalue with type int \&\&\& (equivalent to x)}$
     1945px = &rrrx; $\C{// starting from rrrx, \&rrrx is an rvalue with type int *\&\&\& (\&x)}$
     1946ppx = &&rrrx; $\C{// starting from \&rrrx, \&\&rrrx is an rvalue with type int **\&\& (\&rx)}$
     1947pppx = &&&rrrx; $\C{// starting from \&\&rrrx, \&\&\&rrrx is an rvalue with type int ***\& (\&rrx)}$
     1948ppppx = &&&&rrrx; $\C{// starting from \&\&\&rrrx, \&\&\&\&rrrx is an rvalue with type int **** (\&rrrx)}$
    19621949\end{cfa}
    19631950The following example shows the second rule applied to different \Index{lvalue} contexts:
     
    19651952int x, * px, ** ppx, *** pppx;
    19661953int & rx = x, && rrx = rx, &&& rrrx = rrx ;
    1967 rrrx = 2; §\C{// rrrx is an lvalue with type int \&\&\& (equivalent to x)}§
    1968 &rrrx = px; §\C{// starting from rrrx, \&rrrx is an rvalue with type int *\&\&\& (rx)}§
    1969 &&rrrx = ppx; §\C{// starting from \&rrrx, \&\&rrrx is an rvalue with type int **\&\& (rrx)}§
    1970 &&&rrrx = pppx; §\C{// starting from \&\&rrrx, \&\&\&rrrx is an rvalue with type int ***\& (rrrx)}\CRT§
     1954rrrx = 2; $\C{// rrrx is an lvalue with type int \&\&\& (equivalent to x)}$
     1955&rrrx = px; $\C{// starting from rrrx, \&rrrx is an rvalue with type int *\&\&\& (rx)}$
     1956&&rrrx = ppx; $\C{// starting from \&rrrx, \&\&rrrx is an rvalue with type int **\&\& (rrx)}$
     1957&&&rrrx = pppx; $\C{// starting from \&\&rrrx, \&\&\&rrrx is an rvalue with type int ***\& (rrrx)}\CRT$
    19711958\end{cfa}
    19721959
     
    19811968\begin{cfa}
    19821969int x;
    1983 x + 1; §\C[2.0in]{// lvalue variable (int) converts to rvalue for expression}§
     1970x + 1; $\C[2.0in]{// lvalue variable (int) converts to rvalue for expression}$
    19841971\end{cfa}
    19851972An rvalue has no type qualifiers (©cv©), so the lvalue qualifiers are dropped.
     
    19911978\begin{cfa}
    19921979int x, &r = x, f( int p );
    1993 x = ®r® + f( ®r® ); §\C{// lvalue reference converts to rvalue}§
     1980x = @r@ + f( @r@ ); $\C{// lvalue reference converts to rvalue}$
    19941981\end{cfa}
    19951982An rvalue has no type qualifiers (©cv©), so the reference qualifiers are dropped.
     
    19981985lvalue to reference conversion: \lstinline[deletekeywords=lvalue]@lvalue-type cv1 T@ converts to ©cv2 T &©, which allows implicitly converting variables to references.
    19991986\begin{cfa}
    2000 int x, &r = ®x®, f( int & p ); §\C{// lvalue variable (int) convert to reference (int \&)}§
    2001 f( ®x® ); §\C{// lvalue variable (int) convert to reference (int \&)}§
     1987int x, &r = @x@, f( int & p ); $\C{// lvalue variable (int) convert to reference (int \&)}$
     1988f( @x@ ); $\C{// lvalue variable (int) convert to reference (int \&)}$
    20021989\end{cfa}
    20031990Conversion can restrict a type, where ©cv1© $\le$ ©cv2©, \eg passing an ©int© to a ©const volatile int &©, which has low cost.
     
    20091996\begin{cfa}
    20101997int x, & f( int & p );
    2011 f( ®x + 3® );   §\C[1.5in]{// rvalue parameter (int) implicitly converts to lvalue temporary reference (int \&)}§
    2012 ®&f®(...) = &x; §\C{// rvalue result (int \&) implicitly converts to lvalue temporary reference (int \&)}\CRT§
     1998f( @x + 3@ );   $\C[1.5in]{// rvalue parameter (int) implicitly converts to lvalue temporary reference (int \&)}$
     1999@&f@(...) = &x; $\C{// rvalue result (int \&) implicitly converts to lvalue temporary reference (int \&)}\CRT$
    20132000\end{cfa}
    20142001In both case, modifications to the temporary are inaccessible (\Index{warning}).
     
    21822169The point of the new syntax is to allow returning multiple values from a routine~\cite{Galletly96,CLU}, \eg:
    21832170\begin{cfa}
    2184 ®[ int o1, int o2, char o3 ]® f( int i1, char i2, char i3 ) {
    2185         §\emph{routine body}§
     2171@[ int o1, int o2, char o3 ]@ f( int i1, char i2, char i3 ) {
     2172        $\emph{routine body}$
    21862173}
    21872174\end{cfa}
     
    21942181Declaration qualifiers can only appear at the start of a routine definition, \eg:
    21952182\begin{cfa}
    2196 ®extern® [ int x ] g( int y ) {§\,§}
     2183@extern@ [ int x ] g( int y ) {$\,$}
    21972184\end{cfa}
    21982185Lastly, if there are no output parameters or input parameters, the brackets and/or parentheses must still be specified;
    21992186in both cases the type is assumed to be void as opposed to old style C defaults of int return type and unknown parameter types, respectively, as in:
    22002187\begin{cfa}
    2201 [§\,§] g(); §\C{// no input or output parameters}§
    2202 [ void ] g( void ); §\C{// no input or output parameters}§
     2188[$\,$] g(); $\C{// no input or output parameters}$
     2189[ void ] g( void ); $\C{// no input or output parameters}$
    22032190\end{cfa}
    22042191
     
    22182205\begin{cfa}
    22192206typedef int foo;
    2220 int f( int (* foo) ); §\C{// foo is redefined as a parameter name}§
     2207int f( int (* foo) ); $\C{// foo is redefined as a parameter name}$
    22212208\end{cfa}
    22222209The string ``©int (* foo)©'' declares a C-style named-parameter of type pointer to an integer (the parenthesis are superfluous), while the same string declares a \CFA style unnamed parameter of type routine returning integer with unnamed parameter of type pointer to foo.
     
    22262213C-style declarations can be used to declare parameters for \CFA style routine definitions, \eg:
    22272214\begin{cfa}
    2228 [ int ] f( * int, int * ); §\C{// returns an integer, accepts 2 pointers to integers}§
    2229 [ * int, int * ] f( int ); §\C{// returns 2 pointers to integers, accepts an integer}§
     2215[ int ] f( * int, int * ); $\C{// returns an integer, accepts 2 pointers to integers}$
     2216[ * int, int * ] f( int ); $\C{// returns 2 pointers to integers, accepts an integer}$
    22302217\end{cfa}
    22312218The reason for allowing both declaration styles in the new context is for backwards compatibility with existing preprocessor macros that generate C-style declaration-syntax, as in:
    22322219\begin{cfa}
    22332220#define ptoa( n, d ) int (*n)[ d ]
    2234 int f( ptoa( p, 5 ) ) ... §\C{// expands to int f( int (*p)[ 5 ] )}§
    2235 [ int ] f( ptoa( p, 5 ) ) ... §\C{// expands to [ int ] f( int (*p)[ 5 ] )}§
     2221int f( ptoa( p, 5 ) ) ... $\C{// expands to int f( int (*p)[ 5 ] )}$
     2222[ int ] f( ptoa( p, 5 ) ) ... $\C{// expands to [ int ] f( int (*p)[ 5 ] )}$
    22362223\end{cfa}
    22372224Again, programmers are highly encouraged to use one declaration form or the other, rather than mixing the forms.
     
    22522239\begin{minipage}{\linewidth}
    22532240\begin{cfa}
    2254 ®[ int x, int y ]® f() {
     2241@[ int x, int y ]@ f() {
    22552242        int z;
    22562243        ... x = 0; ... y = z; ...
    2257         ®return;® §\C{// implicitly return x, y}§
     2244        @return;@ $\C{// implicitly return x, y}$
    22582245}
    22592246\end{cfa}
     
    22652252[ int x, int y ] f() {
    22662253        ...
    2267 } §\C{// implicitly return x, y}§
     2254} $\C{// implicitly return x, y}$
    22682255\end{cfa}
    22692256In this case, the current values of ©x© and ©y© are returned to the calling routine just as if a ©return© had been encountered.
     
    22742261[ int x, int y ] f( int, x, int y ) {
    22752262        ...
    2276 } §\C{// implicitly return x, y}§
     2263} $\C{// implicitly return x, y}$
    22772264\end{cfa}
    22782265This notation allows the compiler to eliminate temporary variables in nested routine calls.
    22792266\begin{cfa}
    2280 [ int x, int y ] f( int, x, int y ); §\C{// prototype declaration}§
     2267[ int x, int y ] f( int, x, int y ); $\C{// prototype declaration}$
    22812268int a, b;
    22822269[a, b] = f( f( f( a, b ) ) );
     
    22922279as well, parameter names are optional, \eg:
    22932280\begin{cfa}
    2294 [ int x ] f (); §\C{// returning int with no parameters}§
    2295 [ * int ] g (int y); §\C{// returning pointer to int with int parameter}§
    2296 [ ] h ( int, char ); §\C{// returning no result with int and char parameters}§
    2297 [ * int, int ] j ( int ); §\C{// returning pointer to int and int, with int parameter}§
     2281[ int x ] f (); $\C{// returning int with no parameters}$
     2282[ * int ] g (int y); $\C{// returning pointer to int with int parameter}$
     2283[ ] h ( int, char ); $\C{// returning no result with int and char parameters}$
     2284[ * int, int ] j ( int ); $\C{// returning pointer to int and int, with int parameter}$
    22982285\end{cfa}
    22992286This syntax allows a prototype declaration to be created by cutting and pasting source text from the routine definition header (or vice versa).
    2300 Like C, it is possible to declare multiple routine-prototypes in a single declaration, where the return type is distributed across \emph{all} routine names in the declaration list (see~\VRef{s:AlternativeDeclarations}), \eg:
     2287Like C, it is possible to declare multiple routine-prototypes in a single declaration, where the return type is distributed across \emph{all} routine names in the declaration list \see{\VRef{s:AlternativeDeclarations}}, \eg:
    23012288\begin{cfa}
    23022289C :             const double bar1(), bar2( int ), bar3( double );
    2303 §\CFA§: [const double] foo(), foo( int ), foo( double ) { return 3.0; }
     2290$\CFA$: [const double] foo(), foo( int ), foo( double ) { return 3.0; }
    23042291\end{cfa}
    23052292\CFA allows the last routine in the list to define its body.
     
    23162303The syntax for pointers to \CFA routines specifies the pointer name on the right, \eg:
    23172304\begin{cfa}
    2318 * [ int x ] () fp; §\C{// pointer to routine returning int with no parameters}§
    2319 * [ * int ] (int y) gp; §\C{// pointer to routine returning pointer to int with int parameter}§
    2320 * [ ] (int,char) hp; §\C{// pointer to routine returning no result with int and char parameters}§
    2321 * [ * int,int ] ( int ) jp; §\C{// pointer to routine returning pointer to int and int, with int parameter}§
     2305* [ int x ] () fp; $\C[2.25in]{// pointer to routine returning int with no parameters}$
     2306* [ * int ] (int y) gp; $\C{// pointer to routine returning pointer to int with int parameter}$
     2307* [ ] (int,char) hp; $\C{// pointer to routine returning no result with int and char parameters}$
     2308* [ * int,int ] ( int ) jp; $\C{// pointer to routine returning pointer to int and int, with int parameter}\CRT$
    23222309\end{cfa}
    23232310While parameter names are optional, \emph{a routine name cannot be specified};
    23242311for example, the following is incorrect:
    23252312\begin{cfa}
    2326 * [ int x ] f () fp; §\C{// routine name "f" is not allowed}§
     2313* [ int x ] f () fp; $\C{// routine name "f" is not allowed}$
    23272314\end{cfa}
    23282315
     
    23472334whereas a named (keyword) call may be:
    23482335\begin{cfa}
    2349 p( z : 3, x : 4, y : 7 );  §\C{// rewrite $\Rightarrow$ p( 4, 7, 3 )}§
     2336p( z : 3, x : 4, y : 7 );  $\C{// rewrite \(\Rightarrow\) p( 4, 7, 3 )}$
    23502337\end{cfa}
    23512338Here the order of the arguments is unimportant, and the names of the parameters are used to associate argument values with the corresponding parameters.
     
    23642351For example, the following routine prototypes and definition are all valid.
    23652352\begin{cfa}
    2366 void p( int, int, int ); §\C{// equivalent prototypes}§
     2353void p( int, int, int ); $\C{// equivalent prototypes}$
    23672354void p( int x, int y, int z );
    23682355void p( int y, int x, int z );
    23692356void p( int z, int y, int x );
    2370 void p( int q, int r, int s ) {} §\C{// match with this definition}§
     2357void p( int q, int r, int s ) {} $\C{// match with this definition}$
    23712358\end{cfa}
    23722359Forcing matching parameter names in routine prototypes with corresponding routine definitions is possible, but goes against a strong tradition in C programming.
     
    23802367int f( int x, double y );
    23812368
    2382 f( j : 3, i : 4 ); §\C{// 1st f}§
    2383 f( x : 7, y : 8.1 ); §\C{// 2nd f}§
    2384 f( 4, 5 );  §\C{// ambiguous call}§
     2369f( j : 3, i : 4 ); $\C{// 1st f}$
     2370f( x : 7, y : 8.1 ); $\C{// 2nd f}$
     2371f( 4, 5 );  $\C{// ambiguous call}$
    23852372\end{cfa}
    23862373However, named arguments compound routine resolution in conjunction with conversions:
    23872374\begin{cfa}
    2388 f( i : 3, 5.7 ); §\C{// ambiguous call ?}§
     2375f( i : 3, 5.7 ); $\C{// ambiguous call ?}$
    23892376\end{cfa}
    23902377Depending on the cost associated with named arguments, this call could be resolvable or ambiguous.
     
    24002387the allowable positional calls are:
    24012388\begin{cfa}
    2402 p(); §\C{// rewrite $\Rightarrow$ p( 1, 2, 3 )}§
    2403 p( 4 ); §\C{// rewrite $\Rightarrow$ p( 4, 2, 3 )}§
    2404 p( 4, 4 ); §\C{// rewrite $\Rightarrow$ p( 4, 4, 3 )}§
    2405 p( 4, 4, 4 ); §\C{// rewrite $\Rightarrow$ p( 4, 4, 4 )}§
     2389p(); $\C{// rewrite \(\Rightarrow\) p( 1, 2, 3 )}$
     2390p( 4 ); $\C{// rewrite \(\Rightarrow\) p( 4, 2, 3 )}$
     2391p( 4, 4 ); $\C{// rewrite \(\Rightarrow\) p( 4, 4, 3 )}$
     2392p( 4, 4, 4 ); $\C{// rewrite \(\Rightarrow\) p( 4, 4, 4 )}$
    24062393// empty arguments
    2407 p(  , 4, 4 ); §\C{// rewrite $\Rightarrow$ p( 1, 4, 4 )}§
    2408 p( 4,  , 4 ); §\C{// rewrite $\Rightarrow$ p( 4, 2, 4 )}§
    2409 p( 4, 4,   ); §\C{// rewrite $\Rightarrow$ p( 4, 4, 3 )}§
    2410 p( 4,  ,   ); §\C{// rewrite $\Rightarrow$ p( 4, 2, 3 )}§
    2411 p(  , 4,   ); §\C{// rewrite $\Rightarrow$ p( 1, 4, 3 )}§
    2412 p(  ,  , 4 ); §\C{// rewrite $\Rightarrow$ p( 1, 2, 4 )}§
    2413 p(  ,  ,   ); §\C{// rewrite $\Rightarrow$ p( 1, 2, 3 )}§
     2394p(  , 4, 4 ); $\C{// rewrite \(\Rightarrow\) p( 1, 4, 4 )}$
     2395p( 4,  , 4 ); $\C{// rewrite \(\Rightarrow\) p( 4, 2, 4 )}$
     2396p( 4, 4,   ); $\C{// rewrite \(\Rightarrow\) p( 4, 4, 3 )}$
     2397p( 4,  ,   ); $\C{// rewrite \(\Rightarrow\) p( 4, 2, 3 )}$
     2398p(  , 4,   ); $\C{// rewrite \(\Rightarrow\) p( 1, 4, 3 )}$
     2399p(  ,  , 4 ); $\C{// rewrite \(\Rightarrow\) p( 1, 2, 4 )}$
     2400p(  ,  ,   ); $\C{// rewrite \(\Rightarrow\) p( 1, 2, 3 )}$
    24142401\end{cfa}
    24152402Here the missing arguments are inserted from the default values in the parameter list.
     
    24352422Default values may only appear in a prototype versus definition context:
    24362423\begin{cfa}
    2437 void p( int x, int y = 2, int z = 3 ); §\C{// prototype: allowed}§
    2438 void p( int, int = 2, int = 3 ); §\C{// prototype: allowed}§
    2439 void p( int x, int y = 2, int z = 3 ) {} §\C{// definition: not allowed}§
     2424void p( int x, int y = 2, int z = 3 ); $\C{// prototype: allowed}$
     2425void p( int, int = 2, int = 3 ); $\C{// prototype: allowed}$
     2426void p( int x, int y = 2, int z = 3 ) {} $\C{// definition: not allowed}$
    24402427\end{cfa}
    24412428The reason for this restriction is to allow separate compilation.
     
    24522439\begin{cfa}
    24532440p( int x, int y, int z, ... );
    2454 p( 1, 4, 5, 6, z : 3, y : 2 ); §\C{// assume p( /* positional */, ... , /* named */ );}§
    2455 p( 1, z : 3, y : 2, 4, 5, 6 ); §\C{// assume p( /* positional */, /* named */, ... );}§
     2441p( 1, 4, 5, 6, z : 3, y : 2 ); $\C{// assume p( /* positional */, ... , /* named */ );}$
     2442p( 1, z : 3, y : 2, 4, 5, 6 ); $\C{// assume p( /* positional */, /* named */, ... );}$
    24562443\end{cfa}
    24572444In the first call, it is necessary for the programmer to conceptually rewrite the call, changing named arguments into positional, before knowing where the ellipse arguments begin.
     
    24622449\begin{cfa}
    24632450void p( int x, int y = 2, int z = 3... );
    2464 p( 1, 4, 5, 6, z : 3 ); §\C{// assume p( /* positional */, ... , /* named */ );}§
    2465 p( 1, z : 3, 4, 5, 6 ); §\C{// assume p( /* positional */, /* named */, ... );}§
     2451p( 1, 4, 5, 6, z : 3 ); $\C{// assume p( /* positional */, ... , /* named */ );}$
     2452p( 1, z : 3, 4, 5, 6 ); $\C{// assume p( /* positional */, /* named */, ... );}$
    24662453\end{cfa}
    24672454The first call is an error because arguments 4 and 5 are actually positional not ellipse arguments;
     
    24692456In the second call, the default value for y is implicitly inserted after argument 1 and the named arguments separate the positional and ellipse arguments, making it trivial to read the call.
    24702457For these reasons, \CFA requires named arguments before ellipse arguments.
    2471 Finally, while ellipse arguments are needed for a small set of existing C routines, like printf, the extended \CFA type system largely eliminates the need for ellipse arguments (see Section 24), making much of this discussion moot.
    2472 
    2473 Default arguments and overloading (see Section 24) are complementary.
     2458Finally, while ellipse arguments are needed for a small set of existing C routines, like ©printf©, the extended \CFA type system largely eliminates the need for ellipse arguments \see{\VRef{s:Overloading}}, making much of this discussion moot.
     2459
     2460Default arguments and overloading \see{\VRef{s:Overloading}} are complementary.
    24742461While in theory default arguments can be simulated with overloading, as in:
    24752462\begin{cquote}
     
    24932480Furthermore, overloading cannot handle accessing default arguments in the middle of a positional list, via a missing argument, such as:
    24942481\begin{cfa}
    2495 p( 1, /* default */, 5 ); §\C{// rewrite $\Rightarrow$ p( 1, 2, 5 )}§
     2482p( 1, /* default */, 5 ); $\C{// rewrite \(\Rightarrow\) p( 1, 2, 5 )}$
    24962483\end{cfa}
    24972484
     
    25062493\begin{cfa}
    25072494struct {
    2508         int f1; §\C{// named field}§
    2509         int f2 : 4; §\C{// named field with bit field size}§
    2510         int : 3; §\C{// unnamed field for basic type with bit field size}§
    2511         int ; §\C{// disallowed, unnamed field}§
    2512         int *; §\C{// disallowed, unnamed field}§
    2513         int (*)( int ); §\C{// disallowed, unnamed field}§
     2495        int f1; $\C{// named field}$
     2496        int f2 : 4; $\C{// named field with bit field size}$
     2497        int : 3; $\C{// unnamed field for basic type with bit field size}$
     2498        int ; $\C{// disallowed, unnamed field}$
     2499        int *; $\C{// disallowed, unnamed field}$
     2500        int (*)( int ); $\C{// disallowed, unnamed field}$
    25142501};
    25152502\end{cfa}
     
    25192506\begin{cfa}
    25202507struct {
    2521         int , , ; §\C{// 3 unnamed fields}§
     2508        int , , ; $\C{// 3 unnamed fields}$
    25222509}
    25232510\end{cfa}
     
    25312518\subsection{Type Nesting}
    25322519
    2533 \CFA allows \Index{type nesting}, and type qualification of the nested types (see \VRef[Figure]{f:TypeNestingQualification}), where as C hoists\index{type hoisting} (refactors) nested types into the enclosing scope and has no type qualification.
     2520\CFA allows \Index{type nesting}, and type qualification of the nested types \see{\VRef[Figure]{f:TypeNestingQualification}}, where as C hoists\index{type hoisting} (refactors) nested types into the enclosing scope and has no type qualification.
    25342521\begin{figure}
    25352522\centering
     
    25872574
    25882575int fred() {
    2589         s.t.c = ®S.®R;  // type qualification
    2590         struct ®S.®T t = { ®S.®R, 1, 2 };
    2591         enum ®S.®C c;
    2592         union ®S.T.®U u;
     2576        s.t.c = @S.@R;  // type qualification
     2577        struct @S.@T t = { @S.@R, 1, 2 };
     2578        enum @S.@C c;
     2579        union @S.T.@U u;
    25932580}
    25942581\end{cfa}
     
    26132600const unsigned int size = 5;
    26142601int ia[size];
    2615 ... §\C{// assign values to array ia}§
    2616 qsort( ia, size ); §\C{// sort ascending order using builtin ?<?}§
     2602... $\C{// assign values to array ia}$
     2603qsort( ia, size ); $\C{// sort ascending order using builtin ?<?}$
    26172604{
    2618         ®int ?<?( int x, int y ) { return x > y; }® §\C{// nested routine}§
    2619         qsort( ia, size ); §\C{// sort descending order by local redefinition}§
     2605        @int ?<?( int x, int y ) { return x > y; }@ $\C{// nested routine}$
     2606        qsort( ia, size ); $\C{// sort descending order by local redefinition}$
    26202607}
    26212608\end{cfa}
     
    26252612The following program in undefined in \CFA (and Indexc{gcc})
    26262613\begin{cfa}
    2627 [* [int]( int )] foo() { §\C{// int (* foo())( int )}§
    2628         int ®i® = 7;
     2614[* [int]( int )] foo() { $\C{// int (* foo())( int )}$
     2615        int @i@ = 7;
    26292616        int bar( int p ) {
    2630                 ®i® += 1; §\C{// dependent on local variable}§
    2631                 sout | ®i®;
     2617                @i@ += 1; $\C{// dependent on local variable}$
     2618                sout | @i@;
    26322619        }
    2633         return bar; §\C{// undefined because of local dependence}§
     2620        return bar; $\C{// undefined because of local dependence}$
    26342621}
    26352622int main() {
    2636         * [int]( int ) fp = foo(); §\C{// int (* fp)( int )}§
     2623        * [int]( int ) fp = foo(); $\C{// int (* fp)( int )}$
    26372624        sout | fp( 3 );
    26382625}
     
    26472634In C and \CFA, lists of elements appear in several contexts, such as the parameter list of a routine call.
    26482635\begin{cfa}
    2649 f( ®2, x, 3 + i® ); §\C{// element list}§
     2636f( @2, x, 3 + i@ ); $\C{// element list}$
    26502637\end{cfa}
    26512638A list of elements is called a \newterm{tuple}, and is different from a \Index{comma expression}.
     
    26562643
    26572644In C and most programming languages, functions return at most one value;
    2658 however, many operations have multiple outcomes, some exceptional (see~\VRef{s:ExceptionHandling}).
     2645however, many operations have multiple outcomes, some exceptional \see{\VRef{s:ExceptionHandling}}.
    26592646To emulate functions with multiple return values, \emph{\Index{aggregation}} and/or \emph{\Index{aliasing}} is used.
    26602647
     
    26622649For example, consider C's \Indexc{div} function, which returns the quotient and remainder for a division of an integer value.
    26632650\begin{cfa}
    2664 typedef struct { int quot, rem; } div_t;        §\C[7cm]{// from include stdlib.h}§
     2651typedef struct { int quot, rem; } div_t;        $\C[7cm]{// from include stdlib.h}$
    26652652div_t div( int num, int den );
    2666 div_t qr = div( 13, 5 ); §\C{// return quotient/remainder aggregate}§
    2667 printf( "%d %d\n", qr.quot, qr.rem ); §\C{// print quotient/remainder}§
     2653div_t qr = div( 13, 5 ); $\C{// return quotient/remainder aggregate}$
     2654printf( "%d %d\n", qr.quot, qr.rem ); $\C{// print quotient/remainder}$
    26682655\end{cfa}
    26692656This approach requires a name for the return type and fields, where \Index{naming} is a common programming-language issue.
     
    26752662For example, consider C's \Indexc{modf} function, which returns the integral and fractional part of a floating value.
    26762663\begin{cfa}
    2677 double modf( double x, double * i ); §\C{// from include math.h}§
    2678 double intp, frac = modf( 13.5, &intp ); §\C{// return integral and fractional components}§
    2679 printf( "%g %g\n", intp, frac ); §\C{// print integral/fractional components}§
     2664double modf( double x, double * i ); $\C{// from include math.h}$
     2665double intp, frac = modf( 13.5, &intp ); $\C{// return integral and fractional components}$
     2666printf( "%g %g\n", intp, frac ); $\C{// print integral/fractional components}$
    26802667\end{cfa}
    26812668This approach requires allocating storage for the return values, which complicates the call site with a sequence of variable declarations leading to the call.
     
    27042691When a function call is passed as an argument to another call, the best match of actual arguments to formal parameters is evaluated given all possible expression interpretations in the current scope.
    27052692\begin{cfa}
    2706 void g( int, int ); §\C{// 1}§
    2707 void g( double, double ); §\C{// 2}§
    2708 g( div( 13, 5 ) ); §\C{// select 1}§
    2709 g( modf( 13.5 ) ); §\C{// select 2}§
     2693void g( int, int ); $\C{// 1}$
     2694void g( double, double ); $\C{// 2}$
     2695g( div( 13, 5 ) ); $\C{// select 1}$
     2696g( modf( 13.5 ) ); $\C{// select 2}$
    27102697\end{cfa}
    27112698In this case, there are two overloaded ©g© routines.
     
    27162703The previous examples can be rewritten passing the multiple returned-values directly to the ©printf© function call.
    27172704\begin{cfa}
    2718 [ int, int ] div( int x, int y ); §\C{// from include stdlib}§
    2719 printf( "%d %d\n", div( 13, 5 ) ); §\C{// print quotient/remainder}§
    2720 
    2721 [ double, double ] modf( double x ); §\C{// from include math}§
    2722 printf( "%g %g\n", modf( 13.5 ) ); §\C{// print integral/fractional components}§
     2705[ int, int ] div( int x, int y ); $\C{// from include stdlib}$
     2706printf( "%d %d\n", div( 13, 5 ) ); $\C{// print quotient/remainder}$
     2707
     2708[ double, double ] modf( double x ); $\C{// from include math}$
     2709printf( "%g %g\n", modf( 13.5 ) ); $\C{// print integral/fractional components}$
    27232710\end{cfa}
    27242711This approach provides the benefits of compile-time checking for appropriate return statements as in aggregation, but without the required verbosity of declaring a new named type.
     
    27302717\begin{cfa}
    27312718int quot, rem;
    2732 [ quot, rem ] = div( 13, 5 ); §\C{// assign multiple variables}§
    2733 printf( "%d %d\n", quot, rem ); §\C{// print quotient/remainder}\CRT§
     2719[ quot, rem ] = div( 13, 5 ); $\C{// assign multiple variables}$
     2720printf( "%d %d\n", quot, rem ); $\C{// print quotient/remainder}\CRT$
    27342721\end{cfa}
    27352722Here, the multiple return-values are matched in much the same way as passing multiple return-values to multiple parameters in a call.
     
    27602747In \CFA, it is possible to overcome this restriction by declaring a \newterm{tuple variable}.
    27612748\begin{cfa}
    2762 [int, int] ®qr® = div( 13, 5 ); §\C{// initialize tuple variable}§
    2763 printf( "%d %d\n", ®qr® ); §\C{// print quotient/remainder}§
     2749[int, int] @qr@ = div( 13, 5 ); $\C{// initialize tuple variable}$
     2750printf( "%d %d\n", @qr@ ); $\C{// print quotient/remainder}$
    27642751\end{cfa}
    27652752It is now possible to match the multiple return-values to a single variable, in much the same way as \Index{aggregation}.
     
    27672754One way to access the individual components of a tuple variable is with assignment.
    27682755\begin{cfa}
    2769 [ quot, rem ] = qr; §\C{// assign multiple variables}§
     2756[ quot, rem ] = qr; $\C{// assign multiple variables}$
    27702757\end{cfa}
    27712758
     
    27902777[int, double] * p;
    27912778
    2792 int y = x.0; §\C{// access int component of x}§
    2793 y = f().1; §\C{// access int component of f}§
    2794 p->0 = 5; §\C{// access int component of tuple pointed-to by p}§
    2795 g( x.1, x.0 ); §\C{// rearrange x to pass to g}§
    2796 double z = [ x, f() ].0.1; §\C{// access second component of first component of tuple expression}§
     2779int y = x.0; $\C{// access int component of x}$
     2780y = f().1; $\C{// access int component of f}$
     2781p->0 = 5; $\C{// access int component of tuple pointed-to by p}$
     2782g( x.1, x.0 ); $\C{// rearrange x to pass to g}$
     2783double z = [ x, f() ].0.1; $\C{// access second component of first component of tuple expression}$
    27972784\end{cfa}
    27982785Tuple-index expressions can occur on any tuple-typed expression, including tuple-returning functions, square-bracketed tuple expressions, and other tuple-index expressions, provided the retrieved component is also a tuple.
     
    28012788
    28022789\subsection{Flattening and Structuring}
     2790\label{s:FlatteningStructuring}
    28032791
    28042792As evident in previous examples, tuples in \CFA do not have a rigid structure.
     
    28612849double y;
    28622850[int, double] z;
    2863 [y, x] = 3.14; §\C{// mass assignment}§
    2864 [x, y] = z;                                                         §\C{// multiple assignment}§
    2865 z = 10;                                                         §\C{// mass assignment}§
    2866 z = [x, y]; §\C{// multiple assignment}§
     2851[y, x] = 3.14; $\C{// mass assignment}$
     2852[x, y] = z;                                                         $\C{// multiple assignment}$
     2853z = 10;                                                         $\C{// mass assignment}$
     2854z = [x, y]; $\C{// multiple assignment}$
    28672855\end{cfa}
    28682856Let $L_i$ for $i$ in $[0, n)$ represent each component of the flattened left side, $R_i$ represent each component of the flattened right side of a multiple assignment, and $R$ represent the right side of a mass assignment.
     
    28722860\begin{cfa}
    28732861[ int, int ] x, y, z;
    2874 [ x, y ] = z;                                              §\C{// multiple assignment, invalid 4 != 2}§
     2862[ x, y ] = z;                                              $\C{// multiple assignment, invalid 4 != 2}$
    28752863\end{cfa}
    28762864Multiple assignment assigns $R_i$ to $L_i$ for each $i$.
     
    29082896        double c, d;
    29092897        [ void ] f( [ int, int ] );
    2910         f( [ c, a ] = [ b, d ] = 1.5 ); §\C{// assignments in parameter list}§
     2898        f( [ c, a ] = [ b, d ] = 1.5 ); $\C{// assignments in parameter list}$
    29112899\end{cfa}
    29122900The tuple expression begins with a mass assignment of ©1.5© into ©[b, d]©, which assigns ©1.5© into ©b©, which is truncated to ©1©, and ©1.5© into ©d©, producing the tuple ©[1, 1.5]© as a result.
     
    29212909\begin{cfa}
    29222910struct S;
    2923 void ?{}(S *); §\C{// (1)}§
    2924 void ?{}(S *, int); §\C{// (2)}§
    2925 void ?{}(S * double); §\C{// (3)}§
    2926 void ?{}(S *, S); §\C{// (4)}§
    2927 
    2928 [S, S] x = [3, 6.28]; §\C{// uses (2), (3), specialized constructors}§
    2929 [S, S] y; §\C{// uses (1), (1), default constructor}§
    2930 [S, S] z = x.0; §\C{// uses (4), (4), copy constructor}§
     2911void ?{}(S *); $\C{// (1)}$
     2912void ?{}(S *, int); $\C{// (2)}$
     2913void ?{}(S * double); $\C{// (3)}$
     2914void ?{}(S *, S); $\C{// (4)}$
     2915
     2916[S, S] x = [3, 6.28]; $\C{// uses (2), (3), specialized constructors}$
     2917[S, S] y; $\C{// uses (1), (1), default constructor}$
     2918[S, S] z = x.0; $\C{// uses (4), (4), copy constructor}$
    29312919\end{cfa}
    29322920In this example, ©x© is initialized by the multiple constructor calls ©?{}(&x.0, 3)© and ©?{}(&x.1, 6.28)©, while ©y© is initialized by two default constructor calls ©?{}(&y.0)© and ©?{}(&y.1)©.
     
    29692957A member-access tuple may be used anywhere a tuple can be used, \eg:
    29702958\begin{cfa}
    2971 s.[ y, z, x ] = [ 3, 3.2, 'x' ]; §\C{// equivalent to s.x = 'x', s.y = 3, s.z = 3.2}§
    2972 f( s.[ y, z ] ); §\C{// equivalent to f( s.y, s.z )}§
     2959s.[ y, z, x ] = [ 3, 3.2, 'x' ]; $\C{// equivalent to s.x = 'x', s.y = 3, s.z = 3.2}$
     2960f( s.[ y, z ] ); $\C{// equivalent to f( s.y, s.z )}$
    29732961\end{cfa}
    29742962Note, the fields appearing in a record-field tuple may be specified in any order;
     
    29802968void f( double, long );
    29812969
    2982 f( x.[ 0, 3 ] ); §\C{// f( x.0, x.3 )}§
    2983 x.[ 0, 1 ] = x.[ 1, 0 ]; §\C{// [ x.0, x.1 ] = [ x.1, x.0 ]}§
     2970f( x.[ 0, 3 ] ); $\C{// f( x.0, x.3 )}$
     2971x.[ 0, 1 ] = x.[ 1, 0 ]; $\C{// [ x.0, x.1 ] = [ x.1, x.0 ]}$
    29842972[ long, int, long ] y = x.[ 2, 0, 2 ];
    29852973\end{cfa}
     
    29982986\begin{cfa}
    29992987[ int, float, double ] f();
    3000 [ double, float ] x = f().[ 2, 1 ]; §\C{// f() called once}§
     2988[ double, float ] x = f().[ 2, 1 ]; $\C{// f() called once}$
    30012989\end{cfa}
    30022990
     
    30112999That is, a cast can be used to select the type of an expression when it is ambiguous, as in the call to an overloaded function.
    30123000\begin{cfa}
    3013 int f(); §\C{// (1)}§
    3014 double f(); §\C{// (2)}§
    3015 
    3016 f(); §\C{// ambiguous - (1),(2) both equally viable}§
    3017 (int)f(); §\C{// choose (2)}§
     3001int f(); $\C{// (1)}$
     3002double f(); $\C{// (2)}$
     3003
     3004f(); $\C{// ambiguous - (1),(2) both equally viable}$
     3005(int)f(); $\C{// choose (2)}$
    30183006\end{cfa}
    30193007Since casting is a fundamental operation in \CFA, casts need to be given a meaningful interpretation in the context of tuples.
     
    30233011void g();
    30243012
    3025 (void)f(); §\C{// valid, ignore results}§
    3026 (int)g(); §\C{// invalid, void cannot be converted to int}§
     3013(void)f(); $\C{// valid, ignore results}$
     3014(int)g(); $\C{// invalid, void cannot be converted to int}$
    30273015
    30283016struct A { int x; };
    3029 (struct A)f(); §\C{// invalid, int cannot be converted to A}§
     3017(struct A)f(); $\C{// invalid, int cannot be converted to A}$
    30303018\end{cfa}
    30313019In C, line 4 is a valid cast, which calls ©f© and discards its result.
     
    30433031        [int, [int, int], int] g();
    30443032
    3045         ([int, double])f(); §\C{// (1) valid}§
    3046         ([int, int, int])g(); §\C{// (2) valid}§
    3047         ([void, [int, int]])g(); §\C{// (3) valid}§
    3048         ([int, int, int, int])g(); §\C{// (4) invalid}§
    3049         ([int, [int, int, int]])g(); §\C{// (5) invalid}§
     3033        ([int, double])f(); $\C{// (1) valid}$
     3034        ([int, int, int])g(); $\C{// (2) valid}$
     3035        ([void, [int, int]])g(); $\C{// (3) valid}$
     3036        ([int, int, int, int])g(); $\C{// (4) invalid}$
     3037        ([int, [int, int, int]])g(); $\C{// (5) invalid}$
    30503038\end{cfa}
    30513039
     
    31073095void f([int, int], int, int);
    31083096
    3109 f([0, 0], 0, 0); §\C{// no cost}§
    3110 f(0, 0, 0, 0); §\C{// cost for structuring}§
    3111 f([0, 0,], [0, 0]); §\C{// cost for flattening}§
    3112 f([0, 0, 0], 0); §\C{// cost for flattening and structuring}§
     3097f([0, 0], 0, 0); $\C{// no cost}$
     3098f(0, 0, 0, 0); $\C{// cost for structuring}$
     3099f([0, 0,], [0, 0]); $\C{// cost for flattening}$
     3100f([0, 0, 0], 0); $\C{// cost for flattening and structuring}$
    31133101\end{cfa}
    31143102
     
    31463134The general syntax of a lexical list is:
    31473135\begin{cfa}
    3148 [ §\emph{exprlist}§ ]
     3136[ $\emph{exprlist}$ ]
    31493137\end{cfa}
    31503138where ©$\emph{exprlist}$© is a list of one or more expressions separated by commas.
     
    31583146Tuples are permitted to contain sub-tuples (\ie nesting), such as ©[ [ 14, 21 ], 9 ]©, which is a 2-element tuple whose first element is itself a tuple.
    31593147Note, a tuple is not a record (structure);
    3160 a record denotes a single value with substructure, whereas a tuple is multiple values with no substructure (see flattening coercion in Section 12.1).
     3148a record denotes a single value with substructure, whereas a tuple is multiple values with no substructure \see{flattening coercion in \VRef{s:FlatteningStructuring}}.
    31613149In essence, tuples are largely a compile time phenomenon, having little or no runtime presence.
    31623150
     
    31663154The general syntax of a tuple type is:
    31673155\begin{cfa}
    3168 [ §\emph{typelist}§ ]
     3156[ $\emph{typelist}$ ]
    31693157\end{cfa}
    31703158where ©$\emph{typelist}$© is a list of one or more legal \CFA or C type specifications separated by commas, which may include other tuple type specifications.
     
    31733161[ unsigned int, char ]
    31743162[ double, double, double ]
    3175 [ * int, int * ] §\C{// mix of CFA and ANSI}§
     3163[ * int, int * ] $\C{// mix of CFA and ANSI}$
    31763164[ * [ 5 ] int, * * char, * [ [ int, int ] ] (int, int) ]
    31773165\end{cfa}
     
    31803168Examples of declarations using tuple types are:
    31813169\begin{cfa}
    3182 [ int, int ] x; §\C{// 2 element tuple, each element of type int}§
    3183 * [ char, char ] y; §\C{// pointer to a 2 element tuple}§
     3170[ int, int ] x; $\C{// 2 element tuple, each element of type int}$
     3171* [ char, char ] y; $\C{// pointer to a 2 element tuple}$
    31843172[ [ int, int ] ] z ([ int, int ]);
    31853173\end{cfa}
     
    31983186[ int, int ] w1;
    31993187[ int, int, int ] w2;
    3200 [ void ] f (int, int, int); §\C{// three input parameters of type int}§
    3201 [ void ] g ([ int, int, int ]); §\C{3 element tuple as input}§
     3188[ void ] f (int, int, int); $\C{// three input parameters of type int}$
     3189[ void ] g ([ int, int, int ]); $\C{3 element tuple as input}$
    32023190f( [ 1, 2, 3 ] );
    32033191f( w1, 3 );
     
    32793267[ int, int, int, int ] w = [ 1, 2, 3, 4 ];
    32803268int x = 5;
    3281 [ x, w ] = [ w, x ]; §\C{// all four tuple coercions}§
     3269[ x, w ] = [ w, x ]; $\C{// all four tuple coercions}$
    32823270\end{cfa}
    32833271Starting on the right-hand tuple in the last assignment statement, w is opened, producing a tuple of four values;
     
    32853273This tuple is then flattened, yielding ©[ 1, 2, 3, 4, 5 ]©, which is structured into ©[ 1, [ 2, 3, 4, 5 ] ]© to match the tuple type of the left-hand side.
    32863274The tuple ©[ 2, 3, 4, 5 ]© is then closed to create a tuple value.
    3287 Finally, ©x© is assigned ©1© and ©w© is assigned the tuple value using multiple assignment (see Section 14).
     3275Finally, ©x© is assigned ©1© and ©w© is assigned the tuple value using \Index{multiple assignment} \see{\VRef{s:TupleAssignment}}.
    32883276\begin{rationale}
    32893277A possible additional language extension is to use the structuring coercion for tuples to initialize a complex record with a tuple.
     
    32963284Mass assignment has the following form:
    32973285\begin{cfa}
    3298 [ §\emph{lvalue}§, ... , §\emph{lvalue}§ ] = §\emph{expr}§;
     3286[ $\emph{lvalue}$, ... , $\emph{lvalue}$ ] = $\emph{expr}$;
    32993287\end{cfa}
    33003288\index{lvalue}
     
    33363324Multiple assignment has the following form:
    33373325\begin{cfa}
    3338 [ §\emph{lvalue}§, ... , §\emph{lvalue}§ ] = [ §\emph{expr}§, ... , §\emph{expr}§ ];
     3326[ $\emph{lvalue}$, ... , $\emph{lvalue}$ ] = [ $\emph{expr}$, ... , $\emph{expr}$ ];
    33393327\end{cfa}
    33403328\index{lvalue}
     
    33673355both these examples produce indeterminate results:
    33683356\begin{cfa}
    3369 f( x++, x++ ); §\C{// C routine call with side effects in arguments}§
    3370 [ v1, v2 ] = [ x++, x++ ]; §\C{// side effects in righthand side of multiple assignment}§
     3357f( x++, x++ ); $\C{// C routine call with side effects in arguments}$
     3358[ v1, v2 ] = [ x++, x++ ]; $\C{// side effects in right-hand side of multiple assignment}$
    33713359\end{cfa}
    33723360
     
    33773365Cascade assignment has the following form:
    33783366\begin{cfa}
    3379 §\emph{tuple}§ = §\emph{tuple}§ = ... = §\emph{tuple}§;
     3367$\emph{tuple}$ = $\emph{tuple}$ = ... = $\emph{tuple}$;
    33803368\end{cfa}
    33813369and it has the same parallel semantics as for mass and multiple assignment.
     
    34243412\begin{cfa}
    34253413int x = 1, y = 2, z = 3;
    3426 sout | x ®|® y ®|® z;
     3414sout | x @|@ y @|@ z;
    34273415\end{cfa}
    34283416&
    34293417\begin{cfa}
    34303418
    3431 cout << x ®<< " "® << y ®<< " "® << z << endl;
     3419cout << x @<< " "@ << y @<< " "@ << z << endl;
    34323420\end{cfa}
    34333421&
     
    34383426\\
    34393427\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3440 1® ®2® ®3
     34281@ @2@ @3
    34413429\end{cfa}
    34423430&
    34433431\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3444 1® ®2® ®3
     34321@ @2@ @3
    34453433\end{cfa}
    34463434&
    34473435\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3448 1® ®2® ®3
     34361@ @2@ @3
    34493437\end{cfa}
    34503438\end{tabular}
     
    34543442\begin{cfa}
    34553443[int, [ int, int ] ] t1 = [ 1, [ 2, 3 ] ], t2 = [ 4, [ 5, 6 ] ];
    3456 sout | t1 | t2; §\C{// print tuples}§
     3444sout | t1 | t2; $\C{// print tuples}$
    34573445\end{cfa}
    34583446\begin{cfa}[showspaces=true,aboveskip=0pt]
    3459 1®, ®2®, ®3 4®, ®5®, ®6
     34471@, @2@, @3 4@, @5@, @6
    34603448\end{cfa}
    34613449Finally, \CFA uses the logical-or operator for I/O as it is the lowest-priority \emph{overloadable} operator, other than assignment.
     
    34663454&
    34673455\begin{cfa}
    3468 sout | x * 3 | y + 1 | z << 2 | x == y | ®(®x | y®)® | ®(®x || y®)® | ®(®x > z ? 1 : 2®)®;
     3456sout | x * 3 | y + 1 | z << 2 | x == y | @(@x | y@)@ | @(@x || y@)@ | @(@x > z ? 1 : 2@)@;
    34693457\end{cfa}
    34703458\\
     
    34723460&
    34733461\begin{cfa}
    3474 cout << x * 3 << y + 1 << ®(®z << 2®)® << ®(®x == y®)® << ®(®x | y®)® << ®(®x || y®)® << ®(®x > z ? 1 : 2®)® << endl;
     3462cout << x * 3 << y + 1 << @(@z << 2@)@ << @(@x == y@)@ << @(@x | y@)@ << @(@x || y@)@ << @(@x > z ? 1 : 2@)@ << endl;
    34753463\end{cfa}
    34763464\\
     
    35073495\\
    35083496\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3509 ®1® ®2.5® ®A®
     3497@1@ @2.5@ @A@
    35103498
    35113499
     
    35133501&
    35143502\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3515 ®1® ®2.5® ®A®
     3503@1@ @2.5@ @A@
    35163504
    35173505
     
    35193507&
    35203508\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3521 ®1®
    3522 ®2.5®
    3523 ®A®
     3509@1@
     3510@2.5@
     3511@A@
    35243512\end{cfa}
    35253513\end{tabular}
     
    35573545
    35583546\item
    3559 {\lstset{language=CFA,deletedelim=**[is][]{¢}{¢}}
    3560 A separator does not appear before a C string starting with the (extended) \Index*{ASCII}\index{ASCII!extended} characters: \lstinline[basicstyle=\tt]@,.;!?)]}%¢»@, where \lstinline[basicstyle=\tt]@»@ is a closing citation mark.
    3561 \begin{cfa}[belowskip=0pt]
     3547A separator does not appear before a C string starting with the (extended) \Index*{ASCII}\index{ASCII!extended} characters: \LstStringStyle{,.;!?)]\}\%\textcent\guillemotright}, where \LstStringStyle{\guillemotright} a closing citation mark.
     3548\begin{cfa}
    35623549sout | 1 | ", x" | 2 | ". x" | 3 | "; x" | 4 | "! x" | 5 | "? x" | 6 | "% x"
    3563                 | 7 | "¢ x" | 8 | "» x" | 9 | ") x" | 10 | "] x" | 11 | "} x";
    3564 \end{cfa}
    3565 \begin{cfa}[basicstyle=\tt,showspaces=true,aboveskip=0pt,belowskip=0pt]
    3566 1®,® x 2®.® x 3®;® x 4®!® x 5®?® x 6®%® x 7§\color{red}\textcent§ x 8®»® x 9®)® x 10®]® x 11®}® x
    3567 \end{cfa}}%
    3568 
    3569 \item
    3570 A separator does not appear after a C string ending with the (extended) \Index*{ASCII}\index{ASCII!extended} characters: \lstinline[mathescape=off,basicstyle=\tt]@([{=$£¥¡¿«@, where \lstinline[basicstyle=\tt]@¡¿@ are inverted opening exclamation and question marks, and \lstinline[basicstyle=\tt]@«@ is an opening citation mark.
     3550           | 7 | "$\LstStringStyle{\textcent}$ x" | 8 | "$\LstStringStyle{\guillemotright}$ x" | 9 | ") x" | 10 | "] x" | 11 | "} x";
     3551\end{cfa}
     3552\begin{cfa}[showspaces=true]
     35531@,@ x 2@.@ x 3@;@ x 4@!@ x 5@?@ x 6@%@ x 7$\R{\LstStringStyle{\textcent}}$ x 8$\R{\LstStringStyle{\guillemotright}}$ x 9@)@ x 10@]@ x 11@}@ x
     3554\end{cfa}
     3555
     3556\item
     3557A separator does not appear after a C string ending with the (extended) \Index*{ASCII}\index{ASCII!extended} characters: \LstStringStyle{([\{=\$\textsterling\textyen\textexclamdown\textquestiondown\guillemotleft}, where \LstStringStyle{\textexclamdown\textquestiondown} are inverted opening exclamation and question marks, and \LstStringStyle{\guillemotleft} is an opening citation mark.
    35713558%$
    3572 \begin{cfa}[mathescape=off]
    3573 sout | "x (" | 1 | "x [" | 2 | "x {" | 3 | "x =" | 4 | "x $" | 5 | "x £" | 6 | "x ¥"
    3574                 | 7 | "x ¡" | 8 | "x ¿" | 9 | "x «" | 10;
     3559\begin{cfa}
     3560sout | "x (" | 1 | "x [" | 2 | "x {" | 3 | "x =" | 4 | "x $" | 5 | "x $\LstStringStyle{\textsterling}$" | 6 | "x $\LstStringStyle{\textyen}$"
     3561           | 7 | "x $\LstStringStyle{\textexclamdown}$" | 8 | "x $\LstStringStyle{\textquestiondown}$" | 9 | "x $\LstStringStyle{\guillemotleft}$" | 10;
    35753562\end{cfa}
    35763563%$
    3577 \begin{cfa}[mathescape=off,basicstyle=\tt,showspaces=true,aboveskip=0pt,belowskip=0pt]
    3578 x ®(®1 x ®[®2 x ®{®3 x ®=®4 x ®$®5 x ®£®6 x ®¥®7 x ®¡®8 x ®¿®9 x ®«®10
     3564\begin{cfa}[showspaces=true]
     3565x @(@1 x @[@2 x @{@3 x @=@4 x $\LstStringStyle{\textdollar}$5 x $\R{\LstStringStyle{\textsterling}}$6 x $\R{\LstStringStyle{\textyen}}$7 x $\R{\LstStringStyle{\textexclamdown}}$8 x $\R{\LstStringStyle{\textquestiondown}}$9 x $\R{\LstStringStyle{\guillemotleft}}$10
    35793566\end{cfa}
    35803567%$
    35813568
    35823569\item
    3583 A seperator does not appear before/after a C string starting/ending with the \Index*{ASCII} quote or whitespace characters: \lstinline[basicstyle=\tt,showspaces=true]@`'": \t\v\f\r\n@
    3584 \begin{cfa}[belowskip=0pt]
     3570A seperator does not appear before/after a C string starting/ending with the \Index*{ASCII} quote or whitespace characters: \lstinline[basicstyle=\tt,showspaces=true]{`'": \t\v\f\r\n}
     3571\begin{cfa}
    35853572sout | "x`" | 1 | "`x'" | 2 | "'x\"" | 3 | "\"x:" | 4 | ":x " | 5 | " x\t" | 6 | "\tx";
    35863573\end{cfa}
    3587 \begin{cfa}[basicstyle=\tt,showspaces=true,showtabs=true,aboveskip=0pt,belowskip=0pt]
    3588 x®`®1®`®x§\color{red}\texttt{'}§2§\color{red}\texttt{'}§x§\color{red}\texttt{"}§3§\color{red}\texttt{"}§x®:®4®:®x® ®5® ®x®      ®6®     ®x
     3574\begin{cfa}[showspaces=true,showtabs=true]
     3575x@`@1@`@x$\R{\texttt{'}}$2$\R{\texttt{'}}$x$\R{\texttt{"}}$3$\R{\texttt{"}}$x@:@4@:@x@ @5@ @x@  @6@     @x
    35893576\end{cfa}
    35903577
    35913578\item
    35923579If a space is desired before or after one of the special string start/end characters, simply insert a space.
    3593 \begin{cfa}[belowskip=0pt]
    3594 sout | "x (§\color{red}\texttt{\textvisiblespace}§" | 1 | "§\color{red}\texttt{\textvisiblespace}§) x" | 2 | "§\color{red}\texttt{\textvisiblespace}§, x" | 3 | "§\color{red}\texttt{\textvisiblespace}§:x:§\color{red}\texttt{\textvisiblespace}§" | 4;
    3595 \end{cfa}
    3596 \begin{cfa}[basicstyle=\tt,showspaces=true,showtabs=true,aboveskip=0pt,belowskip=0pt]
    3597 x (® ®1® ®) x 2® ®, x 3® ®:x:® ®4
     3580\begin{cfa}
     3581sout | "x ($\R{\texttt{\textvisiblespace}}$" | 1 | "$\R{\texttt{\textvisiblespace}}$) x" | 2 | "$\R{\texttt{\textvisiblespace}}$, x" | 3 | "$\R{\texttt{\textvisiblespace}}$:x:$\R{\texttt{\textvisiblespace}}$" | 4;
     3582\end{cfa}
     3583\begin{cfa}[showspaces=true,showtabs=true]
     3584x (@ @1@ @) x 2@ @, x 3@ @:x:@ @4
    35983585\end{cfa}
    35993586\end{enumerate}
     
    36083595\Indexc{sepSet}\index{manipulator!sepSet@©sepSet©} and \Indexc{sep}\index{manipulator!sep@©sep©}/\Indexc{sepGet}\index{manipulator!sepGet@©sepGet©} set and get the separator string.
    36093596The separator string can be at most 16 characters including the ©'\0'© string terminator (15 printable characters).
    3610 \begin{cfa}[mathescape=off,belowskip=0pt]
    3611 sepSet( sout, ", $" ); §\C{// set separator from " " to ", \$"}§
    3612 sout | 1 | 2 | 3 | " \"" | ®sep® | "\"";
     3597\begin{cfa}[escapechar=off,belowskip=0pt]
     3598sepSet( sout, ", $" ); $\C{// set separator from " " to ", \$"}$
     3599sout | 1 | 2 | 3 | " \"" | @sep@ | "\"";
    36133600\end{cfa}
    36143601%$
    36153602\begin{cfa}[mathescape=off,showspaces=true,aboveskip=0pt]
    3616 1®, $®2®, $®3 ®", $"®
     36031@, $@2@, $@3 @", $"@
    36173604\end{cfa}
    36183605%$
    36193606\begin{cfa}[belowskip=0pt]
    3620 sepSet( sout, " " ); §\C{// reset separator to " "}§
    3621 sout | 1 | 2 | 3 | " \"" | ®sepGet( sout )® | "\"";
     3607sepSet( sout, " " ); $\C{// reset separator to " "}$
     3608sout | 1 | 2 | 3 | " \"" | @sepGet( sout )@ | "\"";
    36223609\end{cfa}
    36233610\begin{cfa}[showspaces=true,aboveskip=0pt]
    3624 1® ®2® ®3 ®" "®
     36111@ @2@ @3 @" "@
    36253612\end{cfa}
    36263613©sepGet© can be used to store a separator and then restore it:
    36273614\begin{cfa}[belowskip=0pt]
    3628 char store[®sepSize®]; §\C{// sepSize is the maximum separator size}§
    3629 strcpy( store, sepGet( sout ) ); §\C{// copy current separator}§
    3630 sepSet( sout, "_" ); §\C{// change separator to underscore}§
     3615char store[@sepSize@]; $\C{// sepSize is the maximum separator size}$
     3616strcpy( store, sepGet( sout ) ); $\C{// copy current separator}$
     3617sepSet( sout, "_" ); $\C{// change separator to underscore}$
    36313618sout | 1 | 2 | 3;
    36323619\end{cfa}
    36333620\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3634 1®_®2®_®3
     36211@_@2@_@3
    36353622\end{cfa}
    36363623\begin{cfa}[belowskip=0pt]
    3637 sepSet( sout, store ); §\C{// change separator back to original}§
     3624sepSet( sout, store ); $\C{// change separator back to original}$
    36383625sout | 1 | 2 | 3;
    36393626\end{cfa}
    36403627\begin{cfa}[showspaces=true,aboveskip=0pt]
    3641 1® ®2® ®3
     36281@ @2@ @3
    36423629\end{cfa}
    36433630
     
    36463633The tuple separator-string can be at most 16 characters including the ©'\0'© string terminator (15 printable characters).
    36473634\begin{cfa}[belowskip=0pt]
    3648 sepSetTuple( sout, " " ); §\C{// set tuple separator from ", " to " "}§
    3649 sout | t1 | t2 | " \"" | ®sepTuple® | "\"";
     3635sepSetTuple( sout, " " ); $\C{// set tuple separator from ", " to " "}$
     3636sout | t1 | t2 | " \"" | @sepTuple@ | "\"";
    36503637\end{cfa}
    36513638\begin{cfa}[showspaces=true,aboveskip=0pt]
    3652 1 2 3 4 5 6 ®" "®
     36391 2 3 4 5 6 @" "@
    36533640\end{cfa}
    36543641\begin{cfa}[belowskip=0pt]
    3655 sepSetTuple( sout, ", " ); §\C{// reset tuple separator to ", "}§
    3656 sout | t1 | t2 | " \"" | ®sepGetTuple( sout )® | "\"";
     3642sepSetTuple( sout, ", " ); $\C{// reset tuple separator to ", "}$
     3643sout | t1 | t2 | " \"" | @sepGetTuple( sout )@ | "\"";
    36573644\end{cfa}
    36583645\begin{cfa}[showspaces=true,aboveskip=0pt]
    3659 1, 2, 3 4, 5, 6 ®", "®
     36461, 2, 3 4, 5, 6 @", "@
    36603647\end{cfa}
    36613648As for ©sepGet©, ©sepGetTuple© can be use to store a tuple separator and then restore it.
     
    36643651\Indexc{sepDisable}\index{manipulator!sepDisable@©sepDisable©} and \Indexc{sepEnable}\index{manipulator!sepEnable@©sepEnable©} toggle printing the separator.
    36653652\begin{cfa}[belowskip=0pt]
    3666 sout | sepDisable | 1 | 2 | 3; §\C{// turn off implicit separator}§
     3653sout | sepDisable | 1 | 2 | 3; $\C{// turn off implicit separator}$
    36673654\end{cfa}
    36683655\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
     
    36703657\end{cfa}
    36713658\begin{cfa}[belowskip=0pt]
    3672 sout | sepEnable | 1 | 2 | 3; §\C{// turn on implicit separator}§
     3659sout | sepEnable | 1 | 2 | 3; $\C{// turn on implicit separator}$
    36733660\end{cfa}
    36743661\begin{cfa}[mathescape=off,showspaces=true,aboveskip=0pt,belowskip=0pt]
     
    36793666\Indexc{sepOn}\index{manipulator!sepOn@©sepOn©} and \Indexc{sepOff}\index{manipulator!sepOff@©sepOff©} toggle printing the separator with respect to the next printed item, and then return to the global seperator setting.
    36803667\begin{cfa}[belowskip=0pt]
    3681 sout | 1 | sepOff | 2 | 3; §\C{// turn off implicit separator for the next item}§
     3668sout | 1 | sepOff | 2 | 3; $\C{// turn off implicit separator for the next item}$
    36823669\end{cfa}
    36833670\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
     
    36853672\end{cfa}
    36863673\begin{cfa}[belowskip=0pt]
    3687 sout | sepDisable | 1 | sepOn | 2 | 3; §\C{// turn on implicit separator for the next item}§
     3674sout | sepDisable | 1 | sepOn | 2 | 3; $\C{// turn on implicit separator for the next item}$
    36883675\end{cfa}
    36893676\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
     
    36923679The tuple separator also responses to being turned on and off.
    36933680\begin{cfa}[belowskip=0pt]
    3694 sout | t1 | sepOff | t2; §\C{// turn off implicit separator for the next item}§
     3681sout | t1 | sepOff | t2; $\C{// turn off implicit separator for the next item}$
    36953682\end{cfa}
    36963683\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
     
    37003687use ©sep© to accomplish this functionality.
    37013688\begin{cfa}[belowskip=0pt]
    3702 sout | sepOn | 1 | 2 | 3 | sepOn; §\C{// sepOn does nothing at start/end of line}§
     3689sout | sepOn | 1 | 2 | 3 | sepOn; $\C{// sepOn does nothing at start/end of line}$
    37033690\end{cfa}
    37043691\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
     
    37063693\end{cfa}
    37073694\begin{cfa}[belowskip=0pt]
    3708 sout | sep | 1 | 2 | 3 | sep ; §\C{// use sep to print separator at start/end of line}§
     3695sout | sep | 1 | 2 | 3 | sep ; $\C{// use sep to print separator at start/end of line}$
    37093696\end{cfa}
    37103697\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3711 ® ®1 2 3® ®
     3698@ @1 2 3@ @
    37123699\end{cfa}
    37133700\end{enumerate}
     
    37213708\begin{enumerate}[parsep=0pt]
    37223709\item
    3723 \Indexc{nl}\index{manipulator!nl@©nl©} scans characters until the next newline character, i.e., ignore the remaining characters in the line.
     3710\Indexc{nl}\index{manipulator!nl@©nl©} scans characters until the next newline character, \ie ignore the remaining characters in the line.
    37243711\item
    37253712\Indexc{nlOn}\index{manipulator!nlOn@©nlOn©} reads the newline character, when reading single characters.
     
    37293716For example, in:
    37303717\begin{cfa}
    3731 sin | i | ®nl® | j;
    3732 1 ®2®
     3718sin | i | @nl@ | j;
     37191 @2@
    373337203
    37343721\end{cfa}
     
    37403727\Indexc{nl}\index{manipulator!nl@©nl©} inserts a newline.
    37413728\begin{cfa}
    3742 sout | nl; §\C{// only print newline}§
    3743 sout | 2; §\C{// implicit newline}§
    3744 sout | 3 | nl | 4 | nl; §\C{// terminating nl merged with implicit newline}§
    3745 sout | 5 | nl | nl; §\C{// again terminating nl merged with implicit newline}§
    3746 sout | 6; §\C{// implicit newline}§
     3729sout | nl; $\C{// only print newline}$
     3730sout | 2; $\C{// implicit newline}$
     3731sout | 3 | nl | 4 | nl; $\C{// terminating nl merged with implicit newline}$
     3732sout | 5 | nl | nl; $\C{// again terminating nl merged with implicit newline}$
     3733sout | 6; $\C{// implicit newline}$
    37473734
    374837352
     
    377137580b0 0b11011 0b11011 0b11011 0b11011
    37723759sout | bin( -27HH ) | bin( -27H ) | bin( -27 ) | bin( -27L );
    3773 0b11100101 0b1111111111100101 0b11111111111111111111111111100101 0b®(58 1s)®100101
     37600b11100101 0b1111111111100101 0b11111111111111111111111111100101 0b@(58 1s)@100101
    37743761\end{cfa}
    37753762
     
    38103797\begin{cfa}[belowskip=0pt]
    38113798sout | upcase( bin( 27 ) ) | upcase( hex( 27 ) ) | upcase( 27.5e-10 ) | upcase( hex( 27.5 ) );
    3812 0®B®11011 0®X®1®B® 2.75®E®-09 0®X®1.®B®8®P®+4
     37990@B@11011 0@X@1@B@ 2.75@E@-09 0@X@1.@B@8@P@+4
    38133800\end{cfa}
    38143801
     
    38263813\begin{cfa}[belowskip=0pt]
    38273814sout | 0. | nodp( 0. ) | 27.0 | nodp( 27.0 ) | nodp( 27.5 );
    3828 0.0 ®0® 27.0 ®27® 27.5
     38150.0 @0@ 27.0 @27@ 27.5
    38293816\end{cfa}
    38303817
     
    38333820\begin{cfa}[belowskip=0pt]
    38343821sout | sign( 27 ) | sign( -27 ) | sign( 27. ) | sign( -27. ) | sign( 27.5 ) | sign( -27.5 );
    3835 ®+®27 -27 ®+®27.0 -27.0 ®+®27.5 -27.5
     3822@+@27 -27 @+@27.0 -27.0 @+@27.5 -27.5
    38363823\end{cfa}
    38373824
     
    38463833\end{cfa}
    38473834\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3848 ®  ®34 ® ®34 34
    3849 ®  ®4.000000 ® ®4.000000 4.000000
    3850 ®  ®ab ® ®ab ab
     3835@  @34 @ @34 34
     3836@  @4.000000 @ @4.000000 4.000000
     3837@  @ab @ @ab ab
    38513838\end{cfa}
    38523839If the value is larger, it is printed without truncation, ignoring the ©minimum©.
     
    38573844\end{cfa}
    38583845\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3859 3456®7® 345®67® 34®567®
    3860 3456®.® 345®6.® 34®56.®
    3861 abcd®e® abc®de® ab®cde®
     38463456@7@ 345@67@ 34@567@
     38473456@.@ 345@6.@ 34@56.@
     3848abcd@e@ abc@de@ ab@cde@
    38623849\end{cfa}
    38633850
     
    38683855\end{cfa}
    38693856\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3870  ®0®34     ®00®34 ®00000000®34
     3857 @0@34     @00@34 @00000000@34
    38713858\end{cfa}
    38723859If the value is larger, it is printed without truncation, ignoring the ©precision©.
     
    38833870\end{cfa}
    38843871\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3885 ®    ® ®00000000®34
     3872@    @ @00000000@34
    38863873\end{cfa}
    38873874For floating-point types, ©precision© is the minimum number of digits after the decimal point.
     
    38903877\end{cfa}
    38913878\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3892 27.®500®     27.®5®      28. 27.®50000000®
    3893 \end{cfa}
    3894 For the C-string type, ©precision© is the maximum number of printed characters, so the string is truncared if it exceeds the maximum.
     387927.@500@     27.@5@      28. 27.@50000000@
     3880\end{cfa}
     3881For the C-string type, ©precision© is the maximum number of printed characters, so the string is truncated if it exceeds the maximum.
    38953882\begin{cfa}[belowskip=0pt]
    38963883sout | wd( 6,8, "abcd" ) | wd( 6,8, "abcdefghijk" ) | wd( 6,3, "abcd" );
     
    39083895\end{cfa}
    39093896\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3910 234.567 234.5®7®  234.®6®    23®5®
     3897234.567 234.5@7@  234.@6@    23@5@
    39113898\end{cfa}
    39123899If a value's magnitude is greater than ©significant©, the value is printed in scientific notation with the specified number of significant digits.
     
    39153902\end{cfa}
    39163903\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3917 234567. 2.3457®e+05® 2.346®e+05® 2.35®e+05®
     3904234567. 2.3457@e+05@ 2.346@e+05@ 2.35@e+05@
    39183905\end{cfa}
    39193906If ©significant© is greater than ©minimum©, it defines the number of printed characters.
     
    39313918\end{cfa}
    39323919\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    3933 27®  ® 27.000000  27.500000  027  27.500®    ®
     392027@  @ 27.000000  27.500000  027  27.500@    @
    39343921\end{cfa}
    39353922
     
    39383925\begin{cfa}[belowskip=0pt]
    39393926sout | pad0( wd( 4, 27 ) ) | pad0( wd( 4,3, 27 ) ) | pad0( wd( 8,3, 27.5 ) );
    3940 ®00®27  ®0®27 ®00®27.500
     3927@00@27  @0@27 @00@27.500
    39413928\end{cfa}
    39423929\end{enumerate}
     
    40344021\end{cfa}
    40354022\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    4036 ®abc   ®
    4037 ®abc  ®
    4038 ®xx®
     4023@abc   @
     4024@abc  @
     4025@xx@
    40394026\end{cfa}
    40404027
     
    40474034\end{cfa}
    40484035\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    4049 ®abcd1233.456E+2®
     4036@abcd1233.456E+2@
    40504037\end{cfa}
    40514038Note, input ©wdi© cannot be overloaded with output ©wd© because both have the same parameters but return different types.
     
    40604047\end{cfa}
    40614048\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    4062 ®  -75.35e-4® 25
     4049@  -75.35e-4@ 25
    40634050\end{cfa}
    40644051
     
    40724059\end{cfa}
    40734060\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    4074 ®bca®xyz
     4061@bca@xyz
    40754062\end{cfa}
    40764063
     
    40844071\end{cfa}
    40854072\begin{cfa}[showspaces=true,aboveskip=0pt,belowskip=0pt]
    4086 ®xyz®bca
     4073@xyz@bca
    40874074\end{cfa}
    40884075\end{enumerate}
     
    41014088
    41024089A type definition is different from a typedef in C because a typedef just creates an alias for a type,  while Do.s type definition creates a distinct type.
    4103 This means that users can define distinct function overloads for the new type (see Overloading for more information).
     4090This means that users can define distinct function overloads for the new type \see{\VRef{s:Overloading} for more information}.
    41044091For example:
    41054092
     
    42074194\CFA supports C initialization of structures, but it also adds constructors for more advanced initialization.
    42084195Additionally, \CFA adds destructors that are called when a variable is deallocated (variable goes out of scope or object is deleted).
    4209 These functions take a reference to the structure as a parameter (see References for more information).
     4196These functions take a reference to the structure as a parameter \see{\VRef{s:PointerReference} for more information}.
    42104197
    42114198\begin{figure}
     
    42584245
    42594246\section{Overloading}
     4247\label{s:Overloading}
    42604248
    42614249Overloading refers to the capability of a programmer to define and use multiple objects in a program with the same name.
     
    42904278
    42914279
    4292 \subsection{Overloaded Constant}
     4280\subsection{Constant}
    42934281
    42944282The constants 0 and 1 have special meaning.
     
    43294317
    43304318
    4331 \subsection{Variable Overloading}
     4319\subsection{Variable}
     4320\label{s:VariableOverload}
    43324321
    43334322The overload rules of \CFA allow a programmer to define multiple variables with the same name, but different types.
     
    43724361
    43734362
    4374 \subsection{Operator Overloading}
     4363\subsection{Operator}
    43754364
    43764365\CFA also allows operators to be overloaded, to simplify the use of user-defined types.
     
    44684457For example, given
    44694458\begin{cfa}
    4470 auto j = ®...®
     4459auto j = @...@
    44714460\end{cfa}
    44724461and the need to write a routine to compute using ©j©
    44734462\begin{cfa}
    4474 void rtn( ®...® parm );
     4463void rtn( @...@ parm );
    44754464rtn( j );
    44764465\end{cfa}
     
    47134702
    47144703coroutine Fibonacci {
    4715         int fn; §\C{// used for communication}§
     4704        int fn; $\C{// used for communication}$
    47164705};
    47174706void ?{}( Fibonacci * this ) {
     
    47194708}
    47204709void main( Fibonacci * this ) {
    4721         int fn1, fn2; §\C{// retained between resumes}§
    4722         this->fn = 0; §\C{// case 0}§
     4710        int fn1, fn2; $\C{// retained between resumes}$
     4711        this->fn = 0; $\C{// case 0}$
    47234712        fn1 = this->fn;
    4724         suspend(); §\C{// return to last resume}§
    4725 
    4726         this->fn = 1; §\C{// case 1}§
     4713        suspend(); $\C{// return to last resume}$
     4714
     4715        this->fn = 1; $\C{// case 1}$
    47274716        fn2 = fn1;
    47284717        fn1 = this->fn;
    4729         suspend(); §\C{// return to last resume}§
    4730 
    4731         for ( ;; ) { §\C{// general case}§
     4718        suspend(); $\C{// return to last resume}$
     4719
     4720        for ( ;; ) { $\C{// general case}$
    47324721                this->fn = fn1 + fn2;
    47334722                fn2 = fn1;
    47344723                fn1 = this->fn;
    4735                 suspend(); §\C{// return to last resume}§
     4724                suspend(); $\C{// return to last resume}$
    47364725        } // for
    47374726}
    47384727int next( Fibonacci * this ) {
    4739         resume( this ); §\C{// transfer to last suspend}§
     4728        resume( this ); $\C{// transfer to last suspend}$
    47404729        return this->fn;
    47414730}
     
    49644953When building a \CFA module which needs to be callable from C code, users can use the tools to generate a header file suitable for including in these C files with all of the needed declarations.
    49654954
    4966 In order to interoperate with existing C code, \CFA files can still include header files, the contents of which will be enclosed in a C linkage section to indicate C calling conventions (see Interoperability for more information).
     4955In order to interoperate with existing C code, \CFA files can still include header files, the contents of which will be enclosed in a C linkage section to indicate C calling conventions \see{\VRef{s:Interoperability} for more information}.
    49674956
    49684957
     
    56305619\end{cfa}
    56315620&
    5632 \begin{lstlisting}[language=C++]
     5621\begin{C++}
    56335622class Line {
    56345623        float lnth;
     
    56575646Line line1;
    56585647Line line2( 3.4 );
    5659 \end{lstlisting}
     5648\end{C++}
    56605649&
    56615650\begin{lstlisting}[language=Golang]
     
    62826271In \CFA, there are ambiguous cases with dereference and operator identifiers, \eg ©int *?*?()©, where the string ©*?*?© can be interpreted as:
    62836272\begin{cfa}
    6284 *?§\color{red}\textvisiblespace§*? §\C{// dereference operator, dereference operator}§
    6285 *§\color{red}\textvisiblespace§?*? §\C{// dereference, multiplication operator}§
     6273*?$\R{\textvisiblespace}$*? $\C{// dereference operator, dereference operator}$
     6274*$\R{\textvisiblespace}$?*? $\C{// dereference, multiplication operator}$
    62866275\end{cfa}
    62876276By default, the first interpretation is selected, which does not yield a meaningful parse.
     
    62926281The ambiguity occurs when the deference operator has no parameters:
    62936282\begin{cfa}
    6294 *?()§\color{red}\textvisiblespace...§ ;
    6295 *?()§\color{red}\textvisiblespace...§(...) ;
     6283*?()$\R{\textvisiblespace...}$ ;
     6284*?()$\R{\textvisiblespace...}$(...) ;
    62966285\end{cfa}
    62976286requiring arbitrary whitespace look-ahead for the routine-call parameter-list to disambiguate.
     
    63016290The remaining cases are with the increment/decrement operators and conditional expression, \eg:
    63026291\begin{cfa}
    6303 i++?§\color{red}\textvisiblespace...§(...);
    6304 i?++§\color{red}\textvisiblespace...§(...);
     6292i++?$\R{\textvisiblespace...}$(...);
     6293i?++$\R{\textvisiblespace...}$(...);
    63056294\end{cfa}
    63066295requiring arbitrary whitespace look-ahead for the operator parameter-list, even though that interpretation is an incorrect expression (juxtaposed identifiers).
    63076296Therefore, it is necessary to disambiguate these cases with a space:
    63086297\begin{cfa}
    6309 i++§\color{red}\textvisiblespace§? i : 0;
    6310 i?§\color{red}\textvisiblespace§++i : 0;
     6298i++$\R{\textvisiblespace}$? i : 0;
     6299i?$\R{\textvisiblespace}$++i : 0;
    63116300\end{cfa}
    63126301
     
    63216310\begin{description}
    63226311\item[Change:] add new keywords \\
    6323 New keywords are added to \CFA (see~\VRef{s:CFAKeywords}).
     6312New keywords are added to \CFA \see{\VRef{s:CFAKeywords}}.
    63246313\item[Rationale:] keywords added to implement new semantics of \CFA.
    63256314\item[Effect on original feature:] change to semantics of well-defined feature. \\
    63266315Any \Celeven programs using these keywords as identifiers are invalid \CFA programs.
    6327 \item[Difficulty of converting:] keyword clashes are accommodated by syntactic transformations using the \CFA backquote escape-mechanism (see~\VRef{s:BackquoteIdentifiers}).
     6316\item[Difficulty of converting:] keyword clashes are accommodated by syntactic transformations using the \CFA backquote escape-mechanism \see{\VRef{s:BackquoteIdentifiers}}.
    63286317\item[How widely used:] clashes among new \CFA keywords and existing identifiers are rare.
    63296318\end{description}
     
    63356324\eg:
    63366325\begin{cfa}
    6337 x; §\C{// int x}§
    6338 *y; §\C{// int *y}§
    6339 f( p1, p2 ); §\C{// int f( int p1, int p2 );}§
    6340 g( p1, p2 ) int p1, p2; §\C{// int g( int p1, int p2 );}§
     6326x; $\C{// int x}$
     6327*y; $\C{// int *y}$
     6328f( p1, p2 ); $\C{// int f( int p1, int p2 );}$
     6329g( p1, p2 ) int p1, p2; $\C{// int g( int p1, int p2 );}$
    63416330\end{cfa}
    63426331\CFA continues to support K\&R routine definitions:
    63436332\begin{cfa}
    6344 f( a, b, c ) §\C{// default int return}§
    6345         int a, b; char c §\C{// K\&R parameter declarations}§
     6333f( a, b, c ) $\C{// default int return}$
     6334        int a, b; char c $\C{// K\&R parameter declarations}$
    63466335{
    63476336        ...
     
    63626351int rtn( int i );
    63636352int rtn( char c );
    6364 rtn( 'x' ); §\C{// programmer expects 2nd rtn to be called}§
     6353rtn( 'x' ); $\C{// programmer expects 2nd rtn to be called}$
    63656354\end{cfa}
    63666355\item[Rationale:] it is more intuitive for the call to ©rtn© to match the second version of definition of ©rtn© rather than the first.
     
    63846373\item[Change:] make string literals ©const©:
    63856374\begin{cfa}
    6386 char * p = "abc"; §\C{// valid in C, deprecated in \CFA}§
    6387 char * q = expr ? "abc" : "de"; §\C{// valid in C, invalid in \CFA}§
     6375char * p = "abc"; $\C{// valid in C, deprecated in \CFA}$
     6376char * q = expr ? "abc" : "de"; $\C{// valid in C, invalid in \CFA}$
    63886377\end{cfa}
    63896378The type of a string literal is changed from ©[] char© to ©const [] char©.
     
    63926381\begin{cfa}
    63936382char * p = "abc";
    6394 p[0] = 'w'; §\C{// segment fault or change constant literal}§
     6383p[0] = 'w'; $\C{// segment fault or change constant literal}$
    63956384\end{cfa}
    63966385The same problem occurs when passing a string literal to a routine that changes its argument.
     
    64046393\item[Change:] remove \newterm{tentative definitions}, which only occurs at file scope:
    64056394\begin{cfa}
    6406 int i; §\C{// forward definition}§
    6407 int *j = ®&i®; §\C{// forward reference, valid in C, invalid in \CFA}§
    6408 int i = 0; §\C{// definition}§
     6395int i; $\C{// forward definition}$
     6396int *j = @&i@; $\C{// forward reference, valid in C, invalid in \CFA}$
     6397int i = 0; $\C{// definition}$
    64096398\end{cfa}
    64106399is valid in C, and invalid in \CFA because duplicate overloaded object definitions at the same scope level are disallowed.
     
    64126401\begin{cfa}
    64136402struct X { int i; struct X *next; };
    6414 static struct X a; §\C{// forward definition}§
    6415 static struct X b = { 0, ®&a® };§\C{// forward reference, valid in C, invalid in \CFA}§
    6416 static struct X a = { 1, &b }; §\C{// definition}§
     6403static struct X a; $\C{// forward definition}$
     6404static struct X b = { 0, @&a@ };$\C{// forward reference, valid in C, invalid in \CFA}$
     6405static struct X a = { 1, &b }; $\C{// definition}$
    64176406\end{cfa}
    64186407\item[Rationale:] avoids having different initialization rules for builtin types and user-defined types.
     
    64266415\item[Change:] have ©struct© introduce a scope for nested types:
    64276416\begin{cfa}
    6428 enum ®Colour® { R, G, B, Y, C, M };
     6417enum @Colour@ { R, G, B, Y, C, M };
    64296418struct Person {
    6430         enum ®Colour® { R, G, B };      §\C[7cm]{// nested type}§
    6431         struct Face { §\C{// nested type}§
    6432                 ®Colour® Eyes, Hair; §\C{// type defined outside (1 level)}§
     6419        enum @Colour@ { R, G, B };      $\C[7cm]{// nested type}$
     6420        struct Face { $\C{// nested type}$
     6421                @Colour@ Eyes, Hair; $\C{// type defined outside (1 level)}$
    64336422        };
    6434         ®.Colour® shirt; §\C{// type defined outside (top level)}§
    6435         ®Colour® pants; §\C{// type defined same level}§
    6436         Face looks[10]; §\C{// type defined same level}§
     6423        @.Colour@ shirt; $\C{// type defined outside (top level)}$
     6424        @Colour@ pants; $\C{// type defined same level}$
     6425        Face looks[10]; $\C{// type defined same level}$
    64376426};
    6438 ®Colour® c = R; §\C{// type/enum defined same level}§
    6439 Person®.Colour® pc = Person®.®R;§\C{// type/enum defined inside}§
    6440 Person®.®Face pretty; §\C{// type defined inside}\CRT§
     6427@Colour@ c = R; $\C{// type/enum defined same level}$
     6428Person@.Colour@ pc = Person@.@R;$\C{// type/enum defined inside}$
     6429Person@.@Face pretty; $\C{// type defined inside}\CRT$
    64416430\end{cfa}
    64426431In C, the name of the nested types belongs to the same scope as the name of the outermost enclosing structure, \ie the nested types are hoisted to the scope of the outer-most type, which is not useful and confusing.
     
    64556444\item[Difficulty of converting:] Semantic transformation. To make the struct type name visible in the scope of the enclosing struct, the struct tag could be declared in the scope of the enclosing struct, before the enclosing struct is defined. Example:
    64566445\begin{cfa}
    6457 struct Y; §\C{// struct Y and struct X are at the same scope}§
     6446struct Y; $\C{// struct Y and struct X are at the same scope}$
    64586447struct X {
    64596448        struct Y { /* ... */ } y;
     
    64706459\begin{cfa}
    64716460void foo() {
    6472         int * b = malloc( sizeof(int) ); §\C{// implicitly convert void * to int *}§
    6473         char * c = b; §\C{// implicitly convert int * to void *, and then void * to char *}§
     6461        int * b = malloc( sizeof(int) ); $\C{// implicitly convert void * to int *}$
     6462        char * c = b; $\C{// implicitly convert int * to void *, and then void * to char *}$
    64746463}
    64756464\end{cfa}
    64766465\item[Rationale:] increase type safety
    64776466\item[Effect on original feature:] deletion of semantically well-defined feature.
    6478 \item[Difficulty of converting:] requires adding a cast (see \VRef{s:StorageManagement} for better alternatives):
     6467\item[Difficulty of converting:] requires adding a cast \see{\VRef{s:StorageManagement} for better alternatives}:
    64796468\begin{cfa}
    64806469        int * b = (int *)malloc( sizeof(int) );
     
    65866575\end{cquote}
    65876576For the prescribed head-files, \CFA uses header interposition to wraps these includes in an ©extern "C"©;
    6588 hence, names in these include files are not mangled\index{mangling!name} (see~\VRef{s:Interoperability}).
     6577hence, names in these include files are not mangled\index{mangling!name} \see{\VRef{s:Interoperability}}.
    65896578All other C header files must be explicitly wrapped in ©extern "C"© to prevent name mangling.
    65906579This approach is different from \Index*[C++]{\CC{}} where the name-mangling issue is handled internally in C header-files through checks for preprocessor variable ©__cplusplus©, which adds appropriate ©extern "C"© qualifiers.
     
    66496638Type-safe allocation is provided for all C allocation routines and new \CFA allocation routines, \eg in
    66506639\begin{cfa}
    6651 int * ip = (int *)malloc( sizeof(int) );                §\C{// C}§
    6652 int * ip = malloc();                                                    §\C{// \CFA type-safe version of C malloc}§
    6653 int * ip = alloc();                                                             §\C{// \CFA type-safe uniform alloc}§
     6640int * ip = (int *)malloc( sizeof(int) );                $\C{// C}$
     6641int * ip = malloc();                                                    $\C{// \CFA type-safe version of C malloc}$
     6642int * ip = alloc();                                                             $\C{// \CFA type-safe uniform alloc}$
    66546643\end{cfa}
    66556644the latter two allocations determine the allocation size from the type of ©p© (©int©) and cast the pointer to the allocated storage to ©int *©.
     
    66586647\begin{cfa}
    66596648struct S { int i; } __attribute__(( aligned( 128 ) )); // cache-line alignment
    6660 S * sp = malloc();                                                              §\C{// honour type alignment}§
     6649S * sp = malloc();                                                              $\C{// honour type alignment}$
    66616650\end{cfa}
    66626651the storage allocation is implicitly aligned to 128 rather than the default 16.
     
    66736662\CFA memory management extends allocation to support constructors for initialization of allocated storage, \eg in
    66746663\begin{cfa}
    6675 struct S { int i; };                                                    §\C{// cache-line aglinment}§
     6664struct S { int i; };                                                    $\C{// cache-line alignment}$
    66766665void ?{}( S & s, int i ) { s.i = i; }
    66776666// assume ?|? operator for printing an S
    66786667
    6679 S & sp = *®new®( 3 );                                                   §\C{// call constructor after allocation}§
     6668S & sp = *@new@( 3 );                                                   $\C{// call constructor after allocation}$
    66806669sout | sp.i;
    6681 ®delete®( &sp );
    6682 
    6683 S * spa = ®anew®( 10, 5 );                                              §\C{// allocate array and initialize each array element}§
     6670@delete@( &sp );
     6671
     6672S * spa = @anew@( 10, 5 );                                              $\C{// allocate array and initialize each array element}$
    66846673for ( i; 10 ) sout | spa[i] | nonl;
    66856674sout | nl;
    6686 ®adelete®( 10, spa );
     6675@adelete@( 10, spa );
    66876676\end{cfa}
    66886677Allocation routines ©new©/©anew© allocate a variable/array and initialize storage using the allocated type's constructor.
     
    66936682extern "C" {
    66946683        // C unsafe allocation
    6695         void * malloc( size_t size );§\indexc{malloc}§
    6696         void * calloc( size_t dim, size_t size );§\indexc{calloc}§
    6697         void * realloc( void * ptr, size_t size );§\indexc{realloc}§
    6698         void * memalign( size_t align, size_t size );§\indexc{memalign}§
    6699         void * aligned_alloc( size_t align, size_t size );§\indexc{aligned_alloc}§
    6700         int posix_memalign( void ** ptr, size_t align, size_t size );§\indexc{posix_memalign}§
    6701         void * cmemalign( size_t alignment, size_t noOfElems, size_t elemSize );§\indexc{cmemalign}§ // CFA
     6684        void * malloc( size_t size );$\indexc{malloc}$
     6685        void * calloc( size_t dim, size_t size );$\indexc{calloc}$
     6686        void * realloc( void * ptr, size_t size );$\indexc{realloc}$
     6687        void * memalign( size_t align, size_t size );$\indexc{memalign}$
     6688        void * aligned_alloc( size_t align, size_t size );$\indexc{aligned_alloc}$
     6689        int posix_memalign( void ** ptr, size_t align, size_t size );$\indexc{posix_memalign}$
     6690        void * cmemalign( size_t alignment, size_t noOfElems, size_t elemSize );$\indexc{cmemalign}$ // CFA
    67026691
    67036692        // C unsafe initialization/copy
    6704         void * memset( void * dest, int c, size_t size );§\indexc{memset}§
    6705         void * memcpy( void * dest, const void * src, size_t size );§\indexc{memcpy}§
     6693        void * memset( void * dest, int c, size_t size );$\indexc{memset}$
     6694        void * memcpy( void * dest, const void * src, size_t size );$\indexc{memcpy}$
    67066695}
    67076696
     
    67096698
    67106699forall( dtype T | sized(T) ) {
    6711         // §\CFA§ safe equivalents, i.e., implicit size specification
     6700        // $\CFA$ safe equivalents, i.e., implicit size specification
    67126701        T * malloc( void );
    67136702        T * calloc( size_t dim );
     
    67186707        int posix_memalign( T ** ptr, size_t align );
    67196708
    6720         // §\CFA§ safe general allocation, fill, resize, alignment, array
    6721         T * alloc( void );§\indexc{alloc}§                                      §\C[3.5in]{// variable, T size}§
    6722         T * alloc( size_t dim );                                                        §\C{// array[dim], T size elements}§
    6723         T * alloc( T ptr[], size_t dim );                                       §\C{// realloc array[dim], T size elements}§
    6724 
    6725         T * alloc_set( char fill );§\indexc{alloc_set}§         §\C{// variable, T size, fill bytes with value}§
    6726         T * alloc_set( T fill );                                                        §\C{// variable, T size, fill with value}§
    6727         T * alloc_set( size_t dim, char fill );                         §\C{// array[dim], T size elements, fill bytes with value}§
    6728         T * alloc_set( size_t dim, T fill );                            §\C{// array[dim], T size elements, fill elements with value}§
    6729         T * alloc_set( size_t dim, const T fill[] );            §\C{// array[dim], T size elements, fill elements with array}§
    6730         T * alloc_set( T ptr[], size_t dim, char fill );        §\C{// realloc array[dim], T size elements, fill bytes with value}§
    6731 
    6732         T * alloc_align( size_t align );                                        §\C{// aligned variable, T size}§
    6733         T * alloc_align( size_t align, size_t dim );            §\C{// aligned array[dim], T size elements}§
    6734         T * alloc_align( T ptr[], size_t align );                       §\C{// realloc new aligned array}§
    6735         T * alloc_align( T ptr[], size_t align, size_t dim ); §\C{// realloc new aligned array[dim]}§
    6736 
    6737         T * alloc_align_set( size_t align, char fill );         §\C{// aligned variable, T size, fill bytes with value}§
    6738         T * alloc_align_set( size_t align, T fill );            §\C{// aligned variable, T size, fill with value}§
    6739         T * alloc_align_set( size_t align, size_t dim, char fill ); §\C{// aligned array[dim], T size elements, fill bytes with value}§
    6740         T * alloc_align_set( size_t align, size_t dim, T fill ); §\C{// aligned array[dim], T size elements, fill elements with value}§
    6741         T * alloc_align_set( size_t align, size_t dim, const T fill[] ); §\C{// aligned array[dim], T size elements, fill elements with array}§
    6742         T * alloc_align_set( T ptr[], size_t align, size_t dim, char fill ); §\C{// realloc new aligned array[dim], fill new bytes with value}§
    6743 
    6744         // §\CFA§ safe initialization/copy, i.e., implicit size specification
    6745         T * memset( T * dest, char fill );§\indexc{memset}§
    6746         T * memcpy( T * dest, const T * src );§\indexc{memcpy}§
    6747 
    6748         // §\CFA§ safe initialization/copy, i.e., implicit size specification, array types
     6709        // $\CFA$ safe general allocation, fill, resize, alignment, array
     6710        T * alloc( void );$\indexc{alloc}$                                      $\C[3.5in]{// variable, T size}$
     6711        T * alloc( size_t dim );                                                        $\C{// array[dim], T size elements}$
     6712        T * alloc( T ptr[], size_t dim );                                       $\C{// realloc array[dim], T size elements}$
     6713
     6714        T * alloc_set( char fill );$\indexc{alloc_set}$         $\C{// variable, T size, fill bytes with value}$
     6715        T * alloc_set( T fill );                                                        $\C{// variable, T size, fill with value}$
     6716        T * alloc_set( size_t dim, char fill );                         $\C{// array[dim], T size elements, fill bytes with value}$
     6717        T * alloc_set( size_t dim, T fill );                            $\C{// array[dim], T size elements, fill elements with value}$
     6718        T * alloc_set( size_t dim, const T fill[] );            $\C{// array[dim], T size elements, fill elements with array}$
     6719        T * alloc_set( T ptr[], size_t dim, char fill );        $\C{// realloc array[dim], T size elements, fill bytes with value}$
     6720
     6721        T * alloc_align( size_t align );                                        $\C{// aligned variable, T size}$
     6722        T * alloc_align( size_t align, size_t dim );            $\C{// aligned array[dim], T size elements}$
     6723        T * alloc_align( T ptr[], size_t align );                       $\C{// realloc new aligned array}$
     6724        T * alloc_align( T ptr[], size_t align, size_t dim ); $\C{// realloc new aligned array[dim]}$
     6725
     6726        T * alloc_align_set( size_t align, char fill );         $\C{// aligned variable, T size, fill bytes with value}$
     6727        T * alloc_align_set( size_t align, T fill );            $\C{// aligned variable, T size, fill with value}$
     6728        T * alloc_align_set( size_t align, size_t dim, char fill ); $\C{// aligned array[dim], T size elements, fill bytes with value}$
     6729        T * alloc_align_set( size_t align, size_t dim, T fill ); $\C{// aligned array[dim], T size elements, fill elements with value}$
     6730        T * alloc_align_set( size_t align, size_t dim, const T fill[] ); $\C{// aligned array[dim], T size elements, fill elements with array}$
     6731        T * alloc_align_set( T ptr[], size_t align, size_t dim, char fill ); $\C{// realloc new aligned array[dim], fill new bytes with value}$
     6732
     6733        // $\CFA$ safe initialization/copy, i.e., implicit size specification
     6734        T * memset( T * dest, char fill );$\indexc{memset}$
     6735        T * memcpy( T * dest, const T * src );$\indexc{memcpy}$
     6736
     6737        // $\CFA$ safe initialization/copy, i.e., implicit size specification, array types
    67496738        T * amemset( T dest[], char fill, size_t dim );
    67506739        T * amemcpy( T dest[], const T src[], size_t dim );
    67516740}
    67526741
    6753 // §\CFA§ allocation/deallocation and constructor/destructor, non-array types
    6754 forall( dtype T | sized(T), ttype Params | { void ?{}( T &, Params ); } ) T * new( Params p );§\indexc{new}§
    6755 forall( dtype T | sized(T) | { void ^?{}( T & ); } ) void delete( T * ptr );§\indexc{delete}§
     6742// $\CFA$ allocation/deallocation and constructor/destructor, non-array types
     6743forall( dtype T | sized(T), ttype Params | { void ?{}( T &, Params ); } ) T * new( Params p );$\indexc{new}$
     6744forall( dtype T | sized(T) | { void ^?{}( T & ); } ) void delete( T * ptr );$\indexc{delete}$
    67566745forall( dtype T, ttype Params | sized(T) | { void ^?{}( T & ); void delete( Params ); } )
    67576746  void delete( T * ptr, Params rest );
    67586747
    6759 // §\CFA§ allocation/deallocation and constructor/destructor, array types
    6760 forall( dtype T | sized(T), ttype Params | { void ?{}( T &, Params ); } ) T * anew( size_t dim, Params p );§\indexc{anew}§
    6761 forall( dtype T | sized(T) | { void ^?{}( T & ); } ) void adelete( size_t dim, T arr[] );§\indexc{adelete}§
     6748// $\CFA$ allocation/deallocation and constructor/destructor, array types
     6749forall( dtype T | sized(T), ttype Params | { void ?{}( T &, Params ); } ) T * anew( size_t dim, Params p );$\indexc{anew}$
     6750forall( dtype T | sized(T) | { void ^?{}( T & ); } ) void adelete( size_t dim, T arr[] );$\indexc{adelete}$
    67626751forall( dtype T | sized(T) | { void ^?{}( T & ); }, ttype Params | { void adelete( Params ); } )
    67636752  void adelete( size_t dim, T arr[], Params rest );
     
    67696758\leavevmode
    67706759\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6771 int ato( const char * ptr );§\indexc{ato}§
     6760int ato( const char * ptr );$\indexc{ato}$
    67726761unsigned int ato( const char * ptr );
    67736762long int ato( const char * ptr );
     
    68016790\leavevmode
    68026791\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6803 forall( otype T | { int ?<?( T, T ); } ) §\C{// location}§
    6804 T * bsearch( T key, const T * arr, size_t dim );§\indexc{bsearch}§
    6805 
    6806 forall( otype T | { int ?<?( T, T ); } ) §\C{// position}§
     6792forall( otype T | { int ?<?( T, T ); } ) $\C{// location}$
     6793T * bsearch( T key, const T * arr, size_t dim );$\indexc{bsearch}$
     6794
     6795forall( otype T | { int ?<?( T, T ); } ) $\C{// position}$
    68076796unsigned int bsearch( T key, const T * arr, size_t dim );
    68086797
    68096798forall( otype T | { int ?<?( T, T ); } )
    6810 void qsort( const T * arr, size_t dim );§\indexc{qsort}§
     6799void qsort( const T * arr, size_t dim );$\indexc{qsort}$
    68116800
    68126801forall( otype E | { int ?<?( E, E ); } ) {
    6813         E * bsearch( E key, const E * vals, size_t dim );§\indexc{bsearch}§ §\C{// location}§
    6814         size_t bsearch( E key, const E * vals, size_t dim );§\C{// position}§
    6815         E * bsearchl( E key, const E * vals, size_t dim );§\indexc{bsearchl}§
     6802        E * bsearch( E key, const E * vals, size_t dim );$\indexc{bsearch}$ $\C{// location}$
     6803        size_t bsearch( E key, const E * vals, size_t dim );$\C{// position}$
     6804        E * bsearchl( E key, const E * vals, size_t dim );$\indexc{bsearchl}$
    68166805        size_t bsearchl( E key, const E * vals, size_t dim );
    6817         E * bsearchu( E key, const E * vals, size_t dim );§\indexc{bsearchu}§
     6806        E * bsearchu( E key, const E * vals, size_t dim );$\indexc{bsearchu}$
    68186807        size_t bsearchu( E key, const E * vals, size_t dim );
    68196808}
     
    68296818
    68306819forall( otype E | { int ?<?( E, E ); } ) {
    6831         void qsort( E * vals, size_t dim );§\indexc{qsort}§
     6820        void qsort( E * vals, size_t dim );$\indexc{qsort}$
    68326821}
    68336822\end{cfa}
     
    68386827\leavevmode
    68396828\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6840 unsigned char abs( signed char );§\indexc{abs}§
     6829unsigned char abs( signed char );$\indexc{abs}$
    68416830int abs( int );
    68426831unsigned long int abs( long int );
     
    68576846\leavevmode
    68586847\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6859 void srandom( unsigned int seed );§\indexc{srandom}§
    6860 char random( void );§\indexc{random}§
    6861 char random( char u ); §\C{// [0,u)}§
    6862 char random( char l, char u ); §\C{// [l,u)}§
     6848void srandom( unsigned int seed );$\indexc{srandom}$
     6849char random( void );$\indexc{random}$
     6850char random( char u ); $\C{// [0,u)}$
     6851char random( char l, char u ); $\C{// [l,u)}$
    68636852int random( void );
    6864 int random( int u ); §\C{// [0,u)}§
    6865 int random( int l, int u ); §\C{// [l,u)}§
     6853int random( int u ); $\C{// [0,u)}$
     6854int random( int l, int u ); $\C{// [l,u)}$
    68666855unsigned int random( void );
    6867 unsigned int random( unsigned int u ); §\C{// [0,u)}§
    6868 unsigned int random( unsigned int l, unsigned int u ); §\C{// [l,u)}§
     6856unsigned int random( unsigned int u ); $\C{// [0,u)}$
     6857unsigned int random( unsigned int l, unsigned int u ); $\C{// [l,u)}$
    68696858long int random( void );
    6870 long int random( long int u ); §\C{// [0,u)}§
    6871 long int random( long int l, long int u ); §\C{// [l,u)}§
     6859long int random( long int u ); $\C{// [0,u)}$
     6860long int random( long int l, long int u ); $\C{// [l,u)}$
    68726861unsigned long int random( void );
    6873 unsigned long int random( unsigned long int u ); §\C{// [0,u)}§
    6874 unsigned long int random( unsigned long int l, unsigned long int u ); §\C{// [l,u)}§
    6875 float random( void );                                            §\C{// [0.0, 1.0)}§
    6876 double random( void );                                           §\C{// [0.0, 1.0)}§
    6877 float _Complex random( void );                           §\C{// [0.0, 1.0)+[0.0, 1.0)i}§
    6878 double _Complex random( void );                          §\C{// [0.0, 1.0)+[0.0, 1.0)i}§
    6879 long double _Complex random( void );             §\C{// [0.0, 1.0)+[0.0, 1.0)i}§
     6862unsigned long int random( unsigned long int u ); $\C{// [0,u)}$
     6863unsigned long int random( unsigned long int l, unsigned long int u ); $\C{// [l,u)}$
     6864float random( void );                                            $\C{// [0.0, 1.0)}$
     6865double random( void );                                           $\C{// [0.0, 1.0)}$
     6866float _Complex random( void );                           $\C{// [0.0, 1.0)+[0.0, 1.0)i}$
     6867double _Complex random( void );                          $\C{// [0.0, 1.0)+[0.0, 1.0)i}$
     6868long double _Complex random( void );             $\C{// [0.0, 1.0)+[0.0, 1.0)i}$
    68806869\end{cfa}
    68816870
     
    68856874\leavevmode
    68866875\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6887 forall( otype T | { int ?<?( T, T ); } ) T min( T t1, T t2 );§\indexc{min}§
    6888 forall( otype T | { int ?>?( T, T ); } ) T max( T t1, T t2 );§\indexc{max}§
    6889 forall( otype T | { T min( T, T ); T max( T, T ); } ) T clamp( T value, T min_val, T max_val );§\indexc{clamp}§
    6890 forall( otype T ) void swap( T * t1, T * t2 );§\indexc{swap}§
     6876forall( otype T | { int ?<?( T, T ); } ) T min( T t1, T t2 );$\indexc{min}$
     6877forall( otype T | { int ?>?( T, T ); } ) T max( T t1, T t2 );$\indexc{max}$
     6878forall( otype T | { T min( T, T ); T max( T, T ); } ) T clamp( T value, T min_val, T max_val );$\indexc{clamp}$
     6879forall( otype T ) void swap( T * t1, T * t2 );$\indexc{swap}$
    68916880\end{cfa}
    68926881
     
    69026891\leavevmode
    69036892\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6904 float ?%?( float, float );§\indexc{fmod}§
     6893float ?%?( float, float );$\indexc{fmod}$
    69056894float fmod( float, float );
    69066895double ?%?( double, double );
     
    69096898long double fmod( long double, long double );
    69106899
    6911 float remainder( float, float );§\indexc{remainder}§
     6900float remainder( float, float );$\indexc{remainder}$
    69126901double remainder( double, double );
    69136902long double remainder( long double, long double );
    69146903
    6915 float remquo( float, float, int * );§\indexc{remquo}§
     6904float remquo( float, float, int * );$\indexc{remquo}$
    69166905double remquo( double, double, int * );
    69176906long double remquo( long double, long double, int * );
     
    69206909[ int, long double ] remquo( long double, long double );
    69216910
    6922 float div( float, float, int * );§\indexc{div}§ §\C{// alternative name for remquo}§
     6911float div( float, float, int * );$\indexc{div}$ $\C{// alternative name for remquo}$
    69236912double div( double, double, int * );
    69246913long double div( long double, long double, int * );
     
    69276916[ int, long double ] div( long double, long double );
    69286917
    6929 float fma( float, float, float );§\indexc{fma}§
     6918float fma( float, float, float );$\indexc{fma}$
    69306919double fma( double, double, double );
    69316920long double fma( long double, long double, long double );
    69326921
    6933 float fdim( float, float );§\indexc{fdim}§
     6922float fdim( float, float );$\indexc{fdim}$
    69346923double fdim( double, double );
    69356924long double fdim( long double, long double );
    69366925
    6937 float nan( const char * );§\indexc{nan}§
     6926float nan( const char * );$\indexc{nan}$
    69386927double nan( const char * );
    69396928long double nan( const char * );
     
    69456934\leavevmode
    69466935\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6947 float exp( float );§\indexc{exp}§
     6936float exp( float );$\indexc{exp}$
    69486937double exp( double );
    69496938long double exp( long double );
     
    69526941long double _Complex exp( long double _Complex );
    69536942
    6954 float exp2( float );§\indexc{exp2}§
     6943float exp2( float );$\indexc{exp2}$
    69556944double exp2( double );
    69566945long double exp2( long double );
     
    69596948// long double _Complex exp2( long double _Complex );
    69606949
    6961 float expm1( float );§\indexc{expm1}§
     6950float expm1( float );$\indexc{expm1}$
    69626951double expm1( double );
    69636952long double expm1( long double );
    69646953
    6965 float pow( float, float );§\indexc{pow}§
     6954float pow( float, float );$\indexc{pow}$
    69666955double pow( double, double );
    69676956long double pow( long double, long double );
     
    69766965\leavevmode
    69776966\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    6978 float log( float );§\indexc{log}§
     6967float log( float );$\indexc{log}$
    69796968double log( double );
    69806969long double log( long double );
     
    69836972long double _Complex log( long double _Complex );
    69846973
    6985 float log2( float );§\indexc{log2}§
     6974float log2( float );$\indexc{log2}$
    69866975double log2( double );
    69876976long double log2( long double );
     
    69906979// long double _Complex log2( long double _Complex );
    69916980
    6992 float log10( float );§\indexc{log10}§
     6981float log10( float );$\indexc{log10}$
    69936982double log10( double );
    69946983long double log10( long double );
     
    69976986// long double _Complex log10( long double _Complex );
    69986987
    6999 float log1p( float );§\indexc{log1p}§
     6988float log1p( float );$\indexc{log1p}$
    70006989double log1p( double );
    70016990long double log1p( long double );
    70026991
    7003 int ilogb( float );§\indexc{ilogb}§
     6992int ilogb( float );$\indexc{ilogb}$
    70046993int ilogb( double );
    70056994int ilogb( long double );
    70066995
    7007 float logb( float );§\indexc{logb}§
     6996float logb( float );$\indexc{logb}$
    70086997double logb( double );
    70096998long double logb( long double );
    70106999
    7011 float sqrt( float );§\indexc{sqrt}§
     7000float sqrt( float );$\indexc{sqrt}$
    70127001double sqrt( double );
    70137002long double sqrt( long double );
     
    70167005long double _Complex sqrt( long double _Complex );
    70177006
    7018 float cbrt( float );§\indexc{cbrt}§
     7007float cbrt( float );$\indexc{cbrt}$
    70197008double cbrt( double );
    70207009long double cbrt( long double );
    70217010
    7022 float hypot( float, float );§\indexc{hypot}§
     7011float hypot( float, float );$\indexc{hypot}$
    70237012double hypot( double, double );
    70247013long double hypot( long double, long double );
     
    70307019\leavevmode
    70317020\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    7032 float sin( float );§\indexc{sin}§
     7021float sin( float );$\indexc{sin}$
    70337022double sin( double );
    70347023long double sin( long double );
     
    70377026long double _Complex sin( long double _Complex );
    70387027
    7039 float cos( float );§\indexc{cos}§
     7028float cos( float );$\indexc{cos}$
    70407029double cos( double );
    70417030long double cos( long double );
     
    70447033long double _Complex cos( long double _Complex );
    70457034
    7046 float tan( float );§\indexc{tan}§
     7035float tan( float );$\indexc{tan}$
    70477036double tan( double );
    70487037long double tan( long double );
     
    70517040long double _Complex tan( long double _Complex );
    70527041
    7053 float asin( float );§\indexc{asin}§
     7042float asin( float );$\indexc{asin}$
    70547043double asin( double );
    70557044long double asin( long double );
     
    70587047long double _Complex asin( long double _Complex );
    70597048
    7060 float acos( float );§\indexc{acos}§
     7049float acos( float );$\indexc{acos}$
    70617050double acos( double );
    70627051long double acos( long double );
     
    70657054long double _Complex acos( long double _Complex );
    70667055
    7067 float atan( float );§\indexc{atan}§
     7056float atan( float );$\indexc{atan}$
    70687057double atan( double );
    70697058long double atan( long double );
     
    70727061long double _Complex atan( long double _Complex );
    70737062
    7074 float atan2( float, float );§\indexc{atan2}§
     7063float atan2( float, float );$\indexc{atan2}$
    70757064double atan2( double, double );
    70767065long double atan2( long double, long double );
    70777066
    7078 float atan( float, float ); §\C{// alternative name for atan2}§
    7079 double atan( double, double );§\indexc{atan}§
     7067float atan( float, float ); $\C{// alternative name for atan2}$
     7068double atan( double, double );$\indexc{atan}$
    70807069long double atan( long double, long double );
    70817070\end{cfa}
     
    70867075\leavevmode
    70877076\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    7088 float sinh( float );§\indexc{sinh}§
     7077float sinh( float );$\indexc{sinh}$
    70897078double sinh( double );
    70907079long double sinh( long double );
     
    70937082long double _Complex sinh( long double _Complex );
    70947083
    7095 float cosh( float );§\indexc{cosh}§
     7084float cosh( float );$\indexc{cosh}$
    70967085double cosh( double );
    70977086long double cosh( long double );
     
    71007089long double _Complex cosh( long double _Complex );
    71017090
    7102 float tanh( float );§\indexc{tanh}§
     7091float tanh( float );$\indexc{tanh}$
    71037092double tanh( double );
    71047093long double tanh( long double );
     
    71077096long double _Complex tanh( long double _Complex );
    71087097
    7109 float asinh( float );§\indexc{asinh}§
     7098float asinh( float );$\indexc{asinh}$
    71107099double asinh( double );
    71117100long double asinh( long double );
     
    71147103long double _Complex asinh( long double _Complex );
    71157104
    7116 float acosh( float );§\indexc{acosh}§
     7105float acosh( float );$\indexc{acosh}$
    71177106double acosh( double );
    71187107long double acosh( long double );
     
    71217110long double _Complex acosh( long double _Complex );
    71227111
    7123 float atanh( float );§\indexc{atanh}§
     7112float atanh( float );$\indexc{atanh}$
    71247113double atanh( double );
    71257114long double atanh( long double );
     
    71347123\leavevmode
    71357124\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    7136 float erf( float );§\indexc{erf}§
     7125float erf( float );$\indexc{erf}$
    71377126double erf( double );
    71387127long double erf( long double );
     
    71417130long double _Complex erf( long double _Complex );
    71427131
    7143 float erfc( float );§\indexc{erfc}§
     7132float erfc( float );$\indexc{erfc}$
    71447133double erfc( double );
    71457134long double erfc( long double );
     
    71487137long double _Complex erfc( long double _Complex );
    71497138
    7150 float lgamma( float );§\indexc{lgamma}§
     7139float lgamma( float );$\indexc{lgamma}$
    71517140double lgamma( double );
    71527141long double lgamma( long double );
     
    71557144long double lgamma( long double, int * );
    71567145
    7157 float tgamma( float );§\indexc{tgamma}§
     7146float tgamma( float );$\indexc{tgamma}$
    71587147double tgamma( double );
    71597148long double tgamma( long double );
     
    71657154\leavevmode
    71667155\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    7167 float floor( float );§\indexc{floor}§
     7156float floor( float );$\indexc{floor}$
    71687157double floor( double );
    71697158long double floor( long double );
    71707159
    7171 float ceil( float );§\indexc{ceil}§
     7160float ceil( float );$\indexc{ceil}$
    71727161double ceil( double );
    71737162long double ceil( long double );
    71747163
    7175 float trunc( float );§\indexc{trunc}§
     7164float trunc( float );$\indexc{trunc}$
    71767165double trunc( double );
    71777166long double trunc( long double );
    71787167
    7179 float rint( float );§\indexc{rint}§
     7168float rint( float );$\indexc{rint}$
    71807169long double rint( long double );
    71817170long int rint( float );
     
    71867175long long int rint( long double );
    71877176
    7188 long int lrint( float );§\indexc{lrint}§
     7177long int lrint( float );$\indexc{lrint}$
    71897178long int lrint( double );
    71907179long int lrint( long double );
     
    71937182long long int llrint( long double );
    71947183
    7195 float nearbyint( float );§\indexc{nearbyint}§
     7184float nearbyint( float );$\indexc{nearbyint}$
    71967185double nearbyint( double );
    71977186long double nearbyint( long double );
    71987187
    7199 float round( float );§\indexc{round}§
     7188float round( float );$\indexc{round}$
    72007189long double round( long double );
    72017190long int round( float );
     
    72067195long long int round( long double );
    72077196
    7208 long int lround( float );§\indexc{lround}§
     7197long int lround( float );$\indexc{lround}$
    72097198long int lround( double );
    72107199long int lround( long double );
     
    72197208\leavevmode
    72207209\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    7221 float copysign( float, float );§\indexc{copysign}§
     7210float copysign( float, float );$\indexc{copysign}$
    72227211double copysign( double, double );
    72237212long double copysign( long double, long double );
    72247213
    7225 float frexp( float, int * );§\indexc{frexp}§
     7214float frexp( float, int * );$\indexc{frexp}$
    72267215double frexp( double, int * );
    72277216long double frexp( long double, int * );
    72287217
    7229 float ldexp( float, int );§\indexc{ldexp}§
     7218float ldexp( float, int );$\indexc{ldexp}$
    72307219double ldexp( double, int );
    72317220long double ldexp( long double, int );
    72327221
    7233 [ float, float ] modf( float );§\indexc{modf}§
     7222[ float, float ] modf( float );$\indexc{modf}$
    72347223float modf( float, float * );
    72357224[ double, double ] modf( double );
     
    72387227long double modf( long double, long double * );
    72397228
    7240 float nextafter( float, float );§\indexc{nextafter}§
     7229float nextafter( float, float );$\indexc{nextafter}$
    72417230double nextafter( double, double );
    72427231long double nextafter( long double, long double );
    72437232
    7244 float nexttoward( float, long double );§\indexc{nexttoward}§
     7233float nexttoward( float, long double );$\indexc{nexttoward}$
    72457234double nexttoward( double, long double );
    72467235long double nexttoward( long double, long double );
    72477236
    7248 float scalbn( float, int );§\indexc{scalbn}§
     7237float scalbn( float, int );$\indexc{scalbn}$
    72497238double scalbn( double, int );
    72507239long double scalbn( long double, int );
    72517240
    7252 float scalbln( float, long int );§\indexc{scalbln}§
     7241float scalbln( float, long int );$\indexc{scalbln}$
    72537242double scalbln( double, long int );
    72547243long double scalbln( long double, long int );
     
    72677256\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    72687257struct Duration {
    7269         int64_t tv; §\C{// nanoseconds}§
     7258        int64_t tv; $\C{// nanoseconds}$
    72707259};
    72717260
     
    73977386\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    73987387struct Time {
    7399         uint64_t tv; §\C{// nanoseconds since UNIX epoch}§
     7388        uint64_t tv; $\C{// nanoseconds since UNIX epoch}$
    74007389};
    74017390
     
    74687457\begin{cfa}[aboveskip=0pt,belowskip=0pt]
    74697458struct Clock {
    7470         Duration offset; §\C{// for virtual clock: contains offset from real-time}§
    7471         int clocktype; §\C{// implementation only -1 (virtual), CLOCK\_REALTIME}§
     7459        Duration offset; $\C{// for virtual clock: contains offset from real-time}$
     7460        int clocktype; $\C{// implementation only -1 (virtual), CLOCK\_REALTIME}$
    74727461};
    74737462
     
    74777466void ?{}( Clock & clk, Duration adj );
    74787467
    7479 Duration getResNsec(); §\C{// with nanoseconds}§
    7480 Duration getRes(); §\C{// without nanoseconds}§
    7481 
    7482 Time getTimeNsec(); §\C{// with nanoseconds}§
    7483 Time getTime(); §\C{// without nanoseconds}§
     7468Duration getResNsec(); $\C{// with nanoseconds}$
     7469Duration getRes(); $\C{// without nanoseconds}$
     7470
     7471Time getTimeNsec(); $\C{// with nanoseconds}$
     7472Time getTime(); $\C{// without nanoseconds}$
    74847473Time getTime( Clock & clk );
    74857474Time ?()( Clock & clk );
     
    74977486
    74987487\begin{cfa}
    7499 void ?{}( Int * this ); §\C{// constructor/destructor}§
     7488void ?{}( Int * this ); $\C{// constructor/destructor}$
    75007489void ?{}( Int * this, Int init );
    75017490void ?{}( Int * this, zero_t );
     
    75067495void ^?{}( Int * this );
    75077496
    7508 Int ?=?( Int * lhs, Int rhs ); §\C{// assignment}§
     7497Int ?=?( Int * lhs, Int rhs ); $\C{// assignment}$
    75097498Int ?=?( Int * lhs, long int rhs );
    75107499Int ?=?( Int * lhs, unsigned long int rhs );
     
    75237512unsigned long int narrow( Int val );
    75247513
    7525 int ?==?( Int oper1, Int oper2 ); §\C{// comparison}§
     7514int ?==?( Int oper1, Int oper2 ); $\C{// comparison}$
    75267515int ?==?( Int oper1, long int oper2 );
    75277516int ?==?( long int oper2, Int oper1 );
     
    75597548int ?>=?( unsigned long int oper1, Int oper2 );
    75607549
    7561 Int +?( Int oper ); §\C{// arithmetic}§
     7550Int +?( Int oper ); $\C{// arithmetic}$
    75627551Int -?( Int oper );
    75637552Int ~?( Int oper );
     
    76417630Int ?>>=?( Int * lhs, mp_bitcnt_t shift );
    76427631
    7643 Int abs( Int oper ); §\C{// number functions}§
     7632Int abs( Int oper ); $\C{// number functions}$
    76447633Int fact( unsigned long int N );
    76457634Int gcd( Int oper1, Int oper2 );
     
    76537642Int sqrt( Int oper );
    76547643
    7655 forall( dtype istype | istream( istype ) ) istype * ?|?( istype * is, Int * mp );  §\C{// I/O}§
     7644forall( dtype istype | istream( istype ) ) istype * ?|?( istype * is, Int * mp );  $\C{// I/O}$
    76567645forall( dtype ostype | ostream( ostype ) ) ostype * ?|?( ostype * os, Int mp );
    76577646\end{cfa}
     
    76647653\hline
    76657654\begin{cfa}
    7666 #include <gmp>§\indexc{gmp}§
     7655#include <gmp>$\indexc{gmp}$
    76677656int main( void ) {
    76687657        sout | "Factorial Numbers";
     
    76787667&
    76797668\begin{cfa}
    7680 #include <gmp.h>§\indexc{gmp.h}§
     7669#include <gmp.h>$\indexc{gmp.h}$
    76817670int main( void ) {
    7682         ®gmp_printf®( "Factorial Numbers\n" );
    7683         ®mpz_t® fact;
    7684         ®mpz_init_set_ui®( fact, 1 );
    7685         ®gmp_printf®( "%d %Zd\n", 0, fact );
     7671        @gmp_printf@( "Factorial Numbers\n" );
     7672        @mpz_t@ fact;
     7673        @mpz_init_set_ui@( fact, 1 );
     7674        @gmp_printf@( "%d %Zd\n", 0, fact );
    76867675        for ( unsigned int i = 1; i <= 40; i += 1 ) {
    7687                 ®mpz_mul_ui®( fact, fact, i );
    7688                 ®gmp_printf®( "%d %Zd\n", i, fact );
     7676                @mpz_mul_ui@( fact, fact, i );
     7677                @gmp_printf@( "%d %Zd\n", i, fact );
    76897678        }
    76907679}
     
    77517740\begin{cfa}[belowskip=0pt]
    77527741// implementation
    7753 struct Rational {§\indexc{Rational}§
    7754         long int numerator, denominator; §\C{// invariant: denominator > 0}§
     7742struct Rational {$\indexc{Rational}$
     7743        long int numerator, denominator; $\C{// invariant: denominator > 0}$
    77557744}; // Rational
    77567745
    7757 Rational rational(); §\C{// constructors}§
     7746Rational rational(); $\C{// constructors}$
    77587747Rational rational( long int n );
    77597748Rational rational( long int n, long int d );
     
    77617750void ?{}( Rational * r, one_t );
    77627751
    7763 long int numerator( Rational r ); §\C{// numerator/denominator getter/setter}§
     7752long int numerator( Rational r ); $\C{// numerator/denominator getter/setter}$
    77647753long int numerator( Rational r, long int n );
    77657754long int denominator( Rational r );
    77667755long int denominator( Rational r, long int d );
    77677756
    7768 int ?==?( Rational l, Rational r ); §\C{// comparison}§
     7757int ?==?( Rational l, Rational r ); $\C{// comparison}$
    77697758int ?!=?( Rational l, Rational r );
    77707759int ?<?( Rational l, Rational r );
     
    77737762int ?>=?( Rational l, Rational r );
    77747763
    7775 Rational -?( Rational r ); §\C{// arithmetic}§
     7764Rational -?( Rational r ); $\C{// arithmetic}$
    77767765Rational ?+?( Rational l, Rational r );
    77777766Rational ?-?( Rational l, Rational r );
     
    77797768Rational ?/?( Rational l, Rational r );
    77807769
    7781 double widen( Rational r ); §\C{// conversion}§
     7770double widen( Rational r ); $\C{// conversion}$
    77827771Rational narrow( double f, long int md );
    77837772
Note: See TracChangeset for help on using the changeset viewer.