\chapter{Array} \section{Introduction} This chapter describes my contribution of language and library features that provide a length-checked array type, as in: \begin{lstlisting} array(float, 99) x; // x contains 99 floats void f( array(float, 42) & a ) {} f(x); // statically rejected: types are different forall( T, [N] ) void g( array(T, N) & a, int i ) { T elem = a[i]; // dynamically checked: requires 0 <= i < N } g(x, 0); // T is float, N is 99, succeeds g(x, 1000); // T is float, N is 99, dynamic check fails \end{lstlisting} This example first declares @x@ a variable, whose type is an instantiation of the generic type named @array@, with arguments @float@ and @99@. Next, it declares @f@ as a function that expects a length-42 array; the type system rejects the call's attempt to pass @x@ to @f@, because the lengths do not match. Next, the @forall@ annotation on function @g@ introduces @T@ as a familiar type parameter and @N@ as a \emph{dimension} parameter, a new feature that represents a count of elements, as managed by the type system. Because @g@ accepts any length of array; the type system accepts the calls' passing @x@ to @g@, inferring that this length is 99. Just as the caller's code does not need to explain that @T@ is @float@, the safe capture and communication of the value @99@ occurs without programmer involvement. In the case of the second call (which passes the value 1000 for @i@), within the body of @g@, the attempt to subscript @a@ by @i@ fails with a runtime error, since $@i@ \nless @N@$. The type @array@, as seen above, comes from my additions to the \CFA standard library. It is very similar to the built-in array type, which \CFA inherits from C. Its runtime characteristics are often identical, and some features are available in both. \begin{lstlisting} forall( [N] ) void declDemo() { float a1[N]; // built-in type ("C array") array(float, N) a2; // type from library } \end{lstlisting} If a caller instantiates @N@ with 42, then both locally-declared array variables, @a1@ and @a2@, become arrays of 42 elements, each element being a @float@. The two variables have identical size and layout; they both encapsulate 42-float stack allocations, no heap allocations, and no further "bookkeeping" allocations/header. Having the @array@ library type (that of @a2@) is a tactical measure, an early implementation that offers full feature support. A future goal (TODO xref) is to port all of its features into the built-in array type (that of @a1@); then, the library type could be removed, and \CFA would have only one array type. In present state, the built-in array has partial support for the new features. The fully-featured library type is used exclusively in introductory examples; feature support and C compatibility are revisited in sec TODO. Offering the @array@ type, as a distinct alternative from the the C array, is consistent with \CFA's extension philosophy (TODO xref background) to date. A few compatibility-breaking changes to the behaviour of the C array were also made, both as an implementation convenience, and as justified fixes to C's lax treatment. The @array@ type is an opportunity to start from a clean slate and show a cohesive selection of features. A clean slate was an important starting point because it meant not having to deal with every inherited complexity introduced in TODO xref background-array. My contributions are \begin{itemize} \item a type system enhancement that lets polymorphic functions and generic types be parameterized by a numeric value: @forall( [N] )@ \item TODO: general parking... \item identify specific abilities brought by @array@ \item Where there is a gap concerning this feature's readiness for prime-time, identification of specific workable improvements that are likely to close the gap \end{itemize} \section{Definitions and design considerations} \subsection{Dependent typing} \section{Features Added} The present work adds a type @array@ to the \CFA standard library~\cite{Cforall}. This array's length is statically managed and dynamically valued. This static management achieves argument safety and suggests a path to subscript safety as future work (TODO: cross reference). This section presents motivating examples of the new array type's usage and follows up with definitions of the notations that appear. The core of the new array management is tracking all array lengths in the type system. Dynamically valued lengths are represented using type variables. The stratification of type variables preceding object declarations makes a length referenceable everywhere that it is needed. For example, a declaration can share one length, @N@, among a pair of parameters and the return. \lstinput{10-17}{hello-array.cfa} Here, the function @f@ does a pointwise comparison, checking if each pair of numbers is within half a percent of each other, returning the answers in a newly allocated @bool@ array. The array type uses the parameterized length information in its @sizeof@ determination, illustrated in the example's call to @alloc@. That call requests an allocation of type @array(bool, N)@, which the type system deduces from the left-hand side of the initialization, into the return type of the @alloc@ call. Preexisting \CFA behaviour is leveraged here, both in the return-type-only polymorphism, and the @sized(T)@-aware standard-library @alloc@ routine. The new @array@ type plugs into this behaviour by implementing the @sized@/@sizeof@ assertion to have the intuitive meaning. As a result, this design avoids an opportunity for programmer error by making the size/length communication to a called routine implicit, compared with C's @calloc@ (or the low-level \CFA analog @aalloc@), which take an explicit length parameter not managed by the type system. \VRef[Figure]{f:fHarness} shows the harness to use the @f@ function illustrating how dynamic values are fed into the system. Here, the @a@ array is loaded with decreasing values, and the @b@ array with amounts off by a constant, giving relative differences within tolerance at first and out of tolerance later. The program main is run with two different inputs of sequence length. \begin{figure} \lstinput{30-49}{hello-array.cfa} \caption{\lstinline{f} Harness} \label{f:fHarness} \end{figure} The loops in the program main follow the more familiar pattern of using the ordinary variable @n@ to convey the length. The type system implicitly captures this value at the call site (@main@ calling @f@) and makes it available within the callee (@f@'s loop bound). The two parts of the example show @n@ adapting a variable into a type-system managed length (at @main@'s declarations of @a@, @b@, and @result@), @N@ adapting in the opposite direction (at @f@'s loop bound), and a pass-thru use of a managed length (at @f@'s declaration of @ret@). The @forall( ...[N] )@ participates in the user-relevant declaration of the name @N@, which becomes usable in parameter/return declarations and in the function @b@. The present form is chosen to parallel the existing @forall@ forms: \begin{cfa} forall( @[N]@ ) ... // array kind forall( & T ) ... // reference kind (dtype) forall( T ) ... // value kind (otype) \end{cfa} The notation @array(thing, N)@ is a single-dimensional case, giving a generic type instance. In summary: \begin{itemize} \item @[N]@ -- within a forall, declares the type variable @N@ to be a managed length \item $e$ -- a type representing the value of $e$ as a managed length, where $e$ is a @size_t@-typed expression \item N -- an expression of type @size_t@, whose value is the managed length @N@ \item @array( thing, N0, N1, ... )@ -- a type wrapping $\prod_i N_i$ adjacent occurrences of @thing@ objects \end{itemize} Unsigned integers have a special status in this type system. Unlike how C++ allows \begin{lstlisting}[language=c++] template< size_t N, char * msg, typename T >... // declarations \end{lstlisting} \CFA does not accommodate values of any user-provided type. TODO: discuss connection with dependent types. An example of a type error demonstrates argument safety. The running example has @f@ expecting two arrays of the same length. A compile-time error occurs when attempting to call @f@ with arrays whose lengths may differ. \begin{cfa} forall( [M], [N] ) void bad( array(float, M) &a, array(float, N) &b ) { f( a, a ); // ok f( b, b ); // ok f( a, b ); // error } \end{cfa} %\lstinput{60-65}{hello-array.cfa} As is common practice in C, the programmer is free to cast, to assert knowledge not shared with the type system. \begin{cfa} forall( [M], [N] ) void bad_fixed( array(float, M) & a, array(float, N) & b ) { if ( M == N ) { f( a, (array(float, M) &)b ); // cast b to matching type } } \end{cfa} %\lstinput{70-75}{hello-array.cfa} Argument safety and the associated implicit communication of array length work with \CFA's generic types too. \CFA allows aggregate types to be generalized with multiple type parameters, including parameterized element type, so can it be defined over a parameterized length. Doing so gives a refinement of C's ``flexible array member'' pattern, that allows nesting structures with array members anywhere within other structures. \lstinput{10-16}{hello-accordion.cfa} This structure's layout has the starting offset of @cost_contribs@ varying in @Nclients@, and the offset of @total_cost@ varying in both generic parameters. For a function that operates on a @request@ structure, the type system handles this variation transparently. \lstinput{40-47}{hello-accordion.cfa} In the example, different runs of the program result in different offset values being used. \lstinput{60-76}{hello-accordion.cfa} The output values show that @summarize@ and its caller agree on both the offsets (where the callee starts reading @cost_contribs@ and where the callee writes @total_cost@). Yet the call site still says just, ``pass the request.'' \section{Multidimensional implementation} \label{toc:mdimpl} TODO: introduce multidimensional array feature and approaches The new \CFA standard library @array@ datatype supports multidimensional uses more richly than the C array. The new array's multidimensional interface and implementation, follows an array-of-arrays setup, meaning, like C's @float[n][m]@ type, one contiguous object, with coarsely-strided dimensions directly wrapping finely-strided dimensions. This setup is in contrast with the pattern of array of pointers to other allocations representing a sub-array. Beyond what C's type offers, the new array brings direct support for working with a noncontiguous array slice, allowing a program to work with dimension subscripts given in a non-physical order. C and C++ require a programmer with such a need to manage pointer/offset arithmetic manually. Examples are shown using a $5 \times 7$ float array, @a@, loaded with increments of $0.1$ when stepping across the length-7 finely-strided dimension shown on columns, and with increments of $1.0$ when stepping across the length-5 coarsely-strided dimension shown on rows. %\lstinput{120-126}{hello-md.cfa} The memory layout of @a@ has strictly increasing numbers along its 35 contiguous positions. A trivial form of slicing extracts a contiguous inner array, within an array-of-arrays. Like with the C array, a lesser-dimensional array reference can be bound to the result of subscripting a greater-dimensional array, by a prefix of its dimensions. This action first subscripts away the most coarsely strided dimensions, leaving a result that expects to be be subscripted by the more finely strided dimensions. \lstinput{60-66}{hello-md.cfa} \lstinput[aboveskip=0pt]{140-140}{hello-md.cfa} This function declaration is asserting too much knowledge about its parameter @c@, for it to be usable for printing either a row slice or a column slice. Specifically, declaring the parameter @c@ with type @array@ means that @c@ is contiguous. However, the function does not use this fact. For the function to do its job, @c@ need only be of a container type that offers a subscript operator (of type @ptrdiff_t@ $\rightarrow$ @float@), with managed length @N@. The new-array library provides the trait @ix@, so-defined. With it, the original declaration can be generalized, while still implemented with the same body, to the latter declaration: \lstinput{40-44}{hello-md.cfa} \lstinput[aboveskip=0pt]{145-145}{hello-md.cfa} Nontrivial slicing, in this example, means passing a noncontiguous slice to @print1d@. The new-array library provides a ``subscript by all'' operation for this purpose. In a multi-dimensional subscript operation, any dimension given as @all@ is left ``not yet subscripted by a value,'' implementing the @ix@ trait, waiting for such a value. \lstinput{150-151}{hello-md.cfa} The example has shown that @a[2]@ and @a[[2, all]]@ both refer to the same, ``2.*'' slice. Indeed, the various @print1d@ calls under discussion access the entry with value 2.3 as @a[2][3]@, @a[[2,all]][3]@, and @a[[all,3]][2]@. This design preserves (and extends) C array semantics by defining @a[[i,j]]@ to be @a[i][j]@ for numeric subscripts, but also for ``subscripting by all''. That is: \begin{tabular}{cccccl} @a[[2,all]][3]@ & $=$ & @a[2][all][3]@ & $=$ & @a[2][3]@ & (here, @all@ is redundant) \\ @a[[all,3]][2]@ & $=$ & @a[all][3][2]@ & $=$ & @a[2][3]@ & (here, @all@ is effective) \end{tabular} Narrating progress through each of the @-[-][-][-]@ expressions gives, firstly, a definition of @-[all]@, and secondly, a generalization of C's @-[i]@. \noindent Where @all@ is redundant: \begin{tabular}{ll} @a@ & 2-dimensional, want subscripts for coarse then fine \\ @a[2]@ & 1-dimensional, want subscript for fine; lock coarse = 2 \\ @a[2][all]@ & 1-dimensional, want subscript for fine \\ @a[2][all][3]@ & 0-dimensional; lock fine = 3 \end{tabular} \noindent Where @all@ is effective: \begin{tabular}{ll} @a@ & 2-dimensional, want subscripts for coarse then fine \\ @a[all]@ & 2-dimensional, want subscripts for fine then coarse \\ @a[all][3]@ & 1-dimensional, want subscript for coarse; lock fine = 3 \\ @a[all][3][2]@ & 0-dimensional; lock coarse = 2 \end{tabular} The semantics of @-[all]@ is to dequeue from the front of the ``want subscripts'' list and re-enqueue at its back. The semantics of @-[i]@ is to dequeue from the front of the ``want subscripts'' list and lock its value to be @i@. Contiguous arrays, and slices of them, are all realized by the same underlying parameterized type. It includes stride information in its metatdata. The @-[all]@ operation is a conversion from a reference to one instantiation, to a reference to another instantiation. The running example's @all@-effective step, stated more concretely, is: \begin{tabular}{ll} @a@ & : 5 of ( 7 of float each spaced 1 float apart ) each spaced 7 floats apart \\ @a[all]@ & : 7 of ( 5 of float each spaced 7 floats apart ) each spaced 1 float apart \end{tabular} \begin{figure} \includegraphics{measuring-like-layout} \caption{Visualization of subscripting by value and by \lstinline[language=CFA,basicstyle=\ttfamily]{all}, for \lstinline[language=CFA,basicstyle=\ttfamily]{a} of type \lstinline[language=CFA,basicstyle=\ttfamily]{array( float, 5, 7 )}. The horizontal dimension represents memory addresses while vertical layout is conceptual.} \label{fig:subscr-all} \end{figure} \noindent While the latter description implies overlapping elements, Figure \ref{fig:subscr-all} shows that the overlaps only occur with unused spaces between elements. Its depictions of @a[all][...]@ show the navigation of a memory layout with nontrivial strides, that is, with ``spaced \_ floats apart'' values that are greater or smaller than the true count of valid indices times the size of a logically indexed element. Reading from the bottom up, the expression @a[all][3][2]@ shows a float, that is masquerading as a @float[7]@, for the purpose of being arranged among its peers; five such occurrences form @a[all][3]@. The tail of flatter boxes extending to the right of a proper element represents this stretching. At the next level of containment, the structure @a[all][3]@ masquerades as a @float[1]@, for the purpose of being arranged among its peers; seven such occurrences form @a[all]@. The vertical staircase arrangement represents this compression, and resulting overlapping. The new-array library defines types and operations that ensure proper elements are accessed soundly in spite of the overlapping. The private @arpk@ structure (array with explicit packing) is generic over these two types (and more): the contained element, what it is masquerading as. This structure's public interface is the @array(...)@ construction macro and the two subscript operators. Construction by @array@ initializes the masquerading-as type information to be equal to the contained-element information. Subscripting by @all@ rearranges the order of masquerading-as types to achieve, in general, nontrivial striding. Subscripting by a number consumes the masquerading-as size of the contained element type, does normal array stepping according to that size, and returns there element found there, in unmasked form. The @arpk@ structure and its @-[i]@ operator are thus defined as: \begin{lstlisting} forall( ztype(N), // length of current dimension dtype(S) | sized(S), // masquerading-as dtype E_im, // immediate element, often another array dtype E_base // base element, e.g. float, never array ) { struct arpk { S strides[N]; // so that sizeof(this) is N of S }; // expose E_im, stride by S E_im & ?[?]( arpk(N, S, E_im, E_base) & a, ptrdiff_t i ) { return (E_im &) a.strides[i]; } } \end{lstlisting} An instantiation of the @arpk@ generic is given by the @array(E_base, N0, N1, ...)@ expansion, which is @arpk( N0, Rec, Rec, E_base )@, where @Rec@ is @array(E_base, N1, ...)@. In the base case, @array(E_base)@ is just @E_base@. Because this construction uses the same value for the generic parameters @S@ and @E_im@, the resulting layout has trivial strides. Subscripting by @all@, to operate on nontrivial strides, is a dequeue-enqueue operation on the @E_im@ chain, which carries @S@ instantiations, intact, to new positions. Expressed as an operation on types, this rotation is: \begin{eqnarray*} suball( arpk(N, S, E_i, E_b) ) & = & enq( N, S, E_i, E_b ) \\ enq( N, S, E_b, E_b ) & = & arpk( N, S, E_b, E_b ) \\ enq( N, S, arpk(N', S', E_i', E_b), E_b ) & = & arpk( N', S', enq(N, S, E_i', E_b), E_b ) \end{eqnarray*} \section{Bound checks, added and removed} \CFA array subscripting is protected with runtime bound checks. Having dependent typing causes the optimizer to remove more of these bound checks than it would without them. This section provides a demonstration of the effect. The experiment compares the \CFA array system with the padded-room system [TODO:xref] most typically exemplified by Java arrays, but also reflected in the C++ pattern where restricted vector usage models a checked array. The essential feature of this padded-room system is the one-to-one correspondence between array instances and the symbolic bounds on which dynamic checks are based. The experiment compares with the C++ version to keep access to generated assembly code simple. As a control case, a simple loop (with no reused dimension sizes) is seen to get the same optimization treatment in both the \CFA and C++ versions. When the programmer treats the array's bound correctly (making the subscript ``obviously fine''), no dynamic bound check is observed in the program's optimized assembly code. But when the bounds are adjusted, such that the subscript is possibly invalid, the bound check appears in the optimized assembly, ready to catch an occurrence the mistake. TODO: paste source and assembly codes Incorporating reuse among dimension sizes is seen to give \CFA an advantage at being optimized. The case is naive matrix multiplication over a row-major encoding. TODO: paste source codes \section{Comparison with other arrays} \CFA's array is the first lightweight application of dependently-typed bound tracking to an extension of C. Other extensions of C that apply dependently-typed bound tracking are heavyweight, in that the bound tracking is part of a linearly typed ownership system that further helps guarantee statically the validity of every pointer deference. These systems, therefore, ask the programmer to convince the type checker that every pointer dereference is valid. \CFA imposes the lighter-weight obligation, with the more limited guarantee, that initially-declared bounds are respected thereafter. \CFA's array is also the first extension of C to use its tracked bounds to generate the pointer arithmetic implied by advanced allocation patterns. Other bound-tracked extensions of C either forbid certain C patterns entirely, or address the problem of \emph{verifying} that the user's provided pointer arithmetic is self-consistent. The \CFA array, applied to accordion structures [TOD: cross-reference] \emph{implies} the necessary pointer arithmetic, generated automatically, and not appearing at all in a user's program. \subsection{Safety in a padded room} Java's array [TODO:cite] is a straightforward example of assuring safety against undefined behaviour, at a cost of expressiveness for more applied properties. Consider the array parameter declarations in: \begin{tabular}{rl} C & @void f( size_t n, size_t m, float a[n][m] );@ \\ Java & @void f( float[][] a );@ \end{tabular} Java's safety against undefined behaviour assures the callee that, if @a@ is non-null, then @a.length@ is a valid access (say, evaluating to the number $\ell$) and if @i@ is in $[0, \ell)$ then @a[i]@ is a valid access. If a value of @i@ outside this range is used, a runtime error is guaranteed. In these respects, C offers no guarantees at all. Notably, the suggestion that @n@ is the intended size of the first dimension of @a@ is documentation only. Indeed, many might prefer the technically equivalent declarations @float a[][m]@ or @float (*a)[m]@ as emphasizing the ``no guarantees'' nature of an infrequently used language feature, over using the opportunity to explain a programmer intention. Moreover, even if @a[0][0]@ is valid for the purpose intended, C's basic infamous feature is the possibility of an @i@, such that @a[i][0]@ is not valid for the same purpose, and yet, its evaluation does not produce an error. Java's lack of expressiveness for more applied properties means these outcomes are possible: \begin{itemize} \item @a[0][17]@ and @a[2][17]@ are valid accesses, yet @a[1][17]@ is a runtime error, because @a[1]@ is a null pointer \item the same observation, now because @a[1]@ refers to an array of length 5 \item execution times vary, because the @float@ values within @a@ are sometimes stored nearly contiguously, and other times, not at all \end{itemize} C's array has none of these limitations, nor do any of the ``array language'' comparators discussed in this section. This Java level of safety and expressiveness is also exemplified in the C family, with the commonly given advice [TODO:cite example], for C++ programmers to use @std::vector@ in place of the C++ language's array, which is essentially the C array. The advice is that, while a vector is also more powerful (and quirky) than an array, its capabilities include options to preallocate with an upfront size, to use an available bound-checked accessor (@a.at(i)@ in place of @a[i]@), to avoid using @push_back@, and to use a vector of vectors. Used with these restrictions, out-of-bound accesses are stopped, and in-bound accesses never exercise the vector's ability to grow, which is to say, they never make the program slow to reallocate and copy, and they never invalidate the program's other references to the contained values. Allowing this scheme the same referential integrity assumption that \CFA enjoys [TODO:xref], this scheme matches Java's safety and expressiveness exactly. [TODO: decide about going deeper; some of the Java expressiveness concerns have mitigations, up to even more tradeoffs.] \subsection{Levels of dependently typed arrays} The \CFA array and the field of ``array language'' comparators all leverage dependent types to improve on the expressiveness over C and Java, accommodating examples such as: \begin{itemize} \item a \emph{zip}-style operation that consumes two arrays of equal length \item a \emph{map}-style operation whose produced length matches the consumed length \item a formulation of matrix multiplication, where the two operands must agree on a middle dimension, and where the result dimensions match the operands' outer dimensions \end{itemize} Across this field, this expressiveness is not just an available place to document such assumption, but these requirements are strongly guaranteed by default, with varying levels of statically/dynamically checked and ability to opt out. Along the way, the \CFA array also closes the safety gap (with respect to bounds) that Java has over C. Dependent type systems, considered for the purpose of bound-tracking, can be full-strength or restricted. In a full-strength dependent type system, a type can encode an arbitrarily complex predicate, with bound-tracking being an easy example. The tradeoff of this expressiveness is complexity in the checker, even typically, a potential for its nontermination. In a restricted dependent type system (purposed for bound tracking), the goal is to check helpful properties, while keeping the checker well-behaved; the other restricted checkers surveyed here, including \CFA's, always terminate. [TODO: clarify how even Idris type checking terminates] Idris is a current, general-purpose dependently typed programming language. Length checking is a common benchmark for full dependent type systems. Here, the capability being considered is to track lengths that adjust during the execution of a program, such as when an \emph{add} operation produces a collection one element longer than the one on which it started. [TODO: finish explaining what Data.Vect is and then the essence of the comparison] POINTS: here is how our basic checks look (on a system that does not have to compromise); it can also do these other cool checks, but watch how I can mess with its conservativeness and termination Two current, state-of-the-art array languages, Dex\cite{arr:dex:long} and Futhark\cite{arr:futhark:tytheory}, offer offer novel contributions concerning similar, restricted dependent types for tracking array length. Unlike \CFA, both are garbage-collected functional languages. Because they are garbage-collected, referential integrity is built-in, meaning that the heavyweight analysis, that \CFA aims to avoid, is unnecessary. So, like \CFA, the checking in question is a lightweight bounds-only analysis. Like \CFA, their checks that are conservatively limited by forbidding arithmetic in the depended-upon expression. The Futhark work discusses the working language's connection to a lambda calculus, with typing rules and a safety theorem proven in reference to an operational semantics. There is a particular emphasis on an existential type, enabling callee-determined return shapes. Dex uses a novel conception of size, embedding its quantitative information completely into an ordinary type. Futhark and full-strength dependently typed languages treat array sizes are ordinary values. Futhark restricts these expressions syntactically to variables and constants, while a full-strength dependent system does not. CFA's hybrid presentation, @forall( [N] )@, has @N@ belonging to the type system, yet has no instances. Belonging to the type system means it is inferred at a call site and communicated implicitly, like in Dex and unlike in Futhark. Having no instances means there is no type for a variable @i@ that constrains @i@ to be in the range for @N@, unlike Dex, [TODO: verify], but like Futhark. \subsection{Static safety in C extensions} \section{Future Work} \subsection{Declaration syntax} \subsection{Range slicing} \subsection{With a module system} \subsection{With described enumerations} A project in \CFA's current portfolio will improve enumerations. In the incumbent state, \CFA has C's enumerations, unmodified. I will not discuss the core of this project, which has a tall mission already, to improve type safety, maintain appropriate C compatibility and offer more flexibility about storage use. It also has a candidate stretch goal, to adapt \CFA's @forall@ generic system to communicate generalized enumerations: \begin{lstlisting} forall( T | is_enum(T) ) void show_in_context( T val ) { for( T i ) { string decorator = ""; if ( i == val-1 ) decorator = "< ready"; if ( i == val ) decorator = "< go" ; sout | i | decorator; } } enum weekday { mon, tue, wed = 500, thu, fri }; show_in_context( wed ); \end{lstlisting} with output \begin{lstlisting} mon tue < ready wed < go thu fri \end{lstlisting} The details in this presentation aren't meant to be taken too precisely as suggestions for how it should look in \CFA. But the example shows these abilities: \begin{itemize} \item a built-in way (the @is_enum@ trait) for a generic routine to require enumeration-like information about its instantiating type \item an implicit implementation of the trait whenever a user-written enum occurs (@weekday@'s declaration implies @is_enum@) \item a total order over the enumeration constants, with predecessor/successor (@val-1@) available, and valid across gaps in values (@tue == 1 && wed == 500 && tue == wed - 1@) \item a provision for looping (the @for@ form used) over the values of the type. \end{itemize} If \CFA gets such a system for describing the list of values in a type, then \CFA arrays are poised to move from the Futhark level of expressiveness, up to the Dex level. [TODO: introduce Ada in the comparators] In Ada and Dex, an array is conceived as a function whose domain must satisfy only certain structural assumptions, while in C, C++, Java, Futhark and \CFA today, the domain is a prefix of the natural numbers. The generality has obvious aesthetic benefits for programmers working on scheduling resources to weekdays, and for programmers who prefer to count from an initial number of their own choosing. This change of perspective also lets us remove ubiquitous dynamic bound checks. [TODO: xref] discusses how automatically inserted bound checks can often be optimized away. But this approach is unsatisfying to a programmer who believes she has written code in which dynamic checks are unnecessary, but now seeks confirmation. To remove the ubiquitous dynamic checking is to say that an ordinary subscript operation is only valid when it can be statically verified to be in-bound (and so the ordinary subscript is not dynamically checked), and an explicit dynamic check is available when the static criterion is impractical to meet. [TODO, fix confusion: Idris has this arrangement of checks, but still the natural numbers as the domain.] The structural assumptions required for the domain of an array in Dex are given by the trait (there, ``interface'') @Ix@, which says that the parameter @n@ is a type (which could take an argument like @weekday@) that provides two-way conversion with the integers and a report on the number of values. Dex's @Ix@ is analogous the @is_enum@ proposed for \CFA above. \begin{lstlisting} interface Ix n get_size n : Unit -> Int ordinal : n -> Int unsafe_from_ordinal n : Int -> n \end{lstlisting} Dex uses this foundation of a trait (as an array type's domain) to achieve polymorphism over shapes. This flavour of polymorphism lets a function be generic over how many (and the order of) dimensions a caller uses when interacting with arrays communicated with this function. Dex's example is a routine that calculates pointwise differences between two samples. Done with shape polymorphism, one function body is equally applicable to a pair of single-dimensional audio clips (giving a single-dimensional result) and a pair of two-dimensional photographs (giving a two-dimensional result). In both cases, but with respectively dimensioned interpretations of ``size,'' this function requires the argument sizes to match, and it produces a result of the that size. The polymorphism plays out with the pointwise-difference routine advertising a single-dimensional interface whose domain type is generic. In the audio instantiation, the duration-of-clip type argument is used for the domain. In the photograph instantiation, it's the tuple-type of $ \langle \mathrm{img\_wd}, \mathrm{img\_ht} \rangle $. This use of a tuple-as-index is made possible by the built-in rule for implementing @Ix@ on a pair, given @Ix@ implementations for its elements \begin{lstlisting} instance {a b} [Ix a, Ix b] Ix (a & b) get_size = \(). size a * size b ordinal = \(i, j). (ordinal i * size b) + ordinal j unsafe_from_ordinal = \o. bs = size b (unsafe_from_ordinal a (idiv o bs), unsafe_from_ordinal b (rem o bs)) \end{lstlisting} and by a user-provided adapter expression at the call site that shows how to indexing with a tuple is backed by indexing each dimension at a time \begin{lstlisting} img_trans :: (img_wd,img_ht)=>Real img_trans.(i,j) = img.i.j result = pairwise img_trans \end{lstlisting} [TODO: cite as simplification of example from https://openreview.net/pdf?id=rJxd7vsWPS section 4] In the case of adapting this pattern to \CFA, my current work provides an adapter from ``successively subscripted'' to ``subscripted by tuple,'' so it is likely that generalizing my adapter beyond ``subscripted by @ptrdiff_t@'' is sufficient to make a user-provided adapter unnecessary. \subsection{Retire pointer arithmetic} \section{\CFA} XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX \\ moved from background chapter \\ XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX \\ Traditionally, fixing C meant leaving the C-ism alone, while providing a better alternative beside it. (For later: That's what I offer with array.hfa, but in the future-work vision for arrays, the fix includes helping programmers stop accidentally using a broken C-ism.) \subsection{\CFA features interacting with arrays} Prior work on \CFA included making C arrays, as used in C code from the wild, work, if this code is fed into @cfacc@. The quality of this this treatment was fine, with no more or fewer bugs than is typical. More mixed results arose with feeding these ``C'' arrays into preexisting \CFA features. A notable success was with the \CFA @alloc@ function, which type information associated with a polymorphic return type replaces @malloc@'s use of programmer-supplied size information. \begin{cfa} // C, library void * malloc( size_t ); // C, user struct tm * el1 = malloc( sizeof(struct tm) ); struct tm * ar1 = malloc( 10 * sizeof(struct tm) ); // CFA, library forall( T * ) T * alloc(); // CFA, user tm * el2 = alloc(); tm (*ar2)[10] = alloc(); \end{cfa} The alloc polymorphic return compiles into a hidden parameter, which receives a compiler-generated argument. This compiler's argument generation uses type information from the left-hand side of the initialization to obtain the intended type. Using a compiler-produced value eliminates an opportunity for user error. TODO: fix in following: even the alloc call gives bad code gen: verify it was always this way; walk back the wording about things just working here; assignment (rebind) seems to offer workaround, as in bkgd-cfa-arrayinteract.cfa Bringing in another \CFA feature, reference types, both resolves a sore spot of the last example, and gives a first example of an array-interaction bug. In the last example, the choice of ``pointer to array'' @ar2@ breaks a parallel with @ar1@. They are not subscripted in the same way. \begin{cfa} ar1[5]; (*ar2)[5]; \end{cfa} Using ``reference to array'' works at resolving this issue. TODO: discuss connection with Doug-Lea \CC proposal. \begin{cfa} tm (&ar3)[10] = *alloc(); ar3[5]; \end{cfa} The implicit size communication to @alloc@ still works in the same ways as for @ar2@. Using proper array types (@ar2@ and @ar3@) addresses a concern about using raw element pointers (@ar1@), albeit a theoretical one. TODO xref C standard does not claim that @ar1@ may be subscripted, because no stage of interpreting the construction of @ar1@ has it be that ``there is an \emph{array object} here.'' But both @*ar2@ and the referent of @ar3@ are the results of \emph{typed} @alloc@ calls, where the type requested is an array, making the result, much more obviously, an array object. The ``reference to array'' type has its sore spots too. TODO see also @dimexpr-match-c/REFPARAM_CALL@ (under @TRY_BUG_1@) TODO: I fixed a bug associated with using an array as a T. I think. Did I really? What was the bug?