Changeset bf8112b for doc/theses/mike_brooks_MMath
- Timestamp:
- Apr 28, 2026, 5:09:43 AM (20 hours ago)
- Branches:
- master
- Children:
- 5faa3a5
- Parents:
- 9a35b43
- File:
-
- 1 edited
-
doc/theses/mike_brooks_MMath/list.tex (modified) (3 diffs)
Legend:
- Unmodified
- Added
- Removed
-
doc/theses/mike_brooks_MMath/list.tex
r9a35b43 rbf8112b 1379 1379 1380 1380 \VRef[Figure]{fig:plot-list-1ord} gives the first-order effects. 1381 The first breakdown, architecture/size-zone (left), shows the overall performance of all 12 experiment on the two different hardware architectures for small andmedium lists (624 / 4 = 156 experiments per column).1381 The first breakdown, architecture/size-zone (left), shows the overall performance of all configurations, split by the two different hardware architectures and by small \vs medium lists (624 / 4 = 156 experiments per column). 1382 1382 % The relative experiment duration for each experiment is shown as a bar in each column and the black bar in that column shows the average of all 12 experiments. 1383 1383 By inspection of the averages, Intel runs faster than AMD. 1384 1384 Within an architecture, the small zone (lists of 4--16 elements) runs faster than the medium zone (lists of 50--200 elements). 1385 The overall slower execution on the AMD results from its smaller L3cache \vs the larger cache on the Intel.1385 The overall slower execution on the AMD results from its smaller cache \vs the larger cache on the Intel. 1386 1386 (No NUMA effects for these list sizes.) 1387 1387 Specifically, a 20\% standard deviation exists here, between the means of the four physical-effect categories. … … 1393 1393 The second breakdown, use case (middle), shows the overall performance for each of the 12 use cases from \VRef[Figure]{f:ExperimentOperations} (624 / 12 = 52 experiments per column). 1394 1394 % A similar situation comes from \VRef[Figure]{fig:plot-list-1ord}'s second comparison, by use case. 1395 While specific differences do occur, like framework X doing better on stacks than on queues, the overall range of the standard deviation of the individual use cases' means is only 9\%, indicating no unusual cases.1395 The standard deviation of the individual use cases' means is 10\%. 1396 1396 A more detailed analysis occurs in the discussion of \VRef[Figure]{fig:plot-list-2ord}. 1397 1397 % But they are so irrelevant to the issue of picking a winning framework that it is sufficient here to number the use cases opaquely. … … 1401 1401 The third breakdown, framework (right), shows the overall performance of the 4 list implementations (624 / 3.25 = 192). 1402 1402 Here, \CFA runs similarly to \uCpp and LQ-@list@ runs similarly to @tailq@. 1403 The standard deviation of the frameworks' means is 8\%.1403 The standard deviation of the frameworks' means is 7\%. 1404 1404 % Framework choice has, therefore, less impact on your speed than the lottery tickets you already hold. 1405 1405 Now, \CFA/\uCpp run slower than LQ-@list@/@tailq@ by 15\%, a fact explored further in \VRef{s:SweetSoreSpots}. 1406 1406 But so too does use case X typically beat use case II by 38\%. 1407 1407 As does a small size on the Intel typically beat a medium size on the AMD by 66\%. 1408 Hence, architecture and usage pattern s have a significant affect on the specificframework.1408 Hence, architecture and usage pattern have a more significant effect on speed than the selection of a framework. 1409 1409 1410 1410
Note:
See TracChangeset
for help on using the changeset viewer.