Changeset cf85f96
- Timestamp:
- Jun 17, 2021, 10:31:39 PM (4 years ago)
- Branches:
- ADT, ast-experimental, enum, forall-pointer-decay, jacob/cs343-translation, master, new-ast-unique-expr, pthread-emulation, qualifiedEnum
- Children:
- df7597e0
- Parents:
- 12daa43
- Files:
-
- 3 edited
Legend:
- Unmodified
- Added
- Removed
-
libcfa/src/device/cpu.cfa
r12daa43 rcf85f96 256 256 } 257 257 258 struct llc_map_t { 259 raw_cache_instance * raw; 260 unsigned count; 261 unsigned start; 262 }; 263 258 264 // returns an allocate list of all the different distinct last level caches 259 static [* idx_range_t, size_t cnt] distinct_llcs(unsigned cpus, unsigned llc_idx, raw_cache_instance ** raw) {265 static [*llc_map_t, size_t cnt] distinct_llcs(unsigned cpus, unsigned llc_idx, raw_cache_instance ** raw) { 260 266 // Allocate at least one element 261 idx_range_t* ranges = alloc();267 llc_map_t* ranges = alloc(); 262 268 size_t range_cnt = 1; 263 269 264 270 // Initialize with element 0 265 *ranges = raw[0][llc_idx].range; 271 ranges->raw = &raw[0][llc_idx]; 272 ranges->count = 0; 273 ranges->start = -1u; 266 274 267 275 // Go over all other cpus 268 276 CPU_LOOP: for(i; 1~cpus) { 269 277 // Check if the range is already there 270 idx_range_t candidate = raw[i][llc_idx].range;278 raw_cache_instance * candidate = &raw[i][llc_idx]; 271 279 for(j; range_cnt) { 272 idx_range_texist = ranges[j];280 llc_map_t & exist = ranges[j]; 273 281 // If the range is already there just jump to the next cpu 274 if(0 == strcmp(candidate , exist)) continue CPU_LOOP;282 if(0 == strcmp(candidate->range, exist.raw->range)) continue CPU_LOOP; 275 283 } 276 284 277 285 // The range wasn't there, added to the list 278 286 ranges = alloc(range_cnt + 1, ranges`realloc); 279 ranges[range_cnt] = candidate; 287 ranges[range_cnt].raw = candidate; 288 ranges[range_cnt].count = 0; 289 ranges[range_cnt].start = -1u; 280 290 range_cnt++; 281 291 } … … 287 297 struct cpu_pairing_t { 288 298 unsigned cpu; 289 unsigned llc_id;299 unsigned id; 290 300 }; 291 301 292 302 int ?<?( cpu_pairing_t lhs, cpu_pairing_t rhs ) { 293 return lhs. llc_id < rhs.llc_id;294 } 295 296 static [[]cpu_pairing_t] get_cpu_pairings(unsigned cpus, raw_cache_instance ** raw, idx_range_t * maps, size_t map_cnt) {303 return lhs.id < rhs.id; 304 } 305 306 static [[]cpu_pairing_t] get_cpu_pairings(unsigned cpus, raw_cache_instance ** raw, llc_map_t * maps, size_t map_cnt) { 297 307 cpu_pairing_t * pairings = alloc(cpus); 298 308 … … 301 311 idx_range_t want = raw[i][0].range; 302 312 MAP_LOOP: for(j; map_cnt) { 303 if(0 != strcmp(want, maps[j] )) continue MAP_LOOP;304 305 pairings[i]. llc_id = j;313 if(0 != strcmp(want, maps[j].raw->range)) continue MAP_LOOP; 314 315 pairings[i].id = j; 306 316 continue CPU_LOOP; 307 317 } … … 312 322 return pairings; 313 323 } 324 325 #include <fstream.hfa> 314 326 315 327 extern "C" { … … 336 348 337 349 // Find number of distinct cache instances 338 idx_range_t * maps;350 llc_map_t * maps; 339 351 size_t map_cnt; 340 352 [maps, map_cnt] = distinct_llcs(cpus, cache_levels - llc, raw); 341 353 342 354 #if defined(__CFA_WITH_VERIFY__) 355 // Verify that the caches cover the all the cpus 343 356 { 344 unsigned width = 0; 357 unsigned width1 = 0; 358 unsigned width2 = 0; 345 359 for(i; map_cnt) { 346 360 const char * _; 347 width += read_width(maps[i], strlen(maps[i]), &_); 361 width1 += read_width(maps[i].raw->range, strlen(maps[i].raw->range), &_); 362 width2 += maps[i].raw->width; 348 363 } 349 verify(width == cpus); 364 verify(width1 == cpus); 365 verify(width2 == cpus); 350 366 } 351 367 #endif … … 357 373 qsort(pairings, cpus); 358 374 359 unsigned llc_width = raw[0][cache_levels - llc].width; 360 361 // From the mappins build the actual cpu map we want 375 { 376 unsigned it = 0; 377 for(i; cpus) { 378 unsigned llc_id = pairings[i].id; 379 if(maps[llc_id].start == -1u) { 380 maps[llc_id].start = it; 381 it += maps[llc_id].raw->width; 382 /* paranoid */ verify(maps[llc_id].start < it); 383 /* paranoid */ verify(it != -1u); 384 } 385 } 386 /* paranoid */ verify(it == cpus); 387 } 388 389 // From the mappings build the actual cpu map we want 362 390 struct cpu_map_entry_t * entries = alloc(cpus); 363 391 for(i; cpus) { entries[i].count = 0; } 364 392 for(i; cpus) { 393 /* paranoid */ verify(pairings[i].id < map_cnt); 365 394 unsigned c = pairings[i].cpu; 366 entries[c].start = pairings[i].llc_id * llc_width; 367 entries[c].count = llc_width; 395 unsigned llc_id = pairings[i].id; 396 unsigned width = maps[llc_id].raw->width; 397 unsigned start = maps[llc_id].start; 398 unsigned self = start + (maps[llc_id].count++); 399 entries[c].count = width; 400 entries[c].start = start; 401 entries[c].self = self; 368 402 } 369 403 -
libcfa/src/device/cpu.hfa
r12daa43 rcf85f96 17 17 18 18 struct cpu_map_entry_t { 19 unsigned self; 19 20 unsigned start; 20 21 unsigned count; -
tests/device/cpu.cfa
r12daa43 rcf85f96 17 17 #include <fstream.hfa> 18 18 #include <device/cpu.hfa> 19 #include <stdlib.hfa> 20 21 #include <errno.h> 22 #include <stdio.h> 23 #include <string.h> 24 #include <unistd.h> 25 19 26 extern "C" { 27 #include <dirent.h> 28 #include <sys/types.h> 29 #include <sys/stat.h> 20 30 #include <sys/sysinfo.h> 31 #include <fcntl.h> 32 } 33 34 // go through a directory calling fn on each file 35 static int iterate_dir( const char * path, void (*fn)(struct dirent * ent) ) { 36 // open the directory 37 DIR *dir = opendir(path); 38 if(dir == 0p) { return ENOTDIR; } 39 40 // call fn for each 41 struct dirent * ent; 42 while ((ent = readdir(dir)) != 0p) { 43 fn( ent ); 44 } 45 46 // no longer need this 47 closedir(dir); 48 return 0; 49 } 50 51 // count the number of directories with the specified prefix 52 // the directories counted have the form '[prefix]N' where prefix is the parameter 53 // and N is an base 10 integer. 54 static int count_prefix_dirs(const char * path, const char * prefix) { 55 // read the directory and find the cpu count 56 // and make sure everything is as expected 57 int max = -1; 58 int count = 0; 59 void lambda(struct dirent * ent) { 60 // were are looking for prefixX, where X is a number 61 // check that it starts with 'cpu 62 char * s = strstr(ent->d_name, prefix); 63 if(s == 0p) { return; } 64 if(s != ent->d_name) { return; } 65 66 // check that the next part is a number 67 s += strlen(prefix); 68 char * end; 69 long int val = strtol(s, &end, 10); 70 if(*end != '\0' || val < 0) { return; } 71 72 // check that it's a directory 73 if(ent->d_type != DT_DIR) { return; } 74 75 // it's a match! 76 max = max(val, max); 77 count++; 78 } 79 iterate_dir(path, lambda); 80 81 /* paranoid */ verifyf(count == max + 1, "Inconsistent %s count, counted %d, but max %s was %d", prefix, count, prefix, (int)max); 82 83 return count; 84 } 85 86 // Count number of cache *indexes* in the system 87 // cache indexes are distinct from cache level as Data or Instruction cache 88 // can share a level but not an index 89 // PITFALL: assumes all cpus have the same indexes as cpu0 90 static int count_cache_indexes(void) { 91 return count_prefix_dirs("/sys/devices/system/cpu/cpu0/cache", "index"); 92 } 93 94 // read information about a spcficic cache index/cpu file into the output buffer 95 static size_t read_cpuidxinfo_into(unsigned cpu, unsigned idx, const char * file, char * out, size_t out_len) { 96 // Pick the file we want and read it 97 char buf[128]; 98 /* paranoid */ __attribute__((unused)) int len = 99 snprintf(buf, 128, "/sys/devices/system/cpu/cpu%u/cache/index%u/%s", cpu, idx, file); 100 /* paranoid */ verifyf(len > 0, "Could not generate '%s' filename for cpu %u, index %u", file, cpu, idx); 101 102 int fd = open(buf, 0, O_RDONLY); 103 /* paranoid */ verifyf(fd > 0, "Could not open file '%s'", buf); 104 105 ssize_t r = read(fd, out, out_len); 106 /* paranoid */ verifyf(r > 0, "Could not read file '%s'", buf); 107 108 /* paranoid */ __attribute__((unused)) int ret = 109 close(fd); 110 /* paranoid */ verifyf(ret == 0, "Could not close file '%s'", buf); 111 112 out[r-1] = '\0'; 113 return r-1; 114 } 115 116 unsigned find_idx() { 117 int idxs = count_cache_indexes(); 118 119 unsigned found_level = 0; 120 unsigned found = -1u; 121 for(i; idxs) { 122 unsigned idx = idxs - 1 - i; 123 char buf[32]; 124 125 // Level is the cache level: higher means bigger and slower 126 read_cpuidxinfo_into(0, idx, "level", buf, 32); 127 char * end; 128 unsigned long level = strtoul(buf, &end, 10); 129 /* paranoid */ verifyf(level <= 250, "Cpu %u has more than 250 levels of cache, that doesn't sound right", 0); 130 /* paranoid */ verify(*end == '\0'); 131 132 if(found_level < level) { 133 found_level = level; 134 found = idx; 135 } 136 } 137 138 /* paranoid */ verify(found != -1u); 139 return found; 21 140 } 22 141 23 142 int main() { 143 //----------------------------------------------------------------------- 24 144 int ret1 = get_nprocs(); 25 145 int ret2 = cpu_info.hthrd_count; … … 31 151 } 32 152 153 //----------------------------------------------------------------------- 154 // Make sure no one has the same self 155 for(ime; cpu_info.hthrd_count) { 156 unsigned me = cpu_info.llc_map[ime].self; 157 { 158 unsigned s = cpu_info.llc_map[ime].start; 159 unsigned e = s + cpu_info.llc_map[ime].count; 160 if(me < s || me >= e) { 161 sout | "CPU" | ime | "outside of it's own map: " | s | "<=" | me | "<" | e; 162 } 163 } 164 165 166 for(ithem; cpu_info.hthrd_count) { 167 if(ime == ithem) continue; 168 169 unsigned them = cpu_info.llc_map[ithem].self; 170 if(me == them) { 171 sout | "CPU" | ime | "has conflicting self id with" | ithem | "(" | me | ")"; 172 } 173 } 174 } 175 176 177 //----------------------------------------------------------------------- 178 unsigned idx = find_idx(); 179 // For all procs check mapping is consistent 180 for(cpu_me; cpu_info.hthrd_count) { 181 char buf_me[32]; 182 size_t len_me = read_cpuidxinfo_into(cpu_me, idx, "shared_cpu_list", buf_me, 32); 183 for(cpu_them; cpu_info.hthrd_count) { 184 if(cpu_me == cpu_them) continue; 185 char buf_them[32]; 186 size_t len_them = read_cpuidxinfo_into(cpu_them, idx, "shared_cpu_list", buf_them, 32); 187 188 bool match_file = len_them == len_me && 0 == strncmp(buf_them, buf_me, len_me); 189 bool match_info = cpu_info.llc_map[cpu_me].start == cpu_info.llc_map[cpu_them].start && cpu_info.llc_map[cpu_me].count == cpu_info.llc_map[cpu_them].count; 190 191 if(match_file != match_info) { 192 sout | "CPU" | cpu_me | "and" | cpu_them | "have inconsitent file and cpu_info"; 193 sout | cpu_me | ": <" | cpu_info.llc_map[cpu_me ].start | "," | cpu_info.llc_map[cpu_me ].count | "> '" | buf_me | "'"; 194 sout | cpu_me | ": <" | cpu_info.llc_map[cpu_them].start | "," | cpu_info.llc_map[cpu_them].count | "> '" | buf_them | "'"; 195 } 196 } 197 } 33 198 }
Note: See TracChangeset
for help on using the changeset viewer.