Changeset 04b73b6 for libcfa/src/concurrency
- Timestamp:
- Jul 23, 2020, 3:37:05 PM (6 years ago)
- Branches:
- ADT, arm-eh, ast-experimental, enum, forall-pointer-decay, jacob/cs343-translation, master, new-ast, new-ast-unique-expr, pthread-emulation, qualifiedEnum, stuck-waitfor-destruct
- Children:
- f4ec4a90
- Parents:
- f0c3120 (diff), e262b5e (diff)
Note: this is a merge changeset, the changes displayed below correspond to the merge itself.
Use the(diff)links above to see all the changes relative to each parent. - Location:
- libcfa/src/concurrency
- Files:
-
- 7 edited
-
io.cfa (modified) (4 diffs)
-
iocall.cfa (modified) (24 diffs)
-
kernel.cfa (modified) (8 diffs)
-
ready_queue.cfa (modified) (5 diffs)
-
snzi.hfa (modified) (1 diff)
-
stats.cfa (modified) (3 diffs)
-
stats.hfa (modified) (3 diffs)
Legend:
- Unmodified
- Added
- Removed
-
libcfa/src/concurrency/io.cfa
rf0c3120 r04b73b6 16 16 #if defined(__CFA_DEBUG__) 17 17 // #define __CFA_DEBUG_PRINT_IO__ 18 #define __CFA_DEBUG_PRINT_IO_CORE__18 // #define __CFA_DEBUG_PRINT_IO_CORE__ 19 19 #endif 20 20 … … 22 22 #include "bitmanip.hfa" 23 23 24 #if !defined( HAVE_LINUX_IO_URING_H)24 #if !defined(CFA_HAVE_LINUX_IO_URING_H) 25 25 void __kernel_io_startup( cluster &, unsigned, bool ) { 26 26 // Nothing to do without io_uring … … 490 490 static uint32_t __release_consumed_submission( struct __io_data & ring ); 491 491 492 static inline void process(struct io_uring_cqe & cqe, struct __processor_id_t * id ) { 493 struct __io_user_data_t * data = (struct __io_user_data_t *)(uintptr_t)cqe.user_data; 494 __cfadbg_print_safe( io, "Kernel I/O : Syscall completed : cqe %p, result %d for %p\n", data, cqe.res, data->thrd ); 495 496 data->result = cqe.res; 497 if(!id) { unpark( data->thrd __cfaabi_dbg_ctx2 ); } 498 else { __unpark( id, data->thrd __cfaabi_dbg_ctx2 ); } 499 } 500 492 501 // Process a single completion message from the io_uring 493 502 // This is NOT thread-safe … … 538 547 /* paranoid */ verify(&cqe); 539 548 540 struct __io_user_data_t * data = (struct __io_user_data_t *)(uintptr_t)cqe.user_data; 541 __cfadbg_print_safe( io, "Kernel I/O : Performed reading io cqe %p, result %d for %p\n", data, cqe.res, data->thrd ); 542 543 data->result = cqe.res; 544 if(!mask) { unpark( data->thrd __cfaabi_dbg_ctx2 ); } 545 else { __unpark( &ring.poller.slow.id, data->thrd __cfaabi_dbg_ctx2 ); } 549 process( cqe, !mask ? (struct __processor_id_t *)0p : &ring.poller.slow.id ); 546 550 } 547 551 -
libcfa/src/concurrency/iocall.cfa
rf0c3120 r04b73b6 20 20 //============================================================================================= 21 21 22 #if defined( HAVE_LINUX_IO_URING_H)22 #if defined(CFA_HAVE_LINUX_IO_URING_H) 23 23 #include <stdint.h> 24 24 #include <linux/io_uring.h> … … 122 122 #if defined(HAVE_PREADV2) 123 123 ssize_t cfa_preadv2(int fd, const struct iovec *iov, int iovcnt, off_t offset, int flags) { 124 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_READV)124 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_READV) 125 125 return preadv2(fd, iov, iovcnt, offset, flags); 126 126 #else … … 134 134 135 135 ssize_t cfa_preadv2_fixed(int fd, const struct iovec *iov, int iovcnt, off_t offset, int flags) { 136 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_READV)136 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_READV) 137 137 return preadv2(fd, iov, iovcnt, offset, flags); 138 138 #else … … 149 149 #if defined(HAVE_PWRITEV2) 150 150 ssize_t cfa_pwritev2(int fd, const struct iovec *iov, int iovcnt, off_t offset, int flags) { 151 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_WRITEV)151 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_WRITEV) 152 152 return pwritev2(fd, iov, iovcnt, offset, flags); 153 153 #else … … 162 162 163 163 int cfa_fsync(int fd) { 164 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_FSYNC)164 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_FSYNC) 165 165 return fsync(fd); 166 166 #else … … 174 174 175 175 int cfa_sync_file_range(int fd, int64_t offset, int64_t nbytes, unsigned int flags) { 176 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_SYNC_FILE_RANGE)176 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_SYNC_FILE_RANGE) 177 177 return sync_file_range(fd, offset, nbytes, flags); 178 178 #else … … 190 190 191 191 ssize_t cfa_sendmsg(int sockfd, const struct msghdr *msg, int flags) { 192 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_SENDMSG)192 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_SENDMSG) 193 193 return sendmsg(sockfd, msg, flags); 194 194 #else … … 203 203 204 204 ssize_t cfa_recvmsg(int sockfd, struct msghdr *msg, int flags) { 205 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_RECVMSG)205 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_RECVMSG) 206 206 return recvmsg(sockfd, msg, flags); 207 207 #else … … 216 216 217 217 ssize_t cfa_send(int sockfd, const void *buf, size_t len, int flags) { 218 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_SEND)218 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_SEND) 219 219 return send( sockfd, buf, len, flags ); 220 220 #else … … 231 231 232 232 ssize_t cfa_recv(int sockfd, void *buf, size_t len, int flags) { 233 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_RECV)233 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_RECV) 234 234 return recv( sockfd, buf, len, flags ); 235 235 #else … … 246 246 247 247 int cfa_accept4(int sockfd, struct sockaddr *addr, socklen_t *addrlen, int flags) { 248 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_ACCEPT)248 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_ACCEPT) 249 249 return accept4( sockfd, addr, addrlen, flags ); 250 250 #else … … 261 261 262 262 int cfa_connect(int sockfd, const struct sockaddr *addr, socklen_t addrlen) { 263 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_CONNECT)263 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_CONNECT) 264 264 return connect( sockfd, addr, addrlen ); 265 265 #else … … 275 275 276 276 int cfa_fallocate(int fd, int mode, uint64_t offset, uint64_t len) { 277 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_FALLOCATE)277 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_FALLOCATE) 278 278 return fallocate( fd, mode, offset, len ); 279 279 #else … … 290 290 291 291 int cfa_fadvise(int fd, uint64_t offset, uint64_t len, int advice) { 292 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_FADVISE)292 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_FADVISE) 293 293 return posix_fadvise( fd, offset, len, advice ); 294 294 #else … … 305 305 306 306 int cfa_madvise(void *addr, size_t length, int advice) { 307 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_MADVISE)307 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_MADVISE) 308 308 return madvise( addr, length, advice ); 309 309 #else … … 320 320 321 321 int cfa_openat(int dirfd, const char *pathname, int flags, mode_t mode) { 322 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_OPENAT)322 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_OPENAT) 323 323 return openat( dirfd, pathname, flags, mode ); 324 324 #else … … 335 335 336 336 int cfa_close(int fd) { 337 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_CLOSE)337 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_CLOSE) 338 338 return close( fd ); 339 339 #else … … 349 349 struct statx; 350 350 int cfa_statx(int dirfd, const char *pathname, int flags, unsigned int mask, struct statx *statxbuf) { 351 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_STATX)351 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_STATX) 352 352 #if defined(__NR_statx) 353 353 return syscall( __NR_statx, dirfd, pathname, flags, mask, statxbuf ); … … 367 367 368 368 ssize_t cfa_read(int fd, void *buf, size_t count) { 369 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_READ)369 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_READ) 370 370 return read( fd, buf, count ); 371 371 #else … … 379 379 380 380 ssize_t cfa_write(int fd, void *buf, size_t count) { 381 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_WRITE)381 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_WRITE) 382 382 return read( fd, buf, count ); 383 383 #else … … 391 391 392 392 ssize_t cfa_splice(int fd_in, loff_t *off_in, int fd_out, loff_t *off_out, size_t len, unsigned int flags) { 393 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_SPLICE)393 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_SPLICE) 394 394 return splice( fd_in, off_in, fd_out, off_out, len, flags ); 395 395 #else … … 405 405 } 406 406 407 ssize_t cfa_splice(int fd_in, loff_t *off_in, int fd_out, loff_t *off_out, size_t len, unsigned int flags, int in_flags, int out_flags) { 408 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_SPLICE) 409 return splice( fd_in, off_in, fd_out, off_out, len, flags ); 410 #else 411 __submit_prelude 412 413 (*sqe){ IORING_OP_SPLICE, fd_out, 0p, len, off_out }; 414 sqe->splice_fd_in = fd_in; 415 sqe->splice_off_in = off_in; 416 sqe->splice_flags = flags | out_flags; 417 sqe->flags = in_flags; 418 419 __submit_wait 420 #endif 421 } 422 407 423 ssize_t cfa_tee(int fd_in, int fd_out, size_t len, unsigned int flags) { 408 #if !defined( HAVE_LINUX_IO_URING_H) || !defined(IORING_OP_TEE)424 #if !defined(CFA_HAVE_LINUX_IO_URING_H) || !defined(CFA_HAVE_IORING_OP_TEE) 409 425 return tee( fd_in, fd_out, len, flags ); 410 426 #else … … 429 445 430 446 bool has_user_level_blocking( fptr_t func ) { 431 #if defined( HAVE_LINUX_IO_URING_H)447 #if defined(CFA_HAVE_LINUX_IO_URING_H) 432 448 #if defined(HAVE_PREADV2) 433 449 if( /*func == (fptr_t)preadv2 || */ 434 450 func == (fptr_t)cfa_preadv2 ) 435 #define _CFA_IO_FEATURE_ IORING_OP_READV ,436 return IS_DEFINED( IORING_OP_READV);451 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_READV , 452 return IS_DEFINED(CFA_HAVE_IORING_OP_READV); 437 453 #endif 438 454 … … 440 456 if( /*func == (fptr_t)pwritev2 || */ 441 457 func == (fptr_t)cfa_pwritev2 ) 442 #define _CFA_IO_FEATURE_ IORING_OP_WRITEV ,443 return IS_DEFINED( IORING_OP_WRITEV);458 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_WRITEV , 459 return IS_DEFINED(CFA_HAVE_IORING_OP_WRITEV); 444 460 #endif 445 461 446 462 if( /*func == (fptr_t)fsync || */ 447 463 func == (fptr_t)cfa_fsync ) 448 #define _CFA_IO_FEATURE_ IORING_OP_FSYNC ,449 return IS_DEFINED( IORING_OP_FSYNC);464 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_FSYNC , 465 return IS_DEFINED(CFA_HAVE_IORING_OP_FSYNC); 450 466 451 467 if( /*func == (fptr_t)ync_file_range || */ 452 468 func == (fptr_t)cfa_sync_file_range ) 453 #define _CFA_IO_FEATURE_ IORING_OP_SYNC_FILE_RANGE ,454 return IS_DEFINED( IORING_OP_SYNC_FILE_RANGE);469 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_SYNC_FILE_RANGE , 470 return IS_DEFINED(CFA_HAVE_IORING_OP_SYNC_FILE_RANGE); 455 471 456 472 if( /*func == (fptr_t)sendmsg || */ 457 473 func == (fptr_t)cfa_sendmsg ) 458 #define _CFA_IO_FEATURE_ IORING_OP_SENDMSG ,459 return IS_DEFINED( IORING_OP_SENDMSG);474 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_SENDMSG , 475 return IS_DEFINED(CFA_HAVE_IORING_OP_SENDMSG); 460 476 461 477 if( /*func == (fptr_t)recvmsg || */ 462 478 func == (fptr_t)cfa_recvmsg ) 463 #define _CFA_IO_FEATURE_ IORING_OP_RECVMSG ,464 return IS_DEFINED( IORING_OP_RECVMSG);479 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_RECVMSG , 480 return IS_DEFINED(CFA_HAVE_IORING_OP_RECVMSG); 465 481 466 482 if( /*func == (fptr_t)send || */ 467 483 func == (fptr_t)cfa_send ) 468 #define _CFA_IO_FEATURE_ IORING_OP_SEND ,469 return IS_DEFINED( IORING_OP_SEND);484 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_SEND , 485 return IS_DEFINED(CFA_HAVE_IORING_OP_SEND); 470 486 471 487 if( /*func == (fptr_t)recv || */ 472 488 func == (fptr_t)cfa_recv ) 473 #define _CFA_IO_FEATURE_ IORING_OP_RECV ,474 return IS_DEFINED( IORING_OP_RECV);489 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_RECV , 490 return IS_DEFINED(CFA_HAVE_IORING_OP_RECV); 475 491 476 492 if( /*func == (fptr_t)accept4 || */ 477 493 func == (fptr_t)cfa_accept4 ) 478 #define _CFA_IO_FEATURE_ IORING_OP_ACCEPT ,479 return IS_DEFINED( IORING_OP_ACCEPT);494 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_ACCEPT , 495 return IS_DEFINED(CFA_HAVE_IORING_OP_ACCEPT); 480 496 481 497 if( /*func == (fptr_t)connect || */ 482 498 func == (fptr_t)cfa_connect ) 483 #define _CFA_IO_FEATURE_ IORING_OP_CONNECT ,484 return IS_DEFINED( IORING_OP_CONNECT);499 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_CONNECT , 500 return IS_DEFINED(CFA_HAVE_IORING_OP_CONNECT); 485 501 486 502 if( /*func == (fptr_t)fallocate || */ 487 503 func == (fptr_t)cfa_fallocate ) 488 #define _CFA_IO_FEATURE_ IORING_OP_FALLOCATE ,489 return IS_DEFINED( IORING_OP_FALLOCATE);504 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_FALLOCATE , 505 return IS_DEFINED(CFA_HAVE_IORING_OP_FALLOCATE); 490 506 491 507 if( /*func == (fptr_t)posix_fadvise || */ 492 508 func == (fptr_t)cfa_fadvise ) 493 #define _CFA_IO_FEATURE_ IORING_OP_FADVISE ,494 return IS_DEFINED( IORING_OP_FADVISE);509 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_FADVISE , 510 return IS_DEFINED(CFA_HAVE_IORING_OP_FADVISE); 495 511 496 512 if( /*func == (fptr_t)madvise || */ 497 513 func == (fptr_t)cfa_madvise ) 498 #define _CFA_IO_FEATURE_ IORING_OP_MADVISE ,499 return IS_DEFINED( IORING_OP_MADVISE);514 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_MADVISE , 515 return IS_DEFINED(CFA_HAVE_IORING_OP_MADVISE); 500 516 501 517 if( /*func == (fptr_t)openat || */ 502 518 func == (fptr_t)cfa_openat ) 503 #define _CFA_IO_FEATURE_ IORING_OP_OPENAT ,504 return IS_DEFINED( IORING_OP_OPENAT);519 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_OPENAT , 520 return IS_DEFINED(CFA_HAVE_IORING_OP_OPENAT); 505 521 506 522 if( /*func == (fptr_t)close || */ 507 523 func == (fptr_t)cfa_close ) 508 #define _CFA_IO_FEATURE_ IORING_OP_CLOSE ,509 return IS_DEFINED( IORING_OP_CLOSE);524 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_CLOSE , 525 return IS_DEFINED(CFA_HAVE_IORING_OP_CLOSE); 510 526 511 527 if( /*func == (fptr_t)read || */ 512 528 func == (fptr_t)cfa_read ) 513 #define _CFA_IO_FEATURE_ IORING_OP_READ ,514 return IS_DEFINED( IORING_OP_READ);529 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_READ , 530 return IS_DEFINED(CFA_HAVE_IORING_OP_READ); 515 531 516 532 if( /*func == (fptr_t)write || */ 517 533 func == (fptr_t)cfa_write ) 518 #define _CFA_IO_FEATURE_ IORING_OP_WRITE ,519 return IS_DEFINED( IORING_OP_WRITE);534 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_WRITE , 535 return IS_DEFINED(CFA_HAVE_IORING_OP_WRITE); 520 536 521 537 if( /*func == (fptr_t)splice || */ 522 func == (fptr_t)cfa_splice ) 523 #define _CFA_IO_FEATURE_IORING_OP_SPLICE , 524 return IS_DEFINED(IORING_OP_SPLICE); 538 func == (fptr_t)(ssize_t (*)(int, loff_t *, int, loff_t *, size_t, unsigned int))cfa_splice, 539 func == (fptr_t)(ssize_t (*)(int, loff_t *, int, loff_t *, size_t, unsigned int, int, int))cfa_splice ) 540 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_SPLICE , 541 return IS_DEFINED(CFA_HAVE_IORING_OP_SPLICE); 525 542 526 543 if( /*func == (fptr_t)tee || */ 527 544 func == (fptr_t)cfa_tee ) 528 #define _CFA_IO_FEATURE_ IORING_OP_TEE ,529 return IS_DEFINED( IORING_OP_TEE);545 #define _CFA_IO_FEATURE_CFA_HAVE_IORING_OP_TEE , 546 return IS_DEFINED(CFA_HAVE_IORING_OP_TEE); 530 547 #endif 531 548 -
libcfa/src/concurrency/kernel.cfa
rf0c3120 r04b73b6 228 228 static void * __invoke_processor(void * arg); 229 229 230 void ?{}(processor & this, const char name[], cluster & _cltr) with( this ) {230 static init(processor & this, const char name[], cluster & _cltr) with( this ) { 231 231 this.name = name; 232 232 this.cltr = &_cltr; 233 233 id = -1u; 234 terminated{ 0 };235 234 destroyer = 0p; 236 235 do_terminate = false; 237 236 preemption_alarm = 0p; 238 237 pending_preemption = false; 239 runner.proc = &this;240 238 241 239 #if !defined(__CFA_NO_STATISTICS__) … … 244 242 #endif 245 243 246 idle{}; 244 __atomic_fetch_add( &cltr->nprocessors, 1u, __ATOMIC_SEQ_CST ); 245 246 id = doregister((__processor_id_t*)&this); 247 248 // Lock the RWlock so no-one pushes/pops while we are changing the queue 249 uint_fast32_t last_size = ready_mutate_lock(); 250 251 // Adjust the ready queue size 252 ready_queue_grow( cltr ); 253 254 // Unlock the RWlock 255 ready_mutate_unlock( last_size ); 256 257 __cfadbg_print_safe(runtime_core, "Kernel : core %p created\n", &this); 258 } 259 260 // Not a ctor, it just preps the destruction but should not destroy members 261 void deinit(processor & this) { 262 // Lock the RWlock so no-one pushes/pops while we are changing the queue 263 uint_fast32_t last_size = ready_mutate_lock(); 264 265 // Adjust the ready queue size 266 ready_queue_shrink( this.cltr ); 267 268 // Make sure we aren't on the idle queue 269 unsafe_remove( this.cltr->idles, &this ); 270 271 // Unlock the RWlock 272 ready_mutate_unlock( last_size ); 273 274 // Finally we don't need the read_lock any more 275 unregister((__processor_id_t*)&this); 276 } 277 278 void ?{}(processor & this, const char name[], cluster & _cltr) { 279 ( this.idle ){}; 280 ( this.terminated ){ 0 }; 281 ( this.runner ){}; 282 init( this, name, _cltr ); 247 283 248 284 __cfadbg_print_safe(runtime_core, "Kernel : Starting core %p\n", &this); 249 285 250 286 this.stack = __create_pthread( &this.kernel_thread, __invoke_processor, (void *)&this ); 251 __atomic_fetch_add( &cltr->nprocessors, 1u, __ATOMIC_SEQ_CST ); 252 253 __cfadbg_print_safe(runtime_core, "Kernel : core %p created\n", &this); 287 254 288 } 255 289 … … 269 303 270 304 free( this.stack ); 305 306 deinit( this ); 271 307 272 308 __atomic_fetch_sub( &cltr->nprocessors, 1u, __ATOMIC_SEQ_CST ); … … 318 354 319 355 __cfadbg_print_safe(runtime_core, "Kernel : core %p starting\n", this); 320 321 // register the processor unless it's the main thread which is handled in the boot sequence 322 if(this != mainProcessor) { 323 this->id = doregister((__processor_id_t*)this); 324 #if !defined(__CFA_NO_STATISTICS__) 325 if( this->print_halts ) { 326 __cfaabi_bits_print_safe( STDOUT_FILENO, "Processor : %d - %s (%p)\n", this->id, this->name, (void*)this); 327 } 328 #endif 329 330 // Lock the RWlock so no-one pushes/pops while we are changing the queue 331 uint_fast32_t last_size = ready_mutate_lock(); 332 333 // Adjust the ready queue size 334 ready_queue_grow( this->cltr ); 335 336 // Unlock the RWlock 337 ready_mutate_unlock( last_size ); 338 } 356 #if !defined(__CFA_NO_STATISTICS__) 357 if( this->print_halts ) { 358 __cfaabi_bits_print_safe( STDOUT_FILENO, "Processor : %d - %s (%p)\n", this->id, this->name, (void*)this); 359 } 360 #endif 339 361 340 362 { … … 375 397 V( this->terminated ); 376 398 377 // unregister the processor unless it's the main thread which is handled in the boot sequence 378 if(this != mainProcessor) { 379 // Lock the RWlock so no-one pushes/pops while we are changing the queue 380 uint_fast32_t last_size = ready_mutate_lock(); 381 382 // Adjust the ready queue size 383 ready_queue_shrink( this->cltr ); 384 385 // Make sure we aren't on the idle queue 386 #if !defined(__CFA_NO_STATISTICS__) 387 bool removed = 388 #endif 389 unsafe_remove( this->cltr->idles, this ); 390 391 #if !defined(__CFA_NO_STATISTICS__) 392 if(removed) __tls_stats()->ready.sleep.exits++; 393 #endif 394 395 // Unlock the RWlock 396 ready_mutate_unlock( last_size ); 397 398 // Finally we don't need the read_lock any more 399 unregister((__processor_id_t*)this); 400 } 401 else { 399 if(this == mainProcessor) { 402 400 // HACK : the coroutine context switch expects this_thread to be set 403 401 // and it make sense for it to be set in all other cases except here … … 859 857 860 858 void ?{}(processor & this) with( this ) { 861 name = "Main Processor"; 862 cltr = mainCluster; 863 terminated{ 0 }; 864 do_terminate = false; 865 preemption_alarm = 0p; 866 pending_preemption = false; 859 ( this.idle ){}; 860 ( this.terminated ){ 0 }; 861 ( this.runner ){}; 862 init( this, "Main Processor", *mainCluster ); 867 863 kernel_thread = pthread_self(); 868 id = -1u;869 870 #if !defined(__CFA_NO_STATISTICS__)871 print_stats = false;872 print_halts = false;873 #endif874 864 875 865 runner{ &this }; 876 866 __cfadbg_print_safe(runtime_core, "Kernel : constructed main processor context %p\n", &runner); 877 878 __atomic_fetch_add( &cltr->nprocessors, 1u, __ATOMIC_SEQ_CST );879 867 } 880 868 … … 883 871 mainProcessor = (processor *)&storage_mainProcessor; 884 872 (*mainProcessor){}; 885 886 mainProcessor->id = doregister( (__processor_id_t*)mainProcessor);887 873 888 874 //initialize the global state variables … … 944 930 kernel_stop_preemption(); 945 931 946 unregister((__processor_id_t*)mainProcessor);947 948 932 // Destroy the main processor and its context in reverse order of construction 949 933 // These were manually constructed so we need manually destroy them 950 934 void ^?{}(processor & this) with( this ){ 935 deinit( this ); 936 951 937 /* paranoid */ verify( this.do_terminate == true ); 952 938 __atomic_fetch_sub( &cltr->nprocessors, 1u, __ATOMIC_SEQ_CST ); -
libcfa/src/concurrency/ready_queue.cfa
rf0c3120 r04b73b6 186 186 //======================================================================= 187 187 void ?{}(__ready_queue_t & this) with (this) { 188 189 lanes.data = alloc(4); 190 for( i; 4 ) { 191 (lanes.data[i]){}; 192 } 193 lanes.count = 4; 194 snzi{ log2( lanes.count / 8 ) }; 188 lanes.data = 0p; 189 lanes.count = 0; 195 190 } 196 191 197 192 void ^?{}(__ready_queue_t & this) with (this) { 198 verify( 4== lanes.count );193 verify( 0 == lanes.count ); 199 194 verify( !query( snzi ) ); 200 201 ^(snzi){};202 203 for( i; 4 ) {204 ^(lanes.data[i]){};205 }206 195 free(lanes.data); 207 196 } … … 495 484 } 496 485 486 #warning remove when alloc is fixed 487 forall( dtype T | sized(T) ) 488 static inline T * correct_alloc( T ptr[], size_t dim ) { 489 if( dim == 0 ) { 490 free(ptr); 491 return 0p; 492 } 493 T * temp = alloc( dim ); 494 if(ptr) { 495 memcpy( temp, ptr, dim * sizeof(T)); 496 free(ptr); 497 } 498 return temp; 499 } 500 497 501 // Grow the ready queue 498 502 void ready_queue_grow (struct cluster * cltr) { … … 513 517 514 518 // Allocate new array (uses realloc and memcpies the data) 515 lanes.data = alloc(lanes.data, ncount);519 lanes.data = correct_alloc(lanes.data, ncount); 516 520 517 521 // Fix the moved data … … 558 562 size_t ocount = lanes.count; 559 563 // Check that we have some space left 560 if(ocount < 8) abort("Program attempted to destroy more Ready Queues than were created");564 if(ocount < 4) abort("Program attempted to destroy more Ready Queues than were created"); 561 565 562 566 // reduce the actual count so push doesn't use the old queues … … 600 604 601 605 // Allocate new array (uses realloc and memcpies the data) 602 lanes.data = alloc(lanes.data, lanes.count);606 lanes.data = correct_alloc(lanes.data, lanes.count); 603 607 604 608 // Fix the moved data -
libcfa/src/concurrency/snzi.hfa
rf0c3120 r04b73b6 120 120 //-------------------------------------------------- 121 121 // SNZI object 122 void ?{}( __snzi_t & this ) { 123 this.mask = 0; 124 this.root = 0; 125 this.nodes = 0p; 126 } 127 122 128 void ?{}( __snzi_t & this, unsigned depth ) with( this ) { 123 129 mask = (1 << depth) - 1; -
libcfa/src/concurrency/stats.cfa
rf0c3120 r04b73b6 24 24 stats->ready.sleep.exits = 0; 25 25 26 #if defined( HAVE_LINUX_IO_URING_H)26 #if defined(CFA_HAVE_LINUX_IO_URING_H) 27 27 stats->io.submit_q.submit_avg.rdy = 0; 28 28 stats->io.submit_q.submit_avg.csm = 0; … … 59 59 __atomic_fetch_add( &cltr->ready.sleep.exits , proc->ready.sleep.exits , __ATOMIC_SEQ_CST ); 60 60 61 #if defined( HAVE_LINUX_IO_URING_H)61 #if defined(CFA_HAVE_LINUX_IO_URING_H) 62 62 __atomic_fetch_add( &cltr->io.submit_q.submit_avg.rdy , proc->io.submit_q.submit_avg.rdy , __ATOMIC_SEQ_CST ); 63 63 __atomic_fetch_add( &cltr->io.submit_q.submit_avg.csm , proc->io.submit_q.submit_avg.csm , __ATOMIC_SEQ_CST ); … … 121 121 } 122 122 123 #if defined( HAVE_LINUX_IO_URING_H)123 #if defined(CFA_HAVE_LINUX_IO_URING_H) 124 124 if( flags & CFA_STATS_IO ) { 125 125 double avgrdy = ((double)io.submit_q.submit_avg.rdy) / io.submit_q.submit_avg.cnt; -
libcfa/src/concurrency/stats.hfa
rf0c3120 r04b73b6 11 11 enum { 12 12 CFA_STATS_READY_Q = 0x01, 13 #if defined( HAVE_LINUX_IO_URING_H)13 #if defined(CFA_HAVE_LINUX_IO_URING_H) 14 14 CFA_STATS_IO = 0x02, 15 15 #endif … … 64 64 }; 65 65 66 #if defined( HAVE_LINUX_IO_URING_H)66 #if defined(CFA_HAVE_LINUX_IO_URING_H) 67 67 struct __attribute__((aligned(64))) __stats_io_t{ 68 68 struct { … … 99 99 struct __attribute__((aligned(128))) __stats_t { 100 100 __stats_readQ_t ready; 101 #if defined( HAVE_LINUX_IO_URING_H)101 #if defined(CFA_HAVE_LINUX_IO_URING_H) 102 102 __stats_io_t io; 103 103 #endif
Note:
See TracChangeset
for help on using the changeset viewer.