- File:
-
- 1 edited
Legend:
- Unmodified
- Added
- Removed
-
libcfa/src/concurrency/kernel.cfa
rae7adbc4 r7cf3b1d 205 205 // Don't block if we are done 206 206 if( __atomic_load_n(&this->do_terminate, __ATOMIC_SEQ_CST) ) break MAIN_LOOP; 207 208 #if !defined(__CFA_NO_STATISTICS__)209 __tls_stats()->ready.sleep.halts++;210 #endif211 207 212 208 // Push self to idle stack … … 554 550 /* paranoid */ verify( 0x0D15EA5E0D15EA5Ep == thrd->canary ); 555 551 556 const bool local = thrd->state != Start;557 552 if (thrd->preempted == __NO_PREEMPTION) thrd->state = Ready; 558 553 … … 733 728 // Wake a thread from the front if there are any 734 729 static void __wake_one(cluster * this) { 730 eventfd_t val; 731 735 732 /* paranoid */ verify( ! __preemption_enabled() ); 736 733 /* paranoid */ verify( ready_schedule_islocked() ); 737 734 738 735 // Check if there is a sleeping processor 739 // int fd = __atomic_load_n(&this->procs.fd, __ATOMIC_SEQ_CST); 740 int fd = 0; 741 if( __atomic_load_n(&this->procs.fd, __ATOMIC_SEQ_CST) != 0 ) { 742 fd = __atomic_exchange_n(&this->procs.fd, 0, __ATOMIC_RELAXED); 743 } 744 745 // If no one is sleeping, we are done 746 if( fd == 0 ) return; 747 748 // We found a processor, wake it up 749 eventfd_t val; 750 val = 1; 751 eventfd_write( fd, val ); 752 753 #if !defined(__CFA_NO_STATISTICS__) 754 if( kernelTLS().this_stats ) { 755 __tls_stats()->ready.sleep.wakes++; 756 } 757 else { 758 __atomic_fetch_add(&this->stats->ready.sleep.wakes, 1, __ATOMIC_RELAXED); 759 } 760 #endif 736 struct __fd_waitctx * fdp = __atomic_load_n(&this->procs.fdw, __ATOMIC_SEQ_CST); 737 738 // If no one is sleeping: we are done 739 if( fdp == 0p ) return; 740 741 int fd = 1; 742 if( __atomic_load_n(&fdp->fd, __ATOMIC_SEQ_CST) != 1 ) { 743 fd = __atomic_exchange_n(&fdp->fd, 1, __ATOMIC_RELAXED); 744 } 745 746 switch(fd) { 747 case 0: 748 // If the processor isn't ready to sleep then the exchange will already wake it up 749 #if !defined(__CFA_NO_STATISTICS__) 750 if( kernelTLS().this_stats ) { __tls_stats()->ready.sleep.early++; 751 } else { __atomic_fetch_add(&this->stats->ready.sleep.early, 1, __ATOMIC_RELAXED); } 752 #endif 753 break; 754 case 1: 755 // If someone else already said they will wake them: we are done 756 #if !defined(__CFA_NO_STATISTICS__) 757 if( kernelTLS().this_stats ) { __tls_stats()->ready.sleep.seen++; 758 } else { __atomic_fetch_add(&this->stats->ready.sleep.seen, 1, __ATOMIC_RELAXED); } 759 #endif 760 break; 761 default: 762 // If the processor was ready to sleep, we need to wake it up with an actual write 763 val = 1; 764 eventfd_write( fd, val ); 765 766 #if !defined(__CFA_NO_STATISTICS__) 767 if( kernelTLS().this_stats ) { __tls_stats()->ready.sleep.wakes++; 768 } else { __atomic_fetch_add(&this->stats->ready.sleep.wakes, 1, __ATOMIC_RELAXED); } 769 #endif 770 break; 771 } 761 772 762 773 /* paranoid */ verify( ready_schedule_islocked() ); … … 771 782 772 783 __cfadbg_print_safe(runtime_core, "Kernel : waking Processor %p\n", this); 784 785 this->idle_wctx.fd = 1; 773 786 774 787 eventfd_t val; … … 780 793 781 794 static void idle_sleep(processor * this, io_future_t & future, iovec & iov) { 795 // Tell everyone we are ready to go do sleep 796 for() { 797 int expected = this->idle_wctx.fd; 798 799 // Someone already told us to wake-up! No time for a nap. 800 if(expected == 1) { return; } 801 802 // Try to mark that we are going to sleep 803 if(__atomic_compare_exchange_n(&this->idle_wctx.fd, &expected, this->idle_fd, false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST) ) { 804 // Every one agreed, taking a nap 805 break; 806 } 807 } 808 809 782 810 #if !defined(CFA_WITH_IO_URING_IDLE) 783 811 #if !defined(__CFA_NO_STATISTICS__) … … 826 854 827 855 static bool mark_idle(__cluster_proc_list & this, processor & proc) { 856 #if !defined(__CFA_NO_STATISTICS__) 857 __tls_stats()->ready.sleep.halts++; 858 #endif 859 860 proc.idle_wctx.fd = 0; 861 828 862 /* paranoid */ verify( ! __preemption_enabled() ); 829 863 if(!try_lock( this )) return false; … … 833 867 insert_first(this.idles, proc); 834 868 835 __atomic_store_n(&this.fd , proc.idle_fd, __ATOMIC_SEQ_CST);869 __atomic_store_n(&this.fdw, &proc.idle_wctx, __ATOMIC_SEQ_CST); 836 870 unlock( this ); 837 871 /* paranoid */ verify( ! __preemption_enabled() ); … … 849 883 850 884 { 851 int fd= 0;852 if(!this.idles`isEmpty) fd = this.idles`first.idle_fd;853 __atomic_store_n(&this.fd , fd, __ATOMIC_SEQ_CST);885 struct __fd_waitctx * wctx = 0; 886 if(!this.idles`isEmpty) wctx = &this.idles`first.idle_wctx; 887 __atomic_store_n(&this.fdw, wctx, __ATOMIC_SEQ_CST); 854 888 } 855 889
Note:
See TracChangeset
for help on using the changeset viewer.