Changes in / [54dcab1:231b18f]


Ignore:
Files:
15 edited

Legend:

Unmodified
Added
Removed
  • benchmark/readyQ/cycle.cfa

    r54dcab1 r231b18f  
    8484                }
    8585
    86                 printf("Duration (ms)        : %'ld\n", (end - start)`dms);
     86                printf("Duration (ms)        : %'ld\n", (end - start)`ms);
    8787                printf("Number of processors : %'d\n", nprocs);
    8888                printf("Number of threads    : %'d\n", tthreads);
     
    9090                printf("Total Operations(ops): %'15llu\n", global_counter);
    9191                printf("Total blocks         : %'15llu\n", global_blocks);
    92                 printf("Ops per second       : %'18.2lf\n", ((double)global_counter) / (end - start)`ds);
    93                 printf("ns per ops           : %'18.2lf\n", (end - start)`dns / global_counter);
     92                printf("Ops per second       : %'18.2lf\n", ((double)global_counter) / (end - start)`s);
     93                printf("ns per ops           : %'18.2lf\n", ((double)(end - start)`ns) / global_counter);
    9494                printf("Ops per threads      : %'15llu\n", global_counter / tthreads);
    9595                printf("Ops per procs        : %'15llu\n", global_counter / nprocs);
    96                 printf("Ops/sec/procs        : %'18.2lf\n", (((double)global_counter) / nprocs) / (end - start)`ds);
    97                 printf("ns per ops/procs     : %'18.2lf\n", (end - start)`dns / (global_counter / nprocs));
     96                printf("Ops/sec/procs        : %'18.2lf\n", (((double)global_counter) / nprocs) / (end - start)`s);
     97                printf("ns per ops/procs     : %'18.2lf\n", ((double)(end - start)`ns) / (global_counter / nprocs));
    9898                fflush(stdout);
    9999        }
  • benchmark/readyQ/cycle.go

    r54dcab1 r231b18f  
    7272        p.Printf("Cycle size (# thrds) : %d\n", ring_size);
    7373        p.Printf("Total Operations(ops): %15d\n", global_counter)
    74         p.Printf("Ops per second       : %18.2f\n", float64(global_counter) / delta.Seconds())
     74        p.Printf("Yields per second    : %18.2f\n", float64(global_counter) / delta.Seconds())
    7575        p.Printf("ns per ops           : %18.2f\n", float64(delta.Nanoseconds()) / float64(global_counter))
    7676        p.Printf("Ops per threads      : %15d\n", global_counter / uint64(tthreads))
  • benchmark/readyQ/rq_bench.hfa

    r54dcab1 r231b18f  
    8888}
    8989
    90 struct __attribute__((aligned(128))) bench_sem {
     90struct bench_sem {
    9191        struct $thread * volatile ptr;
    9292};
  • benchmark/readyQ/rq_bench.hpp

    r54dcab1 r231b18f  
    7575}
    7676
    77 class __attribute__((aligned(128))) bench_sem {
     77class bench_sem {
    7878        Fibre * volatile ptr = nullptr;
    7979public:
  • libcfa/src/concurrency/io.cfa

    r54dcab1 r231b18f  
    7676
    7777        static inline bool next( __leaderlock_t & this ) {
    78                 /* paranoid */ verify( ! __preemption_enabled() );
     78                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    7979                struct $thread * nextt;
    8080                for() {
     
    168168        // This is NOT thread-safe
    169169        static [int, bool] __drain_io( & struct __io_data ring ) {
    170                 /* paranoid */ verify( ! __preemption_enabled() );
     170                /* paranoid */ verify( !kernelTLS.preemption_state.enabled );
    171171
    172172                unsigned to_submit = 0;
     
    404404                                        return;
    405405                                }
    406                                 /* paranoid */ verify( ! __preemption_enabled() );
     406                                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    407407                                __STATS__( true,
    408408                                        io.submit_q.leader += 1;
     
    442442
    443443                        #if defined(LEADER_LOCK)
    444                                 /* paranoid */ verify( ! __preemption_enabled() );
     444                                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    445445                                next(ring.submit_q.submit_lock);
    446446                        #else
  • libcfa/src/concurrency/io/setup.cfa

    r54dcab1 r231b18f  
    149149                id.full_proc = false;
    150150                id.id = doregister(&id);
    151                 __cfaabi_tls.this_proc_id = &id;
     151                kernelTLS.this_proc_id = &id;
    152152                __cfaabi_dbg_print_safe( "Kernel : IO poller thread starting\n" );
    153153
     
    179179                                __cfadbg_print_safe(io_core, "Kernel I/O : Unparking io poller %p\n", io_ctx);
    180180                                #if !defined( __CFA_NO_STATISTICS__ )
    181                                         __cfaabi_tls.this_stats = io_ctx->self.curr_cluster->stats;
     181                                        kernelTLS.this_stats = io_ctx->self.curr_cluster->stats;
    182182                                #endif
    183183                                post( io_ctx->sem );
  • libcfa/src/concurrency/kernel.cfa

    r54dcab1 r231b18f  
    122122        // Because of a bug, we couldn't initialized the seed on construction
    123123        // Do it here
    124         __cfaabi_tls.rand_seed ^= rdtscl();
    125         __cfaabi_tls.ready_rng.fwd_seed = 25214903917_l64u * (rdtscl() ^ (uintptr_t)&runner);
     124        kernelTLS.rand_seed ^= rdtscl();
     125        kernelTLS.ready_rng.fwd_seed = 25214903917_l64u * (rdtscl() ^ (uintptr_t)&runner);
    126126        __tls_rand_advance_bck();
    127127
     
    217217                // and it make sense for it to be set in all other cases except here
    218218                // fake it
    219                 __cfaabi_tls.this_thread = mainThread;
     219                kernelTLS.this_thread = mainThread;
    220220        }
    221221
     
    230230// from the processor coroutine to the target thread
    231231static void __run_thread(processor * this, $thread * thrd_dst) {
    232         /* paranoid */ verify( ! __preemption_enabled() );
     232        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    233233        /* paranoid */ verifyf( thrd_dst->state == Ready || thrd_dst->preempted != __NO_PREEMPTION, "state : %d, preempted %d\n", thrd_dst->state, thrd_dst->preempted);
    234234        /* paranoid */ verifyf( thrd_dst->link.next == 0p, "Expected null got %p", thrd_dst->link.next );
     
    247247
    248248                // Update global state
    249                 kernelTLS().this_thread = thrd_dst;
    250 
    251                 /* paranoid */ verify( ! __preemption_enabled() );
    252                 /* paranoid */ verify( kernelTLS().this_thread == thrd_dst );
     249                kernelTLS.this_thread = thrd_dst;
     250
     251                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     252                /* paranoid */ verify( kernelTLS.this_thread == thrd_dst );
    253253                /* paranoid */ verify( thrd_dst->context.SP );
    254254                /* paranoid */ verify( thrd_dst->state != Halted );
     
    267267                /* paranoid */ verifyf( ((uintptr_t)thrd_dst->context.SP) < ((uintptr_t)__get_stack(thrd_dst->curr_cor)->base ), "ERROR : Destination $thread %p has been corrupted.\n StackPointer too small.\n", thrd_dst );
    268268                /* paranoid */ verify( thrd_dst->context.SP );
    269                 /* paranoid */ verify( kernelTLS().this_thread == thrd_dst );
    270                 /* paranoid */ verify( ! __preemption_enabled() );
     269                /* paranoid */ verify( kernelTLS.this_thread == thrd_dst );
     270                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    271271
    272272                // Reset global state
    273                 kernelTLS().this_thread = 0p;
     273                kernelTLS.this_thread = 0p;
    274274
    275275                // We just finished running a thread, there are a few things that could have happened.
     
    315315        proc_cor->state = Active;
    316316
    317         /* paranoid */ verify( ! __preemption_enabled() );
     317        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    318318}
    319319
    320320// KERNEL_ONLY
    321321void returnToKernel() {
    322         /* paranoid */ verify( ! __preemption_enabled() );
    323         $coroutine * proc_cor = get_coroutine(kernelTLS().this_processor->runner);
    324         $thread * thrd_src = kernelTLS().this_thread;
     322        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     323        $coroutine * proc_cor = get_coroutine(kernelTLS.this_processor->runner);
     324        $thread * thrd_src = kernelTLS.this_thread;
    325325
    326326        #if !defined(__CFA_NO_STATISTICS__)
    327                 struct processor * last_proc = kernelTLS().this_processor;
     327                struct processor * last_proc = kernelTLS.this_processor;
    328328        #endif
    329329
     
    345345
    346346        #if !defined(__CFA_NO_STATISTICS__)
    347                 if(last_proc != kernelTLS().this_processor) {
     347                if(last_proc != kernelTLS.this_processor) {
    348348                        __tls_stats()->ready.threads.migration++;
    349349                }
    350350        #endif
    351351
    352         /* paranoid */ verify( ! __preemption_enabled() );
     352        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    353353        /* paranoid */ verifyf( ((uintptr_t)thrd_src->context.SP) < ((uintptr_t)__get_stack(thrd_src->curr_cor)->base ), "ERROR : Returning $thread %p has been corrupted.\n StackPointer too small.\n", thrd_src );
    354354        /* paranoid */ verifyf( ((uintptr_t)thrd_src->context.SP) > ((uintptr_t)__get_stack(thrd_src->curr_cor)->limit), "ERROR : Returning $thread %p has been corrupted.\n StackPointer too large.\n", thrd_src );
     
    359359// KERNEL ONLY
    360360void __schedule_thread( $thread * thrd ) {
    361         /* paranoid */ verify( ! __preemption_enabled() );
    362361        /* paranoid */ verify( thrd );
    363362        /* paranoid */ verify( thrd->state != Halted );
    364         /* paranoid */ verify( kernelTLS().this_proc_id );
     363        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     364        /* paranoid */ verify( kernelTLS.this_proc_id );
    365365        /* paranoid */ #if defined( __CFA_WITH_VERIFY__ )
    366366        /* paranoid */  if( thrd->state == Blocked || thrd->state == Start ) assertf( thrd->preempted == __NO_PREEMPTION,
     
    380380        ready_schedule_unlock();
    381381
    382         /* paranoid */ verify( ! __preemption_enabled() );
     382        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    383383}
    384384
    385385// KERNEL ONLY
    386386static inline $thread * __next_thread(cluster * this) with( *this ) {
    387         /* paranoid */ verify( ! __preemption_enabled() );
    388         /* paranoid */ verify( kernelTLS().this_proc_id );
     387        /* paranoid */ verify( kernelTLS.this_proc_id );
     388        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    389389
    390390        ready_schedule_lock();
     
    392392        ready_schedule_unlock();
    393393
    394         /* paranoid */ verify( kernelTLS().this_proc_id );
    395         /* paranoid */ verify( ! __preemption_enabled() );
     394        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     395        /* paranoid */ verify( kernelTLS.this_proc_id );
    396396        return thrd;
    397397}
     
    399399// KERNEL ONLY
    400400static inline $thread * __next_thread_slow(cluster * this) with( *this ) {
    401         /* paranoid */ verify( ! __preemption_enabled() );
    402         /* paranoid */ verify( kernelTLS().this_proc_id );
     401        /* paranoid */ verify( kernelTLS.this_proc_id );
     402        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    403403
    404404        ready_schedule_lock();
     
    406406        ready_schedule_unlock();
    407407
    408         /* paranoid */ verify( kernelTLS().this_proc_id );
    409         /* paranoid */ verify( ! __preemption_enabled() );
     408        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     409        /* paranoid */ verify( kernelTLS.this_proc_id );
    410410        return thrd;
    411411}
     
    414414        if( !thrd ) return;
    415415
     416        /* paranoid */ verify( kernelTLS.this_proc_id );
     417        bool full = kernelTLS.this_proc_id->full_proc;
     418        if(full) disable_interrupts();
     419
     420        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    416421        int old_ticket = __atomic_fetch_add(&thrd->ticket, 1, __ATOMIC_SEQ_CST);
    417422        switch(old_ticket) {
     
    423428                        /* paranoid */ verify( thrd->state == Blocked );
    424429
    425                         {
    426                                 /* paranoid */ verify( publicTLS_get(this_proc_id) );
    427                                 bool full = publicTLS_get(this_proc_id)->full_proc;
    428                                 if(full) disable_interrupts();
    429 
    430                                 /* paranoid */ verify( ! __preemption_enabled() );
    431 
    432                                 // Wake lost the race,
    433                                 __schedule_thread( thrd );
    434 
    435                                 /* paranoid */ verify( ! __preemption_enabled() );
    436 
    437                                 if(full) enable_interrupts( __cfaabi_dbg_ctx );
    438                                 /* paranoid */ verify( publicTLS_get(this_proc_id) );
    439                         }
    440 
     430                        // Wake lost the race,
     431                        __schedule_thread( thrd );
    441432                        break;
    442433                default:
     
    444435                        abort("Thread %p (%s) has mismatch park/unpark\n", thrd, thrd->self_cor.name);
    445436        }
     437        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     438
     439        if(full) enable_interrupts( __cfaabi_dbg_ctx );
     440        /* paranoid */ verify( kernelTLS.this_proc_id );
    446441}
    447442
    448443void park( void ) {
    449         /* paranoid */ verify( __preemption_enabled() );
     444        /* paranoid */ verify( kernelTLS.preemption_state.enabled );
    450445        disable_interrupts();
    451         /* paranoid */ verify( ! __preemption_enabled() );
    452         /* paranoid */ verify( kernelTLS().this_thread->preempted == __NO_PREEMPTION );
     446        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     447        /* paranoid */ verify( kernelTLS.this_thread->preempted == __NO_PREEMPTION );
    453448
    454449        returnToKernel();
    455450
    456         /* paranoid */ verify( ! __preemption_enabled() );
     451        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    457452        enable_interrupts( __cfaabi_dbg_ctx );
    458         /* paranoid */ verify( __preemption_enabled() );
     453        /* paranoid */ verify( kernelTLS.preemption_state.enabled );
    459454
    460455}
     
    465460        // Should never return
    466461        void __cfactx_thrd_leave() {
    467                 $thread * thrd = active_thread();
     462                $thread * thrd = TL_GET( this_thread );
    468463                $monitor * this = &thrd->self_mon;
    469464
     
    478473
    479474                // Leave the thread
    480                 /* paranoid */ verify( ! __preemption_enabled() );
     475                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    481476                returnToKernel();
    482477                abort();
     
    488483// KERNEL ONLY
    489484bool force_yield( __Preemption_Reason reason ) {
    490         /* paranoid */ verify( __preemption_enabled() );
     485        /* paranoid */ verify( kernelTLS.preemption_state.enabled );
    491486        disable_interrupts();
    492         /* paranoid */ verify( ! __preemption_enabled() );
    493 
    494         $thread * thrd = kernelTLS().this_thread;
     487        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     488
     489        $thread * thrd = kernelTLS.this_thread;
    495490        /* paranoid */ verify(thrd->state == Active);
    496491
     
    506501        }
    507502
    508         /* paranoid */ verify( ! __preemption_enabled() );
     503        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    509504        enable_interrupts_noPoll();
    510         /* paranoid */ verify( __preemption_enabled() );
     505        /* paranoid */ verify( kernelTLS.preemption_state.enabled );
    511506
    512507        return preempted;
     
    518513// Wake a thread from the front if there are any
    519514static void __wake_one(cluster * this) {
    520         /* paranoid */ verify( ! __preemption_enabled() );
     515        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    521516        /* paranoid */ verify( ready_schedule_islocked() );
    522517
     
    538533
    539534        /* paranoid */ verify( ready_schedule_islocked() );
    540         /* paranoid */ verify( ! __preemption_enabled() );
     535        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    541536
    542537        return;
     
    548543
    549544        disable_interrupts();
    550                 /* paranoid */ verify( ! __preemption_enabled() );
     545                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    551546                post( this->idle );
    552547        enable_interrupts( __cfaabi_dbg_ctx );
     
    554549
    555550static void push  (__cluster_idles & this, processor & proc) {
    556         /* paranoid */ verify( ! __preemption_enabled() );
     551        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    557552        lock( this );
    558553                this.idle++;
     
    561556                insert_first(this.list, proc);
    562557        unlock( this );
    563         /* paranoid */ verify( ! __preemption_enabled() );
     558        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    564559}
    565560
    566561static void remove(__cluster_idles & this, processor & proc) {
    567         /* paranoid */ verify( ! __preemption_enabled() );
     562        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    568563        lock( this );
    569564                this.idle--;
     
    572567                remove(proc);
    573568        unlock( this );
    574         /* paranoid */ verify( ! __preemption_enabled() );
     569        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    575570}
    576571
     
    616611        }
    617612
    618         return __cfaabi_tls.this_thread;
     613        return kernelTLS.this_thread;
    619614}
    620615
     
    641636
    642637int kernel_abort_lastframe( void ) __attribute__ ((__nothrow__)) {
    643         return get_coroutine(kernelTLS().this_thread) == get_coroutine(mainThread) ? 4 : 2;
     638        return get_coroutine(kernelTLS.this_thread) == get_coroutine(mainThread) ? 4 : 2;
    644639}
    645640
     
    673668        if ( count < 0 ) {
    674669                // queue current task
    675                 append( waiting, active_thread() );
     670                append( waiting, kernelTLS.this_thread );
    676671
    677672                // atomically release spin lock and block
     
    723718                void __cfaabi_dbg_record_lock(__spinlock_t & this, const char prev_name[]) {
    724719                        this.prev_name = prev_name;
    725                         this.prev_thrd = kernelTLS().this_thread;
     720                        this.prev_thrd = kernelTLS.this_thread;
    726721                }
    727722        }
  • libcfa/src/concurrency/kernel.hfa

    r54dcab1 r231b18f  
    275275static inline [cluster *&, cluster *& ] __get( cluster & this ) __attribute__((const)) { return this.node.[next, prev]; }
    276276
    277 static inline struct processor * active_processor() { return publicTLS_get( this_processor ); } // UNSAFE
    278 static inline struct cluster   * active_cluster  () { return publicTLS_get( this_processor )->cltr; }
     277static inline struct processor * active_processor() { return TL_GET( this_processor ); } // UNSAFE
     278static inline struct cluster   * active_cluster  () { return TL_GET( this_processor )->cltr; }
    279279
    280280#if !defined(__CFA_NO_STATISTICS__)
    281         void print_stats_now( cluster & this, int flags );
    282 
    283281        static inline void print_stats_at_exit( cluster & this, int flags ) {
    284282                this.print_stats |= flags;
  • libcfa/src/concurrency/kernel/fwd.hfa

    r54dcab1 r231b18f  
    5555                                uint64_t bck_seed;
    5656                        } ready_rng;
    57                 } __cfaabi_tls __attribute__ ((tls_model ( "initial-exec" )));
     57                } kernelTLS __attribute__ ((tls_model ( "initial-exec" )));
    5858
    59                 extern bool __preemption_enabled();
    6059
    61                 static inline KernelThreadData & kernelTLS( void ) {
    62                         /* paranoid */ verify( ! __preemption_enabled() );
    63                         return __cfaabi_tls;
    64                 }
    65 
    66                 extern uintptr_t __cfatls_get( unsigned long int member );
    67                 // #define publicTLS_get( member ) ((typeof(__cfaabi_tls.member))__cfatls_get( __builtin_offsetof(KernelThreadData, member) ))
    68                 #define publicTLS_get( member ) (__cfaabi_tls.member)
    69                 // extern forall(otype T) T __cfatls_get( T * member, T value );
    70                 // #define publicTLS_set( member, value ) __cfatls_set( (typeof(member)*)__builtin_offsetof(KernelThreadData, member), value );
    7160
    7261                static inline uint64_t __tls_rand() {
    7362                        #if defined(__SIZEOF_INT128__)
    74                                 return __lehmer64( kernelTLS().rand_seed );
     63                                return __lehmer64( kernelTLS.rand_seed );
    7564                        #else
    76                                 return __xorshift64( kernelTLS().rand_seed );
     65                                return __xorshift64( kernelTLS.rand_seed );
    7766                        #endif
    7867                }
     
    8675                static inline unsigned __tls_rand_fwd() {
    8776
    88                         kernelTLS().ready_rng.fwd_seed = (A * kernelTLS().ready_rng.fwd_seed + C) & (M - 1);
    89                         return kernelTLS().ready_rng.fwd_seed >> D;
     77                        kernelTLS.ready_rng.fwd_seed = (A * kernelTLS.ready_rng.fwd_seed + C) & (M - 1);
     78                        return kernelTLS.ready_rng.fwd_seed >> D;
    9079                }
    9180
    9281                static inline unsigned __tls_rand_bck() {
    93                         unsigned int r = kernelTLS().ready_rng.bck_seed >> D;
    94                         kernelTLS().ready_rng.bck_seed = AI * (kernelTLS().ready_rng.bck_seed - C) & (M - 1);
     82                        unsigned int r = kernelTLS.ready_rng.bck_seed >> D;
     83                        kernelTLS.ready_rng.bck_seed = AI * (kernelTLS.ready_rng.bck_seed - C) & (M - 1);
    9584                        return r;
    9685                }
     
    10392
    10493                static inline void __tls_rand_advance_bck(void) {
    105                         kernelTLS().ready_rng.bck_seed = kernelTLS().ready_rng.fwd_seed;
     94                        kernelTLS.ready_rng.bck_seed = kernelTLS.ready_rng.fwd_seed;
    10695                }
    10796        }
    10897
     98        #if 0 // def __ARM_ARCH
     99                // function prototypes are only really used by these macros on ARM
     100                void disable_global_interrupts();
     101                void enable_global_interrupts();
    109102
     103                #define TL_GET( member ) ( { __typeof__( kernelTLS.member ) target; \
     104                        disable_global_interrupts(); \
     105                        target = kernelTLS.member; \
     106                        enable_global_interrupts(); \
     107                        target; } )
     108                #define TL_SET( member, value ) disable_global_interrupts(); \
     109                        kernelTLS.member = value; \
     110                        enable_global_interrupts();
     111        #else
     112                #define TL_GET( member ) kernelTLS.member
     113                #define TL_SET( member, value ) kernelTLS.member = value;
     114        #endif
    110115
    111116        extern void disable_interrupts();
     
    116121                extern void park( void );
    117122                extern void unpark( struct $thread * this );
    118                 static inline struct $thread * active_thread () {
    119                         struct $thread * t = publicTLS_get( this_thread );
    120                         /* paranoid */ verify( t );
    121                         return t;
    122                 }
     123                static inline struct $thread * active_thread () { return TL_GET( this_thread ); }
    123124
    124125                extern bool force_yield( enum __Preemption_Reason );
     
    139140                #if !defined(__CFA_NO_STATISTICS__)
    140141                        static inline struct __stats_t * __tls_stats() {
    141                                 /* paranoid */ verify( ! __preemption_enabled() );
    142                                 /* paranoid */ verify( kernelTLS().this_stats );
    143                                 return kernelTLS().this_stats;
     142                                /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
     143                                /* paranoid */ verify( kernelTLS.this_stats );
     144                                return kernelTLS.this_stats;
    144145                        }
    145146
  • libcfa/src/concurrency/kernel/startup.cfa

    r54dcab1 r231b18f  
    118118//-----------------------------------------------------------------------------
    119119// Global state
    120 thread_local struct KernelThreadData __cfaabi_tls __attribute__ ((tls_model ( "initial-exec" ))) @= {
     120thread_local struct KernelThreadData kernelTLS __attribute__ ((tls_model ( "initial-exec" ))) @= {
    121121        NULL,                                                                                           // cannot use 0p
    122122        NULL,
     
    156156// Kernel boot procedures
    157157static void __kernel_startup(void) {
    158         /* paranoid */ verify( ! __preemption_enabled() );
     158        verify( ! kernelTLS.preemption_state.enabled );
    159159        __cfadbg_print_safe(runtime_core, "Kernel : Starting\n");
    160160
     
    212212
    213213        //initialize the global state variables
    214         __cfaabi_tls.this_processor = mainProcessor;
    215         __cfaabi_tls.this_proc_id   = (__processor_id_t*)mainProcessor;
    216         __cfaabi_tls.this_thread    = mainThread;
     214        kernelTLS.this_processor = mainProcessor;
     215        kernelTLS.this_proc_id   = (__processor_id_t*)mainProcessor;
     216        kernelTLS.this_thread    = mainThread;
    217217
    218218        #if !defined( __CFA_NO_STATISTICS__ )
    219                 __cfaabi_tls.this_stats = (__stats_t *)& storage_mainProcStats;
    220                 __init_stats( __cfaabi_tls.this_stats );
     219                kernelTLS.this_stats = (__stats_t *)& storage_mainProcStats;
     220                __init_stats( kernelTLS.this_stats );
    221221        #endif
    222222
     
    234234        // context. Hence, the main thread does not begin through __cfactx_invoke_thread, like all other threads. The trick here is that
    235235        // mainThread is on the ready queue when this call is made.
    236         __kernel_first_resume( __cfaabi_tls.this_processor );
     236        __kernel_first_resume( kernelTLS.this_processor );
    237237
    238238
     
    251251        __cfadbg_print_safe(runtime_core, "Kernel : Started\n--------------------------------------------------\n\n");
    252252
    253         /* paranoid */ verify( ! __preemption_enabled() );
     253        verify( ! kernelTLS.preemption_state.enabled );
    254254        enable_interrupts( __cfaabi_dbg_ctx );
    255         /* paranoid */ verify( __preemption_enabled() );
    256 
     255        verify( TL_GET( preemption_state.enabled ) );
    257256}
    258257
     
    263262        mainCluster->io.ctxs = 0p;
    264263
    265         /* paranoid */ verify( __preemption_enabled() );
     264        /* paranoid */ verify( TL_GET( preemption_state.enabled ) );
    266265        disable_interrupts();
    267         /* paranoid */ verify( ! __preemption_enabled() );
     266        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    268267
    269268        __cfadbg_print_safe(runtime_core, "\n--------------------------------------------------\nKernel : Shutting down\n");
     
    273272        // which is currently here
    274273        __atomic_store_n(&mainProcessor->do_terminate, true, __ATOMIC_RELEASE);
    275         __kernel_last_resume( __cfaabi_tls.this_processor );
     274        __kernel_last_resume( kernelTLS.this_processor );
    276275        mainThread->self_cor.state = Halted;
    277276
     
    322321                __stats_t local_stats;
    323322                __init_stats( &local_stats );
    324                 __cfaabi_tls.this_stats = &local_stats;
     323                kernelTLS.this_stats = &local_stats;
    325324        #endif
    326325
    327326        processor * proc = (processor *) arg;
    328         __cfaabi_tls.this_processor = proc;
    329         __cfaabi_tls.this_proc_id   = (__processor_id_t*)proc;
    330         __cfaabi_tls.this_thread    = 0p;
    331         __cfaabi_tls.preemption_state.[enabled, disable_count] = [false, 1];
     327        kernelTLS.this_processor = proc;
     328        kernelTLS.this_proc_id   = (__processor_id_t*)proc;
     329        kernelTLS.this_thread    = 0p;
     330        kernelTLS.preemption_state.[enabled, disable_count] = [false, 1];
    332331        // SKULLDUGGERY: We want to create a context for the processor coroutine
    333332        // which is needed for the 2-step context switch. However, there is no reason
     
    341340
    342341        //Set global state
    343         __cfaabi_tls.this_thread = 0p;
     342        kernelTLS.this_thread = 0p;
    344343
    345344        //We now have a proper context from which to schedule threads
     
    371370        $coroutine * dst = get_coroutine(this->runner);
    372371
    373         /* paranoid */ verify( ! __preemption_enabled() );
    374 
    375         __cfaabi_tls.this_thread->curr_cor = dst;
     372        verify( ! kernelTLS.preemption_state.enabled );
     373
     374        kernelTLS.this_thread->curr_cor = dst;
    376375        __stack_prepare( &dst->stack, 65000 );
    377376        __cfactx_start(main, dst, this->runner, __cfactx_invoke_coroutine);
    378377
    379         /* paranoid */ verify( ! __preemption_enabled() );
     378        verify( ! kernelTLS.preemption_state.enabled );
    380379
    381380        dst->last = &src->self_cor;
     
    395394        /* paranoid */ verify(src->state == Active);
    396395
    397         /* paranoid */ verify( ! __preemption_enabled() );
     396        verify( ! kernelTLS.preemption_state.enabled );
    398397}
    399398
     
    403402        $coroutine * dst = get_coroutine(this->runner);
    404403
    405         /* paranoid */ verify( ! __preemption_enabled() );
    406         /* paranoid */ verify( dst->starter == src );
    407         /* paranoid */ verify( dst->context.SP );
     404        verify( ! kernelTLS.preemption_state.enabled );
     405        verify( dst->starter == src );
     406        verify( dst->context.SP );
    408407
    409408        // SKULLDUGGERY in debug the processors check that the
     
    547546
    548547                P( terminated );
    549                 /* paranoid */ verify( active_processor() != &this);
     548                verify( kernelTLS.this_processor != &this);
    550549        }
    551550
     
    697696#if defined(__CFA_WITH_VERIFY__)
    698697static bool verify_fwd_bck_rng(void) {
    699         __cfaabi_tls.ready_rng.fwd_seed = 25214903917_l64u * (rdtscl() ^ (uintptr_t)&verify_fwd_bck_rng);
     698        kernelTLS.ready_rng.fwd_seed = 25214903917_l64u * (rdtscl() ^ (uintptr_t)&verify_fwd_bck_rng);
    700699
    701700        unsigned values[10];
  • libcfa/src/concurrency/kernel_private.hfa

    r54dcab1 r231b18f  
    3838#endif
    3939;
    40 
    41 extern bool __preemption_enabled();
    4240
    4341//release/wake-up the following resources
     
    183181//  creating/destroying queues
    184182static inline void ready_schedule_lock(void) with(*__scheduler_lock) {
    185         /* paranoid */ verify( ! __preemption_enabled() );
    186         /* paranoid */ verify( kernelTLS().this_proc_id );
    187 
    188         unsigned iproc = kernelTLS().this_proc_id->id;
    189         /*paranoid*/ verify(data[iproc].handle == kernelTLS().this_proc_id);
     183        /*paranoid*/ verify( kernelTLS.this_proc_id );
     184
     185        unsigned iproc = kernelTLS.this_proc_id->id;
     186        /*paranoid*/ verify(data[iproc].handle == kernelTLS.this_proc_id);
    190187        /*paranoid*/ verify(iproc < ready);
    191188
     
    210207
    211208static inline void ready_schedule_unlock(void) with(*__scheduler_lock) {
    212         /* paranoid */ verify( ! __preemption_enabled() );
    213         /* paranoid */ verify( kernelTLS().this_proc_id );
    214 
    215         unsigned iproc = kernelTLS().this_proc_id->id;
    216         /*paranoid*/ verify(data[iproc].handle == kernelTLS().this_proc_id);
     209        /*paranoid*/ verify( kernelTLS.this_proc_id );
     210
     211        unsigned iproc = kernelTLS.this_proc_id->id;
     212        /*paranoid*/ verify(data[iproc].handle == kernelTLS.this_proc_id);
    217213        /*paranoid*/ verify(iproc < ready);
    218214        /*paranoid*/ verify(data[iproc].lock);
     
    227223#ifdef __CFA_WITH_VERIFY__
    228224        static inline bool ready_schedule_islocked(void) {
    229                 /* paranoid */ verify( ! __preemption_enabled() );
    230                 /*paranoid*/ verify( kernelTLS().this_proc_id );
    231                 __processor_id_t * proc = kernelTLS().this_proc_id;
     225                /*paranoid*/ verify( kernelTLS.this_proc_id );
     226                __processor_id_t * proc = kernelTLS.this_proc_id;
    232227                return __scheduler_lock->data[proc->id].owned;
    233228        }
  • libcfa/src/concurrency/monitor.cfa

    r54dcab1 r231b18f  
    8282// Enter single monitor
    8383static void __enter( $monitor * this, const __monitor_group_t & group ) {
    84         $thread * thrd = active_thread();
    85 
    8684        // Lock the monitor spinlock
    8785        lock( this->lock __cfaabi_dbg_ctx2 );
     86        // Interrupts disable inside critical section
     87        $thread * thrd = kernelTLS.this_thread;
    8888
    8989        __cfaabi_dbg_print_safe( "Kernel : %10p Entering mon %p (%p)\n", thrd, this, this->owner);
     
    126126                __cfaabi_dbg_print_safe( "Kernel : %10p Entered  mon %p\n", thrd, this);
    127127
    128                 /* paranoid */ verifyf( active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     128                /* paranoid */ verifyf( kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    129129                return;
    130130        }
     
    132132        __cfaabi_dbg_print_safe( "Kernel : %10p Entered  mon %p\n", thrd, this);
    133133
    134         /* paranoid */ verifyf( active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     134        /* paranoid */ verifyf( kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    135135        /* paranoid */ verify( this->lock.lock );
    136136
     
    141141
    142142static void __dtor_enter( $monitor * this, fptr_t func, bool join ) {
    143         $thread * thrd = active_thread();
    144 
    145143        // Lock the monitor spinlock
    146144        lock( this->lock __cfaabi_dbg_ctx2 );
     145        // Interrupts disable inside critical section
     146        $thread * thrd = kernelTLS.this_thread;
    147147
    148148        __cfaabi_dbg_print_safe( "Kernel : %10p Entering dtor for mon %p (%p)\n", thrd, this, this->owner);
     
    155155                __set_owner( this, thrd );
    156156
    157                 verifyf( active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     157                verifyf( kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    158158
    159159                unlock( this->lock );
     
    174174                this->owner = thrd;
    175175
    176                 verifyf( active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     176                verifyf( kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    177177
    178178                unlock( this->lock );
     
    200200
    201201                // Release the next thread
    202                 /* paranoid */ verifyf( urgent->owner->waiting_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     202                /* paranoid */ verifyf( urgent->owner->waiting_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    203203                unpark( urgent->owner->waiting_thread );
    204204
     
    207207
    208208                // Some one was waiting for us, enter
    209                 /* paranoid */ verifyf( active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     209                /* paranoid */ verifyf( kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    210210        }
    211211        else {
     
    224224                park();
    225225
    226                 /* paranoid */ verifyf( active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     226                /* paranoid */ verifyf( kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    227227                return;
    228228        }
     
    237237        lock( this->lock __cfaabi_dbg_ctx2 );
    238238
    239         __cfaabi_dbg_print_safe( "Kernel : %10p Leaving mon %p (%p)\n", active_thread(), this, this->owner);
    240 
    241         /* paranoid */ verifyf( active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     239        __cfaabi_dbg_print_safe( "Kernel : %10p Leaving mon %p (%p)\n", kernelTLS.this_thread, this, this->owner);
     240
     241        /* paranoid */ verifyf( kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    242242
    243243        // Leaving a recursion level, decrement the counter
     
    270270void __dtor_leave( $monitor * this, bool join ) {
    271271        __cfaabi_dbg_debug_do(
    272                 if( active_thread() != this->owner ) {
    273                         abort( "Destroyed monitor %p has inconsistent owner, expected %p got %p.\n", this, active_thread(), this->owner);
     272                if( TL_GET( this_thread ) != this->owner ) {
     273                        abort( "Destroyed monitor %p has inconsistent owner, expected %p got %p.\n", this, TL_GET( this_thread ), this->owner);
    274274                }
    275275                if( this->recursion != 1  && !join ) {
     
    287287        /* paranoid */ verify( this->lock.lock );
    288288        /* paranoid */ verifyf( thrd == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", thrd, this->owner, this->recursion, this );
    289         /* paranoid */ verify( ! __preemption_enabled() );
     289        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    290290        /* paranoid */ verify( thrd->state == Halted );
    291291        /* paranoid */ verify( this->recursion == 1 );
     
    303303        // Unpark the next owner if needed
    304304        /* paranoid */ verifyf( !new_owner || new_owner == this->owner, "Expected owner to be %p, got %p (m: %p)", new_owner, this->owner, this );
    305         /* paranoid */ verify( ! __preemption_enabled() );
     305        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    306306        /* paranoid */ verify( thrd->state == Halted );
    307307        unpark( new_owner );
     
    327327// Sorts monitors before entering
    328328void ?{}( monitor_guard_t & this, $monitor * m [], __lock_size_t count, fptr_t func ) {
    329         $thread * thrd = active_thread();
     329        $thread * thrd = TL_GET( this_thread );
    330330
    331331        // Store current array
     
    362362
    363363        // Restore thread context
    364         active_thread()->monitors = this.prev;
     364        TL_GET( this_thread )->monitors = this.prev;
    365365}
    366366
     
    369369void ?{}( monitor_dtor_guard_t & this, $monitor * m [], fptr_t func, bool join ) {
    370370        // optimization
    371         $thread * thrd = active_thread();
     371        $thread * thrd = TL_GET( this_thread );
    372372
    373373        // Store current array
     
    392392
    393393        // Restore thread context
    394         active_thread()->monitors = this.prev;
     394        TL_GET( this_thread )->monitors = this.prev;
    395395}
    396396
     
    432432
    433433        // Create the node specific to this wait operation
    434         wait_ctx( active_thread(), user_info );
     434        wait_ctx( TL_GET( this_thread ), user_info );
    435435
    436436        // Append the current wait operation to the ones already queued on the condition
     
    483483        //Some more checking in debug
    484484        __cfaabi_dbg_debug_do(
    485                 $thread * this_thrd = active_thread();
     485                $thread * this_thrd = TL_GET( this_thread );
    486486                if ( this.monitor_count != this_thrd->monitors.size ) {
    487487                        abort( "Signal on condition %p made with different number of monitor(s), expected %zi got %zi", &this, this.monitor_count, this_thrd->monitors.size );
     
    531531
    532532        // Create the node specific to this wait operation
    533         wait_ctx_primed( active_thread(), 0 )
     533        wait_ctx_primed( kernelTLS.this_thread, 0 )
    534534
    535535        //save contexts
     
    630630
    631631                                // Create the node specific to this wait operation
    632                                 wait_ctx_primed( active_thread(), 0 );
     632                                wait_ctx_primed( kernelTLS.this_thread, 0 );
    633633
    634634                                // Save monitor states
     
    682682
    683683        // Create the node specific to this wait operation
    684         wait_ctx_primed( active_thread(), 0 );
     684        wait_ctx_primed( kernelTLS.this_thread, 0 );
    685685
    686686        monitor_save;
     
    688688
    689689        for( __lock_size_t i = 0; i < count; i++) {
    690                 verify( monitors[i]->owner == active_thread() );
     690                verify( monitors[i]->owner == kernelTLS.this_thread );
    691691        }
    692692
     
    724724static inline void __set_owner( $monitor * monitors [], __lock_size_t count, $thread * owner ) {
    725725        /* paranoid */ verify ( monitors[0]->lock.lock );
    726         /* paranoid */ verifyf( monitors[0]->owner == active_thread(), "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), monitors[0]->owner, monitors[0]->recursion, monitors[0] );
     726        /* paranoid */ verifyf( monitors[0]->owner == kernelTLS.this_thread, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, monitors[0]->owner, monitors[0]->recursion, monitors[0] );
    727727        monitors[0]->owner        = owner;
    728728        monitors[0]->recursion    = 1;
    729729        for( __lock_size_t i = 1; i < count; i++ ) {
    730730                /* paranoid */ verify ( monitors[i]->lock.lock );
    731                 /* paranoid */ verifyf( monitors[i]->owner == active_thread(), "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), monitors[i]->owner, monitors[i]->recursion, monitors[i] );
     731                /* paranoid */ verifyf( monitors[i]->owner == kernelTLS.this_thread, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, monitors[i]->owner, monitors[i]->recursion, monitors[i] );
    732732                monitors[i]->owner        = owner;
    733733                monitors[i]->recursion    = 0;
     
    755755                //regardless of if we are ready to baton pass,
    756756                //we need to set the monitor as in use
    757                 /* paranoid */ verifyf( !this->owner || active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     757                /* paranoid */ verifyf( !this->owner || kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    758758                __set_owner( this,  urgent->owner->waiting_thread );
    759759
     
    764764        // Get the next thread in the entry_queue
    765765        $thread * new_owner = pop_head( this->entry_queue );
    766         /* paranoid */ verifyf( !this->owner || active_thread() == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", active_thread(), this->owner, this->recursion, this );
     766        /* paranoid */ verifyf( !this->owner || kernelTLS.this_thread == this->owner, "Expected owner to be %p, got %p (r: %i, m: %p)", kernelTLS.this_thread, this->owner, this->recursion, this );
    767767        /* paranoid */ verify( !new_owner || new_owner->link.next == 0p );
    768768        __set_owner( this, new_owner );
     
    892892
    893893static inline void brand_condition( condition & this ) {
    894         $thread * thrd = active_thread();
     894        $thread * thrd = TL_GET( this_thread );
    895895        if( !this.monitors ) {
    896896                // __cfaabi_dbg_print_safe( "Branding\n" );
  • libcfa/src/concurrency/preemption.cfa

    r54dcab1 r231b18f  
    163163// Kernel Signal Tools
    164164//=============================================================================================
     165
    165166// In a user-level threading system, there are handful of thread-local variables where this problem occurs on the ARM.
    166 //
     167// 
    167168// For each kernel thread running user-level threads, there is a flag variable to indicate if interrupts are
    168169// enabled/disabled for that kernel thread. Therefore, this variable is made thread local.
    169 //
     170// 
    170171// For example, this code fragment sets the state of the "interrupt" variable in thread-local memory.
    171 //
     172// 
    172173// _Thread_local volatile int interrupts;
    173174// int main() {
    174175//     interrupts = 0; // disable interrupts }
    175 //
     176// 
    176177// which generates the following code on the ARM
    177 //
     178// 
    178179// (gdb) disassemble main
    179180// Dump of assembler code for function main:
     
    184185//    0x0000000000000620 <+16>: str     wzr, [x1]
    185186//    0x0000000000000624 <+20>: ret
    186 //
     187// 
    187188// The mrs moves a pointer from coprocessor register tpidr_el0 into register x1.  Register w0 is set to 0. The two adds
    188189// increase the TLS pointer with the displacement (offset) 0x10, which is the location in the TSL of variable
     
    192193// the user-level ready-queue it is run on kernel thread M. It now stores 0 into "interrupts" back on kernel thread N,
    193194// turning off interrupt on the wrong kernel thread.
    194 //
     195// 
    195196// On the x86, the following code is generated for the same code fragment.
    196 //
     197// 
    197198// (gdb) disassemble main
    198199// Dump of assembler code for function main:
    199200//    0x0000000000400420 <+0>:  movl   $0x0,%fs:0xfffffffffffffffc
    200201//    0x000000000040042c <+12>: xor    %eax,%eax
    201 //    0x000000000040042e <+14>: retq
    202 //
     202//    0x000000000040042e <+14>: retq   
     203// 
    203204// and there is base-displacement addressing used to atomically reset variable "interrupts" off of the TSL pointer in
    204205// register "fs".
    205 //
     206// 
    206207// Hence, the ARM has base-displacement address for the general purpose registers, BUT not to the coprocessor
    207208// registers. As a result, generating the address for the write into variable "interrupts" is no longer atomic.
    208 //
     209// 
    209210// Note this problem does NOT occur when just using multiple kernel threads because the preemption ALWAYS restarts the
    210211// thread on the same kernel thread.
    211 //
     212// 
    212213// The obvious question is why does ARM use a coprocessor register to store the TSL pointer given that coprocessor
    213214// registers are second-class registers with respect to the instruction set. One possible answer is that they did not
     
    215216// available.
    216217
    217 //----------
    218 // special case for preemption since used often
    219 bool __preemption_enabled() {
    220         // create a assembler label before
    221         // marked as clobber all to avoid movement
    222         asm volatile("__cfaasm_check_before:":::"memory");
    223 
    224         // access tls as normal
    225         bool enabled = __cfaabi_tls.preemption_state.enabled;
    226 
    227         // create a assembler label after
    228         // marked as clobber all to avoid movement
    229         asm volatile("__cfaasm_check_after:":::"memory");
    230         return enabled;
    231 }
    232 
    233 //----------
    234 // Get data from the TLS block
    235 uintptr_t __cfatls_get( unsigned long int offset ) __attribute__((__noinline__)); //no inline to avoid problems
    236 uintptr_t __cfatls_get( unsigned long int offset ) {
    237         // create a assembler label before
    238         // marked as clobber all to avoid movement
    239         asm volatile("__cfaasm_get_before:":::"memory");
    240 
    241         // access tls as normal (except for pointer arithmetic)
    242         uintptr_t val = *(uintptr_t*)((uintptr_t)&__cfaabi_tls + offset);
    243 
    244         // create a assembler label after
    245         // marked as clobber all to avoid movement
    246         asm volatile("__cfaasm_get_after:":::"memory");
    247         return val;
    248 }
     218__cfaabi_dbg_debug_do( static thread_local void * last_interrupt = 0; )
    249219
    250220extern "C" {
    251221        // Disable interrupts by incrementing the counter
    252222        void disable_interrupts() {
    253                 // create a assembler label before
    254                 // marked as clobber all to avoid movement
    255                 asm volatile("__cfaasm_disable_before:":::"memory");
    256 
    257                 with( __cfaabi_tls.preemption_state ) {
     223                with( kernelTLS.preemption_state ) {
    258224                        #if GCC_VERSION > 50000
    259225                        static_assert(__atomic_always_lock_free(sizeof(enabled), &enabled), "Must be lock-free");
     
    272238                        verify( new_val < 65_000u );              // If this triggers someone is disabling interrupts without enabling them
    273239                }
    274 
    275                 // create a assembler label after
    276                 // marked as clobber all to avoid movement
    277                 asm volatile("__cfaasm_disable_after:":::"memory");
    278240        }
    279241
     
    281243        // If counter reaches 0, execute any pending __cfactx_switch
    282244        void enable_interrupts( __cfaabi_dbg_ctx_param ) {
    283                 // create a assembler label before
    284                 // marked as clobber all to avoid movement
    285                 asm volatile("__cfaasm_enable_before:":::"memory");
    286 
    287                 processor   * proc = __cfaabi_tls.this_processor; // Cache the processor now since interrupts can start happening after the atomic store
     245                processor   * proc = kernelTLS.this_processor; // Cache the processor now since interrupts can start happening after the atomic store
    288246                /* paranoid */ verify( proc );
    289247
    290                 with( __cfaabi_tls.preemption_state ){
     248                with( kernelTLS.preemption_state ){
    291249                        unsigned short prev = disable_count;
    292250                        disable_count -= 1;
     
    315273                // For debugging purposes : keep track of the last person to enable the interrupts
    316274                __cfaabi_dbg_debug_do( proc->last_enable = caller; )
    317 
    318                 // create a assembler label after
    319                 // marked as clobber all to avoid movement
    320                 asm volatile("__cfaasm_enable_after:":::"memory");
    321275        }
    322276
     
    324278        // Don't execute any pending __cfactx_switch even if counter reaches 0
    325279        void enable_interrupts_noPoll() {
    326                 // create a assembler label before
    327                 // marked as clobber all to avoid movement
    328                 asm volatile("__cfaasm_nopoll_before:":::"memory");
    329 
    330                 unsigned short prev = __cfaabi_tls.preemption_state.disable_count;
    331                 __cfaabi_tls.preemption_state.disable_count -= 1;
     280                unsigned short prev = kernelTLS.preemption_state.disable_count;
     281                kernelTLS.preemption_state.disable_count -= 1;
    332282                verifyf( prev != 0u, "Incremented from %u\n", prev );                     // If this triggers someone is enabled already enabled interrupts
    333283                if( prev == 1 ) {
    334284                        #if GCC_VERSION > 50000
    335                         static_assert(__atomic_always_lock_free(sizeof(__cfaabi_tls.preemption_state.enabled), &__cfaabi_tls.preemption_state.enabled), "Must be lock-free");
     285                        static_assert(__atomic_always_lock_free(sizeof(kernelTLS.preemption_state.enabled), &kernelTLS.preemption_state.enabled), "Must be lock-free");
    336286                        #endif
    337287                        // Set enabled flag to true
    338288                        // should be atomic to avoid preemption in the middle of the operation.
    339289                        // use memory order RELAXED since there is no inter-thread on this variable requirements
    340                         __atomic_store_n(&__cfaabi_tls.preemption_state.enabled, true, __ATOMIC_RELAXED);
     290                        __atomic_store_n(&kernelTLS.preemption_state.enabled, true, __ATOMIC_RELAXED);
    341291
    342292                        // Signal the compiler that a fence is needed but only for signal handlers
    343293                        __atomic_signal_fence(__ATOMIC_RELEASE);
    344294                }
    345 
    346                 // create a assembler label after
    347                 // marked as clobber all to avoid movement
    348                 asm volatile("__cfaasm_nopoll_after:":::"memory");
    349295        }
    350296}
     
    381327static void timeout( $thread * this ) {
    382328        #if !defined( __CFA_NO_STATISTICS__ )
    383                 kernelTLS().this_stats = this->curr_cluster->stats;
     329                kernelTLS.this_stats = this->curr_cluster->stats;
    384330        #endif
    385331        unpark( this );
     
    392338static inline bool preemption_ready() {
    393339        // Check if preemption is safe
    394         bool ready = __cfaabi_tls.preemption_state.enabled && ! __cfaabi_tls.preemption_state.in_progress;
     340        bool ready = kernelTLS.preemption_state.enabled && ! kernelTLS.preemption_state.in_progress;
    395341
    396342        // Adjust the pending flag accordingly
    397         __cfaabi_tls.this_processor->pending_preemption = !ready;
     343        kernelTLS.this_processor->pending_preemption = !ready;
    398344        return ready;
    399345}
     
    409355
    410356        // Start with preemption disabled until ready
    411         __cfaabi_tls.preemption_state.enabled = false;
    412         __cfaabi_tls.preemption_state.disable_count = 1;
     357        kernelTLS.preemption_state.enabled = false;
     358        kernelTLS.preemption_state.disable_count = 1;
    413359
    414360        // Initialize the event kernel
     
    468414// Kernel Signal Handlers
    469415//=============================================================================================
    470 struct asm_region {
    471         void * before;
    472         void * after;
    473 };
    474 
    475 //-----------------------------------------------------------------------------
    476 // Some assembly required
    477 #if defined( __i386 )
    478         #define __cfaasm_label( label ) \
    479                 ({ \
    480                         struct asm_region region; \
    481                         asm( \
    482                                 "movl $__cfaasm_" #label "_before, %[vb]\n\t" \
    483                                 "movl $__cfaasm_" #label "_after , %[va]\n\t" \
    484                                  : [vb]"=r"(region.before), [vb]"=r"(region.before) \
    485                         ); \
    486                         region; \
    487                 });
    488 #elif defined( __x86_64 )
    489         #ifdef __PIC__
    490                 #define PLT "@PLT"
    491         #else
    492                 #define PLT ""
    493         #endif
    494         #define __cfaasm_label( label ) \
    495                 ({ \
    496                         struct asm_region region; \
    497                         asm( \
    498                                 "movq $__cfaasm_" #label "_before" PLT ", %[vb]\n\t" \
    499                                 "movq $__cfaasm_" #label "_after"  PLT ", %[va]\n\t" \
    500                                  : [vb]"=r"(region.before), [va]"=r"(region.after) \
    501                         ); \
    502                         region; \
    503                 });
    504 #elif defined( __aarch64__ )
    505         #error __cfaasm_label undefined for arm
    506 #else
    507         #error unknown hardware architecture
    508 #endif
    509 
    510 __cfaabi_dbg_debug_do( static thread_local void * last_interrupt = 0; )
    511416
    512417// Context switch signal handler
    513418// Receives SIGUSR1 signal and causes the current thread to yield
    514419static void sigHandler_ctxSwitch( __CFA_SIGPARMS__ ) {
    515         void * ip = (void *)(cxt->uc_mcontext.CFA_REG_IP);
    516         __cfaabi_dbg_debug_do( last_interrupt = ip; )
     420        __cfaabi_dbg_debug_do( last_interrupt = (void *)(cxt->uc_mcontext.CFA_REG_IP); )
    517421
    518422        // SKULLDUGGERY: if a thread creates a processor and the immediately deletes it,
     
    520424        // before the kernel thread has even started running. When that happens, an interrupt
    521425        // with a null 'this_processor' will be caught, just ignore it.
    522         if(! __cfaabi_tls.this_processor ) return;
     426        if(! kernelTLS.this_processor ) return;
    523427
    524428        choose(sfp->si_value.sival_int) {
    525429                case PREEMPT_NORMAL   : ;// Normal case, nothing to do here
    526                 case PREEMPT_TERMINATE: verify( __atomic_load_n( &__cfaabi_tls.this_processor->do_terminate, __ATOMIC_SEQ_CST ) );
     430                case PREEMPT_TERMINATE: verify( __atomic_load_n( &kernelTLS.this_processor->do_terminate, __ATOMIC_SEQ_CST ) );
    527431                default:
    528432                        abort( "internal error, signal value is %d", sfp->si_value.sival_int );
     
    532436        if( !preemption_ready() ) { return; }
    533437
    534         struct asm_region region;
    535         region = __cfaasm_label( get     ); if( ip >= region.before && ip <= region.after ) return;
    536         region = __cfaasm_label( check   ); if( ip >= region.before && ip <= region.after ) return;
    537         region = __cfaasm_label( disable ); if( ip >= region.before && ip <= region.after ) return;
    538         region = __cfaasm_label( enable  ); if( ip >= region.before && ip <= region.after ) return;
    539         region = __cfaasm_label( nopoll  ); if( ip >= region.before && ip <= region.after ) return;
    540 
    541         __cfaabi_dbg_print_buffer_decl( " KERNEL: preempting core %p (%p @ %p).\n", __cfaabi_tls.this_processor, __cfaabi_tls.this_thread, (void *)(cxt->uc_mcontext.CFA_REG_IP) );
     438        __cfaabi_dbg_print_buffer_decl( " KERNEL: preempting core %p (%p @ %p).\n", kernelTLS.this_processor, kernelTLS.this_thread, (void *)(cxt->uc_mcontext.CFA_REG_IP) );
    542439
    543440        // Sync flag : prevent recursive calls to the signal handler
    544         __cfaabi_tls.preemption_state.in_progress = true;
     441        kernelTLS.preemption_state.in_progress = true;
    545442
    546443        // Clear sighandler mask before context switching.
     
    552449        }
    553450
     451        // TODO: this should go in finish action
    554452        // Clear the in progress flag
    555         __cfaabi_tls.preemption_state.in_progress = false;
     453        kernelTLS.preemption_state.in_progress = false;
    556454
    557455        // Preemption can occur here
     
    570468        id.full_proc = false;
    571469        id.id = doregister(&id);
    572         __cfaabi_tls.this_proc_id = &id;
     470        kernelTLS.this_proc_id = &id;
    573471
    574472        // Block sigalrms to control when they arrive
     
    638536
    639537void __cfaabi_check_preemption() {
    640         bool ready = __preemption_enabled();
     538        bool ready = kernelTLS.preemption_state.enabled;
    641539        if(!ready) { abort("Preemption should be ready"); }
    642540
     
    661559#ifdef __CFA_WITH_VERIFY__
    662560bool __cfaabi_dbg_in_kernel() {
    663         return !__preemption_enabled();
     561        return !kernelTLS.preemption_state.enabled;
    664562}
    665563#endif
  • libcfa/src/concurrency/ready_queue.cfa

    r54dcab1 r231b18f  
    150150//  queues or removing them.
    151151uint_fast32_t ready_mutate_lock( void ) with(*__scheduler_lock) {
    152         /* paranoid */ verify( ! __preemption_enabled() );
     152        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    153153
    154154        // Step 1 : lock global lock
     
    166166        }
    167167
    168         /* paranoid */ verify( ! __preemption_enabled() );
     168        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    169169        return s;
    170170}
    171171
    172172void ready_mutate_unlock( uint_fast32_t last_s ) with(*__scheduler_lock) {
    173         /* paranoid */ verify( ! __preemption_enabled() );
     173        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    174174
    175175        // Step 1 : release local locks
     
    188188        __atomic_store_n(&lock, (bool)false, __ATOMIC_RELEASE);
    189189
    190         /* paranoid */ verify( ! __preemption_enabled() );
     190        /* paranoid */ verify( ! kernelTLS.preemption_state.enabled );
    191191}
    192192
     
    252252                preferred =
    253253                        //*
    254                         kernelTLS().this_processor ? kernelTLS().this_processor->id * 4 : -1;
     254                        kernelTLS.this_processor ? kernelTLS.this_processor->id * 4 : -1;
    255255                        /*/
    256256                        thrd->link.preferred * 4;
     
    331331                // Don't bother trying locally too much
    332332                int local_tries = 8;
    333                 preferred = kernelTLS().this_processor->id * 4;
     333                preferred = kernelTLS.this_processor->id * 4;
    334334        #endif
    335335
  • tools/stat.py

    r54dcab1 r231b18f  
    1 #!/usr/bin/python3
     1#!/usr/bin/python
    22
    33import sys
     
    1717                avg = numpy.mean  (content)
    1818                std = numpy.std   (content)
    19                 print("median {0:.1f} avg {1:.1f} stddev {2:.1f}".format( med, avg, std ))
     19                print "median {0:.1f} avg {1:.1f} stddev {2:.1f}".format( med, avg, std )
    2020
    2121
Note: See TracChangeset for help on using the changeset viewer.