source: libcfa/src/concurrency/kernel_private.hfa @ f80f840

ADTarm-ehast-experimentalenumforall-pointer-decayjacob/cs343-translationnew-astnew-ast-unique-exprpthread-emulationqualifiedEnum
Last change on this file since f80f840 was f80f840, checked in by Thierry Delisle <tdelisle@…>, 4 years ago

Merge branch 'master' into relaxed_ready

  • Property mode set to 100644
File size: 6.7 KB
RevLine 
[75f3522]1//
2// Cforall Version 1.0.0 Copyright (C) 2016 University of Waterloo
3//
4// The contents of this file are covered under the licence agreement in the
5// file "LICENCE" distributed with Cforall.
6//
[73abe95]7// kernel_private.hfa --
[75f3522]8//
9// Author           : Thierry Delisle
10// Created On       : Mon Feb 13 12:27:26 2017
[6b0b624]11// Last Modified By : Peter A. Buhr
[1805b1b]12// Last Modified On : Sat Nov 30 19:25:02 2019
13// Update Count     : 8
[75f3522]14//
15
[6b0b624]16#pragma once
[75f3522]17
[58b6d1b]18#include "kernel.hfa"
19#include "thread.hfa"
[75f3522]20
[73abe95]21#include "alarm.hfa"
[fa21ac9]22
[4aa2fb2]23
[75f3522]24//-----------------------------------------------------------------------------
25// Scheduler
[1c273d0]26
27extern "C" {
[2026bb6]28        void disable_interrupts() OPTIONAL_THREAD;
[969b3fe]29        void enable_interrupts_noPoll();
[36982fc]30        void enable_interrupts( __cfaabi_dbg_ctx_param );
[1c273d0]31}
32
[348006f]33void ScheduleThread( thread_desc * );
[1c273d0]34static inline void WakeThread( thread_desc * thrd ) {
35        if( !thrd ) return;
36
[722c0848]37        verify(thrd->state == Inactive);
38
[1c273d0]39        disable_interrupts();
40        ScheduleThread( thrd );
[36982fc]41        enable_interrupts( __cfaabi_dbg_ctx );
[1c273d0]42}
[348006f]43thread_desc * nextThread(cluster * this);
[75f3522]44
[e60e0dc]45//Block current thread and release/wake-up the following resources
[82ff5845]46void BlockInternal(void);
[ea7d2b0]47void BlockInternal(__spinlock_t * lock);
[82ff5845]48void BlockInternal(thread_desc * thrd);
[ea7d2b0]49void BlockInternal(__spinlock_t * lock, thread_desc * thrd);
50void BlockInternal(__spinlock_t * locks [], unsigned short count);
51void BlockInternal(__spinlock_t * locks [], unsigned short count, thread_desc * thrds [], unsigned short thrd_count);
[09800e9]52void BlockInternal(__finish_callback_fptr_t callback);
[ea7d2b0]53void LeaveThread(__spinlock_t * lock, thread_desc * thrd);
[db6f06a]54
[75f3522]55//-----------------------------------------------------------------------------
56// Processor
57void main(processorCtx_t *);
[85b1deb]58
[1805b1b]59void * create_pthread( pthread_t *, void * (*)(void *), void * );
60
[85b1deb]61static inline void wake_fast(processor * this) {
62        __cfaabi_dbg_print_safe("Kernel : Waking up processor %p\n", this);
63        post( this->idleLock );
64}
65
66static inline void wake(processor * this) {
67        disable_interrupts();
68        wake_fast(this);
69        enable_interrupts( __cfaabi_dbg_ctx );
70}
[75f3522]71
[e60e0dc]72struct event_kernel_t {
[fa21ac9]73        alarm_list_t alarms;
[ea7d2b0]74        __spinlock_t lock;
[fa21ac9]75};
76
[e60e0dc]77extern event_kernel_t * event_kernel;
78
[d8548e2]79struct __cfa_kernel_preemption_state_t {
[b69ea6b]80        bool enabled;
81        bool in_progress;
82        unsigned short disable_count;
83};
84
[afc2427]85extern volatile thread_local __cfa_kernel_preemption_state_t preemption_state __attribute__ ((tls_model ( "initial-exec" )));
[c81ebf9]86
[75f3522]87//-----------------------------------------------------------------------------
88// Threads
89extern "C" {
90      forall(dtype T | is_thread(T))
91      void CtxInvokeThread(T * this);
92}
93
[c3acb841]94extern void ThreadCtxSwitch(coroutine_desc * src, coroutine_desc * dst);
[75f3522]95
[f7d6bb0]96__cfaabi_dbg_debug_do(
97        extern void __cfaabi_dbg_thread_register  ( thread_desc * thrd );
98        extern void __cfaabi_dbg_thread_unregister( thread_desc * thrd );
99)
100
[969b3fe]101//-----------------------------------------------------------------------------
102// Utils
[7768b8d]103#define KERNEL_STORAGE(T,X) __attribute((aligned(__alignof__(T)))) static char storage_##X[sizeof(T)]
[969b3fe]104
[21184e3]105static inline uint32_t tls_rand() {
106        kernelTLS.rand_seed ^= kernelTLS.rand_seed << 6;
107        kernelTLS.rand_seed ^= kernelTLS.rand_seed >> 21;
108        kernelTLS.rand_seed ^= kernelTLS.rand_seed << 7;
109        return kernelTLS.rand_seed;
110}
111
[de94a60]112
[a1a17a7]113void doregister( struct cluster & cltr );
114void unregister( struct cluster & cltr );
[de94a60]115
[a1a17a7]116void doregister( struct cluster * cltr, struct thread_desc & thrd );
117void unregister( struct cluster * cltr, struct thread_desc & thrd );
[de94a60]118
[7768b8d]119//=======================================================================
120// Cluster lock API
121//=======================================================================
122struct __attribute__((aligned(64))) __processor_id {
123        processor * volatile handle;
124        volatile bool lock;
125};
126
127// Lock-Free registering/unregistering of threads
128// Register a processor to a given cluster and get its unique id in return
129unsigned doregister( struct cluster * cltr, struct processor * proc );
130
131// Unregister a processor from a given cluster using its id, getting back the original pointer
132void     unregister( struct cluster * cltr, struct processor * proc );
133
134//=======================================================================
135// Reader-writer lock implementation
136// Concurrent with doregister/unregister,
137//    i.e., threads can be added at any point during or between the entry/exit
138static inline void __atomic_acquire(volatile bool * ll) {
139        while( __builtin_expect(__atomic_exchange_n(ll, (bool)true, __ATOMIC_SEQ_CST), false) ) {
140                while(__atomic_load_n(ll, (int)__ATOMIC_RELAXED))
141                        asm volatile("pause");
142        }
143        /* paranoid */ verify(*ll);
144}
145
146static inline bool __atomic_try_acquire(volatile bool * ll) {
[b798713]147        return !__atomic_exchange_n(ll, (bool)true, __ATOMIC_SEQ_CST);
[7768b8d]148}
149
150static inline void __atomic_unlock(volatile bool * ll) {
151        /* paranoid */ verify(*ll);
152        __atomic_store_n(ll, (bool)false, __ATOMIC_RELEASE);
153}
154
155//-----------------------------------------------------------------------
156// Reader side : acquire when using the ready queue to schedule but not
157//  creating/destroying queues
[b798713]158static inline void ready_schedule_lock( struct cluster * cltr, struct processor * proc) with(cltr->ready_lock) {
[7768b8d]159        unsigned iproc = proc->id;
160        /*paranoid*/ verify(data[iproc].handle == proc);
161        /*paranoid*/ verify(iproc < ready);
162
163        // Step 1 : make sure no writer are in the middle of the critical section
164        while(__atomic_load_n(&lock, (int)__ATOMIC_RELAXED))
165                asm volatile("pause");
166
167        // Fence needed because we don't want to start trying to acquire the lock
168        // before we read a false.
169        // Not needed on x86
170        // std::atomic_thread_fence(std::memory_order_seq_cst);
171
172        // Step 2 : acquire our local lock
173        __atomic_acquire( &data[iproc].lock );
174        /*paranoid*/ verify(data[iproc].lock);
175}
176
[b798713]177static inline void ready_schedule_unlock( struct cluster * cltr, struct processor * proc) with(cltr->ready_lock) {
[7768b8d]178        unsigned iproc = proc->id;
179        /*paranoid*/ verify(data[iproc].handle == proc);
180        /*paranoid*/ verify(iproc < ready);
181        /*paranoid*/ verify(data[iproc].lock);
182        __atomic_store_n(&data[iproc].lock, false, __ATOMIC_RELEASE);
183}
184
185//-----------------------------------------------------------------------
186// Writer side : acquire when changing the ready queue, e.g. adding more
187//  queues or removing them.
188uint_fast32_t ready_mutate_lock( struct cluster & cltr );
189
190void ready_mutate_unlock( struct cluster & cltr, uint_fast32_t );
191
[b798713]192//=======================================================================
193// Ready-Queue API
194
195__attribute__((hot)) bool push(struct cluster * cltr, struct thread_desc * thrd);
196__attribute__((hot)) thread_desc * pop(struct cluster * cltr);
197void ready_queue_grow  (struct cluster * cltr);
198void ready_queue_shrink(struct cluster * cltr);
199
200#if !defined(__CFA_NO_STATISTICS__)
201void stats_tls_tally(struct cluster * cltr);
202#else
203static inline void stats_tls_tally(struct cluster * cltr) {}
204#endif
[de94a60]205
[75f3522]206// Local Variables: //
207// mode: c //
208// tab-width: 4 //
[4aa2fb2]209// End: //
Note: See TracBrowser for help on using the repository browser.