source: libcfa/src/concurrency/kernel.hfa @ b798713

ADTarm-ehast-experimentalenumforall-pointer-decayjacob/cs343-translationnew-astnew-ast-unique-exprpthread-emulationqualifiedEnum
Last change on this file since b798713 was b798713, checked in by Thierry Delisle <tdelisle@…>, 4 years ago

Working ready queue

  • Property mode set to 100644
File size: 7.7 KB
Line 
1//
2// Cforall Version 1.0.0 Copyright (C) 2016 University of Waterloo
3//
4// The contents of this file are covered under the licence agreement in the
5// file "LICENCE" distributed with Cforall.
6//
7// kernel --
8//
9// Author           : Thierry Delisle
10// Created On       : Tue Jan 17 12:27:26 2017
11// Last Modified By : Peter A. Buhr
12// Last Modified On : Sat Jun 22 11:39:17 2019
13// Update Count     : 16
14//
15
16#pragma once
17
18#include <stdbool.h>
19
20#include "invoke.h"
21#include "time_t.hfa"
22#include "coroutine.hfa"
23
24extern "C" {
25#include <pthread.h>
26#include <semaphore.h>
27}
28
29//-----------------------------------------------------------------------------
30// Locks
31struct semaphore {
32        __spinlock_t lock;
33        int count;
34        __queue_t(thread_desc) waiting;
35};
36
37void  ?{}(semaphore & this, int count = 1);
38void ^?{}(semaphore & this);
39void   P (semaphore & this);
40void   V (semaphore & this);
41
42
43//-----------------------------------------------------------------------------
44// Processor
45extern struct cluster * mainCluster;
46
47enum FinishOpCode { No_Action, Release, Schedule, Release_Schedule, Release_Multi, Release_Multi_Schedule, Callback };
48
49typedef void (*__finish_callback_fptr_t)(void);
50
51//TODO use union, many of these fields are mutually exclusive (i.e. MULTI vs NOMULTI)
52struct FinishAction {
53        FinishOpCode action_code;
54        /*
55        // Union of possible actions
56        union {
57                // Option 1 : locks and threads
58                struct {
59                        // 1 thread or N thread
60                        union {
61                                thread_desc * thrd;
62                                struct {
63                                        thread_desc ** thrds;
64                                        unsigned short thrd_count;
65                                };
66                        };
67                        // 1 lock or N lock
68                        union {
69                                __spinlock_t * lock;
70                                struct {
71                                        __spinlock_t ** locks;
72                                        unsigned short lock_count;
73                                };
74                        };
75                };
76                // Option 2 : action pointer
77                __finish_callback_fptr_t callback;
78        };
79        /*/
80        thread_desc * thrd;
81        thread_desc ** thrds;
82        unsigned short thrd_count;
83        __spinlock_t * lock;
84        __spinlock_t ** locks;
85        unsigned short lock_count;
86        __finish_callback_fptr_t callback;
87        //*/
88};
89static inline void ?{}(FinishAction & this) {
90        this.action_code = No_Action;
91        this.thrd = NULL;
92        this.lock = NULL;
93}
94static inline void ^?{}(FinishAction &) {}
95
96// Processor
97coroutine processorCtx_t {
98        struct processor * proc;
99};
100
101// Wrapper around kernel threads
102struct processor {
103        // Main state
104        // Coroutine ctx who does keeps the state of the processor
105        struct processorCtx_t runner;
106
107        // Cluster from which to get threads
108        struct cluster * cltr;
109        unsigned int id;
110
111        // Name of the processor
112        const char * name;
113
114        // Handle to pthreads
115        pthread_t kernel_thread;
116
117        // RunThread data
118        // Action to do after a thread is ran
119        struct FinishAction finish;
120
121        // Preemption data
122        // Node which is added in the discrete event simulaiton
123        struct alarm_node_t * preemption_alarm;
124
125        // If true, a preemption was triggered in an unsafe region, the processor must preempt as soon as possible
126        bool pending_preemption;
127
128        // Idle lock
129        __bin_sem_t idleLock;
130
131        // Termination
132        // Set to true to notify the processor should terminate
133        volatile bool do_terminate;
134
135        // Termination synchronisation
136        semaphore terminated;
137
138        // Link lists fields
139        struct __dbg_node_proc {
140                struct processor * next;
141                struct processor * prev;
142        } node;
143
144#ifdef __CFA_DEBUG__
145        // Last function to enable preemption on this processor
146        const char * last_enable;
147#endif
148};
149
150void  ?{}(processor & this, const char * name, struct cluster & cltr);
151void ^?{}(processor & this);
152
153static inline void  ?{}(processor & this)                    { this{ "Anonymous Processor", *mainCluster}; }
154static inline void  ?{}(processor & this, struct cluster & cltr)    { this{ "Anonymous Processor", cltr}; }
155static inline void  ?{}(processor & this, const char * name) { this{name, *mainCluster }; }
156
157static inline [processor *&, processor *& ] __get( processor & this ) {
158        return this.node.[next, prev];
159}
160
161
162//-----------------------------------------------------------------------------
163// Cluster Tools
164struct __processor_id;
165
166// Reader-Writer lock protecting the ready-queue
167struct __clusterRWLock_t {
168        // total cachelines allocated
169        unsigned int max;
170
171        // cachelines currently in use
172        volatile unsigned int alloc;
173
174        // cachelines ready to itereate over
175        // (!= to alloc when thread is in second half of doregister)
176        volatile unsigned int ready;
177
178        // writer lock
179        volatile bool lock;
180
181        // data pointer
182        __processor_id * data;
183};
184
185void  ?{}(__clusterRWLock_t & this);
186void ^?{}(__clusterRWLock_t & this);
187
188// Underlying sub quues of the ready queue
189struct __attribute__((aligned(128))) __intrusive_ready_queue_t {
190        // spin lock protecting the queue
191        volatile bool lock;
192        unsigned int last_id;
193
194        // anchor for the head and the tail of the queue
195        struct __sentinel_t {
196                // Link lists fields
197                // instrusive link field for threads
198                // must be exactly as in thread_desc
199                __thread_desc_link link;
200        } before, after;
201
202        // Optional statistic counters
203        #if !defined(__CFA_NO_SCHED_STATS__)
204                struct __attribute__((aligned(64))) {
205                        // difference between number of push and pops
206                        ssize_t diff;
207
208                        // total number of pushes and pops
209                        size_t  push;
210                        size_t  pop ;
211                } stat;
212        #endif
213};
214
215void  ?{}(__intrusive_ready_queue_t & this);
216void ^?{}(__intrusive_ready_queue_t & this);
217
218typedef unsigned long long __cfa_readyQ_mask_t;
219
220// enum {
221//      __cfa_ready_queue_mask_size = (64 - sizeof(size_t)) / sizeof(size_t),
222//      __cfa_max_ready_queues = __cfa_ready_queue_mask_size * 8 * sizeof(size_t)
223// };
224
225#define __cfa_readyQ_mask_size ((64 - sizeof(size_t)) / sizeof(__cfa_readyQ_mask_t))
226#define __cfa_max_readyQs (__cfa_readyQ_mask_size * 8 * sizeof(__cfa_readyQ_mask_t))
227
228//TODO adjust cache size to ARCHITECTURE
229struct __attribute__((aligned(128))) __ready_queue_t {
230        struct {
231                volatile size_t count;
232                volatile __cfa_readyQ_mask_t mask[ __cfa_readyQ_mask_size ];
233        } empty;
234
235        struct __attribute__((aligned(64))) {
236                __intrusive_ready_queue_t * volatile data;
237                volatile size_t count;
238        } list;
239
240        #if !defined(__CFA_NO_STATISTICS__)
241                __attribute__((aligned(64))) struct {
242                        struct {
243                                struct {
244                                        volatile size_t attempt;
245                                        volatile size_t success;
246                                } push;
247                                struct {
248                                        volatile size_t maskrds;
249                                        volatile size_t attempt;
250                                        volatile size_t success;
251                                } pop;
252                        } pick;
253                        struct {
254                                volatile size_t value;
255                                volatile size_t count;
256                        } full;
257                } global_stats;
258
259        #endif
260};
261
262void  ?{}(__ready_queue_t & this);
263void ^?{}(__ready_queue_t & this);
264
265//-----------------------------------------------------------------------------
266// Cluster
267struct cluster {
268        // Ready queue locks
269        __clusterRWLock_t ready_lock;
270
271        // Ready queue for threads
272        __ready_queue_t ready_queue;
273
274        // Name of the cluster
275        const char * name;
276
277        // Preemption rate on this cluster
278        Duration preemption_rate;
279
280        // List of processors
281        __spinlock_t proc_list_lock;
282        __dllist_t(struct processor) idles;
283
284        // List of threads
285        __spinlock_t thread_list_lock;
286        __dllist_t(struct thread_desc) threads;
287        unsigned int nthreads;
288
289        // Link lists fields
290        struct __dbg_node_cltr {
291                cluster * next;
292                cluster * prev;
293        } node;
294};
295extern Duration default_preemption();
296
297void ?{} (cluster & this, const char * name, Duration preemption_rate);
298void ^?{}(cluster & this);
299
300static inline void ?{} (cluster & this)                           { this{"Anonymous Cluster", default_preemption()}; }
301static inline void ?{} (cluster & this, Duration preemption_rate) { this{"Anonymous Cluster", preemption_rate}; }
302static inline void ?{} (cluster & this, const char * name)        { this{name, default_preemption()}; }
303
304static inline [cluster *&, cluster *& ] __get( cluster & this ) {
305        return this.node.[next, prev];
306}
307
308static inline struct processor * active_processor() { return TL_GET( this_processor ); } // UNSAFE
309static inline struct cluster   * active_cluster  () { return TL_GET( this_processor )->cltr; }
310
311// Local Variables: //
312// mode: c //
313// tab-width: 4 //
314// End: //
Note: See TracBrowser for help on using the repository browser.