source: libcfa/src/concurrency/kernel.hfa @ 708ae38

ADTast-experimentalenumpthread-emulationqualifiedEnum
Last change on this file since 708ae38 was 708ae38, checked in by Thierry Delisle <tdelisle@…>, 3 years ago

Some more cleanup and grow/shrink now readjusts io timestamps.
(They are still unused).

  • Property mode set to 100644
File size: 9.3 KB
Line 
1//
2// Cforall Version 1.0.0 Copyright (C) 2016 University of Waterloo
3//
4// The contents of this file are covered under the licence agreement in the
5// file "LICENCE" distributed with Cforall.
6//
7// kernel -- Header containing the core of the kernel API
8//
9// Author           : Thierry Delisle
10// Created On       : Tue Jan 17 12:27:26 2017
11// Last Modified By : Peter A. Buhr
12// Last Modified On : Tue Feb  4 12:29:26 2020
13// Update Count     : 22
14//
15
16#pragma once
17
18#include "invoke.h"
19#include "time_t.hfa"
20#include "coroutine.hfa"
21
22#include "containers/list.hfa"
23
24extern "C" {
25        #include <bits/pthreadtypes.h>
26        #include <pthread.h>
27        #include <linux/types.h>
28}
29
30#ifdef __CFA_WITH_VERIFY__
31        extern bool __cfaabi_dbg_in_kernel();
32#endif
33
34//-----------------------------------------------------------------------------
35// I/O
36struct cluster;
37struct $io_context;
38struct $io_arbiter;
39
40struct io_context_params {
41        int num_entries;
42};
43
44void  ?{}(io_context_params & this);
45
46//-----------------------------------------------------------------------------
47// Processor
48extern struct cluster * mainCluster;
49
50// Processor id, required for scheduling threads
51
52
53coroutine processorCtx_t {
54        struct processor * proc;
55};
56
57
58struct __fd_waitctx {
59        volatile int fd;
60};
61
62// Wrapper around kernel threads
63struct __attribute__((aligned(128))) processor {
64        // Cluster from which to get threads
65        struct cluster * cltr;
66
67        // Ready Queue state per processor
68        struct {
69                unsigned short its;
70                unsigned short itr;
71                unsigned id;
72                unsigned target;
73                unsigned last;
74                signed   cpu;
75        } rdq;
76
77        // Set to true to notify the processor should terminate
78        volatile bool do_terminate;
79
80        // Coroutine ctx who does keeps the state of the processor
81        struct processorCtx_t runner;
82
83        // Name of the processor
84        const char * name;
85
86        // Handle to pthreads
87        pthread_t kernel_thread;
88
89        // Unique id for the processor (not per cluster)
90        unsigned unique_id;
91
92        struct {
93                $io_context * ctx;
94                volatile bool pending;
95                volatile bool dirty;
96        } io;
97
98        // Preemption data
99        // Node which is added in the discrete event simulaiton
100        struct alarm_node_t * preemption_alarm;
101
102        // If true, a preemption was triggered in an unsafe region, the processor must preempt as soon as possible
103        bool pending_preemption;
104
105        // Idle lock (kernel semaphore)
106        int idle_fd;
107
108        // Idle waitctx
109        struct __fd_waitctx idle_wctx;
110
111        // Termination synchronisation (user semaphore)
112        oneshot terminated;
113
114        // pthread Stack
115        void * stack;
116
117        // Link lists fields
118        inline dlink(processor);
119
120        // special init fields
121        // This is needed for memcached integration
122        // once memcached experiments are done this should probably be removed
123        // it is not a particularly safe scheme as it can make processors less homogeneous
124        struct {
125                thread$ * thrd;
126        } init;
127
128        struct KernelThreadData * local_data;
129
130        #if !defined(__CFA_NO_STATISTICS__)
131                int print_stats;
132                bool print_halts;
133        #endif
134
135#ifdef __CFA_DEBUG__
136        // Last function to enable preemption on this processor
137        const char * last_enable;
138#endif
139};
140P9_EMBEDDED( processor, dlink(processor) )
141
142void  ?{}(processor & this, const char name[], struct cluster & cltr);
143void ^?{}(processor & this);
144
145static inline void  ?{}(processor & this)                        { this{ "Anonymous Processor", *mainCluster}; }
146static inline void  ?{}(processor & this, struct cluster & cltr) { this{ "Anonymous Processor", cltr}; }
147static inline void  ?{}(processor & this, const char name[])     { this{name, *mainCluster}; }
148
149//-----------------------------------------------------------------------------
150// Cluster Tools
151
152// Intrusives lanes which are used by the ready queue
153struct __attribute__((aligned(128))) __intrusive_lane_t;
154void  ?{}(__intrusive_lane_t & this);
155void ^?{}(__intrusive_lane_t & this);
156
157// Aligned timestamps which are used by the ready queue and io subsystem
158struct __attribute__((aligned(128))) __timestamp_t {
159        volatile unsigned long long tv;
160        volatile unsigned long long ma;
161};
162
163static inline void  ?{}(__timestamp_t & this) { this.tv = 0; this.ma = 0; }
164static inline void ^?{}(__timestamp_t &) {}
165
166
167struct __attribute__((aligned(16))) __cache_id_t {
168        volatile unsigned id;
169};
170
171// //TODO adjust cache size to ARCHITECTURE
172// // Structure holding the ready queue
173// struct __ready_queue_t {
174//      // Data tracking the actual lanes
175//      // On a seperate cacheline from the used struct since
176//      // used can change on each push/pop but this data
177//      // only changes on shrink/grow
178//      struct {
179//              // Arary of lanes
180//              __intrusive_lane_t * volatile data;
181
182//              __cache_id_t * volatile caches;
183
184//              // Number of lanes (empty or not)
185//              volatile size_t count;
186//      } lanes;
187// };
188
189// void  ?{}(__ready_queue_t & this);
190// void ^?{}(__ready_queue_t & this);
191
192// Idle Sleep
193struct __cluster_proc_list {
194        // Spin lock protecting the queue
195        __spinlock_t lock;
196
197        // FD to use to wake a processor
198        struct __fd_waitctx * volatile fdw;
199
200        // Total number of processors
201        unsigned total;
202
203        // Total number of idle processors
204        unsigned idle;
205
206        // List of idle processors
207        dlist(processor) idles;
208
209        // List of active processors
210        dlist(processor) actives;
211};
212
213//-----------------------------------------------------------------------------
214// Cluster
215struct __attribute__((aligned(128))) cluster {
216        struct {
217                struct {
218                        // Arary of subqueues
219                        __intrusive_lane_t * volatile data;
220
221                        // Time since subqueues were processed
222                        __timestamp_t * volatile tscs;
223
224                        // Number of subqueue / timestamps
225                        size_t count;
226                } readyQ;
227
228                struct {
229                        // Time since subqueues were processed
230                        __timestamp_t * volatile tscs;
231
232                        // Number of I/O subqueues
233                        size_t count;
234                } io;
235
236                // Cache each kernel thread belongs to
237                __cache_id_t * volatile caches;
238        } sched;
239
240        // // Ready queue for threads
241        // __ready_queue_t ready_queue;
242
243        // Name of the cluster
244        const char * name;
245
246        // Preemption rate on this cluster
247        Duration preemption_rate;
248
249        // List of idle processors
250        __cluster_proc_list procs;
251
252        // List of threads
253        __spinlock_t thread_list_lock;
254        __dllist_t(struct thread$) threads;
255        unsigned int nthreads;
256
257        // Link lists fields
258        struct __dbg_node_cltr {
259                cluster * next;
260                cluster * prev;
261        } node;
262
263        struct {
264                $io_arbiter * arbiter;
265                io_context_params params;
266        } io;
267
268        #if !defined(__CFA_NO_STATISTICS__)
269                struct __stats_t * stats;
270                int print_stats;
271        #endif
272};
273extern Duration default_preemption();
274
275void ?{} (cluster & this, const char name[], Duration preemption_rate, unsigned num_io, const io_context_params & io_params);
276void ^?{}(cluster & this);
277
278static inline void ?{} (cluster & this)                                            { io_context_params default_params;    this{"Anonymous Cluster", default_preemption(), 1, default_params}; }
279static inline void ?{} (cluster & this, Duration preemption_rate)                  { io_context_params default_params;    this{"Anonymous Cluster", preemption_rate, 1, default_params}; }
280static inline void ?{} (cluster & this, const char name[])                         { io_context_params default_params;    this{name, default_preemption(), 1, default_params}; }
281static inline void ?{} (cluster & this, unsigned num_io)                           { io_context_params default_params;    this{"Anonymous Cluster", default_preemption(), num_io, default_params}; }
282static inline void ?{} (cluster & this, Duration preemption_rate, unsigned num_io) { io_context_params default_params;    this{"Anonymous Cluster", preemption_rate, num_io, default_params}; }
283static inline void ?{} (cluster & this, const char name[], unsigned num_io)        { io_context_params default_params;    this{name, default_preemption(), num_io, default_params}; }
284static inline void ?{} (cluster & this, const io_context_params & io_params)                                            { this{"Anonymous Cluster", default_preemption(), 1, io_params}; }
285static inline void ?{} (cluster & this, Duration preemption_rate, const io_context_params & io_params)                  { this{"Anonymous Cluster", preemption_rate, 1, io_params}; }
286static inline void ?{} (cluster & this, const char name[], const io_context_params & io_params)                         { this{name, default_preemption(), 1, io_params}; }
287static inline void ?{} (cluster & this, unsigned num_io, const io_context_params & io_params)                           { this{"Anonymous Cluster", default_preemption(), num_io, io_params}; }
288static inline void ?{} (cluster & this, Duration preemption_rate, unsigned num_io, const io_context_params & io_params) { this{"Anonymous Cluster", preemption_rate, num_io, io_params}; }
289static inline void ?{} (cluster & this, const char name[], unsigned num_io, const io_context_params & io_params)        { this{name, default_preemption(), num_io, io_params}; }
290
291static inline [cluster *&, cluster *& ] __get( cluster & this ) __attribute__((const)) { return this.node.[next, prev]; }
292
293static inline struct processor * active_processor() { return publicTLS_get( this_processor ); } // UNSAFE
294static inline struct cluster   * active_cluster  () { return publicTLS_get( this_processor )->cltr; }
295
296#if !defined(__CFA_NO_STATISTICS__)
297        void print_stats_now( cluster & this, int flags );
298
299        static inline void print_stats_at_exit( cluster & this, int flags ) {
300                this.print_stats |= flags;
301        }
302
303        static inline void print_stats_at_exit( processor & this, int flags ) {
304                this.print_stats |= flags;
305        }
306
307        void print_halts( processor & this );
308#endif
309
310// Local Variables: //
311// mode: c //
312// tab-width: 4 //
313// End: //
Note: See TracBrowser for help on using the repository browser.