source: libcfa/src/concurrency/kernel.hfa @ 6502a2b

ADTarm-ehast-experimentalenumforall-pointer-decayjacob/cs343-translationnew-astnew-ast-unique-exprpthread-emulationqualifiedEnum
Last change on this file since 6502a2b was 6502a2b, checked in by Thierry Delisle <tdelisle@…>, 4 years ago

Tentative fix to the deadlock in cluster shutdown

  • Property mode set to 100644
File size: 6.6 KB
Line 
1//
2// Cforall Version 1.0.0 Copyright (C) 2016 University of Waterloo
3//
4// The contents of this file are covered under the licence agreement in the
5// file "LICENCE" distributed with Cforall.
6//
7// kernel --
8//
9// Author           : Thierry Delisle
10// Created On       : Tue Jan 17 12:27:26 2017
11// Last Modified By : Peter A. Buhr
12// Last Modified On : Tue Feb  4 12:29:26 2020
13// Update Count     : 22
14//
15
16#pragma once
17
18#include <stdbool.h>
19#include <stdint.h>
20
21#include "invoke.h"
22#include "time_t.hfa"
23#include "coroutine.hfa"
24
25extern "C" {
26#include <pthread.h>
27#include <semaphore.h>
28}
29
30//-----------------------------------------------------------------------------
31// Locks
32struct semaphore {
33        __spinlock_t lock;
34        int count;
35        __queue_t($thread) waiting;
36};
37
38void  ?{}(semaphore & this, int count = 1);
39void ^?{}(semaphore & this);
40void   P (semaphore & this);
41bool   V (semaphore & this);
42bool   V (semaphore & this, unsigned count);
43
44
45//-----------------------------------------------------------------------------
46// Processor
47extern struct cluster * mainCluster;
48
49// Processor
50coroutine processorCtx_t {
51        struct processor * proc;
52};
53
54// Wrapper around kernel threads
55struct processor {
56        // Main state
57        // Coroutine ctx who does keeps the state of the processor
58        struct processorCtx_t runner;
59
60        // Cluster from which to get threads
61        struct cluster * cltr;
62
63        // Name of the processor
64        const char * name;
65
66        // Handle to pthreads
67        pthread_t kernel_thread;
68
69        // RunThread data
70        // Action to do after a thread is ran
71        $thread * destroyer;
72
73        // Preemption data
74        // Node which is added in the discrete event simulaiton
75        struct alarm_node_t * preemption_alarm;
76
77        // If true, a preemption was triggered in an unsafe region, the processor must preempt as soon as possible
78        bool pending_preemption;
79
80        // Idle lock (kernel semaphore)
81        __bin_sem_t idle;
82
83        // Termination
84        // Set to true to notify the processor should terminate
85        volatile bool do_terminate;
86
87        // Termination synchronisation (user semaphore)
88        semaphore terminated;
89
90        // pthread Stack
91        void * stack;
92
93        // Link lists fields
94        struct __dbg_node_proc {
95                struct processor * next;
96                struct processor * prev;
97        } node;
98
99#ifdef __CFA_DEBUG__
100        // Last function to enable preemption on this processor
101        const char * last_enable;
102#endif
103};
104
105void  ?{}(processor & this, const char name[], struct cluster & cltr);
106void ^?{}(processor & this);
107
108static inline void  ?{}(processor & this)                    { this{ "Anonymous Processor", *mainCluster}; }
109static inline void  ?{}(processor & this, struct cluster & cltr)    { this{ "Anonymous Processor", cltr}; }
110static inline void  ?{}(processor & this, const char name[]) { this{name, *mainCluster }; }
111
112static inline [processor *&, processor *& ] __get( processor & this ) __attribute__((const)) { return this.node.[next, prev]; }
113
114//-----------------------------------------------------------------------------
115// I/O
116#if defined(HAVE_LINUX_IO_URING_H)
117struct io_uring_sq {
118        // Head and tail of the ring (associated with array)
119        volatile uint32_t * head;
120        volatile uint32_t * tail;
121
122        // The actual kernel ring which uses head/tail
123        // indexes into the sqes arrays
124        uint32_t * array;
125
126        // number of entries and mask to go with it
127        const uint32_t * num;
128        const uint32_t * mask;
129
130        // Submission flags (Not sure what for)
131        uint32_t * flags;
132
133        // number of sqes not submitted (whatever that means)
134        uint32_t * dropped;
135
136        // Like head/tail but not seen by the kernel
137        volatile uint32_t alloc;
138        volatile uint32_t ready;
139
140        __spinlock_t lock;
141
142        // A buffer of sqes (not the actual ring)
143        struct io_uring_sqe * sqes;
144
145        // The location and size of the mmaped area
146        void * ring_ptr;
147        size_t ring_sz;
148
149        // Statistics
150        #if !defined(__CFA_NO_STATISTICS__)
151                struct {
152                        struct {
153                                unsigned long long int val;
154                                unsigned long long int cnt;
155                        } submit_avg;
156                } stats;
157        #endif
158};
159
160struct io_uring_cq {
161        // Head and tail of the ring
162        volatile uint32_t * head;
163        volatile uint32_t * tail;
164
165        // number of entries and mask to go with it
166        const uint32_t * mask;
167        const uint32_t * num;
168
169        // number of cqes not submitted (whatever that means)
170        uint32_t * overflow;
171
172        // the kernel ring
173        struct io_uring_cqe * cqes;
174
175        // The location and size of the mmaped area
176        void * ring_ptr;
177        size_t ring_sz;
178
179        // Statistics
180        #if !defined(__CFA_NO_STATISTICS__)
181                struct {
182                        struct {
183                                unsigned long long int val;
184                                unsigned long long int cnt;
185                        } completed_avg;
186                } stats;
187        #endif
188};
189
190#if defined(__CFA_IO_POLLING_USER__)
191        struct __io_poller_fast {
192                struct io_ring * ring;
193                bool waiting;
194                $thread thrd;
195        };
196#endif
197
198struct io_ring {
199        struct io_uring_sq submit_q;
200        struct io_uring_cq completion_q;
201        uint32_t flags;
202        int fd;
203        semaphore submit;
204        volatile bool done;
205        struct {
206                struct {
207                        void * stack;
208                        pthread_t kthrd;
209                } slow;
210                #if defined(__CFA_IO_POLLING_USER__)
211                        __io_poller_fast fast;
212                        __bin_sem_t sem;
213                #endif
214        } poller;
215};
216#endif
217
218//-----------------------------------------------------------------------------
219// Cluster
220struct cluster {
221        // Ready queue locks
222        __spinlock_t ready_queue_lock;
223
224        // Ready queue for threads
225        __queue_t($thread) ready_queue;
226
227        // Name of the cluster
228        const char * name;
229
230        // Preemption rate on this cluster
231        Duration preemption_rate;
232
233        // List of processors
234        __spinlock_t idle_lock;
235        __dllist_t(struct processor) procs;
236        __dllist_t(struct processor) idles;
237        unsigned int nprocessors;
238
239        // List of threads
240        __spinlock_t thread_list_lock;
241        __dllist_t(struct $thread) threads;
242        unsigned int nthreads;
243
244        // Link lists fields
245        struct __dbg_node_cltr {
246                cluster * next;
247                cluster * prev;
248        } node;
249
250        #if defined(HAVE_LINUX_IO_URING_H)
251                struct io_ring io;
252        #endif
253
254        #if !defined(__CFA_NO_STATISTICS__)
255                bool print_stats;
256        #endif
257};
258extern Duration default_preemption();
259
260void ?{} (cluster & this, const char name[], Duration preemption_rate);
261void ^?{}(cluster & this);
262
263static inline void ?{} (cluster & this)                           { this{"Anonymous Cluster", default_preemption()}; }
264static inline void ?{} (cluster & this, Duration preemption_rate) { this{"Anonymous Cluster", preemption_rate}; }
265static inline void ?{} (cluster & this, const char name[])        { this{name, default_preemption()}; }
266
267static inline [cluster *&, cluster *& ] __get( cluster & this ) __attribute__((const)) { return this.node.[next, prev]; }
268
269static inline struct processor * active_processor() { return TL_GET( this_processor ); } // UNSAFE
270static inline struct cluster   * active_cluster  () { return TL_GET( this_processor )->cltr; }
271
272#if !defined(__CFA_NO_STATISTICS__)
273        static inline void print_stats_at_exit( cluster & this ) {
274                this.print_stats = true;
275        }
276#endif
277
278// Local Variables: //
279// mode: c //
280// tab-width: 4 //
281// End: //
Note: See TracBrowser for help on using the repository browser.