source: libcfa/src/concurrency/locks.hfa@ 145dcd5

ADT ast-experimental enum forall-pointer-decay pthread-emulation qualifiedEnum
Last change on this file since 145dcd5 was 7636fcc, checked in by Thierry Delisle <tdelisle@…>, 4 years ago

Removed fast_lock from testing and added deprecation attribute to it.

  • Property mode set to 100644
File size: 13.0 KB
RevLine 
[ab1b971]1//
2// Cforall Version 1.0.0 Copyright (C) 2021 University of Waterloo
3//
4// The contents of this file are covered under the licence agreement in the
5// file "LICENCE" distributed with Cforall.
6//
7// locks.hfa -- PUBLIC
8// Runtime locks that used with the runtime thread system.
9//
10// Author : Colby Alexander Parsons
11// Created On : Thu Jan 21 19:46:50 2021
12// Last Modified By :
13// Last Modified On :
14// Update Count :
15//
16
[f4e35326]17#pragma once
18
[848439f]19#include <stdbool.h>
[5a46e09]20#include <stdio.h>
[848439f]21
[ab1b971]22#include "bits/weakso_locks.hfa"
[f4ec5e45]23#include "containers/queueLockFree.hfa"
[82f4063]24#include "containers/list.hfa"
[f4ec5e45]25
[07033ce]26#include "limits.hfa"
[f4ec5e45]27#include "thread.hfa"
[848439f]28
29#include "time_t.hfa"
30#include "time.hfa"
31
[f4ec5e45]32//-----------------------------------------------------------------------------
33// Semaphores
34
35// '0-nary' semaphore
36// Similar to a counting semaphore except the value of one is never reached
37// as a consequence, a V() that would bring the value to 1 *spins* until
38// a P consumes it
39struct Semaphore0nary {
40 __spinlock_t lock; // needed to protect
[e84ab3d]41 mpsc_queue(thread$) queue;
[f4ec5e45]42};
43
[e84ab3d]44static inline bool P(Semaphore0nary & this, thread$ * thrd) {
[82f4063]45 /* paranoid */ verify(!thrd`next);
46 /* paranoid */ verify(!(&(*thrd)`next));
[f4ec5e45]47
48 push(this.queue, thrd);
49 return true;
50}
51
52static inline bool P(Semaphore0nary & this) {
[e84ab3d]53 thread$ * thrd = active_thread();
[f4ec5e45]54 P(this, thrd);
55 park();
56 return true;
57}
58
[e84ab3d]59static inline thread$ * V(Semaphore0nary & this, bool doUnpark = true) {
60 thread$ * next;
[f4ec5e45]61 lock(this.lock __cfaabi_dbg_ctx2);
62 for (;;) {
63 next = pop(this.queue);
64 if (next) break;
65 Pause();
66 }
67 unlock(this.lock);
68
69 if (doUnpark) unpark(next);
70 return next;
71}
72
73// Wrapper used on top of any sempahore to avoid potential locking
74struct BinaryBenaphore {
75 volatile ssize_t counter;
76};
77
78static inline {
79 void ?{}(BinaryBenaphore & this) { this.counter = 0; }
80 void ?{}(BinaryBenaphore & this, zero_t) { this.counter = 0; }
81 void ?{}(BinaryBenaphore & this, one_t ) { this.counter = 1; }
82
83 // returns true if no blocking needed
[e20eaf4]84 bool P(BinaryBenaphore & this) {
85 return __atomic_fetch_sub(&this.counter, 1, __ATOMIC_SEQ_CST) > 0;
86 }
87
[f4ec5e45]88 bool tryP(BinaryBenaphore & this) {
89 ssize_t c = this.counter;
[07033ce]90 /* paranoid */ verify( c > MIN );
[f4ec5e45]91 return (c >= 1) && __atomic_compare_exchange_n(&this.counter, &c, c-1, false, __ATOMIC_SEQ_CST, __ATOMIC_RELAXED);
92 }
93
94 // returns true if notify needed
95 bool V(BinaryBenaphore & this) {
96 ssize_t c = 0;
97 for () {
[07033ce]98 /* paranoid */ verify( this.counter < MAX );
[f4ec5e45]99 if (__atomic_compare_exchange_n(&this.counter, &c, c+1, false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST)) {
100 if (c == 0) return true;
101 /* paranoid */ verify(c < 0);
102 return false;
103 } else {
104 if (c == 1) return true;
105 /* paranoid */ verify(c < 1);
106 Pause();
107 }
108 }
109 }
110}
111
112// Binary Semaphore based on the BinaryBenaphore on top of the 0-nary Semaphore
113struct ThreadBenaphore {
114 BinaryBenaphore ben;
115 Semaphore0nary sem;
116};
117
118static inline void ?{}(ThreadBenaphore & this) {}
119static inline void ?{}(ThreadBenaphore & this, zero_t) { (this.ben){ 0 }; }
120static inline void ?{}(ThreadBenaphore & this, one_t ) { (this.ben){ 1 }; }
121
[e20eaf4]122static inline bool P(ThreadBenaphore & this) { return P(this.ben) ? false : P(this.sem); }
[f4ec5e45]123static inline bool tryP(ThreadBenaphore & this) { return tryP(this.ben); }
124static inline bool P(ThreadBenaphore & this, bool wait) { return wait ? P(this) : tryP(this); }
125
[e84ab3d]126static inline thread$ * V(ThreadBenaphore & this, bool doUnpark = true) {
[e20eaf4]127 if (V(this.ben)) return 0p;
[f4ec5e45]128 return V(this.sem, doUnpark);
129}
130
131//-----------------------------------------------------------------------------
132// Semaphore
133struct semaphore {
134 __spinlock_t lock;
135 int count;
[e84ab3d]136 __queue_t(thread$) waiting;
[f4ec5e45]137};
138
139void ?{}(semaphore & this, int count = 1);
140void ^?{}(semaphore & this);
141bool P (semaphore & this);
142bool V (semaphore & this);
143bool V (semaphore & this, unsigned count);
[e84ab3d]144thread$ * V (semaphore & this, bool );
[f4ec5e45]145
[ab1b971]146//----------
147struct single_acquisition_lock {
148 inline blocking_lock;
149};
150
151static inline void ?{}( single_acquisition_lock & this ) {((blocking_lock &)this){ false, false };}
152static inline void ^?{}( single_acquisition_lock & this ) {}
[22b7579]153static inline void lock ( single_acquisition_lock & this ) { lock ( (blocking_lock &)this ); }
154static inline bool try_lock ( single_acquisition_lock & this ) { return try_lock( (blocking_lock &)this ); }
155static inline void unlock ( single_acquisition_lock & this ) { unlock ( (blocking_lock &)this ); }
156static inline size_t on_wait ( single_acquisition_lock & this ) { return on_wait ( (blocking_lock &)this ); }
157static inline void on_wakeup( single_acquisition_lock & this, size_t v ) { on_wakeup ( (blocking_lock &)this, v ); }
[e84ab3d]158static inline void on_notify( single_acquisition_lock & this, struct thread$ * t ) { on_notify( (blocking_lock &)this, t ); }
[ab1b971]159
160//----------
161struct owner_lock {
162 inline blocking_lock;
163};
164
165static inline void ?{}( owner_lock & this ) {((blocking_lock &)this){ true, true };}
166static inline void ^?{}( owner_lock & this ) {}
[f19497c]167static inline void lock ( owner_lock & this ) { lock ( (blocking_lock &)this ); }
[d27b6be]168static inline bool try_lock ( owner_lock & this ) { return try_lock( (blocking_lock &)this ); }
[f19497c]169static inline void unlock ( owner_lock & this ) { unlock ( (blocking_lock &)this ); }
[22b7579]170static inline size_t on_wait ( owner_lock & this ) { return on_wait ( (blocking_lock &)this ); }
171static inline void on_wakeup( owner_lock & this, size_t v ) { on_wakeup ( (blocking_lock &)this, v ); }
[e84ab3d]172static inline void on_notify( owner_lock & this, struct thread$ * t ) { on_notify( (blocking_lock &)this, t ); }
[ab1b971]173
[f4ec5e45]174struct fast_lock {
[e84ab3d]175 thread$ * volatile owner;
[f4ec5e45]176 ThreadBenaphore sem;
177};
178
[7636fcc]179static inline void ?{}(fast_lock & this) __attribute__((deprecated("use linear_backoff_then_block_lock instead")));
[6ba6846]180static inline void ?{}(fast_lock & this) { this.owner = 0p; }
181
[e84ab3d]182static inline bool $try_lock(fast_lock & this, thread$ * thrd) {
183 thread$ * exp = 0p;
[f4ec5e45]184 return __atomic_compare_exchange_n(&this.owner, &exp, thrd, false, __ATOMIC_SEQ_CST, __ATOMIC_RELAXED);
185}
186
[7636fcc]187static inline void lock( fast_lock & this ) __attribute__((deprecated("use linear_backoff_then_block_lock instead"), artificial));
[f4ec5e45]188static inline void lock( fast_lock & this ) {
[e84ab3d]189 thread$ * thrd = active_thread();
[f4ec5e45]190 /* paranoid */verify(thrd != this.owner);
191
192 for (;;) {
193 if ($try_lock(this, thrd)) return;
194 P(this.sem);
195 }
196}
197
[7636fcc]198static inline bool try_lock( fast_lock & this ) __attribute__((deprecated("use linear_backoff_then_block_lock instead"), artificial));
[22b7579]199static inline bool try_lock ( fast_lock & this ) {
[e84ab3d]200 thread$ * thrd = active_thread();
[f4ec5e45]201 /* paranoid */ verify(thrd != this.owner);
202 return $try_lock(this, thrd);
203}
204
[7636fcc]205static inline thread$ * unlock( fast_lock & this ) __attribute__((deprecated("use linear_backoff_then_block_lock instead"), artificial));
[e84ab3d]206static inline thread$ * unlock( fast_lock & this ) {
[ddd473f]207 /* paranoid */ verify(active_thread() == this.owner);
[f4ec5e45]208
[22b7579]209 // open 'owner' before unlocking anyone
210 // so new and unlocked threads don't park incorrectly.
[c7c178b]211 // This may require additional fencing on ARM.
[22b7579]212 this.owner = 0p;
[f4ec5e45]213
[198e335]214 return V(this.sem);
[f4ec5e45]215}
216
[22b7579]217static inline size_t on_wait( fast_lock & this ) { unlock(this); return 0; }
218static inline void on_wakeup( fast_lock & this, size_t ) { lock(this); }
[e84ab3d]219static inline void on_notify( fast_lock &, struct thread$ * t ) { unpark(t); }
[f4ec5e45]220
221struct mcs_node {
222 mcs_node * volatile next;
223 single_sem sem;
224};
225
[8f5576d5]226static inline void ?{}(mcs_node & this) { this.next = 0p; }
[f4ec5e45]227
228static inline mcs_node * volatile & ?`next ( mcs_node * node ) {
229 return node->next;
230}
231
232struct mcs_lock {
233 mcs_queue(mcs_node) queue;
234};
235
236static inline void lock(mcs_lock & l, mcs_node & n) {
237 if(push(l.queue, &n))
238 wait(n.sem);
239}
240
241static inline void unlock(mcs_lock & l, mcs_node & n) {
242 mcs_node * next = advance(l.queue, &n);
243 if(next) post(next->sem);
244}
245
[5a46e09]246struct linear_backoff_then_block_lock {
247 // Spin lock used for mutual exclusion
248 __spinlock_t spinlock;
249
250 // Current thread owning the lock
[e84ab3d]251 struct thread$ * owner;
[5a46e09]252
253 // List of blocked threads
[e84ab3d]254 dlist( thread$ ) blocked_threads;
[5a46e09]255
256 // Used for comparing and exchanging
257 volatile size_t lock_value;
258
259 // used for linear backoff spinning
260 int spin_start;
261 int spin_end;
262 int spin_count;
263
264 // after unsuccessful linear backoff yield this many times
265 int yield_count;
266};
267
268static inline void ?{}( linear_backoff_then_block_lock & this, int spin_start, int spin_end, int spin_count, int yield_count ) {
269 this.spinlock{};
270 this.blocked_threads{};
271 this.lock_value = 0;
272 this.spin_start = spin_start;
273 this.spin_end = spin_end;
274 this.spin_count = spin_count;
275 this.yield_count = yield_count;
276}
[55ad35c]277static inline void ?{}( linear_backoff_then_block_lock & this ) { this{4, 1024, 16, 0}; }
[5a46e09]278static inline void ^?{}( linear_backoff_then_block_lock & this ) {}
[eba9d27]279static inline void ?{}( linear_backoff_then_block_lock & this, linear_backoff_then_block_lock this2 ) = void;
280static inline void ?=?( linear_backoff_then_block_lock & this, linear_backoff_then_block_lock this2 ) = void;
[5a46e09]281
282static inline bool internal_try_lock(linear_backoff_then_block_lock & this, size_t & compare_val) with(this) {
283 if (__atomic_compare_exchange_n(&lock_value, &compare_val, 1, false, __ATOMIC_ACQUIRE, __ATOMIC_RELAXED)) {
284 owner = active_thread();
285 return true;
286 }
287 return false;
288}
289
290static inline bool try_lock(linear_backoff_then_block_lock & this) { size_t compare_val = 0; return internal_try_lock(this, compare_val); }
291
292static inline bool try_lock_contention(linear_backoff_then_block_lock & this) with(this) {
293 if (__atomic_exchange_n(&lock_value, 2, __ATOMIC_ACQUIRE) == 0) {
294 owner = active_thread();
295 return true;
296 }
297 return false;
298}
299
300static inline bool block(linear_backoff_then_block_lock & this) with(this) {
301 lock( spinlock __cfaabi_dbg_ctx2 );
302 if (lock_value != 2) {
303 unlock( spinlock );
304 return true;
305 }
306 insert_last( blocked_threads, *active_thread() );
307 unlock( spinlock );
308 park( );
309 return true;
310}
311
312static inline bool lock(linear_backoff_then_block_lock & this) with(this) {
313 // if owner just return
314 if (active_thread() == owner) return true;
315 size_t compare_val = 0;
316 int spin = spin_start;
317 // linear backoff
318 for( ;; ) {
319 compare_val = 0;
320 if (internal_try_lock(this, compare_val)) return true;
321 if (2 == compare_val) break;
322 for (int i = 0; i < spin; i++) Pause();
323 if (spin >= spin_end) break;
324 spin += spin;
325 }
326
[b7763da]327 if(2 != compare_val && try_lock_contention(this)) return true;
328 // block until signalled
329 while (block(this)) if(try_lock_contention(this)) return true;
[e84ab3d]330
[b7763da]331 // this should never be reached as block(this) always returns true
332 return false;
333}
334
[5a46e09]335static inline void unlock(linear_backoff_then_block_lock & this) with(this) {
336 verify(lock_value > 0);
337 owner = 0p;
338 if (__atomic_exchange_n(&lock_value, 0, __ATOMIC_RELEASE) == 1) return;
339 lock( spinlock __cfaabi_dbg_ctx2 );
[e84ab3d]340 thread$ * t = &try_pop_front( blocked_threads );
[5a46e09]341 unlock( spinlock );
342 unpark( t );
343}
344
[e84ab3d]345static inline void on_notify(linear_backoff_then_block_lock & this, struct thread$ * t ) { unpark(t); }
[dcad80a]346static inline size_t on_wait(linear_backoff_then_block_lock & this) { unlock(this); return 0; }
[bbe3719]347static inline void on_wakeup(linear_backoff_then_block_lock & this, size_t recursion ) { lock(this); }
[5a46e09]348
[ac5816d]349//-----------------------------------------------------------------------------
350// is_blocking_lock
[fd54fef]351trait is_blocking_lock(L & | sized(L)) {
[ac5816d]352 // For synchronization locks to use when acquiring
[e84ab3d]353 void on_notify( L &, struct thread$ * );
[ac5816d]354
355 // For synchronization locks to use when releasing
[22b7579]356 size_t on_wait( L & );
[ac5816d]357
358 // to set recursion count after getting signalled;
[22b7579]359 void on_wakeup( L &, size_t recursion );
[ac5816d]360};
[848439f]361
[ac5816d]362//-----------------------------------------------------------------------------
[82f4063]363// // info_thread
364// // the info thread is a wrapper around a thread used
365// // to store extra data for use in the condition variable
[fd54fef]366forall(L & | is_blocking_lock(L)) {
[ac5816d]367 struct info_thread;
[c131a02]368
[82f4063]369 // // for use by sequence
370 // info_thread(L) *& Back( info_thread(L) * this );
371 // info_thread(L) *& Next( info_thread(L) * this );
[848439f]372}
373
[ac5816d]374//-----------------------------------------------------------------------------
375// Synchronization Locks
[fd54fef]376forall(L & | is_blocking_lock(L)) {
[eeb5023]377 struct condition_variable {
[848439f]378 // Spin lock used for mutual exclusion
379 __spinlock_t lock;
380
381 // List of blocked threads
[82f4063]382 dlist( info_thread(L) ) blocked_threads;
[848439f]383
384 // Count of current blocked threads
385 int count;
386 };
[e84ab3d]387
[848439f]388
[ac5816d]389 void ?{}( condition_variable(L) & this );
[848439f]390 void ^?{}( condition_variable(L) & this );
391
[eeb5023]392 bool notify_one( condition_variable(L) & this );
393 bool notify_all( condition_variable(L) & this );
[848439f]394
[eeb5023]395 uintptr_t front( condition_variable(L) & this );
[848439f]396
[ac5816d]397 bool empty ( condition_variable(L) & this );
398 int counter( condition_variable(L) & this );
[848439f]399
[eeb5023]400 void wait( condition_variable(L) & this );
401 void wait( condition_variable(L) & this, uintptr_t info );
[dff1fd1]402 bool wait( condition_variable(L) & this, Duration duration );
403 bool wait( condition_variable(L) & this, uintptr_t info, Duration duration );
[848439f]404
[eeb5023]405 void wait( condition_variable(L) & this, L & l );
406 void wait( condition_variable(L) & this, L & l, uintptr_t info );
[dff1fd1]407 bool wait( condition_variable(L) & this, L & l, Duration duration );
408 bool wait( condition_variable(L) & this, L & l, uintptr_t info, Duration duration );
[f4ec5e45]409}
Note: See TracBrowser for help on using the repository browser.