Index: libcfa/src/concurrency/io.cfa
===================================================================
--- libcfa/src/concurrency/io.cfa	(revision 7425720f5598dc04de75c544eb585d8a3195aeed)
+++ libcfa/src/concurrency/io.cfa	(revision 1e6ffb44376c08c30c95a0192ae6b14cc2c7fdef)
@@ -169,10 +169,9 @@
 
 			struct io_future_t * future = (struct io_future_t *)(uintptr_t)cqe.user_data;
-			__cfadbg_print_safe( io, "Kernel I/O : Syscall completed : cqe %p, result %d for %p\n", &cqe, cqe.res, future );
+			// __cfadbg_print_safe( io, "Kernel I/O : Syscall completed : cqe %p, result %d for %p\n", &cqe, cqe.res, future );
 
 			__kernel_unpark( fulfil( *future, cqe.res, false ), UNPARK_LOCAL );
 		}
 
-		__cfadbg_print_safe(io, "Kernel I/O : %u completed\n", count);
 		unsigned long long ts_next = ctx->cq.ts = rdtscl();
 
@@ -181,4 +180,5 @@
 		__atomic_store_n( ctx->cq.head, head + count, __ATOMIC_SEQ_CST );
 
+		__cfadbg_print_safe(io, "Kernel I/O : %u completed age %llu\n", count, ts_next);
 		/* paranoid */ verify( ready_schedule_islocked() );
 		/* paranoid */ verify( ! __preemption_enabled() );
@@ -230,5 +230,5 @@
 					const unsigned long long cutoff = calc_cutoff(ctsc, ctx->cq.id, ctxs_count, io.data, io.tscs, __shard_factor.io);
 					const unsigned long long age = moving_average(ctsc, io.tscs[target].tv, io.tscs[target].ma);
-					// __cfadbg_print_safe(ready_queue, "Kernel : Help attempt on %u from %u, age %'llu vs cutoff %'llu, %s\n", target, this, age, cutoff, age > cutoff ? "yes" : "no");
+					__cfadbg_print_safe(io, "Kernel I/O: Help attempt on %u from %u, age %'llu vs cutoff %'llu, %s\n", target, this, age, cutoff, age > cutoff ? "yes" : "no");
 					if(age <= cutoff) break HELP;
 
@@ -301,5 +301,5 @@
 		struct io_uring_sqe * sqes = ctx->sq.sqes;
 		for(i; want) {
-			__cfadbg_print_safe(io, "Kernel I/O : filling loop\n");
+			// __cfadbg_print_safe(io, "Kernel I/O : filling loop\n");
 			out_sqes[i] = &sqes[idxs[i]];
 		}
@@ -319,5 +319,5 @@
 		// copy all the indexes we want from the available list
 		for(i; want) {
-			__cfadbg_print_safe(io, "Kernel I/O : allocating loop\n");
+			// __cfadbg_print_safe(io, "Kernel I/O : allocating loop\n");
 			idxs[i] = sq.free_ring.array[(fhead + i) & mask];
 		}
@@ -336,5 +336,5 @@
 	// sqe == &sqes[idx]
 	struct $io_context * cfa_io_allocate(struct io_uring_sqe * sqes[], __u32 idxs[], __u32 want) {
-		__cfadbg_print_safe(io, "Kernel I/O : attempting to allocate %u\n", want);
+		// __cfadbg_print_safe(io, "Kernel I/O : attempting to allocate %u\n", want);
 
 		disable_interrupts();
@@ -344,5 +344,5 @@
 		/* paranoid */ verify( ctx );
 
-		__cfadbg_print_safe(io, "Kernel I/O : attempting to fast allocation\n");
+		// __cfadbg_print_safe(io, "Kernel I/O : attempting to fast allocation\n");
 
 		// We can proceed to the fast path
@@ -352,5 +352,5 @@
 			enable_interrupts();
 
-			__cfadbg_print_safe(io, "Kernel I/O : fast allocation successful from ring %d\n", ctx->fd);
+			// __cfadbg_print_safe(io, "Kernel I/O : fast allocation successful from ring %d\n", ctx->fd);
 
 			__fill( sqes, want, idxs, ctx );
@@ -367,9 +367,9 @@
 		/* paranoid */ verify( ioarb );
 
-		__cfadbg_print_safe(io, "Kernel I/O : falling back on arbiter for allocation\n");
+		// __cfadbg_print_safe(io, "Kernel I/O : falling back on arbiter for allocation\n");
 
 		struct $io_context * ret = __ioarbiter_allocate(*ioarb, idxs, want);
 
-		__cfadbg_print_safe(io, "Kernel I/O : slow allocation completed from ring %d\n", ret->fd);
+		// __cfadbg_print_safe(io, "Kernel I/O : slow allocation completed from ring %d\n", ret->fd);
 
 		__fill( sqes, want, idxs,ret );
@@ -388,5 +388,5 @@
 		// Add the sqes to the array
 		for( i; have ) {
-			__cfadbg_print_safe(io, "Kernel I/O : __submit loop\n");
+			// __cfadbg_print_safe(io, "Kernel I/O : __submit loop\n");
 			sq.kring.array[ (tail + i) & mask ] = idxs[i];
 		}
@@ -415,5 +415,5 @@
 
 	void cfa_io_submit( struct $io_context * inctx, __u32 idxs[], __u32 have, bool lazy ) __attribute__((nonnull (1))) {
-		__cfadbg_print_safe(io, "Kernel I/O : attempting to submit %u (%s)\n", have, lazy ? "lazy" : "eager");
+		// __cfadbg_print_safe(io, "Kernel I/O : attempting to submit %u (%s)\n", have, lazy ? "lazy" : "eager");
 
 		disable_interrupts();
@@ -432,5 +432,5 @@
 			enable_interrupts();
 
-			__cfadbg_print_safe(io, "Kernel I/O : submitted on fast path\n");
+			// __cfadbg_print_safe(io, "Kernel I/O : submitted on fast path\n");
 			return;
 		}
@@ -440,5 +440,5 @@
 		enable_interrupts();
 
-		__cfadbg_print_safe(io, "Kernel I/O : falling back on arbiter for submission\n");
+		// __cfadbg_print_safe(io, "Kernel I/O : falling back on arbiter for submission\n");
 
 		__ioarbiter_submit(inctx, idxs, have, lazy);
@@ -484,5 +484,5 @@
 		// go through the range and release the sqes
 		for( i; count ) {
-			__cfadbg_print_safe(io, "Kernel I/O : release loop\n");
+			// __cfadbg_print_safe(io, "Kernel I/O : release loop\n");
 			__u32 idx = ctx.sq.kring.array[ (phead + i) & mask ];
 			ctx.sq.free_ring.array[ (ftail + i) & mask ] = idx;
@@ -524,5 +524,5 @@
 
 	static $io_context * __ioarbiter_allocate( $io_arbiter & this, __u32 idxs[], __u32 want ) {
-		__cfadbg_print_safe(io, "Kernel I/O : arbiter allocating\n");
+		// __cfadbg_print_safe(io, "Kernel I/O : arbiter allocating\n");
 
 		__STATS__( false, io.alloc.block += 1; )
