1 | #define _GNU_SOURCE |
---|
2 | |
---|
3 | #include <errno.h> |
---|
4 | #include <signal.h> |
---|
5 | #include <stdio.h> |
---|
6 | #include <string.h> |
---|
7 | #include <unistd.h> |
---|
8 | extern "C" { |
---|
9 | #include <sched.h> |
---|
10 | #include <signal.h> |
---|
11 | #include <sys/eventfd.h> |
---|
12 | #include <sys/socket.h> |
---|
13 | #include <netinet/in.h> |
---|
14 | } |
---|
15 | |
---|
16 | #include <fstream.hfa> |
---|
17 | #include <kernel.hfa> |
---|
18 | #include <locks.hfa> |
---|
19 | #include <iofwd.hfa> |
---|
20 | #include <stats.hfa> |
---|
21 | #include <time.hfa> |
---|
22 | #include <thread.hfa> |
---|
23 | |
---|
24 | #include "filecache.hfa" |
---|
25 | #include "options.hfa" |
---|
26 | #include "socket.hfa" |
---|
27 | #include "printer.hfa" |
---|
28 | #include "worker.hfa" |
---|
29 | |
---|
30 | extern void register_fixed_files( cluster &, int *, unsigned count ); |
---|
31 | |
---|
32 | Duration default_preemption() { |
---|
33 | return 0; |
---|
34 | } |
---|
35 | |
---|
36 | //============================================================================================= |
---|
37 | // Globals |
---|
38 | //============================================================================================= |
---|
39 | void ?{}( ServerCluster & this ) { |
---|
40 | (this.self){ "Server Cluster", options.clopts.params }; |
---|
41 | |
---|
42 | cpu_set_t fullset; |
---|
43 | CPU_ZERO(&fullset); |
---|
44 | int ret = sched_getaffinity(getpid(), sizeof(fullset), &fullset); |
---|
45 | if( ret != 0 ) abort | "sched_getaffinity failed with" | errno | strerror( errno ); |
---|
46 | int cnt = CPU_COUNT(&fullset); |
---|
47 | |
---|
48 | this.procs = alloc(options.clopts.nprocs); |
---|
49 | for(i; options.clopts.nprocs) { |
---|
50 | (this.procs[i]){ "Benchmark Processor", this.self }; |
---|
51 | |
---|
52 | // int c = 0; |
---|
53 | // int n = 1 + (i % cnt); |
---|
54 | // for(int j = 0; j < CPU_SETSIZE; j++) { |
---|
55 | // if(CPU_ISSET(j, &fullset)) n--; |
---|
56 | // if(n == 0) { |
---|
57 | // c = j; |
---|
58 | // break; |
---|
59 | // } |
---|
60 | // } |
---|
61 | // cpu_set_t localset; |
---|
62 | // CPU_ZERO(&localset); |
---|
63 | // CPU_SET(c, &localset); |
---|
64 | // ret = pthread_setaffinity_np(this.procs[i].kernel_thread, sizeof(localset), &localset); |
---|
65 | // if( ret != 0 ) abort | "sched_getaffinity failed with" | ret | strerror( ret ); |
---|
66 | |
---|
67 | #if !defined(__CFA_NO_STATISTICS__) |
---|
68 | if( options.clopts.procstats ) { |
---|
69 | print_stats_at_exit( *this.procs, this.self.print_stats ); |
---|
70 | } |
---|
71 | if( options.clopts.viewhalts ) { |
---|
72 | print_halts( *this.procs ); |
---|
73 | } |
---|
74 | #endif |
---|
75 | } |
---|
76 | |
---|
77 | #if !defined(__CFA_NO_STATISTICS__) |
---|
78 | print_stats_at_exit( this.self, CFA_STATS_READY_Q | CFA_STATS_IO ); |
---|
79 | #endif |
---|
80 | |
---|
81 | options.clopts.instance[options.clopts.cltr_cnt] = &this.self; |
---|
82 | options.clopts.cltr_cnt++; |
---|
83 | } |
---|
84 | |
---|
85 | void ^?{}( ServerCluster & this ) { |
---|
86 | for(i; options.clopts.nprocs) { |
---|
87 | ^(this.procs[i]){}; |
---|
88 | } |
---|
89 | free(this.procs); |
---|
90 | |
---|
91 | ^(this.self){}; |
---|
92 | } |
---|
93 | |
---|
94 | extern void init_protocol(void); |
---|
95 | extern void deinit_protocol(void); |
---|
96 | |
---|
97 | //============================================================================================= |
---|
98 | // REUSEPORT |
---|
99 | //============================================================================================= |
---|
100 | |
---|
101 | size_t sockarr_size; |
---|
102 | struct __attribute__((aligned(128))) Q { |
---|
103 | mpsc_queue(PendingRead) q; |
---|
104 | }; |
---|
105 | |
---|
106 | //============================================================================================= |
---|
107 | // Termination |
---|
108 | //============================================================================================= |
---|
109 | |
---|
110 | int closefd; |
---|
111 | void cleanstop(int) { |
---|
112 | eventfd_t buffer = 1; |
---|
113 | char * buffer_s = (char*)&buffer; |
---|
114 | int ret = write(closefd, buffer_s, sizeof(buffer)); |
---|
115 | if(ret < 0) abort( "eventfd write error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
116 | return; |
---|
117 | } |
---|
118 | |
---|
119 | //============================================================================================= |
---|
120 | // Main |
---|
121 | //=============================================================================================' |
---|
122 | int main( int argc, char * argv[] ) { |
---|
123 | int ret; |
---|
124 | __sighandler_t s = 1p; |
---|
125 | signal(SIGPIPE, s); |
---|
126 | |
---|
127 | //=================== |
---|
128 | // Parse args |
---|
129 | parse_options(argc, argv); |
---|
130 | |
---|
131 | //=================== |
---|
132 | // Setup non-interactive termination |
---|
133 | if(!options.interactive) { |
---|
134 | closefd = eventfd(0, 0); |
---|
135 | if(closefd < 0) abort( "eventfd error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
136 | |
---|
137 | sighandler_t prev = signal(SIGTERM, cleanstop); |
---|
138 | intptr_t prev_workaround = (intptr_t) prev; |
---|
139 | // can't use SIG_ERR it crashes the compiler |
---|
140 | if(prev_workaround == -1) abort( "signal setup error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
141 | |
---|
142 | sout | "Signal termination ready"; |
---|
143 | } |
---|
144 | |
---|
145 | //=================== |
---|
146 | // Open Files |
---|
147 | if( options.file_cache.path ) { |
---|
148 | sout | "Filling cache from" | options.file_cache.path; |
---|
149 | fill_cache( options.file_cache.path ); |
---|
150 | } |
---|
151 | |
---|
152 | //=================== |
---|
153 | // Open Socket |
---|
154 | sout | getpid() | ": Listening on port" | options.socket.port; |
---|
155 | |
---|
156 | struct sockaddr_in address; |
---|
157 | int addrlen = prepaddr(address); |
---|
158 | |
---|
159 | int server_fd; |
---|
160 | |
---|
161 | //=================== |
---|
162 | // Run Server Cluster |
---|
163 | { |
---|
164 | int pipe_cnt = options.clopts.nworkers * 2; |
---|
165 | // int pipe_cnt = 0; |
---|
166 | int pipe_off; |
---|
167 | int * fds; |
---|
168 | [fds, pipe_off] = filefds( pipe_cnt ); |
---|
169 | for(i; 0 ~ pipe_cnt ~ 2) { |
---|
170 | int ret = pipe(&fds[pipe_off + i]); |
---|
171 | if( ret < 0 ) { abort( "pipe error: (%d) %s\n", (int)errno, strerror(errno) ); } |
---|
172 | } |
---|
173 | |
---|
174 | // if(options.file_cache.path && options.file_cache.fixed_fds) { |
---|
175 | // register_fixed_files(cl, fds, pipe_off); |
---|
176 | // } |
---|
177 | |
---|
178 | { |
---|
179 | // Stats printer makes a copy so this needs to persist longer than normal |
---|
180 | connection ** conns; |
---|
181 | AcceptWorker * aworkers = 0p; |
---|
182 | ChannelWorker * cworkers = 0p; |
---|
183 | Acceptor * acceptors = 0p; |
---|
184 | Q * queues = 0p; |
---|
185 | ServerCluster cl[options.clopts.nclusters]; |
---|
186 | |
---|
187 | if(options.stats) { |
---|
188 | stats_thrd = alloc(); |
---|
189 | (*stats_thrd){ cl }; |
---|
190 | } else { |
---|
191 | stats_thrd = 0p; |
---|
192 | } |
---|
193 | |
---|
194 | init_protocol(); |
---|
195 | { |
---|
196 | int nacceptors = options.clopts.nprocs * options.clopts.nclusters; |
---|
197 | conns = alloc(options.clopts.nworkers); |
---|
198 | if(options.socket.reuseport) { |
---|
199 | queues = alloc(nacceptors); |
---|
200 | acceptors = alloc(nacceptors); |
---|
201 | sout | "Creating" | nacceptors | "Acceptors"; |
---|
202 | for(i; nacceptors) { |
---|
203 | (acceptors[i]){ i % options.clopts.nclusters }; |
---|
204 | } |
---|
205 | for(i; nacceptors) { |
---|
206 | (queues[i]){}; |
---|
207 | { |
---|
208 | acceptors[i].sockfd = listener(address, addrlen); |
---|
209 | acceptors[i].addr = (struct sockaddr *)&address; |
---|
210 | acceptors[i].addrlen = (socklen_t*)&addrlen; |
---|
211 | acceptors[i].flags = 0; |
---|
212 | acceptors[i].queue = &queues[i].q; |
---|
213 | } |
---|
214 | unpark( acceptors[i] ); |
---|
215 | } |
---|
216 | |
---|
217 | cworkers = anew(options.clopts.nworkers); |
---|
218 | for(i; options.clopts.nworkers) { |
---|
219 | { |
---|
220 | cworkers[i].conn.pipe[0] = fds[pipe_off + (i * 2) + 0]; |
---|
221 | cworkers[i].conn.pipe[1] = fds[pipe_off + (i * 2) + 1]; |
---|
222 | cworkers[i].queue = &queues[i % nacceptors].q; |
---|
223 | conns[i] = &cworkers[i].conn; |
---|
224 | } |
---|
225 | unpark( cworkers[i] ); |
---|
226 | } |
---|
227 | } |
---|
228 | else { |
---|
229 | server_fd = listener(address, addrlen); |
---|
230 | aworkers = anew(options.clopts.nworkers); |
---|
231 | for(i; options.clopts.nworkers) { |
---|
232 | // if( options.file_cache.fixed_fds ) { |
---|
233 | // workers[i].pipe[0] = pipe_off + (i * 2) + 0; |
---|
234 | // workers[i].pipe[1] = pipe_off + (i * 2) + 1; |
---|
235 | // } |
---|
236 | // else |
---|
237 | { |
---|
238 | aworkers[i].conn.pipe[0] = fds[pipe_off + (i * 2) + 0]; |
---|
239 | aworkers[i].conn.pipe[1] = fds[pipe_off + (i * 2) + 1]; |
---|
240 | aworkers[i].sockfd = server_fd; |
---|
241 | aworkers[i].addr = (struct sockaddr *)&address; |
---|
242 | aworkers[i].addrlen = (socklen_t*)&addrlen; |
---|
243 | aworkers[i].flags = 0; |
---|
244 | conns[i] = &aworkers[i].conn; |
---|
245 | } |
---|
246 | unpark( aworkers[i] ); |
---|
247 | } |
---|
248 | } |
---|
249 | sout | options.clopts.nworkers | "workers started on" | options.clopts.nprocs | "processors /" | options.clopts.nclusters | "clusters"; |
---|
250 | for(i; options.clopts.nclusters) { |
---|
251 | sout | options.clopts.thrd_cnt[i] | nonl; |
---|
252 | } |
---|
253 | sout | nl; |
---|
254 | { |
---|
255 | if(options.interactive) { |
---|
256 | char buffer[128]; |
---|
257 | for() { |
---|
258 | int ret = cfa_read(0, buffer, 128, 0); |
---|
259 | if(ret == 0) break; |
---|
260 | if(ret < 0) abort( "main read error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
261 | sout | "User wrote '" | "" | nonl; |
---|
262 | write(sout, buffer, ret - 1); |
---|
263 | sout | "'"; |
---|
264 | } |
---|
265 | } |
---|
266 | else { |
---|
267 | char buffer[sizeof(eventfd_t)]; |
---|
268 | int ret = cfa_read(closefd, buffer, sizeof(eventfd_t), 0); |
---|
269 | if(ret < 0) abort( "main read error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
270 | } |
---|
271 | |
---|
272 | sout | "Shutdown received"; |
---|
273 | } |
---|
274 | |
---|
275 | //=================== |
---|
276 | // Close Socket and join |
---|
277 | if(options.socket.reuseport) { |
---|
278 | sout | "Notifying connections..." | nonl; flush( sout ); |
---|
279 | for(i; nacceptors) { |
---|
280 | acceptors[i].done = true; |
---|
281 | } |
---|
282 | for(i; options.clopts.nworkers) { |
---|
283 | cworkers[i].done = true; |
---|
284 | } |
---|
285 | sout | "done"; |
---|
286 | |
---|
287 | sout | "Shutting down Socket..." | nonl; flush( sout ); |
---|
288 | for(i; nacceptors) { |
---|
289 | ret = shutdown( acceptors[i].sockfd, SHUT_RD ); |
---|
290 | if( ret < 0 ) { |
---|
291 | abort( "shutdown1 error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
292 | } |
---|
293 | } |
---|
294 | sout | "done"; |
---|
295 | |
---|
296 | sout | "Closing Socket..." | nonl; flush( sout ); |
---|
297 | for(i; nacceptors) { |
---|
298 | ret = close( acceptors[i].sockfd ); |
---|
299 | if( ret < 0) { |
---|
300 | abort( "close socket error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
301 | } |
---|
302 | } |
---|
303 | sout | "done"; |
---|
304 | |
---|
305 | //=================== |
---|
306 | // Close Files |
---|
307 | if( options.file_cache.path ) { |
---|
308 | sout | "Closing open files..." | nonl; flush( sout ); |
---|
309 | close_cache(); |
---|
310 | sout | "done"; |
---|
311 | } |
---|
312 | |
---|
313 | sout | "Stopping accept threads..." | nonl; flush( sout ); |
---|
314 | for(i; nacceptors) { |
---|
315 | join(acceptors[i]); |
---|
316 | } |
---|
317 | sout | "done"; |
---|
318 | |
---|
319 | sout | "Draining worker queues..." | nonl; flush( sout ); |
---|
320 | for(i; nacceptors) { |
---|
321 | PendingRead * p = 0p; |
---|
322 | while(p = pop(queues[i].q)) { |
---|
323 | fulfil(p->f, -ECONNRESET); |
---|
324 | } |
---|
325 | } |
---|
326 | sout | "done"; |
---|
327 | |
---|
328 | sout | "Stopping worker threads..." | nonl; flush( sout ); |
---|
329 | for(i; options.clopts.nworkers) { |
---|
330 | for(j; 2) { |
---|
331 | ret = close(cworkers[i].conn.pipe[j]); |
---|
332 | if(ret < 0) abort( "close pipe %d error: (%d) %s\n", j, (int)errno, strerror(errno) ); |
---|
333 | } |
---|
334 | join(cworkers[i]); |
---|
335 | } |
---|
336 | } |
---|
337 | else { |
---|
338 | sout | "Notifying connections..." | nonl; flush( sout ); |
---|
339 | for(i; options.clopts.nworkers) { |
---|
340 | aworkers[i].done = true; |
---|
341 | } |
---|
342 | sout | "done"; |
---|
343 | |
---|
344 | sout | "Shutting down Socket..." | nonl; flush( sout ); |
---|
345 | ret = shutdown( server_fd, SHUT_RD ); |
---|
346 | if( ret < 0 ) { |
---|
347 | abort( "shutdown2 error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
348 | } |
---|
349 | sout | "done"; |
---|
350 | |
---|
351 | sout | "Closing Socket..." | nonl; flush( sout ); |
---|
352 | ret = close( server_fd ); |
---|
353 | if(ret < 0) { |
---|
354 | abort( "close socket error: (%d) %s\n", (int)errno, strerror(errno) ); |
---|
355 | } |
---|
356 | sout | "done"; |
---|
357 | |
---|
358 | //=================== |
---|
359 | // Close Files |
---|
360 | if( options.file_cache.path ) { |
---|
361 | sout | "Closing open files..." | nonl; flush( sout ); |
---|
362 | close_cache(); |
---|
363 | sout | "done"; |
---|
364 | } |
---|
365 | |
---|
366 | sout | "Stopping connection threads..." | nonl; flush( sout ); |
---|
367 | for(i; options.clopts.nworkers) { |
---|
368 | for(j; 2) { |
---|
369 | ret = close(aworkers[i].conn.pipe[j]); |
---|
370 | if(ret < 0) abort( "close pipe %d error: (%d) %s\n", j, (int)errno, strerror(errno) ); |
---|
371 | } |
---|
372 | join(aworkers[i]); |
---|
373 | } |
---|
374 | } |
---|
375 | } |
---|
376 | sout | "done"; |
---|
377 | |
---|
378 | sout | "Stopping protocol threads..." | nonl; flush( sout ); |
---|
379 | deinit_protocol(); |
---|
380 | sout | "done"; |
---|
381 | |
---|
382 | sout | "Stopping printer threads..." | nonl; flush( sout ); |
---|
383 | if(stats_thrd) { |
---|
384 | notify_one(stats_thrd->var); |
---|
385 | } |
---|
386 | delete(stats_thrd); |
---|
387 | sout | "done"; |
---|
388 | |
---|
389 | // Now that the stats printer is stopped, we can reclaim this |
---|
390 | adelete(aworkers); |
---|
391 | adelete(cworkers); |
---|
392 | adelete(acceptors); |
---|
393 | adelete(queues); |
---|
394 | free(conns); |
---|
395 | |
---|
396 | sout | "Stopping processors/clusters..." | nonl; flush( sout ); |
---|
397 | } |
---|
398 | sout | "done"; |
---|
399 | |
---|
400 | free(fds); |
---|
401 | |
---|
402 | sout | "Stopping processors..." | nonl; flush( sout ); |
---|
403 | } |
---|
404 | sout | "done"; |
---|
405 | } |
---|
406 | |
---|
407 | const size_t zipf_sizes[] = { 102, 204, 307, 409, 512, 614, 716, 819, 921, 1024, 2048, 3072, 4096, 5120, 6144, 7168, 8192, 9216, 10240, 20480, 30720, 40960, 51200, 61440, 71680, 81920, 92160, 102400, 204800, 307200, 409600, 512000, 614400, 716800, 819200, 921600 }; |
---|
408 | static_assert(zipf_cnts == sizeof(zipf_sizes) / sizeof(zipf_sizes[0])); |
---|