xref: /freebsd/contrib/unbound/libunbound/libworker.c (revision 46d2f61818f594174cafe31ee338c6e083fa1876)
1 /*
2  * libunbound/worker.c - worker thread or process that resolves
3  *
4  * Copyright (c) 2007, NLnet Labs. All rights reserved.
5  *
6  * This software is open source.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  *
12  * Redistributions of source code must retain the above copyright notice,
13  * this list of conditions and the following disclaimer.
14  *
15  * Redistributions in binary form must reproduce the above copyright notice,
16  * this list of conditions and the following disclaimer in the documentation
17  * and/or other materials provided with the distribution.
18  *
19  * Neither the name of the NLNET LABS nor the names of its contributors may
20  * be used to endorse or promote products derived from this software without
21  * specific prior written permission.
22  *
23  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
24  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
25  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
26  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
27  * HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
28  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
29  * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
30  * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
31  * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
32  * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
33  * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
34  */
35 
36 /**
37  * \file
38  *
39  * This file contains the worker process or thread that performs
40  * the DNS resolving and validation. The worker is called by a procedure
41  * and if in the background continues until exit, if in the foreground
42  * returns from the procedure when done.
43  */
44 #include "config.h"
45 #ifdef HAVE_SSL
46 #include <openssl/ssl.h>
47 #endif
48 #include "libunbound/libworker.h"
49 #include "libunbound/context.h"
50 #include "libunbound/unbound.h"
51 #include "libunbound/worker.h"
52 #include "libunbound/unbound-event.h"
53 #include "services/outside_network.h"
54 #include "services/mesh.h"
55 #include "services/localzone.h"
56 #include "services/cache/rrset.h"
57 #include "services/outbound_list.h"
58 #include "services/authzone.h"
59 #include "util/fptr_wlist.h"
60 #include "util/module.h"
61 #include "util/regional.h"
62 #include "util/random.h"
63 #include "util/config_file.h"
64 #include "util/netevent.h"
65 #include "util/proxy_protocol.h"
66 #include "util/storage/lookup3.h"
67 #include "util/storage/slabhash.h"
68 #include "util/net_help.h"
69 #include "util/data/dname.h"
70 #include "util/data/msgreply.h"
71 #include "util/data/msgencode.h"
72 #include "util/tube.h"
73 #include "sldns/sbuffer.h"
74 #include "sldns/str2wire.h"
75 #ifdef USE_DNSTAP
76 #include "dnstap/dtstream.h"
77 #endif
78 
79 #ifdef HAVE_TARGETCONDITIONALS_H
80 #include <TargetConditionals.h>
81 #endif
82 
83 #if (defined(TARGET_OS_TV) && TARGET_OS_TV) || (defined(TARGET_OS_WATCH) && TARGET_OS_WATCH)
84 #undef HAVE_FORK
85 #endif
86 
87 /** handle new query command for bg worker */
88 static void handle_newq(struct libworker* w, uint8_t* buf, uint32_t len);
89 
90 /** delete libworker env */
91 static void
libworker_delete_env(struct libworker * w)92 libworker_delete_env(struct libworker* w)
93 {
94 	if(w->env) {
95 		outside_network_quit_prepare(w->back);
96 		mesh_delete(w->env->mesh);
97 		context_release_alloc(w->ctx, w->env->alloc,
98 			!w->is_bg || w->is_bg_thread);
99 		sldns_buffer_free(w->env->scratch_buffer);
100 		regional_destroy(w->env->scratch);
101 		ub_randfree(w->env->rnd);
102 		free(w->env);
103 	}
104 #ifdef HAVE_SSL
105 	SSL_CTX_free(w->sslctx);
106 #endif
107 	outside_network_delete(w->back);
108 }
109 
110 /** delete libworker struct */
111 static void
libworker_delete(struct libworker * w)112 libworker_delete(struct libworker* w)
113 {
114 	if(!w) return;
115 	libworker_delete_env(w);
116 	comm_base_delete(w->base);
117 	free(w);
118 }
119 
120 void
libworker_delete_event(struct libworker * w)121 libworker_delete_event(struct libworker* w)
122 {
123 	if(!w) return;
124 	libworker_delete_env(w);
125 	comm_base_delete_no_base(w->base);
126 	free(w);
127 }
128 
129 /** setup fresh libworker struct */
130 static struct libworker*
libworker_setup(struct ub_ctx * ctx,int is_bg,struct ub_event_base * eb)131 libworker_setup(struct ub_ctx* ctx, int is_bg, struct ub_event_base* eb)
132 {
133 	struct libworker* w = (struct libworker*)calloc(1, sizeof(*w));
134 	struct config_file* cfg = ctx->env->cfg;
135 	int* ports;
136 	int numports;
137 	if(!w) return NULL;
138 	w->is_bg = is_bg;
139 	w->ctx = ctx;
140 	w->env = (struct module_env*)malloc(sizeof(*w->env));
141 	if(!w->env) {
142 		free(w);
143 		return NULL;
144 	}
145 	*w->env = *ctx->env;
146 	w->env->alloc = context_obtain_alloc(ctx, !w->is_bg || w->is_bg_thread);
147 	if(!w->env->alloc) {
148 		libworker_delete(w);
149 		return NULL;
150 	}
151 	w->thread_num = w->env->alloc->thread_num;
152 	alloc_set_id_cleanup(w->env->alloc, &libworker_alloc_cleanup, w);
153 	if(!w->is_bg || w->is_bg_thread) {
154 		lock_basic_lock(&ctx->cfglock);
155 	}
156 	w->env->scratch = regional_create_custom(cfg->msg_buffer_size);
157 	w->env->scratch_buffer = sldns_buffer_new(cfg->msg_buffer_size);
158 #ifdef HAVE_SSL
159 	w->sslctx = connect_sslctx_create(NULL, NULL,
160 		cfg->tls_cert_bundle, cfg->tls_win_cert);
161 	if(!w->sslctx) {
162 		/* to make the setup fail after unlock */
163 		sldns_buffer_free(w->env->scratch_buffer);
164 		w->env->scratch_buffer = NULL;
165 	}
166 #endif
167 	if(!w->is_bg || w->is_bg_thread) {
168 		lock_basic_unlock(&ctx->cfglock);
169 	}
170 	if(!w->env->scratch || !w->env->scratch_buffer) {
171 		libworker_delete(w);
172 		return NULL;
173 	}
174 	w->env->worker = (struct worker*)w;
175 	w->env->probe_timer = NULL;
176 	if(!w->is_bg || w->is_bg_thread) {
177 		lock_basic_lock(&ctx->cfglock);
178 	}
179 	if(!(w->env->rnd = ub_initstate(ctx->seed_rnd))) {
180 		if(!w->is_bg || w->is_bg_thread) {
181 			lock_basic_unlock(&ctx->cfglock);
182 		}
183 		libworker_delete(w);
184 		return NULL;
185 	}
186 	if(!w->is_bg || w->is_bg_thread) {
187 		lock_basic_unlock(&ctx->cfglock);
188 	}
189 	if(1) {
190 		/* primitive lockout for threading: if it overwrites another
191 		 * thread it is like wiping the cache (which is likely empty
192 		 * at the start) */
193 		/* note we are holding the ctx lock in normal threaded
194 		 * cases so that is solved properly, it is only for many ctx
195 		 * in different threads that this may clash */
196 		static int done_raninit = 0;
197 		if(!done_raninit) {
198 			done_raninit = 1;
199 			hash_set_raninit((uint32_t)ub_random(w->env->rnd));
200 		}
201 	}
202 
203 	if(eb)
204 		w->base = comm_base_create_event(eb);
205 	else	w->base = comm_base_create(0);
206 	if(!w->base) {
207 		libworker_delete(w);
208 		return NULL;
209 	}
210 	w->env->worker_base = w->base;
211 	if(!w->is_bg || w->is_bg_thread) {
212 		lock_basic_lock(&ctx->cfglock);
213 	}
214 	numports = cfg_condense_ports(cfg, &ports);
215 	if(numports == 0) {
216 		if(!w->is_bg || w->is_bg_thread) {
217 			lock_basic_unlock(&ctx->cfglock);
218 		}
219 		libworker_delete(w);
220 		return NULL;
221 	}
222 	w->back = outside_network_create(w->base, cfg->msg_buffer_size,
223 		(size_t)cfg->outgoing_num_ports, cfg->out_ifs,
224 		cfg->num_out_ifs, cfg->do_ip4, cfg->do_ip6,
225 		cfg->do_tcp?cfg->outgoing_num_tcp:0, cfg->ip_dscp,
226 		w->env->infra_cache, w->env->rnd, cfg->use_caps_bits_for_id,
227 		ports, numports, cfg->unwanted_threshold,
228 		cfg->outgoing_tcp_mss, &libworker_alloc_cleanup, w,
229 		cfg->do_udp || cfg->udp_upstream_without_downstream, w->sslctx,
230 		cfg->delay_close, cfg->tls_use_sni, NULL, cfg->udp_connect,
231 		cfg->max_reuse_tcp_queries, cfg->tcp_reuse_timeout,
232 		cfg->tcp_auth_query_timeout);
233 	w->env->outnet = w->back;
234 	if(!w->is_bg || w->is_bg_thread) {
235 		lock_basic_unlock(&ctx->cfglock);
236 	}
237 	free(ports);
238 	if(!w->back) {
239 		libworker_delete(w);
240 		return NULL;
241 	}
242 	w->env->mesh = mesh_create(&ctx->mods, w->env);
243 	if(!w->env->mesh) {
244 		libworker_delete(w);
245 		return NULL;
246 	}
247 	w->env->send_query = &libworker_send_query;
248 	w->env->detach_subs = &mesh_detach_subs;
249 	w->env->attach_sub = &mesh_attach_sub;
250 	w->env->add_sub = &mesh_add_sub;
251 	w->env->kill_sub = &mesh_state_delete;
252 	w->env->detect_cycle = &mesh_detect_cycle;
253 	comm_base_timept(w->base, &w->env->now, &w->env->now_tv);
254 	pp_init(&sldns_write_uint16, &sldns_write_uint32);
255 	return w;
256 }
257 
libworker_create_event(struct ub_ctx * ctx,struct ub_event_base * eb)258 struct libworker* libworker_create_event(struct ub_ctx* ctx,
259 	struct ub_event_base* eb)
260 {
261 	return libworker_setup(ctx, 0, eb);
262 }
263 
264 /** handle cancel command for bg worker */
265 static void
handle_cancel(struct libworker * w,uint8_t * buf,uint32_t len)266 handle_cancel(struct libworker* w, uint8_t* buf, uint32_t len)
267 {
268 	struct ctx_query* q;
269 	if(w->is_bg_thread) {
270 		lock_basic_lock(&w->ctx->cfglock);
271 		q = context_deserialize_cancel(w->ctx, buf, len);
272 		lock_basic_unlock(&w->ctx->cfglock);
273 	} else {
274 		q = context_deserialize_cancel(w->ctx, buf, len);
275 	}
276 	if(!q) {
277 		/* probably simply lookup failed, i.e. the message had been
278 		 * processed and answered before the cancel arrived */
279 		return;
280 	}
281 	q->cancelled = 1;
282 	free(buf);
283 }
284 
285 /** do control command coming into bg server */
286 static void
libworker_do_cmd(struct libworker * w,uint8_t * msg,uint32_t len)287 libworker_do_cmd(struct libworker* w, uint8_t* msg, uint32_t len)
288 {
289 	switch(context_serial_getcmd(msg, len)) {
290 		default:
291 		case UB_LIBCMD_ANSWER:
292 			log_err("unknown command for bg worker %d",
293 				(int)context_serial_getcmd(msg, len));
294 			/* and fall through to quit */
295 			ATTR_FALLTHROUGH
296 			/* fallthrough */
297 		case UB_LIBCMD_QUIT:
298 			free(msg);
299 			comm_base_exit(w->base);
300 			break;
301 		case UB_LIBCMD_NEWQUERY:
302 			handle_newq(w, msg, len);
303 			break;
304 		case UB_LIBCMD_CANCEL:
305 			handle_cancel(w, msg, len);
306 			break;
307 	}
308 }
309 
310 /** handle control command coming into server */
311 void
libworker_handle_control_cmd(struct tube * ATTR_UNUSED (tube),uint8_t * msg,size_t len,int err,void * arg)312 libworker_handle_control_cmd(struct tube* ATTR_UNUSED(tube),
313 	uint8_t* msg, size_t len, int err, void* arg)
314 {
315 	struct libworker* w = (struct libworker*)arg;
316 
317 	if(err != 0) {
318 		free(msg);
319 		/* it is of no use to go on, exit */
320 		comm_base_exit(w->base);
321 		return;
322 	}
323 	libworker_do_cmd(w, msg, len); /* also frees the buf */
324 }
325 
326 /** the background thread func */
327 static void*
libworker_dobg(void * arg)328 libworker_dobg(void* arg)
329 {
330 	/* setup */
331 	uint32_t m;
332 	struct libworker* w = (struct libworker*)arg;
333 	struct ub_ctx* ctx;
334 	if(!w) {
335 		log_err("libunbound bg worker init failed, nomem");
336 		return NULL;
337 	}
338 	ctx = w->ctx;
339 	log_thread_set(&w->thread_num);
340 #ifdef THREADS_DISABLED
341 	/* we are forked */
342 	w->is_bg_thread = 0;
343 	/* close non-used parts of the pipes */
344 	tube_close_write(ctx->qq_pipe);
345 	tube_close_read(ctx->rr_pipe);
346 #endif
347 	if(!tube_setup_bg_listen(ctx->qq_pipe, w->base,
348 		libworker_handle_control_cmd, w)) {
349 		log_err("libunbound bg worker init failed, no bglisten");
350 		return NULL;
351 	}
352 	if(!tube_setup_bg_write(ctx->rr_pipe, w->base)) {
353 		log_err("libunbound bg worker init failed, no bgwrite");
354 		return NULL;
355 	}
356 
357 	/* do the work */
358 	comm_base_dispatch(w->base);
359 
360 	/* cleanup */
361 	m = UB_LIBCMD_QUIT;
362 	w->want_quit = 1;
363 	tube_remove_bg_listen(w->ctx->qq_pipe);
364 	tube_remove_bg_write(w->ctx->rr_pipe);
365 	libworker_delete(w);
366 	(void)tube_write_msg(ctx->rr_pipe, (uint8_t*)&m,
367 		(uint32_t)sizeof(m), 0);
368 #ifdef THREADS_DISABLED
369 	/* close pipes from forked process before exit */
370 	tube_close_read(ctx->qq_pipe);
371 	tube_close_write(ctx->rr_pipe);
372 #endif
373 	return NULL;
374 }
375 
libworker_bg(struct ub_ctx * ctx)376 int libworker_bg(struct ub_ctx* ctx)
377 {
378 	struct libworker* w;
379 	/* fork or threadcreate */
380 	lock_basic_lock(&ctx->cfglock);
381 	if(ctx->dothread) {
382 		lock_basic_unlock(&ctx->cfglock);
383 		w = libworker_setup(ctx, 1, NULL);
384 		if(!w) return UB_NOMEM;
385 		w->is_bg_thread = 1;
386 		ctx->thread_worker = w;
387 #ifdef ENABLE_LOCK_CHECKS
388 		w->thread_num = 1; /* for nicer DEBUG checklocks */
389 #endif
390 		ub_thread_create(&ctx->bg_tid, libworker_dobg, w);
391 	} else {
392 		lock_basic_unlock(&ctx->cfglock);
393 #ifndef HAVE_FORK
394 		/* no fork on windows */
395 		return UB_FORKFAIL;
396 #else /* HAVE_FORK */
397 		switch((ctx->bg_pid=fork())) {
398 			case 0:
399 				w = libworker_setup(ctx, 1, NULL);
400 				if(!w) fatal_exit("out of memory");
401 				/* close non-used parts of the pipes */
402 				tube_close_write(ctx->qq_pipe);
403 				tube_close_read(ctx->rr_pipe);
404 				(void)libworker_dobg(w);
405 				exit(0);
406 				break;
407 			case -1:
408 				return UB_FORKFAIL;
409 			default:
410 				/* close non-used parts, so that the worker
411 				 * bgprocess gets 'pipe closed' when the
412 				 * main process exits */
413 				tube_close_read(ctx->qq_pipe);
414 				tube_close_write(ctx->rr_pipe);
415 				break;
416 		}
417 #endif /* HAVE_FORK */
418 	}
419 	return UB_NOERROR;
420 }
421 
422 /** insert canonname */
423 static int
fill_canon(struct ub_result * res,uint8_t * s)424 fill_canon(struct ub_result* res, uint8_t* s)
425 {
426 	char buf[255+2];
427 	dname_str(s, buf);
428 	res->canonname = strdup(buf);
429 	return res->canonname != 0;
430 }
431 
432 /** fill data into result */
433 static int
fill_res(struct ub_result * res,struct ub_packed_rrset_key * answer,uint8_t * finalcname,struct query_info * rq,struct reply_info * rep)434 fill_res(struct ub_result* res, struct ub_packed_rrset_key* answer,
435 	uint8_t* finalcname, struct query_info* rq, struct reply_info* rep)
436 {
437 	size_t i;
438 	struct packed_rrset_data* data;
439 	res->ttl = 0;
440 	if(!answer) {
441 		if(finalcname) {
442 			if(!fill_canon(res, finalcname))
443 				return 0; /* out of memory */
444 		}
445 		if(rep->rrset_count != 0)
446 			res->ttl = (int)rep->ttl;
447 		res->data = (char**)calloc(1, sizeof(char*));
448 		if(!res->data)
449 			return 0; /* out of memory */
450 		res->len = (int*)calloc(1, sizeof(int));
451 		if(!res->len) {
452 			free(res->data);
453 			res->data = NULL;
454 			return 0; /* out of memory */
455 		}
456 		return 1;
457 	}
458 	data = (struct packed_rrset_data*)answer->entry.data;
459 	if(query_dname_compare(rq->qname, answer->rk.dname) != 0) {
460 		if(!fill_canon(res, answer->rk.dname))
461 			return 0; /* out of memory */
462 	} else	res->canonname = NULL;
463 	res->data = (char**)calloc(data->count+1, sizeof(char*));
464 	if(!res->data)
465 		return 0; /* out of memory */
466 	res->len = (int*)calloc(data->count+1, sizeof(int));
467 	if(!res->len) {
468 		free(res->data);
469 		res->data = NULL;
470 		return 0; /* out of memory */
471 	}
472 	for(i=0; i<data->count; i++) {
473 		/* remove rdlength from rdata */
474 		res->len[i] = (int)(data->rr_len[i] - 2);
475 		res->data[i] = memdup(data->rr_data[i]+2, (size_t)res->len[i]);
476 		if(!res->data[i]) {
477 			size_t j;
478 			for(j=0; j<i; j++) {
479 				free(res->data[j]);
480 				res->data[j] = NULL;
481 			}
482 			free(res->data);
483 			res->data = NULL;
484 			free(res->len);
485 			res->len = NULL;
486 			return 0; /* out of memory */
487 		}
488 	}
489 	/* ttl for positive answers, from CNAME and answer RRs */
490 	if(data->count != 0) {
491 		size_t j;
492 		res->ttl = (int)data->ttl;
493 		for(j=0; j<rep->an_numrrsets; j++) {
494 			struct packed_rrset_data* d =
495 				(struct packed_rrset_data*)rep->rrsets[j]->
496 				entry.data;
497 			if((int)d->ttl < res->ttl)
498 				res->ttl = (int)d->ttl;
499 		}
500 	}
501 	/* ttl for negative answers */
502 	if(data->count == 0 && rep->rrset_count != 0)
503 		res->ttl = (int)rep->ttl;
504 	res->data[data->count] = NULL;
505 	res->len[data->count] = 0;
506 	return 1;
507 }
508 
509 /** fill result from parsed message, on error fills servfail */
510 void
libworker_enter_result(struct ub_result * res,sldns_buffer * buf,struct regional * temp,enum sec_status msg_security)511 libworker_enter_result(struct ub_result* res, sldns_buffer* buf,
512 	struct regional* temp, enum sec_status msg_security)
513 {
514 	struct query_info rq;
515 	struct reply_info* rep;
516 	res->rcode = LDNS_RCODE_SERVFAIL;
517 	rep = parse_reply_in_temp_region(buf, temp, &rq);
518 	if(!rep) {
519 		log_err("cannot parse buf");
520 		return; /* error parsing buf, or out of memory */
521 	}
522 	if(!fill_res(res, reply_find_answer_rrset(&rq, rep),
523 		reply_find_final_cname_target(&rq, rep), &rq, rep))
524 		return; /* out of memory */
525 	/* rcode, havedata, nxdomain, secure, bogus */
526 	res->rcode = (int)FLAGS_GET_RCODE(rep->flags);
527 	if(res->data && res->data[0])
528 		res->havedata = 1;
529 	if(res->rcode == LDNS_RCODE_NXDOMAIN)
530 		res->nxdomain = 1;
531 	if(msg_security == sec_status_secure)
532 		res->secure = 1;
533 	if(msg_security == sec_status_bogus ||
534 		msg_security == sec_status_secure_sentinel_fail)
535 		res->bogus = 1;
536 }
537 
538 /** fillup fg results */
539 static void
libworker_fillup_fg(struct ctx_query * q,int rcode,sldns_buffer * buf,enum sec_status s,char * why_bogus,int was_ratelimited)540 libworker_fillup_fg(struct ctx_query* q, int rcode, sldns_buffer* buf,
541 	enum sec_status s, char* why_bogus, int was_ratelimited)
542 {
543 	q->res->was_ratelimited = was_ratelimited;
544 	if(why_bogus)
545 		q->res->why_bogus = strdup(why_bogus);
546 	if(rcode != 0) {
547 		q->res->rcode = rcode;
548 		q->msg_security = s;
549 		return;
550 	}
551 
552 	q->res->rcode = LDNS_RCODE_SERVFAIL;
553 	q->msg_security = sec_status_unchecked;
554 	q->msg = memdup(sldns_buffer_begin(buf), sldns_buffer_limit(buf));
555 	q->msg_len = sldns_buffer_limit(buf);
556 	if(!q->msg) {
557 		return; /* the error is in the rcode */
558 	}
559 
560 	/* canonname and results */
561 	q->msg_security = s;
562 	libworker_enter_result(q->res, buf, q->w->env->scratch, s);
563 }
564 
565 void
libworker_fg_done_cb(void * arg,int rcode,sldns_buffer * buf,enum sec_status s,char * why_bogus,int was_ratelimited)566 libworker_fg_done_cb(void* arg, int rcode, sldns_buffer* buf, enum sec_status s,
567 	char* why_bogus, int was_ratelimited)
568 {
569 	struct ctx_query* q = (struct ctx_query*)arg;
570 	/* fg query is done; exit comm base */
571 	comm_base_exit(q->w->base);
572 
573 	libworker_fillup_fg(q, rcode, buf, s, why_bogus, was_ratelimited);
574 }
575 
576 /** setup qinfo and edns */
577 static int
setup_qinfo_edns(struct libworker * w,struct ctx_query * q,struct query_info * qinfo,struct edns_data * edns)578 setup_qinfo_edns(struct libworker* w, struct ctx_query* q,
579 	struct query_info* qinfo, struct edns_data* edns)
580 {
581 	qinfo->qtype = (uint16_t)q->res->qtype;
582 	qinfo->qclass = (uint16_t)q->res->qclass;
583 	qinfo->local_alias = NULL;
584 	qinfo->qname = sldns_str2wire_dname(q->res->qname, &qinfo->qname_len);
585 	if(!qinfo->qname) {
586 		return 0;
587 	}
588 	edns->edns_present = 1;
589 	edns->ext_rcode = 0;
590 	edns->edns_version = 0;
591 	edns->bits = EDNS_DO;
592 	edns->opt_list_in = NULL;
593 	edns->opt_list_out = NULL;
594 	edns->opt_list_inplace_cb_out = NULL;
595 	edns->padding_block_size = 0;
596 	edns->cookie_present = 0;
597 	edns->cookie_valid = 0;
598 	if(sldns_buffer_capacity(w->back->udp_buff) < 65535)
599 		edns->udp_size = (uint16_t)sldns_buffer_capacity(
600 			w->back->udp_buff);
601 	else	edns->udp_size = 65535;
602 	return 1;
603 }
604 
libworker_fg(struct ub_ctx * ctx,struct ctx_query * q)605 int libworker_fg(struct ub_ctx* ctx, struct ctx_query* q)
606 {
607 	struct libworker* w = libworker_setup(ctx, 0, NULL);
608 	uint16_t qflags, qid;
609 	struct query_info qinfo;
610 	struct edns_data edns;
611 	if(!w)
612 		return UB_INITFAIL;
613 	if(!setup_qinfo_edns(w, q, &qinfo, &edns)) {
614 		libworker_delete(w);
615 		return UB_SYNTAX;
616 	}
617 	qid = 0;
618 	qflags = BIT_RD;
619 	q->w = w;
620 	/* see if there is a fixed answer */
621 	sldns_buffer_write_u16_at(w->back->udp_buff, 0, qid);
622 	sldns_buffer_write_u16_at(w->back->udp_buff, 2, qflags);
623 	if(local_zones_answer(ctx->local_zones, w->env, &qinfo, &edns,
624 		w->back->udp_buff, w->env->scratch, NULL, NULL, 0, NULL, 0,
625 		NULL, 0, NULL, 0, NULL)) {
626 		regional_free_all(w->env->scratch);
627 		libworker_fillup_fg(q, LDNS_RCODE_NOERROR,
628 			w->back->udp_buff, sec_status_insecure, NULL, 0);
629 		libworker_delete(w);
630 		free(qinfo.qname);
631 		return UB_NOERROR;
632 	}
633 	if(ctx->env->auth_zones && auth_zones_answer(ctx->env->auth_zones,
634 		w->env, &qinfo, &edns, NULL, w->back->udp_buff, w->env->scratch)) {
635 		regional_free_all(w->env->scratch);
636 		libworker_fillup_fg(q, LDNS_RCODE_NOERROR,
637 			w->back->udp_buff, sec_status_insecure, NULL, 0);
638 		libworker_delete(w);
639 		free(qinfo.qname);
640 		return UB_NOERROR;
641 	}
642 	/* process new query */
643 	if(!mesh_new_callback(w->env->mesh, &qinfo, qflags, &edns,
644 		w->back->udp_buff, qid, libworker_fg_done_cb, q, 0)) {
645 		free(qinfo.qname);
646 		return UB_NOMEM;
647 	}
648 	free(qinfo.qname);
649 
650 	/* wait for reply */
651 	comm_base_dispatch(w->base);
652 
653 	libworker_delete(w);
654 	return UB_NOERROR;
655 }
656 
657 void
libworker_event_done_cb(void * arg,int rcode,sldns_buffer * buf,enum sec_status s,char * why_bogus,int was_ratelimited)658 libworker_event_done_cb(void* arg, int rcode, sldns_buffer* buf,
659 	enum sec_status s, char* why_bogus, int was_ratelimited)
660 {
661 	struct ctx_query* q = (struct ctx_query*)arg;
662 	ub_event_callback_type cb = q->cb_event;
663 	void* cb_arg = q->cb_arg;
664 	int cancelled = q->cancelled;
665 
666 	/* delete it now */
667 	struct ub_ctx* ctx = q->w->ctx;
668 	lock_basic_lock(&ctx->cfglock);
669 	(void)rbtree_delete(&ctx->queries, q->node.key);
670 	ctx->num_async--;
671 	context_query_delete(q);
672 	lock_basic_unlock(&ctx->cfglock);
673 
674 	if(!cancelled) {
675 		/* call callback */
676 		int sec = 0;
677 		if(s == sec_status_bogus)
678 			sec = 1;
679 		else if(s == sec_status_secure)
680 			sec = 2;
681 		(*cb)(cb_arg, rcode, (buf?(void*)sldns_buffer_begin(buf):NULL),
682 			(buf?(int)sldns_buffer_limit(buf):0), sec, why_bogus, was_ratelimited);
683 	}
684 }
685 
libworker_attach_mesh(struct ub_ctx * ctx,struct ctx_query * q,int * async_id)686 int libworker_attach_mesh(struct ub_ctx* ctx, struct ctx_query* q,
687 	int* async_id)
688 {
689 	struct libworker* w = ctx->event_worker;
690 	uint16_t qflags, qid;
691 	struct query_info qinfo;
692 	struct edns_data edns;
693 	if(!w)
694 		return UB_INITFAIL;
695 	if(!setup_qinfo_edns(w, q, &qinfo, &edns))
696 		return UB_SYNTAX;
697 	qid = 0;
698 	qflags = BIT_RD;
699 	q->w = w;
700 	/* see if there is a fixed answer */
701 	sldns_buffer_write_u16_at(w->back->udp_buff, 0, qid);
702 	sldns_buffer_write_u16_at(w->back->udp_buff, 2, qflags);
703 	if(local_zones_answer(ctx->local_zones, w->env, &qinfo, &edns,
704 		w->back->udp_buff, w->env->scratch, NULL, NULL, 0, NULL, 0,
705 		NULL, 0, NULL, 0, NULL)) {
706 		regional_free_all(w->env->scratch);
707 		free(qinfo.qname);
708 		libworker_event_done_cb(q, LDNS_RCODE_NOERROR,
709 			w->back->udp_buff, sec_status_insecure, NULL, 0);
710 		return UB_NOERROR;
711 	}
712 	if(ctx->env->auth_zones && auth_zones_answer(ctx->env->auth_zones,
713 		w->env, &qinfo, &edns, NULL, w->back->udp_buff, w->env->scratch)) {
714 		regional_free_all(w->env->scratch);
715 		free(qinfo.qname);
716 		libworker_event_done_cb(q, LDNS_RCODE_NOERROR,
717 			w->back->udp_buff, sec_status_insecure, NULL, 0);
718 		return UB_NOERROR;
719 	}
720 	/* process new query */
721 	if(async_id)
722 		*async_id = q->querynum;
723 	if(!mesh_new_callback(w->env->mesh, &qinfo, qflags, &edns,
724 		w->back->udp_buff, qid, libworker_event_done_cb, q, 0)) {
725 		free(qinfo.qname);
726 		return UB_NOMEM;
727 	}
728 	free(qinfo.qname);
729 	return UB_NOERROR;
730 }
731 
732 /** add result to the bg worker result queue */
733 static void
add_bg_result(struct libworker * w,struct ctx_query * q,sldns_buffer * pkt,int err,char * reason,int was_ratelimited)734 add_bg_result(struct libworker* w, struct ctx_query* q, sldns_buffer* pkt,
735 	int err, char* reason, int was_ratelimited)
736 {
737 	uint8_t* msg = NULL;
738 	uint32_t len = 0;
739 
740 	if(w->want_quit) {
741 		context_query_delete(q);
742 		return;
743 	}
744 	/* serialize and delete unneeded q */
745 	if(w->is_bg_thread) {
746 		lock_basic_lock(&w->ctx->cfglock);
747 		if(reason)
748 			q->res->why_bogus = strdup(reason);
749 		q->res->was_ratelimited = was_ratelimited;
750 		if(pkt) {
751 			q->msg_len = sldns_buffer_remaining(pkt);
752 			q->msg = memdup(sldns_buffer_begin(pkt), q->msg_len);
753 			if(!q->msg) {
754 				msg = context_serialize_answer(q, UB_NOMEM, NULL, &len);
755 			} else {
756 				msg = context_serialize_answer(q, err, NULL, &len);
757 			}
758 		} else {
759 			msg = context_serialize_answer(q, err, NULL, &len);
760 		}
761 		lock_basic_unlock(&w->ctx->cfglock);
762 	} else {
763 		if(reason)
764 			q->res->why_bogus = strdup(reason);
765 		q->res->was_ratelimited = was_ratelimited;
766 		msg = context_serialize_answer(q, err, pkt, &len);
767 		(void)rbtree_delete(&w->ctx->queries, q->node.key);
768 		w->ctx->num_async--;
769 		context_query_delete(q);
770 	}
771 
772 	if(!msg) {
773 		log_err("out of memory for async answer");
774 		return;
775 	}
776 	if(!tube_queue_item(w->ctx->rr_pipe, msg, len)) {
777 		log_err("out of memory for async answer");
778 		return;
779 	}
780 }
781 
782 void
libworker_bg_done_cb(void * arg,int rcode,sldns_buffer * buf,enum sec_status s,char * why_bogus,int was_ratelimited)783 libworker_bg_done_cb(void* arg, int rcode, sldns_buffer* buf, enum sec_status s,
784 	char* why_bogus, int was_ratelimited)
785 {
786 	struct ctx_query* q = (struct ctx_query*)arg;
787 
788 	if(q->cancelled || q->w->back->want_to_quit) {
789 		if(q->w->is_bg_thread) {
790 			/* delete it now */
791 			struct ub_ctx* ctx = q->w->ctx;
792 			lock_basic_lock(&ctx->cfglock);
793 			(void)rbtree_delete(&ctx->queries, q->node.key);
794 			ctx->num_async--;
795 			context_query_delete(q);
796 			lock_basic_unlock(&ctx->cfglock);
797 		}
798 		/* cancelled, do not give answer */
799 		return;
800 	}
801 	q->msg_security = s;
802 	if(!buf) {
803 		buf = q->w->env->scratch_buffer;
804 	}
805 	if(rcode != 0) {
806 		error_encode(buf, rcode, NULL, 0, BIT_RD, NULL);
807 	}
808 	add_bg_result(q->w, q, buf, UB_NOERROR, why_bogus, was_ratelimited);
809 }
810 
811 
812 /** handle new query command for bg worker */
813 static void
handle_newq(struct libworker * w,uint8_t * buf,uint32_t len)814 handle_newq(struct libworker* w, uint8_t* buf, uint32_t len)
815 {
816 	uint16_t qflags, qid;
817 	struct query_info qinfo;
818 	struct edns_data edns;
819 	struct ctx_query* q;
820 	if(w->is_bg_thread) {
821 		lock_basic_lock(&w->ctx->cfglock);
822 		q = context_lookup_new_query(w->ctx, buf, len);
823 		lock_basic_unlock(&w->ctx->cfglock);
824 	} else {
825 		q = context_deserialize_new_query(w->ctx, buf, len);
826 	}
827 	free(buf);
828 	if(!q) {
829 		log_err("failed to deserialize newq");
830 		return;
831 	}
832 	if(!setup_qinfo_edns(w, q, &qinfo, &edns)) {
833 		add_bg_result(w, q, NULL, UB_SYNTAX, NULL, 0);
834 		return;
835 	}
836 	qid = 0;
837 	qflags = BIT_RD;
838 	/* see if there is a fixed answer */
839 	sldns_buffer_write_u16_at(w->back->udp_buff, 0, qid);
840 	sldns_buffer_write_u16_at(w->back->udp_buff, 2, qflags);
841 	if(local_zones_answer(w->ctx->local_zones, w->env, &qinfo, &edns,
842 		w->back->udp_buff, w->env->scratch, NULL, NULL, 0, NULL, 0,
843 		NULL, 0, NULL, 0, NULL)) {
844 		regional_free_all(w->env->scratch);
845 		q->msg_security = sec_status_insecure;
846 		add_bg_result(w, q, w->back->udp_buff, UB_NOERROR, NULL, 0);
847 		free(qinfo.qname);
848 		return;
849 	}
850 	if(w->ctx->env->auth_zones && auth_zones_answer(w->ctx->env->auth_zones,
851 		w->env, &qinfo, &edns, NULL, w->back->udp_buff, w->env->scratch)) {
852 		regional_free_all(w->env->scratch);
853 		q->msg_security = sec_status_insecure;
854 		add_bg_result(w, q, w->back->udp_buff, UB_NOERROR, NULL, 0);
855 		free(qinfo.qname);
856 		return;
857 	}
858 	q->w = w;
859 	/* process new query */
860 	if(!mesh_new_callback(w->env->mesh, &qinfo, qflags, &edns,
861 		w->back->udp_buff, qid, libworker_bg_done_cb, q, 0)) {
862 		add_bg_result(w, q, NULL, UB_NOMEM, NULL, 0);
863 	}
864 	free(qinfo.qname);
865 }
866 
libworker_alloc_cleanup(void * arg)867 void libworker_alloc_cleanup(void* arg)
868 {
869 	struct libworker* w = (struct libworker*)arg;
870 	slabhash_clear(&w->env->rrset_cache->table);
871         slabhash_clear(w->env->msg_cache);
872 }
873 
libworker_send_query(struct query_info * qinfo,uint16_t flags,int dnssec,int want_dnssec,int nocaps,int check_ratelimit,struct sockaddr_storage * addr,socklen_t addrlen,uint8_t * zone,size_t zonelen,int tcp_upstream,int ssl_upstream,char * tls_auth_name,struct module_qstate * q,int * was_ratelimited)874 struct outbound_entry* libworker_send_query(struct query_info* qinfo,
875 	uint16_t flags, int dnssec, int want_dnssec, int nocaps,
876 	int check_ratelimit,
877 	struct sockaddr_storage* addr, socklen_t addrlen, uint8_t* zone,
878 	size_t zonelen, int tcp_upstream, int ssl_upstream, char* tls_auth_name,
879 	struct module_qstate* q, int* was_ratelimited)
880 {
881 	struct libworker* w = (struct libworker*)q->env->worker;
882 	struct outbound_entry* e = (struct outbound_entry*)regional_alloc(
883 		q->region, sizeof(*e));
884 	if(!e)
885 		return NULL;
886 	e->qstate = q;
887 	e->qsent = outnet_serviced_query(w->back, qinfo, flags, dnssec,
888 		want_dnssec, nocaps, check_ratelimit, tcp_upstream, ssl_upstream,
889 		tls_auth_name, addr, addrlen, zone, zonelen, q,
890 		libworker_handle_service_reply, e, w->back->udp_buff, q->env,
891 		was_ratelimited);
892 	if(!e->qsent) {
893 		return NULL;
894 	}
895 	return e;
896 }
897 
898 int
libworker_handle_service_reply(struct comm_point * c,void * arg,int error,struct comm_reply * reply_info)899 libworker_handle_service_reply(struct comm_point* c, void* arg, int error,
900         struct comm_reply* reply_info)
901 {
902 	struct outbound_entry* e = (struct outbound_entry*)arg;
903 	struct libworker* lw = (struct libworker*)e->qstate->env->worker;
904 
905 	if(error != 0) {
906 		mesh_report_reply(lw->env->mesh, e, reply_info, error);
907 		return 0;
908 	}
909 	/* sanity check. */
910 	if(!LDNS_QR_WIRE(sldns_buffer_begin(c->buffer))
911 		|| LDNS_OPCODE_WIRE(sldns_buffer_begin(c->buffer)) !=
912 			LDNS_PACKET_QUERY
913 		|| LDNS_QDCOUNT(sldns_buffer_begin(c->buffer)) > 1) {
914 		/* error becomes timeout for the module as if this reply
915 		 * never arrived. */
916 		mesh_report_reply(lw->env->mesh, e, reply_info,
917 			NETEVENT_TIMEOUT);
918 		return 0;
919 	}
920 	mesh_report_reply(lw->env->mesh,  e, reply_info, NETEVENT_NOERROR);
921 	return 0;
922 }
923 
924 /* --- fake callbacks for fptr_wlist to work --- */
worker_handle_control_cmd(struct tube * ATTR_UNUSED (tube),uint8_t * ATTR_UNUSED (buffer),size_t ATTR_UNUSED (len),int ATTR_UNUSED (error),void * ATTR_UNUSED (arg))925 void worker_handle_control_cmd(struct tube* ATTR_UNUSED(tube),
926 	uint8_t* ATTR_UNUSED(buffer), size_t ATTR_UNUSED(len),
927 	int ATTR_UNUSED(error), void* ATTR_UNUSED(arg))
928 {
929 	log_assert(0);
930 }
931 
worker_handle_request(struct comm_point * ATTR_UNUSED (c),void * ATTR_UNUSED (arg),int ATTR_UNUSED (error),struct comm_reply * ATTR_UNUSED (repinfo))932 int worker_handle_request(struct comm_point* ATTR_UNUSED(c),
933 	void* ATTR_UNUSED(arg), int ATTR_UNUSED(error),
934         struct comm_reply* ATTR_UNUSED(repinfo))
935 {
936 	log_assert(0);
937 	return 0;
938 }
939 
worker_handle_service_reply(struct comm_point * ATTR_UNUSED (c),void * ATTR_UNUSED (arg),int ATTR_UNUSED (error),struct comm_reply * ATTR_UNUSED (reply_info))940 int worker_handle_service_reply(struct comm_point* ATTR_UNUSED(c),
941 	void* ATTR_UNUSED(arg), int ATTR_UNUSED(error),
942         struct comm_reply* ATTR_UNUSED(reply_info))
943 {
944 	log_assert(0);
945 	return 0;
946 }
947 
remote_accept_callback(struct comm_point * ATTR_UNUSED (c),void * ATTR_UNUSED (arg),int ATTR_UNUSED (error),struct comm_reply * ATTR_UNUSED (repinfo))948 int remote_accept_callback(struct comm_point* ATTR_UNUSED(c),
949 	void* ATTR_UNUSED(arg), int ATTR_UNUSED(error),
950         struct comm_reply* ATTR_UNUSED(repinfo))
951 {
952 	log_assert(0);
953 	return 0;
954 }
955 
remote_control_callback(struct comm_point * ATTR_UNUSED (c),void * ATTR_UNUSED (arg),int ATTR_UNUSED (error),struct comm_reply * ATTR_UNUSED (repinfo))956 int remote_control_callback(struct comm_point* ATTR_UNUSED(c),
957 	void* ATTR_UNUSED(arg), int ATTR_UNUSED(error),
958         struct comm_reply* ATTR_UNUSED(repinfo))
959 {
960 	log_assert(0);
961 	return 0;
962 }
963 
worker_sighandler(int ATTR_UNUSED (sig),void * ATTR_UNUSED (arg))964 void worker_sighandler(int ATTR_UNUSED(sig), void* ATTR_UNUSED(arg))
965 {
966 	log_assert(0);
967 }
968 
worker_send_query(struct query_info * ATTR_UNUSED (qinfo),uint16_t ATTR_UNUSED (flags),int ATTR_UNUSED (dnssec),int ATTR_UNUSED (want_dnssec),int ATTR_UNUSED (nocaps),int ATTR_UNUSED (check_ratelimit),struct sockaddr_storage * ATTR_UNUSED (addr),socklen_t ATTR_UNUSED (addrlen),uint8_t * ATTR_UNUSED (zone),size_t ATTR_UNUSED (zonelen),int ATTR_UNUSED (tcp_upstream),int ATTR_UNUSED (ssl_upstream),char * ATTR_UNUSED (tls_auth_name),struct module_qstate * ATTR_UNUSED (q),int * ATTR_UNUSED (was_ratelimited))969 struct outbound_entry* worker_send_query(struct query_info* ATTR_UNUSED(qinfo),
970 	uint16_t ATTR_UNUSED(flags), int ATTR_UNUSED(dnssec),
971 	int ATTR_UNUSED(want_dnssec), int ATTR_UNUSED(nocaps),
972 	int ATTR_UNUSED(check_ratelimit),
973 	struct sockaddr_storage* ATTR_UNUSED(addr), socklen_t ATTR_UNUSED(addrlen),
974 	uint8_t* ATTR_UNUSED(zone), size_t ATTR_UNUSED(zonelen), int ATTR_UNUSED(tcp_upstream),
975 	int ATTR_UNUSED(ssl_upstream), char* ATTR_UNUSED(tls_auth_name),
976 	struct module_qstate* ATTR_UNUSED(q), int* ATTR_UNUSED(was_ratelimited))
977 {
978 	log_assert(0);
979 	return 0;
980 }
981 
982 void
worker_alloc_cleanup(void * ATTR_UNUSED (arg))983 worker_alloc_cleanup(void* ATTR_UNUSED(arg))
984 {
985 	log_assert(0);
986 }
987 
worker_stat_timer_cb(void * ATTR_UNUSED (arg))988 void worker_stat_timer_cb(void* ATTR_UNUSED(arg))
989 {
990 	log_assert(0);
991 }
992 
worker_probe_timer_cb(void * ATTR_UNUSED (arg))993 void worker_probe_timer_cb(void* ATTR_UNUSED(arg))
994 {
995 	log_assert(0);
996 }
997 
worker_start_accept(void * ATTR_UNUSED (arg))998 void worker_start_accept(void* ATTR_UNUSED(arg))
999 {
1000 	log_assert(0);
1001 }
1002 
worker_stop_accept(void * ATTR_UNUSED (arg))1003 void worker_stop_accept(void* ATTR_UNUSED(arg))
1004 {
1005 	log_assert(0);
1006 }
1007 
order_lock_cmp(const void * ATTR_UNUSED (e1),const void * ATTR_UNUSED (e2))1008 int order_lock_cmp(const void* ATTR_UNUSED(e1), const void* ATTR_UNUSED(e2))
1009 {
1010 	log_assert(0);
1011 	return 0;
1012 }
1013 
1014 int
codeline_cmp(const void * ATTR_UNUSED (a),const void * ATTR_UNUSED (b))1015 codeline_cmp(const void* ATTR_UNUSED(a), const void* ATTR_UNUSED(b))
1016 {
1017 	log_assert(0);
1018 	return 0;
1019 }
1020 
replay_var_compare(const void * ATTR_UNUSED (a),const void * ATTR_UNUSED (b))1021 int replay_var_compare(const void* ATTR_UNUSED(a), const void* ATTR_UNUSED(b))
1022 {
1023         log_assert(0);
1024         return 0;
1025 }
1026 
remote_get_opt_ssl(char * ATTR_UNUSED (str),void * ATTR_UNUSED (arg))1027 void remote_get_opt_ssl(char* ATTR_UNUSED(str), void* ATTR_UNUSED(arg))
1028 {
1029         log_assert(0);
1030 }
1031 
1032 #ifdef UB_ON_WINDOWS
1033 void
worker_win_stop_cb(int ATTR_UNUSED (fd),short ATTR_UNUSED (ev),void * ATTR_UNUSED (arg))1034 worker_win_stop_cb(int ATTR_UNUSED(fd), short ATTR_UNUSED(ev), void*
1035         ATTR_UNUSED(arg)) {
1036         log_assert(0);
1037 }
1038 
1039 void
wsvc_cron_cb(void * ATTR_UNUSED (arg))1040 wsvc_cron_cb(void* ATTR_UNUSED(arg))
1041 {
1042         log_assert(0);
1043 }
1044 #endif /* UB_ON_WINDOWS */
1045 
1046 #ifdef USE_DNSTAP
dtio_tap_callback(int ATTR_UNUSED (fd),short ATTR_UNUSED (ev),void * ATTR_UNUSED (arg))1047 void dtio_tap_callback(int ATTR_UNUSED(fd), short ATTR_UNUSED(ev),
1048 	void* ATTR_UNUSED(arg))
1049 {
1050 	log_assert(0);
1051 }
1052 #endif
1053 
1054 #ifdef USE_DNSTAP
dtio_mainfdcallback(int ATTR_UNUSED (fd),short ATTR_UNUSED (ev),void * ATTR_UNUSED (arg))1055 void dtio_mainfdcallback(int ATTR_UNUSED(fd), short ATTR_UNUSED(ev),
1056 	void* ATTR_UNUSED(arg))
1057 {
1058 	log_assert(0);
1059 }
1060 #endif
1061 
1062 #ifdef HAVE_NGTCP2
doq_client_event_cb(int ATTR_UNUSED (fd),short ATTR_UNUSED (ev),void * ATTR_UNUSED (arg))1063 void doq_client_event_cb(int ATTR_UNUSED(fd), short ATTR_UNUSED(ev),
1064 	void* ATTR_UNUSED(arg))
1065 {
1066 	log_assert(0);
1067 }
1068 #endif
1069 
1070 #ifdef HAVE_NGTCP2
doq_client_timer_cb(int ATTR_UNUSED (fd),short ATTR_UNUSED (ev),void * ATTR_UNUSED (arg))1071 void doq_client_timer_cb(int ATTR_UNUSED(fd), short ATTR_UNUSED(ev),
1072 	void* ATTR_UNUSED(arg))
1073 {
1074 	log_assert(0);
1075 }
1076 #endif
1077