/src/unbound/services/cache/dns.c
Line | Count | Source (jump to first uncovered line) |
1 | | /* |
2 | | * services/cache/dns.c - Cache services for DNS using msg and rrset caches. |
3 | | * |
4 | | * Copyright (c) 2007, NLnet Labs. All rights reserved. |
5 | | * |
6 | | * This software is open source. |
7 | | * |
8 | | * Redistribution and use in source and binary forms, with or without |
9 | | * modification, are permitted provided that the following conditions |
10 | | * are met: |
11 | | * |
12 | | * Redistributions of source code must retain the above copyright notice, |
13 | | * this list of conditions and the following disclaimer. |
14 | | * |
15 | | * Redistributions in binary form must reproduce the above copyright notice, |
16 | | * this list of conditions and the following disclaimer in the documentation |
17 | | * and/or other materials provided with the distribution. |
18 | | * |
19 | | * Neither the name of the NLNET LABS nor the names of its contributors may |
20 | | * be used to endorse or promote products derived from this software without |
21 | | * specific prior written permission. |
22 | | * |
23 | | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
24 | | * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
25 | | * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
26 | | * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
27 | | * HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
28 | | * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED |
29 | | * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
30 | | * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF |
31 | | * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING |
32 | | * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS |
33 | | * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
34 | | */ |
35 | | |
36 | | /** |
37 | | * \file |
38 | | * |
39 | | * This file contains the DNS cache. |
40 | | */ |
41 | | #include "config.h" |
42 | | #include "iterator/iter_delegpt.h" |
43 | | #include "iterator/iter_utils.h" |
44 | | #include "validator/val_nsec.h" |
45 | | #include "validator/val_utils.h" |
46 | | #include "services/cache/dns.h" |
47 | | #include "services/cache/rrset.h" |
48 | | #include "util/data/msgparse.h" |
49 | | #include "util/data/msgreply.h" |
50 | | #include "util/data/packed_rrset.h" |
51 | | #include "util/data/dname.h" |
52 | | #include "util/module.h" |
53 | | #include "util/net_help.h" |
54 | | #include "util/regional.h" |
55 | | #include "util/config_file.h" |
56 | | #include "sldns/sbuffer.h" |
57 | | |
58 | | /** store rrsets in the rrset cache. |
59 | | * @param env: module environment with caches. |
60 | | * @param rep: contains list of rrsets to store. |
61 | | * @param now: current time. |
62 | | * @param leeway: during prefetch how much leeway to update TTLs. |
63 | | * This makes rrsets (other than type NS) timeout sooner so they get |
64 | | * updated with a new full TTL. |
65 | | * Type NS does not get this, because it must not be refreshed from the |
66 | | * child domain, but keep counting down properly. |
67 | | * @param pside: if from parentside discovered NS, so that its NS is okay |
68 | | * in a prefetch situation to be updated (without becoming sticky). |
69 | | * @param qrep: update rrsets here if cache is better |
70 | | * @param region: for qrep allocs. |
71 | | * @param qstarttime: time when delegations were looked up, this is perhaps |
72 | | * earlier than the time in now. The time is used to determine if RRsets |
73 | | * of type NS have expired, so that they can only be updated using |
74 | | * lookups of delegation points that did not use them, since they had |
75 | | * expired then. |
76 | | */ |
77 | | static void |
78 | | store_rrsets(struct module_env* env, struct reply_info* rep, time_t now, |
79 | | time_t leeway, int pside, struct reply_info* qrep, |
80 | | struct regional* region, time_t qstarttime) |
81 | 0 | { |
82 | 0 | size_t i; |
83 | 0 | time_t ttl, min_ttl = rep->ttl; |
84 | | /* see if rrset already exists in cache, if not insert it. */ |
85 | 0 | for(i=0; i<rep->rrset_count; i++) { |
86 | 0 | rep->ref[i].key = rep->rrsets[i]; |
87 | 0 | rep->ref[i].id = rep->rrsets[i]->id; |
88 | | /* update ref if it was in the cache */ |
89 | 0 | switch(rrset_cache_update(env->rrset_cache, &rep->ref[i], |
90 | 0 | env->alloc, ((ntohs(rep->ref[i].key->rk.type)== |
91 | 0 | LDNS_RR_TYPE_NS && !pside)?qstarttime:now) + leeway)) { |
92 | 0 | case 0: /* ref unchanged, item inserted */ |
93 | 0 | break; |
94 | 0 | case 2: /* ref updated, cache is superior */ |
95 | 0 | if(region) { |
96 | 0 | struct ub_packed_rrset_key* ck; |
97 | 0 | lock_rw_rdlock(&rep->ref[i].key->entry.lock); |
98 | | /* if deleted rrset, do not copy it */ |
99 | 0 | if(rep->ref[i].key->id == 0 || |
100 | 0 | rep->ref[i].id != rep->ref[i].key->id) |
101 | 0 | ck = NULL; |
102 | 0 | else ck = packed_rrset_copy_region( |
103 | 0 | rep->ref[i].key, region, now); |
104 | 0 | lock_rw_unlock(&rep->ref[i].key->entry.lock); |
105 | 0 | if(ck) { |
106 | | /* use cached copy if memory allows */ |
107 | 0 | qrep->rrsets[i] = ck; |
108 | 0 | } |
109 | 0 | } |
110 | | /* no break: also copy key item */ |
111 | | /* the line below is matched by gcc regex and silences |
112 | | * the fallthrough warning */ |
113 | 0 | ATTR_FALLTHROUGH |
114 | | /* fallthrough */ |
115 | 0 | case 1: /* ref updated, item inserted */ |
116 | 0 | rep->rrsets[i] = rep->ref[i].key; |
117 | | /* ref was updated; make sure the message ttl is |
118 | | * updated to the minimum of the current rrsets. */ |
119 | 0 | lock_rw_rdlock(&rep->ref[i].key->entry.lock); |
120 | | /* if deleted, skip ttl update. */ |
121 | 0 | if(rep->ref[i].key->id != 0 && |
122 | 0 | rep->ref[i].id == rep->ref[i].key->id) { |
123 | 0 | ttl = ((struct packed_rrset_data*) |
124 | 0 | rep->rrsets[i]->entry.data)->ttl; |
125 | 0 | if(ttl < min_ttl) min_ttl = ttl; |
126 | 0 | } |
127 | 0 | lock_rw_unlock(&rep->ref[i].key->entry.lock); |
128 | 0 | } |
129 | 0 | } |
130 | 0 | if(min_ttl < rep->ttl) { |
131 | 0 | rep->ttl = min_ttl; |
132 | 0 | rep->prefetch_ttl = PREFETCH_TTL_CALC(rep->ttl); |
133 | 0 | rep->serve_expired_ttl = rep->ttl + SERVE_EXPIRED_TTL; |
134 | 0 | } |
135 | 0 | } |
136 | | |
137 | | /** delete message from message cache */ |
138 | | void |
139 | | msg_cache_remove(struct module_env* env, uint8_t* qname, size_t qnamelen, |
140 | | uint16_t qtype, uint16_t qclass, uint16_t flags) |
141 | 0 | { |
142 | 0 | struct query_info k; |
143 | 0 | hashvalue_type h; |
144 | |
|
145 | 0 | k.qname = qname; |
146 | 0 | k.qname_len = qnamelen; |
147 | 0 | k.qtype = qtype; |
148 | 0 | k.qclass = qclass; |
149 | 0 | k.local_alias = NULL; |
150 | 0 | h = query_info_hash(&k, flags); |
151 | 0 | slabhash_remove(env->msg_cache, h, &k); |
152 | 0 | } |
153 | | |
154 | | void |
155 | | dns_cache_store_msg(struct module_env* env, struct query_info* qinfo, |
156 | | hashvalue_type hash, struct reply_info* rep, time_t leeway, int pside, |
157 | | struct reply_info* qrep, uint32_t flags, struct regional* region, |
158 | | time_t qstarttime) |
159 | 0 | { |
160 | 0 | struct msgreply_entry* e; |
161 | 0 | time_t ttl = rep->ttl; |
162 | 0 | size_t i; |
163 | | |
164 | | /* store RRsets */ |
165 | 0 | for(i=0; i<rep->rrset_count; i++) { |
166 | 0 | rep->ref[i].key = rep->rrsets[i]; |
167 | 0 | rep->ref[i].id = rep->rrsets[i]->id; |
168 | 0 | } |
169 | | |
170 | | /* there was a reply_info_sortref(rep) here but it seems to be |
171 | | * unnecessary, because the cache gets locked per rrset. */ |
172 | 0 | reply_info_set_ttls(rep, *env->now); |
173 | 0 | store_rrsets(env, rep, *env->now, leeway, pside, qrep, region, |
174 | 0 | qstarttime); |
175 | 0 | if(ttl == 0 && !(flags & DNSCACHE_STORE_ZEROTTL)) { |
176 | | /* we do not store the message, but we did store the RRs, |
177 | | * which could be useful for delegation information */ |
178 | 0 | verbose(VERB_ALGO, "TTL 0: dropped msg from cache"); |
179 | 0 | reply_info_delete(rep, NULL); |
180 | | /* if the message is in the cache, remove that msg, |
181 | | * so that the TTL 0 response can be returned for future |
182 | | * responses (i.e. don't get answered from |
183 | | * cache, but instead go to recursion to get this TTL0 |
184 | | * response). |
185 | | * Possible messages that could be in the cache: |
186 | | * - SERVFAIL |
187 | | * - NXDOMAIN |
188 | | * - NODATA |
189 | | * - an older record that is expired |
190 | | * - an older record that did not yet expire */ |
191 | 0 | msg_cache_remove(env, qinfo->qname, qinfo->qname_len, |
192 | 0 | qinfo->qtype, qinfo->qclass, flags); |
193 | 0 | return; |
194 | 0 | } |
195 | | |
196 | | /* store msg in the cache */ |
197 | 0 | reply_info_sortref(rep); |
198 | 0 | if(!(e = query_info_entrysetup(qinfo, rep, hash))) { |
199 | 0 | log_err("store_msg: malloc failed"); |
200 | 0 | reply_info_delete(rep, NULL); |
201 | 0 | return; |
202 | 0 | } |
203 | 0 | slabhash_insert(env->msg_cache, hash, &e->entry, rep, env->alloc); |
204 | 0 | } |
205 | | |
206 | | /** find closest NS or DNAME and returns the rrset (locked) */ |
207 | | static struct ub_packed_rrset_key* |
208 | | find_closest_of_type(struct module_env* env, uint8_t* qname, size_t qnamelen, |
209 | | uint16_t qclass, time_t now, uint16_t searchtype, int stripfront, |
210 | | int noexpiredabove, uint8_t* expiretop, size_t expiretoplen) |
211 | 0 | { |
212 | 0 | struct ub_packed_rrset_key *rrset; |
213 | 0 | uint8_t lablen; |
214 | |
|
215 | 0 | if(stripfront) { |
216 | | /* strip off so that DNAMEs have strict subdomain match */ |
217 | 0 | lablen = *qname; |
218 | 0 | qname += lablen + 1; |
219 | 0 | qnamelen -= lablen + 1; |
220 | 0 | } |
221 | | |
222 | | /* snip off front part of qname until the type is found */ |
223 | 0 | while(qnamelen > 0) { |
224 | 0 | if((rrset = rrset_cache_lookup(env->rrset_cache, qname, |
225 | 0 | qnamelen, searchtype, qclass, 0, now, 0))) { |
226 | 0 | uint8_t* origqname = qname; |
227 | 0 | size_t origqnamelen = qnamelen; |
228 | 0 | if(!noexpiredabove) |
229 | 0 | return rrset; |
230 | | /* if expiretop set, do not look above it, but |
231 | | * qname is equal, so the just found result is also |
232 | | * the nonexpired above part. */ |
233 | 0 | if(expiretop && qnamelen == expiretoplen && |
234 | 0 | query_dname_compare(qname, expiretop)==0) |
235 | 0 | return rrset; |
236 | | /* check for expiry, but we have to let go of the rrset |
237 | | * for the lock ordering */ |
238 | 0 | lock_rw_unlock(&rrset->entry.lock); |
239 | | /* the rrset_cache_expired_above function always takes |
240 | | * off one label (if qnamelen>0) and returns the final |
241 | | * qname where it searched, so we can continue from |
242 | | * there turning the O N*N search into O N. */ |
243 | 0 | if(!rrset_cache_expired_above(env->rrset_cache, &qname, |
244 | 0 | &qnamelen, searchtype, qclass, now, expiretop, |
245 | 0 | expiretoplen)) { |
246 | | /* we want to return rrset, but it may be |
247 | | * gone from cache, if so, just loop like |
248 | | * it was not in the cache in the first place. |
249 | | */ |
250 | 0 | if((rrset = rrset_cache_lookup(env-> |
251 | 0 | rrset_cache, origqname, origqnamelen, |
252 | 0 | searchtype, qclass, 0, now, 0))) { |
253 | 0 | return rrset; |
254 | 0 | } |
255 | 0 | } |
256 | 0 | log_nametypeclass(VERB_ALGO, "ignoring rrset because expired rrsets exist above it", origqname, searchtype, qclass); |
257 | 0 | continue; |
258 | 0 | } |
259 | | |
260 | | /* snip off front label */ |
261 | 0 | lablen = *qname; |
262 | 0 | qname += lablen + 1; |
263 | 0 | qnamelen -= lablen + 1; |
264 | 0 | } |
265 | 0 | return NULL; |
266 | 0 | } |
267 | | |
268 | | /** add addr to additional section */ |
269 | | static void |
270 | | addr_to_additional(struct ub_packed_rrset_key* rrset, struct regional* region, |
271 | | struct dns_msg* msg, time_t now) |
272 | 0 | { |
273 | 0 | if((msg->rep->rrsets[msg->rep->rrset_count] = |
274 | 0 | packed_rrset_copy_region(rrset, region, now))) { |
275 | 0 | msg->rep->ar_numrrsets++; |
276 | 0 | msg->rep->rrset_count++; |
277 | 0 | } |
278 | 0 | } |
279 | | |
280 | | /** lookup message in message cache */ |
281 | | struct msgreply_entry* |
282 | | msg_cache_lookup(struct module_env* env, uint8_t* qname, size_t qnamelen, |
283 | | uint16_t qtype, uint16_t qclass, uint16_t flags, time_t now, int wr) |
284 | 0 | { |
285 | 0 | struct lruhash_entry* e; |
286 | 0 | struct query_info k; |
287 | 0 | hashvalue_type h; |
288 | |
|
289 | 0 | k.qname = qname; |
290 | 0 | k.qname_len = qnamelen; |
291 | 0 | k.qtype = qtype; |
292 | 0 | k.qclass = qclass; |
293 | 0 | k.local_alias = NULL; |
294 | 0 | h = query_info_hash(&k, flags); |
295 | 0 | e = slabhash_lookup(env->msg_cache, h, &k, wr); |
296 | |
|
297 | 0 | if(!e) return NULL; |
298 | 0 | if( now > ((struct reply_info*)e->data)->ttl ) { |
299 | 0 | lock_rw_unlock(&e->lock); |
300 | 0 | return NULL; |
301 | 0 | } |
302 | 0 | return (struct msgreply_entry*)e->key; |
303 | 0 | } |
304 | | |
305 | | /** find and add A and AAAA records for nameservers in delegpt */ |
306 | | static int |
307 | | find_add_addrs(struct module_env* env, uint16_t qclass, |
308 | | struct regional* region, struct delegpt* dp, time_t now, |
309 | | struct dns_msg** msg) |
310 | 0 | { |
311 | 0 | struct delegpt_ns* ns; |
312 | 0 | struct msgreply_entry* neg; |
313 | 0 | struct ub_packed_rrset_key* akey; |
314 | 0 | for(ns = dp->nslist; ns; ns = ns->next) { |
315 | 0 | akey = rrset_cache_lookup(env->rrset_cache, ns->name, |
316 | 0 | ns->namelen, LDNS_RR_TYPE_A, qclass, 0, now, 0); |
317 | 0 | if(akey) { |
318 | 0 | if(!delegpt_add_rrset_A(dp, region, akey, 0, NULL)) { |
319 | 0 | lock_rw_unlock(&akey->entry.lock); |
320 | 0 | return 0; |
321 | 0 | } |
322 | 0 | if(msg) |
323 | 0 | addr_to_additional(akey, region, *msg, now); |
324 | 0 | lock_rw_unlock(&akey->entry.lock); |
325 | 0 | } else { |
326 | | /* BIT_CD on false because delegpt lookup does |
327 | | * not use dns64 translation */ |
328 | 0 | neg = msg_cache_lookup(env, ns->name, ns->namelen, |
329 | 0 | LDNS_RR_TYPE_A, qclass, 0, now, 0); |
330 | 0 | if(neg) { |
331 | 0 | delegpt_add_neg_msg(dp, neg); |
332 | 0 | lock_rw_unlock(&neg->entry.lock); |
333 | 0 | } |
334 | 0 | } |
335 | 0 | akey = rrset_cache_lookup(env->rrset_cache, ns->name, |
336 | 0 | ns->namelen, LDNS_RR_TYPE_AAAA, qclass, 0, now, 0); |
337 | 0 | if(akey) { |
338 | 0 | if(!delegpt_add_rrset_AAAA(dp, region, akey, 0, NULL)) { |
339 | 0 | lock_rw_unlock(&akey->entry.lock); |
340 | 0 | return 0; |
341 | 0 | } |
342 | 0 | if(msg) |
343 | 0 | addr_to_additional(akey, region, *msg, now); |
344 | 0 | lock_rw_unlock(&akey->entry.lock); |
345 | 0 | } else { |
346 | | /* BIT_CD on false because delegpt lookup does |
347 | | * not use dns64 translation */ |
348 | 0 | neg = msg_cache_lookup(env, ns->name, ns->namelen, |
349 | 0 | LDNS_RR_TYPE_AAAA, qclass, 0, now, 0); |
350 | | /* Because recursion for lookup uses BIT_CD, check |
351 | | * for that so it stops the recursion lookup, if a |
352 | | * negative answer is cached. Because the cache uses |
353 | | * the CD flag for type AAAA. */ |
354 | 0 | if(!neg) |
355 | 0 | neg = msg_cache_lookup(env, ns->name, ns->namelen, |
356 | 0 | LDNS_RR_TYPE_AAAA, qclass, BIT_CD, now, 0); |
357 | 0 | if(neg) { |
358 | 0 | delegpt_add_neg_msg(dp, neg); |
359 | 0 | lock_rw_unlock(&neg->entry.lock); |
360 | 0 | } |
361 | 0 | } |
362 | 0 | } |
363 | 0 | return 1; |
364 | 0 | } |
365 | | |
366 | | /** find and add A and AAAA records for missing nameservers in delegpt */ |
367 | | int |
368 | | cache_fill_missing(struct module_env* env, uint16_t qclass, |
369 | | struct regional* region, struct delegpt* dp, uint32_t flags) |
370 | 0 | { |
371 | 0 | struct delegpt_ns* ns; |
372 | 0 | struct msgreply_entry* neg; |
373 | 0 | struct ub_packed_rrset_key* akey; |
374 | 0 | time_t now = *env->now; |
375 | 0 | for(ns = dp->nslist; ns; ns = ns->next) { |
376 | 0 | if(ns->cache_lookup_count > ITERATOR_NAME_CACHELOOKUP_MAX) |
377 | 0 | continue; |
378 | 0 | ns->cache_lookup_count++; |
379 | 0 | akey = rrset_cache_lookup(env->rrset_cache, ns->name, |
380 | 0 | ns->namelen, LDNS_RR_TYPE_A, qclass, flags, now, 0); |
381 | 0 | if(akey) { |
382 | 0 | if(!delegpt_add_rrset_A(dp, region, akey, ns->lame, |
383 | 0 | NULL)) { |
384 | 0 | lock_rw_unlock(&akey->entry.lock); |
385 | 0 | return 0; |
386 | 0 | } |
387 | 0 | log_nametypeclass(VERB_ALGO, "found in cache", |
388 | 0 | ns->name, LDNS_RR_TYPE_A, qclass); |
389 | 0 | lock_rw_unlock(&akey->entry.lock); |
390 | 0 | } else { |
391 | | /* BIT_CD on false because delegpt lookup does |
392 | | * not use dns64 translation */ |
393 | 0 | neg = msg_cache_lookup(env, ns->name, ns->namelen, |
394 | 0 | LDNS_RR_TYPE_A, qclass, 0, now, 0); |
395 | 0 | if(neg) { |
396 | 0 | delegpt_add_neg_msg(dp, neg); |
397 | 0 | lock_rw_unlock(&neg->entry.lock); |
398 | 0 | } |
399 | 0 | } |
400 | 0 | akey = rrset_cache_lookup(env->rrset_cache, ns->name, |
401 | 0 | ns->namelen, LDNS_RR_TYPE_AAAA, qclass, flags, now, 0); |
402 | 0 | if(akey) { |
403 | 0 | if(!delegpt_add_rrset_AAAA(dp, region, akey, ns->lame, |
404 | 0 | NULL)) { |
405 | 0 | lock_rw_unlock(&akey->entry.lock); |
406 | 0 | return 0; |
407 | 0 | } |
408 | 0 | log_nametypeclass(VERB_ALGO, "found in cache", |
409 | 0 | ns->name, LDNS_RR_TYPE_AAAA, qclass); |
410 | 0 | lock_rw_unlock(&akey->entry.lock); |
411 | 0 | } else { |
412 | | /* BIT_CD on false because delegpt lookup does |
413 | | * not use dns64 translation */ |
414 | 0 | neg = msg_cache_lookup(env, ns->name, ns->namelen, |
415 | 0 | LDNS_RR_TYPE_AAAA, qclass, 0, now, 0); |
416 | | /* Because recursion for lookup uses BIT_CD, check |
417 | | * for that so it stops the recursion lookup, if a |
418 | | * negative answer is cached. Because the cache uses |
419 | | * the CD flag for type AAAA. */ |
420 | 0 | if(!neg) |
421 | 0 | neg = msg_cache_lookup(env, ns->name, ns->namelen, |
422 | 0 | LDNS_RR_TYPE_AAAA, qclass, BIT_CD, now, 0); |
423 | 0 | if(neg) { |
424 | 0 | delegpt_add_neg_msg(dp, neg); |
425 | 0 | lock_rw_unlock(&neg->entry.lock); |
426 | 0 | } |
427 | 0 | } |
428 | 0 | } |
429 | 0 | return 1; |
430 | 0 | } |
431 | | |
432 | | /** find and add DS or NSEC to delegation msg */ |
433 | | static void |
434 | | find_add_ds(struct module_env* env, struct regional* region, |
435 | | struct dns_msg* msg, struct delegpt* dp, time_t now) |
436 | 0 | { |
437 | | /* Lookup the DS or NSEC at the delegation point. */ |
438 | 0 | struct ub_packed_rrset_key* rrset = rrset_cache_lookup( |
439 | 0 | env->rrset_cache, dp->name, dp->namelen, LDNS_RR_TYPE_DS, |
440 | 0 | msg->qinfo.qclass, 0, now, 0); |
441 | 0 | if(!rrset) { |
442 | | /* NOTE: this won't work for alternate NSEC schemes |
443 | | * (opt-in, NSEC3) */ |
444 | 0 | rrset = rrset_cache_lookup(env->rrset_cache, dp->name, |
445 | 0 | dp->namelen, LDNS_RR_TYPE_NSEC, msg->qinfo.qclass, |
446 | 0 | 0, now, 0); |
447 | | /* Note: the PACKED_RRSET_NSEC_AT_APEX flag is not used. |
448 | | * since this is a referral, we need the NSEC at the parent |
449 | | * side of the zone cut, not the NSEC at apex side. */ |
450 | 0 | if(rrset && nsec_has_type(rrset, LDNS_RR_TYPE_DS)) { |
451 | 0 | lock_rw_unlock(&rrset->entry.lock); |
452 | 0 | rrset = NULL; /* discard wrong NSEC */ |
453 | 0 | } |
454 | 0 | } |
455 | 0 | if(rrset) { |
456 | | /* add it to auth section. This is the second rrset. */ |
457 | 0 | if((msg->rep->rrsets[msg->rep->rrset_count] = |
458 | 0 | packed_rrset_copy_region(rrset, region, now))) { |
459 | 0 | msg->rep->ns_numrrsets++; |
460 | 0 | msg->rep->rrset_count++; |
461 | 0 | } |
462 | 0 | lock_rw_unlock(&rrset->entry.lock); |
463 | 0 | } |
464 | 0 | } |
465 | | |
466 | | struct dns_msg* |
467 | | dns_msg_create(uint8_t* qname, size_t qnamelen, uint16_t qtype, |
468 | | uint16_t qclass, struct regional* region, size_t capacity) |
469 | 0 | { |
470 | 0 | struct dns_msg* msg = (struct dns_msg*)regional_alloc(region, |
471 | 0 | sizeof(struct dns_msg)); |
472 | 0 | if(!msg) |
473 | 0 | return NULL; |
474 | 0 | msg->qinfo.qname = regional_alloc_init(region, qname, qnamelen); |
475 | 0 | if(!msg->qinfo.qname) |
476 | 0 | return NULL; |
477 | 0 | msg->qinfo.qname_len = qnamelen; |
478 | 0 | msg->qinfo.qtype = qtype; |
479 | 0 | msg->qinfo.qclass = qclass; |
480 | 0 | msg->qinfo.local_alias = NULL; |
481 | | /* non-packed reply_info, because it needs to grow the array */ |
482 | 0 | msg->rep = (struct reply_info*)regional_alloc_zero(region, |
483 | 0 | sizeof(struct reply_info)-sizeof(struct rrset_ref)); |
484 | 0 | if(!msg->rep) |
485 | 0 | return NULL; |
486 | 0 | if(capacity > RR_COUNT_MAX) |
487 | 0 | return NULL; /* integer overflow protection */ |
488 | 0 | msg->rep->flags = BIT_QR; /* with QR, no AA */ |
489 | 0 | msg->rep->qdcount = 1; |
490 | 0 | msg->rep->reason_bogus = LDNS_EDE_NONE; |
491 | 0 | msg->rep->rrsets = (struct ub_packed_rrset_key**) |
492 | 0 | regional_alloc(region, |
493 | 0 | capacity*sizeof(struct ub_packed_rrset_key*)); |
494 | 0 | if(!msg->rep->rrsets) |
495 | 0 | return NULL; |
496 | 0 | return msg; |
497 | 0 | } |
498 | | |
499 | | int |
500 | | dns_msg_authadd(struct dns_msg* msg, struct regional* region, |
501 | | struct ub_packed_rrset_key* rrset, time_t now) |
502 | 0 | { |
503 | 0 | if(!(msg->rep->rrsets[msg->rep->rrset_count++] = |
504 | 0 | packed_rrset_copy_region(rrset, region, now))) |
505 | 0 | return 0; |
506 | 0 | msg->rep->ns_numrrsets++; |
507 | 0 | return 1; |
508 | 0 | } |
509 | | |
510 | | int |
511 | | dns_msg_ansadd(struct dns_msg* msg, struct regional* region, |
512 | | struct ub_packed_rrset_key* rrset, time_t now) |
513 | 0 | { |
514 | 0 | if(!(msg->rep->rrsets[msg->rep->rrset_count++] = |
515 | 0 | packed_rrset_copy_region(rrset, region, now))) |
516 | 0 | return 0; |
517 | 0 | msg->rep->an_numrrsets++; |
518 | 0 | return 1; |
519 | 0 | } |
520 | | |
521 | | struct delegpt* |
522 | | dns_cache_find_delegation(struct module_env* env, uint8_t* qname, |
523 | | size_t qnamelen, uint16_t qtype, uint16_t qclass, |
524 | | struct regional* region, struct dns_msg** msg, time_t now, |
525 | | int noexpiredabove, uint8_t* expiretop, size_t expiretoplen) |
526 | 0 | { |
527 | | /* try to find closest NS rrset */ |
528 | 0 | struct ub_packed_rrset_key* nskey; |
529 | 0 | struct packed_rrset_data* nsdata; |
530 | 0 | struct delegpt* dp; |
531 | |
|
532 | 0 | nskey = find_closest_of_type(env, qname, qnamelen, qclass, now, |
533 | 0 | LDNS_RR_TYPE_NS, 0, noexpiredabove, expiretop, expiretoplen); |
534 | 0 | if(!nskey) /* hope the caller has hints to prime or something */ |
535 | 0 | return NULL; |
536 | 0 | nsdata = (struct packed_rrset_data*)nskey->entry.data; |
537 | | /* got the NS key, create delegation point */ |
538 | 0 | dp = delegpt_create(region); |
539 | 0 | if(!dp || !delegpt_set_name(dp, region, nskey->rk.dname)) { |
540 | 0 | lock_rw_unlock(&nskey->entry.lock); |
541 | 0 | log_err("find_delegation: out of memory"); |
542 | 0 | return NULL; |
543 | 0 | } |
544 | | /* create referral message */ |
545 | 0 | if(msg) { |
546 | | /* allocate the array to as much as we could need: |
547 | | * NS rrset + DS/NSEC rrset + |
548 | | * A rrset for every NS RR |
549 | | * AAAA rrset for every NS RR |
550 | | */ |
551 | 0 | *msg = dns_msg_create(qname, qnamelen, qtype, qclass, region, |
552 | 0 | 2 + nsdata->count*2); |
553 | 0 | if(!*msg || !dns_msg_authadd(*msg, region, nskey, now)) { |
554 | 0 | lock_rw_unlock(&nskey->entry.lock); |
555 | 0 | log_err("find_delegation: out of memory"); |
556 | 0 | return NULL; |
557 | 0 | } |
558 | 0 | } |
559 | 0 | if(!delegpt_rrset_add_ns(dp, region, nskey, 0)) |
560 | 0 | log_err("find_delegation: addns out of memory"); |
561 | 0 | lock_rw_unlock(&nskey->entry.lock); /* first unlock before next lookup*/ |
562 | | /* find and add DS/NSEC (if any) */ |
563 | 0 | if(msg) |
564 | 0 | find_add_ds(env, region, *msg, dp, now); |
565 | | /* find and add A entries */ |
566 | 0 | if(!find_add_addrs(env, qclass, region, dp, now, msg)) |
567 | 0 | log_err("find_delegation: addrs out of memory"); |
568 | 0 | return dp; |
569 | 0 | } |
570 | | |
571 | | /** allocate dns_msg from query_info and reply_info */ |
572 | | static struct dns_msg* |
573 | | gen_dns_msg(struct regional* region, struct query_info* q, size_t num) |
574 | 0 | { |
575 | 0 | struct dns_msg* msg = (struct dns_msg*)regional_alloc(region, |
576 | 0 | sizeof(struct dns_msg)); |
577 | 0 | if(!msg) |
578 | 0 | return NULL; |
579 | 0 | memcpy(&msg->qinfo, q, sizeof(struct query_info)); |
580 | 0 | msg->qinfo.qname = regional_alloc_init(region, q->qname, q->qname_len); |
581 | 0 | if(!msg->qinfo.qname) |
582 | 0 | return NULL; |
583 | | /* allocate replyinfo struct and rrset key array separately */ |
584 | 0 | msg->rep = (struct reply_info*)regional_alloc(region, |
585 | 0 | sizeof(struct reply_info) - sizeof(struct rrset_ref)); |
586 | 0 | if(!msg->rep) |
587 | 0 | return NULL; |
588 | 0 | msg->rep->reason_bogus = LDNS_EDE_NONE; |
589 | 0 | msg->rep->reason_bogus_str = NULL; |
590 | 0 | if(num > RR_COUNT_MAX) |
591 | 0 | return NULL; /* integer overflow protection */ |
592 | 0 | msg->rep->rrsets = (struct ub_packed_rrset_key**) |
593 | 0 | regional_alloc(region, |
594 | 0 | num * sizeof(struct ub_packed_rrset_key*)); |
595 | 0 | if(!msg->rep->rrsets) |
596 | 0 | return NULL; |
597 | 0 | return msg; |
598 | 0 | } |
599 | | |
600 | | struct dns_msg* |
601 | | tomsg(struct module_env* env, struct query_info* q, struct reply_info* r, |
602 | | struct regional* region, time_t now, int allow_expired, |
603 | | struct regional* scratch) |
604 | 0 | { |
605 | 0 | struct dns_msg* msg; |
606 | 0 | size_t i; |
607 | 0 | int is_expired = 0; |
608 | 0 | time_t now_control = now; |
609 | 0 | if(now > r->ttl) { |
610 | | /* Check if we are allowed to serve expired */ |
611 | 0 | if(!allow_expired || !reply_info_can_answer_expired(r, now)) |
612 | 0 | return NULL; |
613 | | /* Change the current time so we can pass the below TTL checks when |
614 | | * serving expired data. */ |
615 | 0 | now_control = r->ttl - env->cfg->serve_expired_reply_ttl; |
616 | 0 | is_expired = 1; |
617 | 0 | } |
618 | | |
619 | 0 | msg = gen_dns_msg(region, q, r->rrset_count); |
620 | 0 | if(!msg) return NULL; |
621 | 0 | msg->rep->flags = r->flags; |
622 | 0 | msg->rep->qdcount = r->qdcount; |
623 | 0 | msg->rep->ttl = is_expired |
624 | 0 | ?SERVE_EXPIRED_REPLY_TTL |
625 | 0 | :r->ttl - now; |
626 | 0 | if(r->prefetch_ttl > now) |
627 | 0 | msg->rep->prefetch_ttl = r->prefetch_ttl - now; |
628 | 0 | else |
629 | 0 | msg->rep->prefetch_ttl = PREFETCH_TTL_CALC(msg->rep->ttl); |
630 | 0 | msg->rep->serve_expired_ttl = msg->rep->ttl + SERVE_EXPIRED_TTL; |
631 | 0 | msg->rep->serve_expired_norec_ttl = 0; |
632 | 0 | msg->rep->security = r->security; |
633 | 0 | msg->rep->an_numrrsets = r->an_numrrsets; |
634 | 0 | msg->rep->ns_numrrsets = r->ns_numrrsets; |
635 | 0 | msg->rep->ar_numrrsets = r->ar_numrrsets; |
636 | 0 | msg->rep->rrset_count = r->rrset_count; |
637 | 0 | msg->rep->authoritative = r->authoritative; |
638 | 0 | msg->rep->reason_bogus = r->reason_bogus; |
639 | 0 | if(r->reason_bogus_str) { |
640 | 0 | msg->rep->reason_bogus_str = regional_strdup(region, r->reason_bogus_str); |
641 | 0 | } |
642 | |
|
643 | 0 | if(!rrset_array_lock(r->ref, r->rrset_count, now_control)) { |
644 | 0 | return NULL; |
645 | 0 | } |
646 | 0 | if(r->an_numrrsets > 0 && (r->rrsets[0]->rk.type == htons( |
647 | 0 | LDNS_RR_TYPE_CNAME) || r->rrsets[0]->rk.type == htons( |
648 | 0 | LDNS_RR_TYPE_DNAME)) && !reply_check_cname_chain(q, r)) { |
649 | | /* cname chain is now invalid, reconstruct msg */ |
650 | 0 | rrset_array_unlock(r->ref, r->rrset_count); |
651 | 0 | return NULL; |
652 | 0 | } |
653 | 0 | if(r->security == sec_status_secure && !reply_all_rrsets_secure(r)) { |
654 | | /* message rrsets have changed status, revalidate */ |
655 | 0 | rrset_array_unlock(r->ref, r->rrset_count); |
656 | 0 | return NULL; |
657 | 0 | } |
658 | 0 | for(i=0; i<msg->rep->rrset_count; i++) { |
659 | 0 | msg->rep->rrsets[i] = packed_rrset_copy_region(r->rrsets[i], |
660 | 0 | region, now); |
661 | 0 | if(!msg->rep->rrsets[i]) { |
662 | 0 | rrset_array_unlock(r->ref, r->rrset_count); |
663 | 0 | return NULL; |
664 | 0 | } |
665 | 0 | } |
666 | 0 | if(env) |
667 | 0 | rrset_array_unlock_touch(env->rrset_cache, scratch, r->ref, |
668 | 0 | r->rrset_count); |
669 | 0 | else |
670 | 0 | rrset_array_unlock(r->ref, r->rrset_count); |
671 | 0 | return msg; |
672 | 0 | } |
673 | | |
674 | | struct dns_msg* |
675 | | dns_msg_deepcopy_region(struct dns_msg* origin, struct regional* region) |
676 | 0 | { |
677 | 0 | size_t i; |
678 | 0 | struct dns_msg* res = NULL; |
679 | 0 | res = gen_dns_msg(region, &origin->qinfo, origin->rep->rrset_count); |
680 | 0 | if(!res) return NULL; |
681 | 0 | *res->rep = *origin->rep; |
682 | 0 | if(origin->rep->reason_bogus_str) { |
683 | 0 | res->rep->reason_bogus_str = regional_strdup(region, |
684 | 0 | origin->rep->reason_bogus_str); |
685 | 0 | } |
686 | 0 | for(i=0; i<res->rep->rrset_count; i++) { |
687 | 0 | res->rep->rrsets[i] = packed_rrset_copy_region( |
688 | 0 | origin->rep->rrsets[i], region, 0); |
689 | 0 | if(!res->rep->rrsets[i]) { |
690 | 0 | return NULL; |
691 | 0 | } |
692 | 0 | } |
693 | 0 | return res; |
694 | 0 | } |
695 | | |
696 | | /** synthesize RRset-only response from cached RRset item */ |
697 | | static struct dns_msg* |
698 | | rrset_msg(struct ub_packed_rrset_key* rrset, struct regional* region, |
699 | | time_t now, struct query_info* q) |
700 | 0 | { |
701 | 0 | struct dns_msg* msg; |
702 | 0 | struct packed_rrset_data* d = (struct packed_rrset_data*) |
703 | 0 | rrset->entry.data; |
704 | 0 | if(now > d->ttl) |
705 | 0 | return NULL; |
706 | 0 | msg = gen_dns_msg(region, q, 1); /* only the CNAME (or other) RRset */ |
707 | 0 | if(!msg) |
708 | 0 | return NULL; |
709 | 0 | msg->rep->flags = BIT_QR; /* reply, no AA, no error */ |
710 | 0 | msg->rep->authoritative = 0; /* reply stored in cache can't be authoritative */ |
711 | 0 | msg->rep->qdcount = 1; |
712 | 0 | msg->rep->ttl = d->ttl - now; |
713 | 0 | msg->rep->prefetch_ttl = PREFETCH_TTL_CALC(msg->rep->ttl); |
714 | 0 | msg->rep->serve_expired_ttl = msg->rep->ttl + SERVE_EXPIRED_TTL; |
715 | 0 | msg->rep->serve_expired_norec_ttl = 0; |
716 | 0 | msg->rep->security = sec_status_unchecked; |
717 | 0 | msg->rep->an_numrrsets = 1; |
718 | 0 | msg->rep->ns_numrrsets = 0; |
719 | 0 | msg->rep->ar_numrrsets = 0; |
720 | 0 | msg->rep->rrset_count = 1; |
721 | 0 | msg->rep->reason_bogus = LDNS_EDE_NONE; |
722 | 0 | msg->rep->rrsets[0] = packed_rrset_copy_region(rrset, region, now); |
723 | 0 | if(!msg->rep->rrsets[0]) /* copy CNAME */ |
724 | 0 | return NULL; |
725 | 0 | return msg; |
726 | 0 | } |
727 | | |
728 | | /** synthesize DNAME+CNAME response from cached DNAME item */ |
729 | | static struct dns_msg* |
730 | | synth_dname_msg(struct ub_packed_rrset_key* rrset, struct regional* region, |
731 | | time_t now, struct query_info* q, enum sec_status* sec_status) |
732 | 0 | { |
733 | 0 | struct dns_msg* msg; |
734 | 0 | struct ub_packed_rrset_key* ck; |
735 | 0 | struct packed_rrset_data* newd, *d = (struct packed_rrset_data*) |
736 | 0 | rrset->entry.data; |
737 | 0 | uint8_t* newname, *dtarg = NULL; |
738 | 0 | size_t newlen, dtarglen; |
739 | 0 | if(now > d->ttl) |
740 | 0 | return NULL; |
741 | | /* only allow validated (with DNSSEC) DNAMEs used from cache |
742 | | * for insecure DNAMEs, query again. */ |
743 | 0 | *sec_status = d->security; |
744 | | /* return sec status, so the status of the CNAME can be checked |
745 | | * by the calling routine. */ |
746 | 0 | msg = gen_dns_msg(region, q, 2); /* DNAME + CNAME RRset */ |
747 | 0 | if(!msg) |
748 | 0 | return NULL; |
749 | 0 | msg->rep->flags = BIT_QR; /* reply, no AA, no error */ |
750 | 0 | msg->rep->authoritative = 0; /* reply stored in cache can't be authoritative */ |
751 | 0 | msg->rep->qdcount = 1; |
752 | 0 | msg->rep->ttl = d->ttl - now; |
753 | 0 | msg->rep->prefetch_ttl = PREFETCH_TTL_CALC(msg->rep->ttl); |
754 | 0 | msg->rep->serve_expired_ttl = msg->rep->ttl + SERVE_EXPIRED_TTL; |
755 | 0 | msg->rep->serve_expired_norec_ttl = 0; |
756 | 0 | msg->rep->security = sec_status_unchecked; |
757 | 0 | msg->rep->an_numrrsets = 1; |
758 | 0 | msg->rep->ns_numrrsets = 0; |
759 | 0 | msg->rep->ar_numrrsets = 0; |
760 | 0 | msg->rep->rrset_count = 1; |
761 | 0 | msg->rep->reason_bogus = LDNS_EDE_NONE; |
762 | 0 | msg->rep->rrsets[0] = packed_rrset_copy_region(rrset, region, now); |
763 | 0 | if(!msg->rep->rrsets[0]) /* copy DNAME */ |
764 | 0 | return NULL; |
765 | | /* synth CNAME rrset */ |
766 | 0 | get_cname_target(rrset, &dtarg, &dtarglen); |
767 | 0 | if(!dtarg) |
768 | 0 | return NULL; |
769 | 0 | newlen = q->qname_len + dtarglen - rrset->rk.dname_len; |
770 | 0 | if(newlen > LDNS_MAX_DOMAINLEN) { |
771 | 0 | msg->rep->flags |= LDNS_RCODE_YXDOMAIN; |
772 | 0 | return msg; |
773 | 0 | } |
774 | 0 | newname = (uint8_t*)regional_alloc(region, newlen); |
775 | 0 | if(!newname) |
776 | 0 | return NULL; |
777 | | /* new name is concatenation of qname front (without DNAME owner) |
778 | | * and DNAME target name */ |
779 | 0 | memcpy(newname, q->qname, q->qname_len-rrset->rk.dname_len); |
780 | 0 | memmove(newname+(q->qname_len-rrset->rk.dname_len), dtarg, dtarglen); |
781 | | /* create rest of CNAME rrset */ |
782 | 0 | ck = (struct ub_packed_rrset_key*)regional_alloc(region, |
783 | 0 | sizeof(struct ub_packed_rrset_key)); |
784 | 0 | if(!ck) |
785 | 0 | return NULL; |
786 | 0 | memset(&ck->entry, 0, sizeof(ck->entry)); |
787 | 0 | msg->rep->rrsets[1] = ck; |
788 | 0 | ck->entry.key = ck; |
789 | 0 | ck->rk.type = htons(LDNS_RR_TYPE_CNAME); |
790 | 0 | ck->rk.rrset_class = rrset->rk.rrset_class; |
791 | 0 | ck->rk.flags = 0; |
792 | 0 | ck->rk.dname = regional_alloc_init(region, q->qname, q->qname_len); |
793 | 0 | if(!ck->rk.dname) |
794 | 0 | return NULL; |
795 | 0 | ck->rk.dname_len = q->qname_len; |
796 | 0 | ck->entry.hash = rrset_key_hash(&ck->rk); |
797 | 0 | newd = (struct packed_rrset_data*)regional_alloc_zero(region, |
798 | 0 | sizeof(struct packed_rrset_data) + sizeof(size_t) + |
799 | 0 | sizeof(uint8_t*) + sizeof(time_t) + sizeof(uint16_t) |
800 | 0 | + newlen); |
801 | 0 | if(!newd) |
802 | 0 | return NULL; |
803 | 0 | ck->entry.data = newd; |
804 | 0 | newd->ttl = d->ttl - now; /* RFC6672: synth CNAME TTL == DNAME TTL */ |
805 | 0 | newd->count = 1; |
806 | 0 | newd->rrsig_count = 0; |
807 | 0 | newd->trust = rrset_trust_ans_noAA; |
808 | 0 | newd->rr_len = (size_t*)((uint8_t*)newd + |
809 | 0 | sizeof(struct packed_rrset_data)); |
810 | 0 | newd->rr_len[0] = newlen + sizeof(uint16_t); |
811 | 0 | packed_rrset_ptr_fixup(newd); |
812 | 0 | newd->rr_ttl[0] = newd->ttl; |
813 | 0 | msg->rep->ttl = newd->ttl; |
814 | 0 | msg->rep->prefetch_ttl = PREFETCH_TTL_CALC(newd->ttl); |
815 | 0 | msg->rep->serve_expired_ttl = newd->ttl + SERVE_EXPIRED_TTL; |
816 | 0 | sldns_write_uint16(newd->rr_data[0], newlen); |
817 | 0 | memmove(newd->rr_data[0] + sizeof(uint16_t), newname, newlen); |
818 | 0 | msg->rep->an_numrrsets ++; |
819 | 0 | msg->rep->rrset_count ++; |
820 | 0 | return msg; |
821 | 0 | } |
822 | | |
823 | | /** Fill TYPE_ANY response with some data from cache */ |
824 | | static struct dns_msg* |
825 | | fill_any(struct module_env* env, |
826 | | uint8_t* qname, size_t qnamelen, uint16_t qtype, uint16_t qclass, |
827 | | struct regional* region) |
828 | 0 | { |
829 | 0 | time_t now = *env->now; |
830 | 0 | struct dns_msg* msg = NULL; |
831 | 0 | uint16_t lookup[] = {LDNS_RR_TYPE_A, LDNS_RR_TYPE_AAAA, |
832 | 0 | LDNS_RR_TYPE_MX, LDNS_RR_TYPE_SOA, LDNS_RR_TYPE_NS, |
833 | 0 | LDNS_RR_TYPE_DNAME, 0}; |
834 | 0 | int i, num=6; /* number of RR types to look up */ |
835 | 0 | log_assert(lookup[num] == 0); |
836 | |
|
837 | 0 | if(env->cfg->deny_any) { |
838 | | /* return empty message */ |
839 | 0 | msg = dns_msg_create(qname, qnamelen, qtype, qclass, |
840 | 0 | region, 0); |
841 | 0 | if(!msg) { |
842 | 0 | return NULL; |
843 | 0 | } |
844 | | /* set NOTIMPL for RFC 8482 */ |
845 | 0 | msg->rep->flags |= LDNS_RCODE_NOTIMPL; |
846 | 0 | msg->rep->security = sec_status_indeterminate; |
847 | 0 | return msg; |
848 | 0 | } |
849 | | |
850 | 0 | for(i=0; i<num; i++) { |
851 | | /* look up this RR for inclusion in type ANY response */ |
852 | 0 | struct ub_packed_rrset_key* rrset = rrset_cache_lookup( |
853 | 0 | env->rrset_cache, qname, qnamelen, lookup[i], |
854 | 0 | qclass, 0, now, 0); |
855 | 0 | struct packed_rrset_data *d; |
856 | 0 | if(!rrset) |
857 | 0 | continue; |
858 | | |
859 | | /* only if rrset from answer section */ |
860 | 0 | d = (struct packed_rrset_data*)rrset->entry.data; |
861 | 0 | if(d->trust == rrset_trust_add_noAA || |
862 | 0 | d->trust == rrset_trust_auth_noAA || |
863 | 0 | d->trust == rrset_trust_add_AA || |
864 | 0 | d->trust == rrset_trust_auth_AA) { |
865 | 0 | lock_rw_unlock(&rrset->entry.lock); |
866 | 0 | continue; |
867 | 0 | } |
868 | | |
869 | | /* create msg if none */ |
870 | 0 | if(!msg) { |
871 | 0 | msg = dns_msg_create(qname, qnamelen, qtype, qclass, |
872 | 0 | region, (size_t)(num-i)); |
873 | 0 | if(!msg) { |
874 | 0 | lock_rw_unlock(&rrset->entry.lock); |
875 | 0 | return NULL; |
876 | 0 | } |
877 | 0 | } |
878 | | |
879 | | /* add RRset to response */ |
880 | 0 | if(!dns_msg_ansadd(msg, region, rrset, now)) { |
881 | 0 | lock_rw_unlock(&rrset->entry.lock); |
882 | 0 | return NULL; |
883 | 0 | } |
884 | 0 | lock_rw_unlock(&rrset->entry.lock); |
885 | 0 | } |
886 | 0 | return msg; |
887 | 0 | } |
888 | | |
889 | | struct dns_msg* |
890 | | dns_cache_lookup(struct module_env* env, |
891 | | uint8_t* qname, size_t qnamelen, uint16_t qtype, uint16_t qclass, |
892 | | uint16_t flags, struct regional* region, struct regional* scratch, |
893 | | int no_partial, uint8_t* dpname, size_t dpnamelen) |
894 | 0 | { |
895 | 0 | struct lruhash_entry* e; |
896 | 0 | struct query_info k; |
897 | 0 | hashvalue_type h; |
898 | 0 | time_t now = *env->now; |
899 | 0 | struct ub_packed_rrset_key* rrset; |
900 | | |
901 | | /* lookup first, this has both NXdomains and ANSWER responses */ |
902 | 0 | k.qname = qname; |
903 | 0 | k.qname_len = qnamelen; |
904 | 0 | k.qtype = qtype; |
905 | 0 | k.qclass = qclass; |
906 | 0 | k.local_alias = NULL; |
907 | 0 | h = query_info_hash(&k, flags); |
908 | 0 | e = slabhash_lookup(env->msg_cache, h, &k, 0); |
909 | 0 | if(e) { |
910 | 0 | struct msgreply_entry* key = (struct msgreply_entry*)e->key; |
911 | 0 | struct reply_info* data = (struct reply_info*)e->data; |
912 | 0 | struct dns_msg* msg = tomsg(env, &key->key, data, region, now, 0, |
913 | 0 | scratch); |
914 | 0 | if(msg) { |
915 | 0 | lock_rw_unlock(&e->lock); |
916 | 0 | return msg; |
917 | 0 | } |
918 | | /* could be msg==NULL; due to TTL or not all rrsets available */ |
919 | 0 | lock_rw_unlock(&e->lock); |
920 | 0 | } |
921 | | |
922 | | /* see if a DNAME exists. Checked for first, to enforce that DNAMEs |
923 | | * are more important, the CNAME is resynthesized and thus |
924 | | * consistent with the DNAME */ |
925 | 0 | if(!no_partial && |
926 | 0 | (rrset=find_closest_of_type(env, qname, qnamelen, qclass, now, |
927 | 0 | LDNS_RR_TYPE_DNAME, 1, 0, NULL, 0))) { |
928 | | /* synthesize a DNAME+CNAME message based on this */ |
929 | 0 | enum sec_status sec_status = sec_status_unchecked; |
930 | 0 | struct dns_msg* msg = synth_dname_msg(rrset, region, now, &k, |
931 | 0 | &sec_status); |
932 | 0 | if(msg) { |
933 | 0 | struct ub_packed_rrset_key* cname_rrset; |
934 | 0 | lock_rw_unlock(&rrset->entry.lock); |
935 | | /* now, after unlocking the DNAME rrset lock, |
936 | | * check the sec_status, and see if we need to look |
937 | | * up the CNAME record associated before it can |
938 | | * be used */ |
939 | | /* normally, only secure DNAMEs allowed from cache*/ |
940 | 0 | if(sec_status == sec_status_secure) |
941 | 0 | return msg; |
942 | | /* but if we have a CNAME cached with this name, then we |
943 | | * have previously already allowed this name to pass. |
944 | | * the next cache lookup is going to fetch that CNAME itself, |
945 | | * but it is better to have the (unsigned)DNAME + CNAME in |
946 | | * that case */ |
947 | 0 | cname_rrset = rrset_cache_lookup( |
948 | 0 | env->rrset_cache, qname, qnamelen, |
949 | 0 | LDNS_RR_TYPE_CNAME, qclass, 0, now, 0); |
950 | 0 | if(cname_rrset) { |
951 | | /* CNAME already synthesized by |
952 | | * synth_dname_msg routine, so we can |
953 | | * straight up return the msg */ |
954 | 0 | lock_rw_unlock(&cname_rrset->entry.lock); |
955 | 0 | return msg; |
956 | 0 | } |
957 | 0 | } else { |
958 | 0 | lock_rw_unlock(&rrset->entry.lock); |
959 | 0 | } |
960 | 0 | } |
961 | | |
962 | | /* see if we have CNAME for this domain, |
963 | | * but not for DS records (which are part of the parent) */ |
964 | 0 | if(!no_partial && qtype != LDNS_RR_TYPE_DS && |
965 | 0 | (rrset=rrset_cache_lookup(env->rrset_cache, qname, qnamelen, |
966 | 0 | LDNS_RR_TYPE_CNAME, qclass, 0, now, 0))) { |
967 | 0 | uint8_t* wc = NULL; |
968 | 0 | size_t wl; |
969 | | /* if the rrset is not a wildcard expansion, with wcname */ |
970 | | /* because, if we return that CNAME rrset on its own, it is |
971 | | * missing the NSEC or NSEC3 proof */ |
972 | 0 | if(!(val_rrset_wildcard(rrset, &wc, &wl) && wc != NULL)) { |
973 | 0 | struct dns_msg* msg = rrset_msg(rrset, region, now, &k); |
974 | 0 | if(msg) { |
975 | 0 | lock_rw_unlock(&rrset->entry.lock); |
976 | 0 | return msg; |
977 | 0 | } |
978 | 0 | } |
979 | 0 | lock_rw_unlock(&rrset->entry.lock); |
980 | 0 | } |
981 | | |
982 | | /* construct DS, DNSKEY messages from rrset cache. */ |
983 | 0 | if((qtype == LDNS_RR_TYPE_DS || qtype == LDNS_RR_TYPE_DNSKEY) && |
984 | 0 | (rrset=rrset_cache_lookup(env->rrset_cache, qname, qnamelen, |
985 | 0 | qtype, qclass, 0, now, 0))) { |
986 | | /* if the rrset is from the additional section, and the |
987 | | * signatures have fallen off, then do not synthesize a msg |
988 | | * instead, allow a full query for signed results to happen. |
989 | | * Forego all rrset data from additional section, because |
990 | | * some signatures may not be present and cause validation |
991 | | * failure. |
992 | | */ |
993 | 0 | struct packed_rrset_data *d = (struct packed_rrset_data*) |
994 | 0 | rrset->entry.data; |
995 | 0 | if(d->trust != rrset_trust_add_noAA && |
996 | 0 | d->trust != rrset_trust_add_AA && |
997 | 0 | (qtype == LDNS_RR_TYPE_DS || |
998 | 0 | (d->trust != rrset_trust_auth_noAA |
999 | 0 | && d->trust != rrset_trust_auth_AA) )) { |
1000 | 0 | struct dns_msg* msg = rrset_msg(rrset, region, now, &k); |
1001 | 0 | if(msg) { |
1002 | 0 | lock_rw_unlock(&rrset->entry.lock); |
1003 | 0 | return msg; |
1004 | 0 | } |
1005 | 0 | } |
1006 | 0 | lock_rw_unlock(&rrset->entry.lock); |
1007 | 0 | } |
1008 | | |
1009 | | /* stop downwards cache search on NXDOMAIN. |
1010 | | * Empty nonterminals are NOERROR, so an NXDOMAIN for foo |
1011 | | * means bla.foo also does not exist. The DNSSEC proofs are |
1012 | | * the same. We search upwards for NXDOMAINs. */ |
1013 | 0 | if(env->cfg->harden_below_nxdomain) { |
1014 | 0 | while(!dname_is_root(k.qname)) { |
1015 | 0 | if(dpname && dpnamelen |
1016 | 0 | && !dname_subdomain_c(k.qname, dpname)) |
1017 | 0 | break; /* no synth nxdomain above the stub */ |
1018 | 0 | dname_remove_label(&k.qname, &k.qname_len); |
1019 | 0 | h = query_info_hash(&k, flags); |
1020 | 0 | e = slabhash_lookup(env->msg_cache, h, &k, 0); |
1021 | 0 | if(!e && k.qtype != LDNS_RR_TYPE_A && |
1022 | 0 | env->cfg->qname_minimisation) { |
1023 | 0 | k.qtype = LDNS_RR_TYPE_A; |
1024 | 0 | h = query_info_hash(&k, flags); |
1025 | 0 | e = slabhash_lookup(env->msg_cache, h, &k, 0); |
1026 | 0 | } |
1027 | 0 | if(e) { |
1028 | 0 | struct reply_info* data = (struct reply_info*)e->data; |
1029 | 0 | struct dns_msg* msg; |
1030 | 0 | if(FLAGS_GET_RCODE(data->flags) == LDNS_RCODE_NXDOMAIN |
1031 | 0 | && data->security == sec_status_secure |
1032 | 0 | && (data->an_numrrsets == 0 || |
1033 | 0 | ntohs(data->rrsets[0]->rk.type) != LDNS_RR_TYPE_CNAME) |
1034 | 0 | && (msg=tomsg(env, &k, data, region, now, 0, scratch))) { |
1035 | 0 | lock_rw_unlock(&e->lock); |
1036 | 0 | msg->qinfo.qname=qname; |
1037 | 0 | msg->qinfo.qname_len=qnamelen; |
1038 | | /* check that DNSSEC really works out */ |
1039 | 0 | msg->rep->security = sec_status_unchecked; |
1040 | 0 | iter_scrub_nxdomain(msg); |
1041 | 0 | return msg; |
1042 | 0 | } |
1043 | 0 | lock_rw_unlock(&e->lock); |
1044 | 0 | } |
1045 | 0 | k.qtype = qtype; |
1046 | 0 | } |
1047 | 0 | } |
1048 | | |
1049 | | /* fill common RR types for ANY response to avoid requery */ |
1050 | 0 | if(qtype == LDNS_RR_TYPE_ANY) { |
1051 | 0 | return fill_any(env, qname, qnamelen, qtype, qclass, region); |
1052 | 0 | } |
1053 | | |
1054 | 0 | return NULL; |
1055 | 0 | } |
1056 | | |
1057 | | int |
1058 | | dns_cache_store(struct module_env* env, struct query_info* msgqinf, |
1059 | | struct reply_info* msgrep, int is_referral, time_t leeway, int pside, |
1060 | | struct regional* region, uint32_t flags, time_t qstarttime, |
1061 | | int is_valrec) |
1062 | 0 | { |
1063 | 0 | struct reply_info* rep = NULL; |
1064 | 0 | if(SERVE_EXPIRED) { |
1065 | | /* We are serving expired records. Before caching, check if a |
1066 | | * useful expired record exists. */ |
1067 | 0 | struct msgreply_entry* e = msg_cache_lookup(env, |
1068 | 0 | msgqinf->qname, msgqinf->qname_len, msgqinf->qtype, |
1069 | 0 | msgqinf->qclass, flags, 0, 1); |
1070 | 0 | if(e) { |
1071 | 0 | struct reply_info* cached = e->entry.data; |
1072 | 0 | if(cached->ttl < *env->now |
1073 | 0 | && reply_info_could_use_expired(cached, *env->now) |
1074 | | /* If we are validating make sure only |
1075 | | * validating modules can update such messages. |
1076 | | * In that case don't cache it and let a |
1077 | | * subsequent module handle the caching. For |
1078 | | * example, the iterator should not replace an |
1079 | | * expired secure answer with a fresh unchecked |
1080 | | * one and let the validator manage caching. */ |
1081 | 0 | && cached->security != sec_status_bogus |
1082 | 0 | && (env->need_to_validate && |
1083 | 0 | msgrep->security == sec_status_unchecked) |
1084 | | /* Exceptions to that rule are: |
1085 | | * o recursions that don't need validation but |
1086 | | * need to update the cache for coherence |
1087 | | * (delegation information while iterating, |
1088 | | * DNSKEY and DS lookups from validator) |
1089 | | * o explicit RRSIG queries that are not |
1090 | | * validated. */ |
1091 | 0 | && !is_valrec |
1092 | 0 | && msgqinf->qtype != LDNS_RR_TYPE_RRSIG) { |
1093 | 0 | if((int)FLAGS_GET_RCODE(msgrep->flags) != |
1094 | 0 | LDNS_RCODE_NOERROR && |
1095 | 0 | (int)FLAGS_GET_RCODE(msgrep->flags) != |
1096 | 0 | LDNS_RCODE_NXDOMAIN) { |
1097 | | /* The current response has an |
1098 | | * erroneous rcode. Adjust norec time |
1099 | | * so that additional lookups are not |
1100 | | * performed for some time. */ |
1101 | 0 | verbose(VERB_ALGO, "set " |
1102 | 0 | "serve-expired-norec-ttl for " |
1103 | 0 | "response in cache"); |
1104 | 0 | cached->serve_expired_norec_ttl = |
1105 | 0 | NORR_TTL + *env->now; |
1106 | 0 | if(env->cfg->serve_expired_ttl_reset && |
1107 | 0 | cached->serve_expired_ttl |
1108 | 0 | < *env->now + |
1109 | 0 | env->cfg->serve_expired_ttl) { |
1110 | | /* Reset serve-expired-ttl for |
1111 | | * valid response in cache. */ |
1112 | 0 | verbose(VERB_ALGO, "reset " |
1113 | 0 | "serve-expired-ttl " |
1114 | 0 | "for response in cache"); |
1115 | 0 | cached->serve_expired_ttl = |
1116 | 0 | *env->now + |
1117 | 0 | env->cfg->serve_expired_ttl; |
1118 | 0 | } |
1119 | 0 | } |
1120 | 0 | verbose(VERB_ALGO, "a validated expired entry " |
1121 | 0 | "could be overwritten, skip caching " |
1122 | 0 | "the new message at this stage"); |
1123 | 0 | lock_rw_unlock(&e->entry.lock); |
1124 | 0 | return 1; |
1125 | 0 | } |
1126 | 0 | lock_rw_unlock(&e->entry.lock); |
1127 | 0 | } |
1128 | 0 | } |
1129 | | /* alloc, malloc properly (not in region, like msg is) */ |
1130 | 0 | rep = reply_info_copy(msgrep, env->alloc, NULL); |
1131 | 0 | if(!rep) |
1132 | 0 | return 0; |
1133 | | /* ttl must be relative ;i.e. 0..86400 not time(0)+86400. |
1134 | | * the env->now is added to message and RRsets in this routine. */ |
1135 | | /* the leeway is used to invalidate other rrsets earlier */ |
1136 | 0 | if(is_referral) { |
1137 | | /* store rrsets */ |
1138 | 0 | struct rrset_ref ref; |
1139 | 0 | size_t i; |
1140 | 0 | for(i=0; i<rep->rrset_count; i++) { |
1141 | 0 | packed_rrset_ttl_add((struct packed_rrset_data*) |
1142 | 0 | rep->rrsets[i]->entry.data, *env->now); |
1143 | 0 | ref.key = rep->rrsets[i]; |
1144 | 0 | ref.id = rep->rrsets[i]->id; |
1145 | | /*ignore ret: it was in the cache, ref updated */ |
1146 | | /* no leeway for typeNS */ |
1147 | 0 | (void)rrset_cache_update(env->rrset_cache, &ref, |
1148 | 0 | env->alloc, |
1149 | 0 | ((ntohs(ref.key->rk.type)==LDNS_RR_TYPE_NS |
1150 | 0 | && !pside) ? qstarttime:*env->now + leeway)); |
1151 | 0 | } |
1152 | 0 | reply_info_delete(rep, NULL); |
1153 | 0 | return 1; |
1154 | 0 | } else { |
1155 | | /* store msg, and rrsets */ |
1156 | 0 | struct query_info qinf; |
1157 | 0 | hashvalue_type h; |
1158 | |
|
1159 | 0 | qinf = *msgqinf; |
1160 | 0 | qinf.qname = memdup(msgqinf->qname, msgqinf->qname_len); |
1161 | 0 | if(!qinf.qname) { |
1162 | 0 | reply_info_parsedelete(rep, env->alloc); |
1163 | 0 | return 0; |
1164 | 0 | } |
1165 | | /* fixup flags to be sensible for a reply based on the cache */ |
1166 | | /* this module means that RA is available. It is an answer QR. |
1167 | | * Not AA from cache. Not CD in cache (depends on client bit). */ |
1168 | 0 | rep->flags |= (BIT_RA | BIT_QR); |
1169 | 0 | rep->flags &= ~(BIT_AA | BIT_CD); |
1170 | 0 | h = query_info_hash(&qinf, (uint16_t)flags); |
1171 | 0 | dns_cache_store_msg(env, &qinf, h, rep, leeway, pside, msgrep, |
1172 | 0 | flags, region, qstarttime); |
1173 | | /* qname is used inside query_info_entrysetup, and set to |
1174 | | * NULL. If it has not been used, free it. free(0) is safe. */ |
1175 | 0 | free(qinf.qname); |
1176 | 0 | } |
1177 | 0 | return 1; |
1178 | 0 | } |
1179 | | |
1180 | | int |
1181 | | dns_cache_prefetch_adjust(struct module_env* env, struct query_info* qinfo, |
1182 | | time_t adjust, uint16_t flags) |
1183 | 0 | { |
1184 | 0 | struct msgreply_entry* msg; |
1185 | 0 | msg = msg_cache_lookup(env, qinfo->qname, qinfo->qname_len, |
1186 | 0 | qinfo->qtype, qinfo->qclass, flags, *env->now, 1); |
1187 | 0 | if(msg) { |
1188 | 0 | struct reply_info* rep = (struct reply_info*)msg->entry.data; |
1189 | 0 | if(rep) { |
1190 | 0 | rep->prefetch_ttl += adjust; |
1191 | 0 | lock_rw_unlock(&msg->entry.lock); |
1192 | 0 | return 1; |
1193 | 0 | } |
1194 | 0 | lock_rw_unlock(&msg->entry.lock); |
1195 | 0 | } |
1196 | 0 | return 0; |
1197 | 0 | } |