Line | Count | Source (jump to first uncovered line) |
1 | | // Copyright 2007 The RE2 Authors. All Rights Reserved. |
2 | | // Use of this source code is governed by a BSD-style |
3 | | // license that can be found in the LICENSE file. |
4 | | |
5 | | // Compile regular expression to Prog. |
6 | | // |
7 | | // Prog and Inst are defined in prog.h. |
8 | | // This file's external interface is just Regexp::CompileToProg. |
9 | | // The Compiler class defined in this file is private. |
10 | | |
11 | | #include <stdint.h> |
12 | | #include <string.h> |
13 | | #include <unordered_map> |
14 | | #include <utility> |
15 | | |
16 | | #include "util/logging.h" |
17 | | #include "util/utf.h" |
18 | | #include "re2/pod_array.h" |
19 | | #include "re2/prog.h" |
20 | | #include "re2/re2.h" |
21 | | #include "re2/regexp.h" |
22 | | #include "re2/walker-inl.h" |
23 | | |
24 | | namespace re2 { |
25 | | |
26 | | // List of pointers to Inst* that need to be filled in (patched). |
27 | | // Because the Inst* haven't been filled in yet, |
28 | | // we can use the Inst* word to hold the list's "next" pointer. |
29 | | // It's kind of sleazy, but it works well in practice. |
30 | | // See http://swtch.com/~rsc/regexp/regexp1.html for inspiration. |
31 | | // |
32 | | // Because the out and out1 fields in Inst are no longer pointers, |
33 | | // we can't use pointers directly here either. Instead, head refers |
34 | | // to inst_[head>>1].out (head&1 == 0) or inst_[head>>1].out1 (head&1 == 1). |
35 | | // head == 0 represents the NULL list. This is okay because instruction #0 |
36 | | // is always the fail instruction, which never appears on a list. |
37 | | struct PatchList { |
38 | | // Returns patch list containing just p. |
39 | 0 | static PatchList Mk(uint32_t p) { |
40 | 0 | return {p, p}; |
41 | 0 | } |
42 | | |
43 | | // Patches all the entries on l to have value p. |
44 | | // Caller must not ever use patch list again. |
45 | 0 | static void Patch(Prog::Inst* inst0, PatchList l, uint32_t p) { |
46 | 0 | while (l.head != 0) { |
47 | 0 | Prog::Inst* ip = &inst0[l.head>>1]; |
48 | 0 | if (l.head&1) { |
49 | 0 | l.head = ip->out1(); |
50 | 0 | ip->out1_ = p; |
51 | 0 | } else { |
52 | 0 | l.head = ip->out(); |
53 | 0 | ip->set_out(p); |
54 | 0 | } |
55 | 0 | } |
56 | 0 | } |
57 | | |
58 | | // Appends two patch lists and returns result. |
59 | 0 | static PatchList Append(Prog::Inst* inst0, PatchList l1, PatchList l2) { |
60 | 0 | if (l1.head == 0) |
61 | 0 | return l2; |
62 | 0 | if (l2.head == 0) |
63 | 0 | return l1; |
64 | 0 | Prog::Inst* ip = &inst0[l1.tail>>1]; |
65 | 0 | if (l1.tail&1) |
66 | 0 | ip->out1_ = l2.head; |
67 | 0 | else |
68 | 0 | ip->set_out(l2.head); |
69 | 0 | return {l1.head, l2.tail}; |
70 | 0 | } |
71 | | |
72 | | uint32_t head; |
73 | | uint32_t tail; // for constant-time append |
74 | | }; |
75 | | |
76 | | static const PatchList kNullPatchList = {0, 0}; |
77 | | |
78 | | // Compiled program fragment. |
79 | | struct Frag { |
80 | | uint32_t begin; |
81 | | PatchList end; |
82 | | bool nullable; |
83 | | |
84 | 0 | Frag() : begin(0), end(kNullPatchList), nullable(false) {} |
85 | | Frag(uint32_t begin, PatchList end, bool nullable) |
86 | 0 | : begin(begin), end(end), nullable(nullable) {} |
87 | | }; |
88 | | |
89 | | // Input encodings. |
90 | | enum Encoding { |
91 | | kEncodingUTF8 = 1, // UTF-8 (0-10FFFF) |
92 | | kEncodingLatin1, // Latin-1 (0-FF) |
93 | | }; |
94 | | |
95 | | class Compiler : public Regexp::Walker<Frag> { |
96 | | public: |
97 | | explicit Compiler(); |
98 | | ~Compiler(); |
99 | | |
100 | | // Compiles Regexp to a new Prog. |
101 | | // Caller is responsible for deleting Prog when finished with it. |
102 | | // If reversed is true, compiles for walking over the input |
103 | | // string backward (reverses all concatenations). |
104 | | static Prog *Compile(Regexp* re, bool reversed, int64_t max_mem); |
105 | | |
106 | | // Compiles alternation of all the re to a new Prog. |
107 | | // Each re has a match with an id equal to its index in the vector. |
108 | | static Prog* CompileSet(Regexp* re, RE2::Anchor anchor, int64_t max_mem); |
109 | | |
110 | | // Interface for Regexp::Walker, which helps traverse the Regexp. |
111 | | // The walk is purely post-recursive: given the machines for the |
112 | | // children, PostVisit combines them to create the machine for |
113 | | // the current node. The child_args are Frags. |
114 | | // The Compiler traverses the Regexp parse tree, visiting |
115 | | // each node in depth-first order. It invokes PreVisit before |
116 | | // visiting the node's children and PostVisit after visiting |
117 | | // the children. |
118 | | Frag PreVisit(Regexp* re, Frag parent_arg, bool* stop); |
119 | | Frag PostVisit(Regexp* re, Frag parent_arg, Frag pre_arg, Frag* child_args, |
120 | | int nchild_args); |
121 | | Frag ShortVisit(Regexp* re, Frag parent_arg); |
122 | | Frag Copy(Frag arg); |
123 | | |
124 | | // Given fragment a, returns a+ or a+?; a* or a*?; a? or a?? |
125 | | Frag Plus(Frag a, bool nongreedy); |
126 | | Frag Star(Frag a, bool nongreedy); |
127 | | Frag Quest(Frag a, bool nongreedy); |
128 | | |
129 | | // Given fragment a, returns (a) capturing as \n. |
130 | | Frag Capture(Frag a, int n); |
131 | | |
132 | | // Given fragments a and b, returns ab; a|b |
133 | | Frag Cat(Frag a, Frag b); |
134 | | Frag Alt(Frag a, Frag b); |
135 | | |
136 | | // Returns a fragment that can't match anything. |
137 | | Frag NoMatch(); |
138 | | |
139 | | // Returns a fragment that matches the empty string. |
140 | | Frag Match(int32_t id); |
141 | | |
142 | | // Returns a no-op fragment. |
143 | | Frag Nop(); |
144 | | |
145 | | // Returns a fragment matching the byte range lo-hi. |
146 | | Frag ByteRange(int lo, int hi, bool foldcase); |
147 | | |
148 | | // Returns a fragment matching an empty-width special op. |
149 | | Frag EmptyWidth(EmptyOp op); |
150 | | |
151 | | // Adds n instructions to the program. |
152 | | // Returns the index of the first one. |
153 | | // Returns -1 if no more instructions are available. |
154 | | int AllocInst(int n); |
155 | | |
156 | | // Rune range compiler. |
157 | | |
158 | | // Begins a new alternation. |
159 | | void BeginRange(); |
160 | | |
161 | | // Adds a fragment matching the rune range lo-hi. |
162 | | void AddRuneRange(Rune lo, Rune hi, bool foldcase); |
163 | | void AddRuneRangeLatin1(Rune lo, Rune hi, bool foldcase); |
164 | | void AddRuneRangeUTF8(Rune lo, Rune hi, bool foldcase); |
165 | | void Add_80_10ffff(); |
166 | | |
167 | | // New suffix that matches the byte range lo-hi, then goes to next. |
168 | | int UncachedRuneByteSuffix(uint8_t lo, uint8_t hi, bool foldcase, int next); |
169 | | int CachedRuneByteSuffix(uint8_t lo, uint8_t hi, bool foldcase, int next); |
170 | | |
171 | | // Returns true iff the suffix is cached. |
172 | | bool IsCachedRuneByteSuffix(int id); |
173 | | |
174 | | // Adds a suffix to alternation. |
175 | | void AddSuffix(int id); |
176 | | |
177 | | // Adds a suffix to the trie starting from the given root node. |
178 | | // Returns zero iff allocating an instruction fails. Otherwise, returns |
179 | | // the current root node, which might be different from what was given. |
180 | | int AddSuffixRecursive(int root, int id); |
181 | | |
182 | | // Finds the trie node for the given suffix. Returns a Frag in order to |
183 | | // distinguish between pointing at the root node directly (end.head == 0) |
184 | | // and pointing at an Alt's out1 or out (end.head&1 == 1 or 0, respectively). |
185 | | Frag FindByteRange(int root, int id); |
186 | | |
187 | | // Compares two ByteRanges and returns true iff they are equal. |
188 | | bool ByteRangeEqual(int id1, int id2); |
189 | | |
190 | | // Returns the alternation of all the added suffixes. |
191 | | Frag EndRange(); |
192 | | |
193 | | // Single rune. |
194 | | Frag Literal(Rune r, bool foldcase); |
195 | | |
196 | | void Setup(Regexp::ParseFlags flags, int64_t max_mem, RE2::Anchor anchor); |
197 | | Prog* Finish(Regexp* re); |
198 | | |
199 | | // Returns .* where dot = any byte |
200 | | Frag DotStar(); |
201 | | |
202 | | private: |
203 | | Prog* prog_; // Program being built. |
204 | | bool failed_; // Did we give up compiling? |
205 | | Encoding encoding_; // Input encoding |
206 | | bool reversed_; // Should program run backward over text? |
207 | | |
208 | | PODArray<Prog::Inst> inst_; |
209 | | int ninst_; // Number of instructions used. |
210 | | int max_ninst_; // Maximum number of instructions. |
211 | | |
212 | | int64_t max_mem_; // Total memory budget. |
213 | | |
214 | | std::unordered_map<uint64_t, int> rune_cache_; |
215 | | Frag rune_range_; |
216 | | |
217 | | RE2::Anchor anchor_; // anchor mode for RE2::Set |
218 | | |
219 | | Compiler(const Compiler&) = delete; |
220 | | Compiler& operator=(const Compiler&) = delete; |
221 | | }; |
222 | | |
223 | 0 | Compiler::Compiler() { |
224 | 0 | prog_ = new Prog(); |
225 | 0 | failed_ = false; |
226 | 0 | encoding_ = kEncodingUTF8; |
227 | 0 | reversed_ = false; |
228 | 0 | ninst_ = 0; |
229 | 0 | max_ninst_ = 1; // make AllocInst for fail instruction okay |
230 | 0 | max_mem_ = 0; |
231 | 0 | int fail = AllocInst(1); |
232 | 0 | inst_[fail].InitFail(); |
233 | 0 | max_ninst_ = 0; // Caller must change |
234 | 0 | } |
235 | | |
236 | 0 | Compiler::~Compiler() { |
237 | 0 | delete prog_; |
238 | 0 | } |
239 | | |
240 | 0 | int Compiler::AllocInst(int n) { |
241 | 0 | if (failed_ || ninst_ + n > max_ninst_) { |
242 | 0 | failed_ = true; |
243 | 0 | return -1; |
244 | 0 | } |
245 | | |
246 | 0 | if (ninst_ + n > inst_.size()) { |
247 | 0 | int cap = inst_.size(); |
248 | 0 | if (cap == 0) |
249 | 0 | cap = 8; |
250 | 0 | while (ninst_ + n > cap) |
251 | 0 | cap *= 2; |
252 | 0 | PODArray<Prog::Inst> inst(cap); |
253 | 0 | if (inst_.data() != NULL) |
254 | 0 | memmove(inst.data(), inst_.data(), ninst_*sizeof inst_[0]); |
255 | 0 | memset(inst.data() + ninst_, 0, (cap - ninst_)*sizeof inst_[0]); |
256 | 0 | inst_ = std::move(inst); |
257 | 0 | } |
258 | 0 | int id = ninst_; |
259 | 0 | ninst_ += n; |
260 | 0 | return id; |
261 | 0 | } |
262 | | |
263 | | // These routines are somewhat hard to visualize in text -- |
264 | | // see http://swtch.com/~rsc/regexp/regexp1.html for |
265 | | // pictures explaining what is going on here. |
266 | | |
267 | | // Returns an unmatchable fragment. |
268 | 0 | Frag Compiler::NoMatch() { |
269 | 0 | return Frag(); |
270 | 0 | } |
271 | | |
272 | | // Is a an unmatchable fragment? |
273 | 0 | static bool IsNoMatch(Frag a) { |
274 | 0 | return a.begin == 0; |
275 | 0 | } |
276 | | |
277 | | // Given fragments a and b, returns fragment for ab. |
278 | 0 | Frag Compiler::Cat(Frag a, Frag b) { |
279 | 0 | if (IsNoMatch(a) || IsNoMatch(b)) |
280 | 0 | return NoMatch(); |
281 | | |
282 | | // Elide no-op. |
283 | 0 | Prog::Inst* begin = &inst_[a.begin]; |
284 | 0 | if (begin->opcode() == kInstNop && |
285 | 0 | a.end.head == (a.begin << 1) && |
286 | 0 | begin->out() == 0) { |
287 | | // in case refs to a somewhere |
288 | 0 | PatchList::Patch(inst_.data(), a.end, b.begin); |
289 | 0 | return b; |
290 | 0 | } |
291 | | |
292 | | // To run backward over string, reverse all concatenations. |
293 | 0 | if (reversed_) { |
294 | 0 | PatchList::Patch(inst_.data(), b.end, a.begin); |
295 | 0 | return Frag(b.begin, a.end, b.nullable && a.nullable); |
296 | 0 | } |
297 | | |
298 | 0 | PatchList::Patch(inst_.data(), a.end, b.begin); |
299 | 0 | return Frag(a.begin, b.end, a.nullable && b.nullable); |
300 | 0 | } |
301 | | |
302 | | // Given fragments for a and b, returns fragment for a|b. |
303 | 0 | Frag Compiler::Alt(Frag a, Frag b) { |
304 | | // Special case for convenience in loops. |
305 | 0 | if (IsNoMatch(a)) |
306 | 0 | return b; |
307 | 0 | if (IsNoMatch(b)) |
308 | 0 | return a; |
309 | | |
310 | 0 | int id = AllocInst(1); |
311 | 0 | if (id < 0) |
312 | 0 | return NoMatch(); |
313 | | |
314 | 0 | inst_[id].InitAlt(a.begin, b.begin); |
315 | 0 | return Frag(id, PatchList::Append(inst_.data(), a.end, b.end), |
316 | 0 | a.nullable || b.nullable); |
317 | 0 | } |
318 | | |
319 | | // When capturing submatches in like-Perl mode, a kOpAlt Inst |
320 | | // treats out_ as the first choice, out1_ as the second. |
321 | | // |
322 | | // For *, +, and ?, if out_ causes another repetition, |
323 | | // then the operator is greedy. If out1_ is the repetition |
324 | | // (and out_ moves forward), then the operator is non-greedy. |
325 | | |
326 | | // Given a fragment for a, returns a fragment for a+ or a+? (if nongreedy) |
327 | 0 | Frag Compiler::Plus(Frag a, bool nongreedy) { |
328 | 0 | int id = AllocInst(1); |
329 | 0 | if (id < 0) |
330 | 0 | return NoMatch(); |
331 | 0 | PatchList pl; |
332 | 0 | if (nongreedy) { |
333 | 0 | inst_[id].InitAlt(0, a.begin); |
334 | 0 | pl = PatchList::Mk(id << 1); |
335 | 0 | } else { |
336 | 0 | inst_[id].InitAlt(a.begin, 0); |
337 | 0 | pl = PatchList::Mk((id << 1) | 1); |
338 | 0 | } |
339 | 0 | PatchList::Patch(inst_.data(), a.end, id); |
340 | 0 | return Frag(a.begin, pl, a.nullable); |
341 | 0 | } |
342 | | |
343 | | // Given a fragment for a, returns a fragment for a* or a*? (if nongreedy) |
344 | 0 | Frag Compiler::Star(Frag a, bool nongreedy) { |
345 | | // When the subexpression is nullable, one Alt isn't enough to guarantee |
346 | | // correct priority ordering within the transitive closure. The simplest |
347 | | // solution is to handle it as (a+)? instead, which adds the second Alt. |
348 | 0 | if (a.nullable) |
349 | 0 | return Quest(Plus(a, nongreedy), nongreedy); |
350 | | |
351 | 0 | int id = AllocInst(1); |
352 | 0 | if (id < 0) |
353 | 0 | return NoMatch(); |
354 | 0 | PatchList pl; |
355 | 0 | if (nongreedy) { |
356 | 0 | inst_[id].InitAlt(0, a.begin); |
357 | 0 | pl = PatchList::Mk(id << 1); |
358 | 0 | } else { |
359 | 0 | inst_[id].InitAlt(a.begin, 0); |
360 | 0 | pl = PatchList::Mk((id << 1) | 1); |
361 | 0 | } |
362 | 0 | PatchList::Patch(inst_.data(), a.end, id); |
363 | 0 | return Frag(id, pl, true); |
364 | 0 | } |
365 | | |
366 | | // Given a fragment for a, returns a fragment for a? or a?? (if nongreedy) |
367 | 0 | Frag Compiler::Quest(Frag a, bool nongreedy) { |
368 | 0 | if (IsNoMatch(a)) |
369 | 0 | return Nop(); |
370 | 0 | int id = AllocInst(1); |
371 | 0 | if (id < 0) |
372 | 0 | return NoMatch(); |
373 | 0 | PatchList pl; |
374 | 0 | if (nongreedy) { |
375 | 0 | inst_[id].InitAlt(0, a.begin); |
376 | 0 | pl = PatchList::Mk(id << 1); |
377 | 0 | } else { |
378 | 0 | inst_[id].InitAlt(a.begin, 0); |
379 | 0 | pl = PatchList::Mk((id << 1) | 1); |
380 | 0 | } |
381 | 0 | return Frag(id, PatchList::Append(inst_.data(), pl, a.end), true); |
382 | 0 | } |
383 | | |
384 | | // Returns a fragment for the byte range lo-hi. |
385 | 0 | Frag Compiler::ByteRange(int lo, int hi, bool foldcase) { |
386 | 0 | int id = AllocInst(1); |
387 | 0 | if (id < 0) |
388 | 0 | return NoMatch(); |
389 | 0 | inst_[id].InitByteRange(lo, hi, foldcase, 0); |
390 | 0 | return Frag(id, PatchList::Mk(id << 1), false); |
391 | 0 | } |
392 | | |
393 | | // Returns a no-op fragment. Sometimes unavoidable. |
394 | 0 | Frag Compiler::Nop() { |
395 | 0 | int id = AllocInst(1); |
396 | 0 | if (id < 0) |
397 | 0 | return NoMatch(); |
398 | 0 | inst_[id].InitNop(0); |
399 | 0 | return Frag(id, PatchList::Mk(id << 1), true); |
400 | 0 | } |
401 | | |
402 | | // Returns a fragment that signals a match. |
403 | 0 | Frag Compiler::Match(int32_t match_id) { |
404 | 0 | int id = AllocInst(1); |
405 | 0 | if (id < 0) |
406 | 0 | return NoMatch(); |
407 | 0 | inst_[id].InitMatch(match_id); |
408 | 0 | return Frag(id, kNullPatchList, false); |
409 | 0 | } |
410 | | |
411 | | // Returns a fragment matching a particular empty-width op (like ^ or $) |
412 | 0 | Frag Compiler::EmptyWidth(EmptyOp empty) { |
413 | 0 | int id = AllocInst(1); |
414 | 0 | if (id < 0) |
415 | 0 | return NoMatch(); |
416 | 0 | inst_[id].InitEmptyWidth(empty, 0); |
417 | 0 | return Frag(id, PatchList::Mk(id << 1), true); |
418 | 0 | } |
419 | | |
420 | | // Given a fragment a, returns a fragment with capturing parens around a. |
421 | 0 | Frag Compiler::Capture(Frag a, int n) { |
422 | 0 | if (IsNoMatch(a)) |
423 | 0 | return NoMatch(); |
424 | 0 | int id = AllocInst(2); |
425 | 0 | if (id < 0) |
426 | 0 | return NoMatch(); |
427 | 0 | inst_[id].InitCapture(2*n, a.begin); |
428 | 0 | inst_[id+1].InitCapture(2*n+1, 0); |
429 | 0 | PatchList::Patch(inst_.data(), a.end, id+1); |
430 | |
|
431 | 0 | return Frag(id, PatchList::Mk((id+1) << 1), a.nullable); |
432 | 0 | } |
433 | | |
434 | | // A Rune is a name for a Unicode code point. |
435 | | // Returns maximum rune encoded by UTF-8 sequence of length len. |
436 | 0 | static int MaxRune(int len) { |
437 | 0 | int b; // number of Rune bits in len-byte UTF-8 sequence (len < UTFmax) |
438 | 0 | if (len == 1) |
439 | 0 | b = 7; |
440 | 0 | else |
441 | 0 | b = 8-(len+1) + 6*(len-1); |
442 | 0 | return (1<<b) - 1; // maximum Rune for b bits. |
443 | 0 | } |
444 | | |
445 | | // The rune range compiler caches common suffix fragments, |
446 | | // which are very common in UTF-8 (e.g., [80-bf]). |
447 | | // The fragment suffixes are identified by their start |
448 | | // instructions. NULL denotes the eventual end match. |
449 | | // The Frag accumulates in rune_range_. Caching common |
450 | | // suffixes reduces the UTF-8 "." from 32 to 24 instructions, |
451 | | // and it reduces the corresponding one-pass NFA from 16 nodes to 8. |
452 | | |
453 | 0 | void Compiler::BeginRange() { |
454 | 0 | rune_cache_.clear(); |
455 | 0 | rune_range_.begin = 0; |
456 | 0 | rune_range_.end = kNullPatchList; |
457 | 0 | } |
458 | | |
459 | | int Compiler::UncachedRuneByteSuffix(uint8_t lo, uint8_t hi, bool foldcase, |
460 | 0 | int next) { |
461 | 0 | Frag f = ByteRange(lo, hi, foldcase); |
462 | 0 | if (next != 0) { |
463 | 0 | PatchList::Patch(inst_.data(), f.end, next); |
464 | 0 | } else { |
465 | 0 | rune_range_.end = PatchList::Append(inst_.data(), rune_range_.end, f.end); |
466 | 0 | } |
467 | 0 | return f.begin; |
468 | 0 | } |
469 | | |
470 | | static uint64_t MakeRuneCacheKey(uint8_t lo, uint8_t hi, bool foldcase, |
471 | 0 | int next) { |
472 | 0 | return (uint64_t)next << 17 | |
473 | 0 | (uint64_t)lo << 9 | |
474 | 0 | (uint64_t)hi << 1 | |
475 | 0 | (uint64_t)foldcase; |
476 | 0 | } |
477 | | |
478 | | int Compiler::CachedRuneByteSuffix(uint8_t lo, uint8_t hi, bool foldcase, |
479 | 0 | int next) { |
480 | 0 | uint64_t key = MakeRuneCacheKey(lo, hi, foldcase, next); |
481 | 0 | std::unordered_map<uint64_t, int>::const_iterator it = rune_cache_.find(key); |
482 | 0 | if (it != rune_cache_.end()) |
483 | 0 | return it->second; |
484 | 0 | int id = UncachedRuneByteSuffix(lo, hi, foldcase, next); |
485 | 0 | rune_cache_[key] = id; |
486 | 0 | return id; |
487 | 0 | } |
488 | | |
489 | 0 | bool Compiler::IsCachedRuneByteSuffix(int id) { |
490 | 0 | uint8_t lo = inst_[id].lo_; |
491 | 0 | uint8_t hi = inst_[id].hi_; |
492 | 0 | bool foldcase = inst_[id].foldcase() != 0; |
493 | 0 | int next = inst_[id].out(); |
494 | |
|
495 | 0 | uint64_t key = MakeRuneCacheKey(lo, hi, foldcase, next); |
496 | 0 | return rune_cache_.find(key) != rune_cache_.end(); |
497 | 0 | } |
498 | | |
499 | 0 | void Compiler::AddSuffix(int id) { |
500 | 0 | if (failed_) |
501 | 0 | return; |
502 | | |
503 | 0 | if (rune_range_.begin == 0) { |
504 | 0 | rune_range_.begin = id; |
505 | 0 | return; |
506 | 0 | } |
507 | | |
508 | 0 | if (encoding_ == kEncodingUTF8) { |
509 | | // Build a trie in order to reduce fanout. |
510 | 0 | rune_range_.begin = AddSuffixRecursive(rune_range_.begin, id); |
511 | 0 | return; |
512 | 0 | } |
513 | | |
514 | 0 | int alt = AllocInst(1); |
515 | 0 | if (alt < 0) { |
516 | 0 | rune_range_.begin = 0; |
517 | 0 | return; |
518 | 0 | } |
519 | 0 | inst_[alt].InitAlt(rune_range_.begin, id); |
520 | 0 | rune_range_.begin = alt; |
521 | 0 | } |
522 | | |
523 | 0 | int Compiler::AddSuffixRecursive(int root, int id) { |
524 | 0 | DCHECK(inst_[root].opcode() == kInstAlt || |
525 | 0 | inst_[root].opcode() == kInstByteRange); |
526 | |
|
527 | 0 | Frag f = FindByteRange(root, id); |
528 | 0 | if (IsNoMatch(f)) { |
529 | 0 | int alt = AllocInst(1); |
530 | 0 | if (alt < 0) |
531 | 0 | return 0; |
532 | 0 | inst_[alt].InitAlt(root, id); |
533 | 0 | return alt; |
534 | 0 | } |
535 | | |
536 | 0 | int br; |
537 | 0 | if (f.end.head == 0) |
538 | 0 | br = root; |
539 | 0 | else if (f.end.head&1) |
540 | 0 | br = inst_[f.begin].out1(); |
541 | 0 | else |
542 | 0 | br = inst_[f.begin].out(); |
543 | |
|
544 | 0 | if (IsCachedRuneByteSuffix(br)) { |
545 | | // We can't fiddle with cached suffixes, so make a clone of the head. |
546 | 0 | int byterange = AllocInst(1); |
547 | 0 | if (byterange < 0) |
548 | 0 | return 0; |
549 | 0 | inst_[byterange].InitByteRange(inst_[br].lo(), inst_[br].hi(), |
550 | 0 | inst_[br].foldcase(), inst_[br].out()); |
551 | | |
552 | | // Ensure that the parent points to the clone, not to the original. |
553 | | // Note that this could leave the head unreachable except via the cache. |
554 | 0 | br = byterange; |
555 | 0 | if (f.end.head == 0) |
556 | 0 | root = br; |
557 | 0 | else if (f.end.head&1) |
558 | 0 | inst_[f.begin].out1_ = br; |
559 | 0 | else |
560 | 0 | inst_[f.begin].set_out(br); |
561 | 0 | } |
562 | | |
563 | 0 | int out = inst_[id].out(); |
564 | 0 | if (!IsCachedRuneByteSuffix(id)) { |
565 | | // The head should be the instruction most recently allocated, so free it |
566 | | // instead of leaving it unreachable. |
567 | 0 | DCHECK_EQ(id, ninst_-1); |
568 | 0 | inst_[id].out_opcode_ = 0; |
569 | 0 | inst_[id].out1_ = 0; |
570 | 0 | ninst_--; |
571 | 0 | } |
572 | |
|
573 | 0 | out = AddSuffixRecursive(inst_[br].out(), out); |
574 | 0 | if (out == 0) |
575 | 0 | return 0; |
576 | | |
577 | 0 | inst_[br].set_out(out); |
578 | 0 | return root; |
579 | 0 | } |
580 | | |
581 | 0 | bool Compiler::ByteRangeEqual(int id1, int id2) { |
582 | 0 | return inst_[id1].lo() == inst_[id2].lo() && |
583 | 0 | inst_[id1].hi() == inst_[id2].hi() && |
584 | 0 | inst_[id1].foldcase() == inst_[id2].foldcase(); |
585 | 0 | } |
586 | | |
587 | 0 | Frag Compiler::FindByteRange(int root, int id) { |
588 | 0 | if (inst_[root].opcode() == kInstByteRange) { |
589 | 0 | if (ByteRangeEqual(root, id)) |
590 | 0 | return Frag(root, kNullPatchList, false); |
591 | 0 | else |
592 | 0 | return NoMatch(); |
593 | 0 | } |
594 | | |
595 | 0 | while (inst_[root].opcode() == kInstAlt) { |
596 | 0 | int out1 = inst_[root].out1(); |
597 | 0 | if (ByteRangeEqual(out1, id)) |
598 | 0 | return Frag(root, PatchList::Mk((root << 1) | 1), false); |
599 | | |
600 | | // CharClass is a sorted list of ranges, so if out1 of the root Alt wasn't |
601 | | // what we're looking for, then we can stop immediately. Unfortunately, we |
602 | | // can't short-circuit the search in reverse mode. |
603 | 0 | if (!reversed_) |
604 | 0 | return NoMatch(); |
605 | | |
606 | 0 | int out = inst_[root].out(); |
607 | 0 | if (inst_[out].opcode() == kInstAlt) |
608 | 0 | root = out; |
609 | 0 | else if (ByteRangeEqual(out, id)) |
610 | 0 | return Frag(root, PatchList::Mk(root << 1), false); |
611 | 0 | else |
612 | 0 | return NoMatch(); |
613 | 0 | } |
614 | | |
615 | 0 | LOG(DFATAL) << "should never happen"; |
616 | 0 | return NoMatch(); |
617 | 0 | } |
618 | | |
619 | 0 | Frag Compiler::EndRange() { |
620 | 0 | return rune_range_; |
621 | 0 | } |
622 | | |
623 | | // Converts rune range lo-hi into a fragment that recognizes |
624 | | // the bytes that would make up those runes in the current |
625 | | // encoding (Latin 1 or UTF-8). |
626 | | // This lets the machine work byte-by-byte even when |
627 | | // using multibyte encodings. |
628 | | |
629 | 0 | void Compiler::AddRuneRange(Rune lo, Rune hi, bool foldcase) { |
630 | 0 | switch (encoding_) { |
631 | 0 | default: |
632 | 0 | case kEncodingUTF8: |
633 | 0 | AddRuneRangeUTF8(lo, hi, foldcase); |
634 | 0 | break; |
635 | 0 | case kEncodingLatin1: |
636 | 0 | AddRuneRangeLatin1(lo, hi, foldcase); |
637 | 0 | break; |
638 | 0 | } |
639 | 0 | } |
640 | | |
641 | 0 | void Compiler::AddRuneRangeLatin1(Rune lo, Rune hi, bool foldcase) { |
642 | | // Latin-1 is easy: runes *are* bytes. |
643 | 0 | if (lo > hi || lo > 0xFF) |
644 | 0 | return; |
645 | 0 | if (hi > 0xFF) |
646 | 0 | hi = 0xFF; |
647 | 0 | AddSuffix(UncachedRuneByteSuffix(static_cast<uint8_t>(lo), |
648 | 0 | static_cast<uint8_t>(hi), foldcase, 0)); |
649 | 0 | } |
650 | | |
651 | 0 | void Compiler::Add_80_10ffff() { |
652 | | // The 80-10FFFF (Runeself-Runemax) rune range occurs frequently enough |
653 | | // (for example, for /./ and /[^a-z]/) that it is worth simplifying: by |
654 | | // permitting overlong encodings in E0 and F0 sequences and code points |
655 | | // over 10FFFF in F4 sequences, the size of the bytecode and the number |
656 | | // of equivalence classes are reduced significantly. |
657 | 0 | int id; |
658 | 0 | if (reversed_) { |
659 | | // Prefix factoring matters, but we don't have to handle it here |
660 | | // because the rune range trie logic takes care of that already. |
661 | 0 | id = UncachedRuneByteSuffix(0xC2, 0xDF, false, 0); |
662 | 0 | id = UncachedRuneByteSuffix(0x80, 0xBF, false, id); |
663 | 0 | AddSuffix(id); |
664 | |
|
665 | 0 | id = UncachedRuneByteSuffix(0xE0, 0xEF, false, 0); |
666 | 0 | id = UncachedRuneByteSuffix(0x80, 0xBF, false, id); |
667 | 0 | id = UncachedRuneByteSuffix(0x80, 0xBF, false, id); |
668 | 0 | AddSuffix(id); |
669 | |
|
670 | 0 | id = UncachedRuneByteSuffix(0xF0, 0xF4, false, 0); |
671 | 0 | id = UncachedRuneByteSuffix(0x80, 0xBF, false, id); |
672 | 0 | id = UncachedRuneByteSuffix(0x80, 0xBF, false, id); |
673 | 0 | id = UncachedRuneByteSuffix(0x80, 0xBF, false, id); |
674 | 0 | AddSuffix(id); |
675 | 0 | } else { |
676 | | // Suffix factoring matters - and we do have to handle it here. |
677 | 0 | int cont1 = UncachedRuneByteSuffix(0x80, 0xBF, false, 0); |
678 | 0 | id = UncachedRuneByteSuffix(0xC2, 0xDF, false, cont1); |
679 | 0 | AddSuffix(id); |
680 | |
|
681 | 0 | int cont2 = UncachedRuneByteSuffix(0x80, 0xBF, false, cont1); |
682 | 0 | id = UncachedRuneByteSuffix(0xE0, 0xEF, false, cont2); |
683 | 0 | AddSuffix(id); |
684 | |
|
685 | 0 | int cont3 = UncachedRuneByteSuffix(0x80, 0xBF, false, cont2); |
686 | 0 | id = UncachedRuneByteSuffix(0xF0, 0xF4, false, cont3); |
687 | 0 | AddSuffix(id); |
688 | 0 | } |
689 | 0 | } |
690 | | |
691 | 0 | void Compiler::AddRuneRangeUTF8(Rune lo, Rune hi, bool foldcase) { |
692 | 0 | if (lo > hi) |
693 | 0 | return; |
694 | | |
695 | | // Pick off 80-10FFFF as a common special case. |
696 | 0 | if (lo == 0x80 && hi == 0x10ffff) { |
697 | 0 | Add_80_10ffff(); |
698 | 0 | return; |
699 | 0 | } |
700 | | |
701 | | // Split range into same-length sized ranges. |
702 | 0 | for (int i = 1; i < UTFmax; i++) { |
703 | 0 | Rune max = MaxRune(i); |
704 | 0 | if (lo <= max && max < hi) { |
705 | 0 | AddRuneRangeUTF8(lo, max, foldcase); |
706 | 0 | AddRuneRangeUTF8(max+1, hi, foldcase); |
707 | 0 | return; |
708 | 0 | } |
709 | 0 | } |
710 | | |
711 | | // ASCII range is always a special case. |
712 | 0 | if (hi < Runeself) { |
713 | 0 | AddSuffix(UncachedRuneByteSuffix(static_cast<uint8_t>(lo), |
714 | 0 | static_cast<uint8_t>(hi), foldcase, 0)); |
715 | 0 | return; |
716 | 0 | } |
717 | | |
718 | | // Split range into sections that agree on leading bytes. |
719 | 0 | for (int i = 1; i < UTFmax; i++) { |
720 | 0 | uint32_t m = (1<<(6*i)) - 1; // last i bytes of a UTF-8 sequence |
721 | 0 | if ((lo & ~m) != (hi & ~m)) { |
722 | 0 | if ((lo & m) != 0) { |
723 | 0 | AddRuneRangeUTF8(lo, lo|m, foldcase); |
724 | 0 | AddRuneRangeUTF8((lo|m)+1, hi, foldcase); |
725 | 0 | return; |
726 | 0 | } |
727 | 0 | if ((hi & m) != m) { |
728 | 0 | AddRuneRangeUTF8(lo, (hi&~m)-1, foldcase); |
729 | 0 | AddRuneRangeUTF8(hi&~m, hi, foldcase); |
730 | 0 | return; |
731 | 0 | } |
732 | 0 | } |
733 | 0 | } |
734 | | |
735 | | // Finally. Generate byte matching equivalent for lo-hi. |
736 | 0 | uint8_t ulo[UTFmax], uhi[UTFmax]; |
737 | 0 | int n = runetochar(reinterpret_cast<char*>(ulo), &lo); |
738 | 0 | int m = runetochar(reinterpret_cast<char*>(uhi), &hi); |
739 | 0 | (void)m; // USED(m) |
740 | 0 | DCHECK_EQ(n, m); |
741 | | |
742 | | // The logic below encodes this thinking: |
743 | | // |
744 | | // 1. When we have built the whole suffix, we know that it cannot |
745 | | // possibly be a suffix of anything longer: in forward mode, nothing |
746 | | // else can occur before the leading byte; in reverse mode, nothing |
747 | | // else can occur after the last continuation byte or else the leading |
748 | | // byte would have to change. Thus, there is no benefit to caching |
749 | | // the first byte of the suffix whereas there is a cost involved in |
750 | | // cloning it if it begins a common prefix, which is fairly likely. |
751 | | // |
752 | | // 2. Conversely, the last byte of the suffix cannot possibly be a |
753 | | // prefix of anything because next == 0, so we will never want to |
754 | | // clone it, but it is fairly likely to be a common suffix. Perhaps |
755 | | // more so in reverse mode than in forward mode because the former is |
756 | | // "converging" towards lower entropy, but caching is still worthwhile |
757 | | // for the latter in cases such as 80-BF. |
758 | | // |
759 | | // 3. Handling the bytes between the first and the last is less |
760 | | // straightforward and, again, the approach depends on whether we are |
761 | | // "converging" towards lower entropy: in forward mode, a single byte |
762 | | // is unlikely to be part of a common suffix whereas a byte range |
763 | | // is more likely so; in reverse mode, a byte range is unlikely to |
764 | | // be part of a common suffix whereas a single byte is more likely |
765 | | // so. The same benefit versus cost argument applies here. |
766 | 0 | int id = 0; |
767 | 0 | if (reversed_) { |
768 | 0 | for (int i = 0; i < n; i++) { |
769 | | // In reverse UTF-8 mode: cache the leading byte; don't cache the last |
770 | | // continuation byte; cache anything else iff it's a single byte (XX-XX). |
771 | 0 | if (i == 0 || (ulo[i] == uhi[i] && i != n-1)) |
772 | 0 | id = CachedRuneByteSuffix(ulo[i], uhi[i], false, id); |
773 | 0 | else |
774 | 0 | id = UncachedRuneByteSuffix(ulo[i], uhi[i], false, id); |
775 | 0 | } |
776 | 0 | } else { |
777 | 0 | for (int i = n-1; i >= 0; i--) { |
778 | | // In forward UTF-8 mode: don't cache the leading byte; cache the last |
779 | | // continuation byte; cache anything else iff it's a byte range (XX-YY). |
780 | 0 | if (i == n-1 || (ulo[i] < uhi[i] && i != 0)) |
781 | 0 | id = CachedRuneByteSuffix(ulo[i], uhi[i], false, id); |
782 | 0 | else |
783 | 0 | id = UncachedRuneByteSuffix(ulo[i], uhi[i], false, id); |
784 | 0 | } |
785 | 0 | } |
786 | 0 | AddSuffix(id); |
787 | 0 | } |
788 | | |
789 | | // Should not be called. |
790 | 0 | Frag Compiler::Copy(Frag arg) { |
791 | | // We're using WalkExponential; there should be no copying. |
792 | 0 | failed_ = true; |
793 | 0 | LOG(DFATAL) << "Compiler::Copy called!"; |
794 | 0 | return NoMatch(); |
795 | 0 | } |
796 | | |
797 | | // Visits a node quickly; called once WalkExponential has |
798 | | // decided to cut this walk short. |
799 | 0 | Frag Compiler::ShortVisit(Regexp* re, Frag) { |
800 | 0 | failed_ = true; |
801 | 0 | return NoMatch(); |
802 | 0 | } |
803 | | |
804 | | // Called before traversing a node's children during the walk. |
805 | 0 | Frag Compiler::PreVisit(Regexp* re, Frag, bool* stop) { |
806 | | // Cut off walk if we've already failed. |
807 | 0 | if (failed_) |
808 | 0 | *stop = true; |
809 | |
|
810 | 0 | return Frag(); // not used by caller |
811 | 0 | } |
812 | | |
813 | 0 | Frag Compiler::Literal(Rune r, bool foldcase) { |
814 | 0 | switch (encoding_) { |
815 | 0 | default: |
816 | 0 | return Frag(); |
817 | | |
818 | 0 | case kEncodingLatin1: |
819 | 0 | return ByteRange(r, r, foldcase); |
820 | | |
821 | 0 | case kEncodingUTF8: { |
822 | 0 | if (r < Runeself) // Make common case fast. |
823 | 0 | return ByteRange(r, r, foldcase); |
824 | 0 | uint8_t buf[UTFmax]; |
825 | 0 | int n = runetochar(reinterpret_cast<char*>(buf), &r); |
826 | 0 | Frag f = ByteRange((uint8_t)buf[0], buf[0], false); |
827 | 0 | for (int i = 1; i < n; i++) |
828 | 0 | f = Cat(f, ByteRange((uint8_t)buf[i], buf[i], false)); |
829 | 0 | return f; |
830 | 0 | } |
831 | 0 | } |
832 | 0 | } |
833 | | |
834 | | // Called after traversing the node's children during the walk. |
835 | | // Given their frags, build and return the frag for this re. |
836 | | Frag Compiler::PostVisit(Regexp* re, Frag, Frag, Frag* child_frags, |
837 | 0 | int nchild_frags) { |
838 | | // If a child failed, don't bother going forward, especially |
839 | | // since the child_frags might contain Frags with NULLs in them. |
840 | 0 | if (failed_) |
841 | 0 | return NoMatch(); |
842 | | |
843 | | // Given the child fragments, return the fragment for this node. |
844 | 0 | switch (re->op()) { |
845 | 0 | case kRegexpRepeat: |
846 | | // Should not see; code at bottom of function will print error |
847 | 0 | break; |
848 | | |
849 | 0 | case kRegexpNoMatch: |
850 | 0 | return NoMatch(); |
851 | | |
852 | 0 | case kRegexpEmptyMatch: |
853 | 0 | return Nop(); |
854 | | |
855 | 0 | case kRegexpHaveMatch: { |
856 | 0 | Frag f = Match(re->match_id()); |
857 | 0 | if (anchor_ == RE2::ANCHOR_BOTH) { |
858 | | // Append \z or else the subexpression will effectively be unanchored. |
859 | | // Complemented by the UNANCHORED case in CompileSet(). |
860 | 0 | f = Cat(EmptyWidth(kEmptyEndText), f); |
861 | 0 | } |
862 | 0 | return f; |
863 | 0 | } |
864 | | |
865 | 0 | case kRegexpConcat: { |
866 | 0 | Frag f = child_frags[0]; |
867 | 0 | for (int i = 1; i < nchild_frags; i++) |
868 | 0 | f = Cat(f, child_frags[i]); |
869 | 0 | return f; |
870 | 0 | } |
871 | | |
872 | 0 | case kRegexpAlternate: { |
873 | 0 | Frag f = child_frags[0]; |
874 | 0 | for (int i = 1; i < nchild_frags; i++) |
875 | 0 | f = Alt(f, child_frags[i]); |
876 | 0 | return f; |
877 | 0 | } |
878 | | |
879 | 0 | case kRegexpStar: |
880 | 0 | return Star(child_frags[0], (re->parse_flags()&Regexp::NonGreedy) != 0); |
881 | | |
882 | 0 | case kRegexpPlus: |
883 | 0 | return Plus(child_frags[0], (re->parse_flags()&Regexp::NonGreedy) != 0); |
884 | | |
885 | 0 | case kRegexpQuest: |
886 | 0 | return Quest(child_frags[0], (re->parse_flags()&Regexp::NonGreedy) != 0); |
887 | | |
888 | 0 | case kRegexpLiteral: |
889 | 0 | return Literal(re->rune(), (re->parse_flags()&Regexp::FoldCase) != 0); |
890 | | |
891 | 0 | case kRegexpLiteralString: { |
892 | | // Concatenation of literals. |
893 | 0 | if (re->nrunes() == 0) |
894 | 0 | return Nop(); |
895 | 0 | Frag f; |
896 | 0 | for (int i = 0; i < re->nrunes(); i++) { |
897 | 0 | Frag f1 = Literal(re->runes()[i], |
898 | 0 | (re->parse_flags()&Regexp::FoldCase) != 0); |
899 | 0 | if (i == 0) |
900 | 0 | f = f1; |
901 | 0 | else |
902 | 0 | f = Cat(f, f1); |
903 | 0 | } |
904 | 0 | return f; |
905 | 0 | } |
906 | | |
907 | 0 | case kRegexpAnyChar: |
908 | 0 | BeginRange(); |
909 | 0 | AddRuneRange(0, Runemax, false); |
910 | 0 | return EndRange(); |
911 | | |
912 | 0 | case kRegexpAnyByte: |
913 | 0 | return ByteRange(0x00, 0xFF, false); |
914 | | |
915 | 0 | case kRegexpCharClass: { |
916 | 0 | CharClass* cc = re->cc(); |
917 | 0 | if (cc->empty()) { |
918 | | // This can't happen. |
919 | 0 | failed_ = true; |
920 | 0 | LOG(DFATAL) << "No ranges in char class"; |
921 | 0 | return NoMatch(); |
922 | 0 | } |
923 | | |
924 | | // ASCII case-folding optimization: if the char class |
925 | | // behaves the same on A-Z as it does on a-z, |
926 | | // discard any ranges wholly contained in A-Z |
927 | | // and mark the other ranges as foldascii. |
928 | | // This reduces the size of a program for |
929 | | // (?i)abc from 3 insts per letter to 1 per letter. |
930 | 0 | bool foldascii = cc->FoldsASCII(); |
931 | | |
932 | | // Character class is just a big OR of the different |
933 | | // character ranges in the class. |
934 | 0 | BeginRange(); |
935 | 0 | for (CharClass::iterator i = cc->begin(); i != cc->end(); ++i) { |
936 | | // ASCII case-folding optimization (see above). |
937 | 0 | if (foldascii && 'A' <= i->lo && i->hi <= 'Z') |
938 | 0 | continue; |
939 | | |
940 | | // If this range contains all of A-Za-z or none of it, |
941 | | // the fold flag is unnecessary; don't bother. |
942 | 0 | bool fold = foldascii; |
943 | 0 | if ((i->lo <= 'A' && 'z' <= i->hi) || i->hi < 'A' || 'z' < i->lo || |
944 | 0 | ('Z' < i->lo && i->hi < 'a')) |
945 | 0 | fold = false; |
946 | |
|
947 | 0 | AddRuneRange(i->lo, i->hi, fold); |
948 | 0 | } |
949 | 0 | return EndRange(); |
950 | 0 | } |
951 | | |
952 | 0 | case kRegexpCapture: |
953 | | // If this is a non-capturing parenthesis -- (?:foo) -- |
954 | | // just use the inner expression. |
955 | 0 | if (re->cap() < 0) |
956 | 0 | return child_frags[0]; |
957 | 0 | return Capture(child_frags[0], re->cap()); |
958 | | |
959 | 0 | case kRegexpBeginLine: |
960 | 0 | return EmptyWidth(reversed_ ? kEmptyEndLine : kEmptyBeginLine); |
961 | | |
962 | 0 | case kRegexpEndLine: |
963 | 0 | return EmptyWidth(reversed_ ? kEmptyBeginLine : kEmptyEndLine); |
964 | | |
965 | 0 | case kRegexpBeginText: |
966 | 0 | return EmptyWidth(reversed_ ? kEmptyEndText : kEmptyBeginText); |
967 | | |
968 | 0 | case kRegexpEndText: |
969 | 0 | return EmptyWidth(reversed_ ? kEmptyBeginText : kEmptyEndText); |
970 | | |
971 | 0 | case kRegexpWordBoundary: |
972 | 0 | return EmptyWidth(kEmptyWordBoundary); |
973 | | |
974 | 0 | case kRegexpNoWordBoundary: |
975 | 0 | return EmptyWidth(kEmptyNonWordBoundary); |
976 | 0 | } |
977 | 0 | failed_ = true; |
978 | 0 | LOG(DFATAL) << "Missing case in Compiler: " << re->op(); |
979 | 0 | return NoMatch(); |
980 | 0 | } |
981 | | |
982 | | // Is this regexp required to start at the beginning of the text? |
983 | | // Only approximate; can return false for complicated regexps like (\Aa|\Ab), |
984 | | // but handles (\A(a|b)). Could use the Walker to write a more exact one. |
985 | 0 | static bool IsAnchorStart(Regexp** pre, int depth) { |
986 | 0 | Regexp* re = *pre; |
987 | 0 | Regexp* sub; |
988 | | // The depth limit makes sure that we don't overflow |
989 | | // the stack on a deeply nested regexp. As the comment |
990 | | // above says, IsAnchorStart is conservative, so returning |
991 | | // a false negative is okay. The exact limit is somewhat arbitrary. |
992 | 0 | if (re == NULL || depth >= 4) |
993 | 0 | return false; |
994 | 0 | switch (re->op()) { |
995 | 0 | default: |
996 | 0 | break; |
997 | 0 | case kRegexpConcat: |
998 | 0 | if (re->nsub() > 0) { |
999 | 0 | sub = re->sub()[0]->Incref(); |
1000 | 0 | if (IsAnchorStart(&sub, depth+1)) { |
1001 | 0 | PODArray<Regexp*> subcopy(re->nsub()); |
1002 | 0 | subcopy[0] = sub; // already have reference |
1003 | 0 | for (int i = 1; i < re->nsub(); i++) |
1004 | 0 | subcopy[i] = re->sub()[i]->Incref(); |
1005 | 0 | *pre = Regexp::Concat(subcopy.data(), re->nsub(), re->parse_flags()); |
1006 | 0 | re->Decref(); |
1007 | 0 | return true; |
1008 | 0 | } |
1009 | 0 | sub->Decref(); |
1010 | 0 | } |
1011 | 0 | break; |
1012 | 0 | case kRegexpCapture: |
1013 | 0 | sub = re->sub()[0]->Incref(); |
1014 | 0 | if (IsAnchorStart(&sub, depth+1)) { |
1015 | 0 | *pre = Regexp::Capture(sub, re->parse_flags(), re->cap()); |
1016 | 0 | re->Decref(); |
1017 | 0 | return true; |
1018 | 0 | } |
1019 | 0 | sub->Decref(); |
1020 | 0 | break; |
1021 | 0 | case kRegexpBeginText: |
1022 | 0 | *pre = Regexp::LiteralString(NULL, 0, re->parse_flags()); |
1023 | 0 | re->Decref(); |
1024 | 0 | return true; |
1025 | 0 | } |
1026 | 0 | return false; |
1027 | 0 | } |
1028 | | |
1029 | | // Is this regexp required to start at the end of the text? |
1030 | | // Only approximate; can return false for complicated regexps like (a\z|b\z), |
1031 | | // but handles ((a|b)\z). Could use the Walker to write a more exact one. |
1032 | 0 | static bool IsAnchorEnd(Regexp** pre, int depth) { |
1033 | 0 | Regexp* re = *pre; |
1034 | 0 | Regexp* sub; |
1035 | | // The depth limit makes sure that we don't overflow |
1036 | | // the stack on a deeply nested regexp. As the comment |
1037 | | // above says, IsAnchorEnd is conservative, so returning |
1038 | | // a false negative is okay. The exact limit is somewhat arbitrary. |
1039 | 0 | if (re == NULL || depth >= 4) |
1040 | 0 | return false; |
1041 | 0 | switch (re->op()) { |
1042 | 0 | default: |
1043 | 0 | break; |
1044 | 0 | case kRegexpConcat: |
1045 | 0 | if (re->nsub() > 0) { |
1046 | 0 | sub = re->sub()[re->nsub() - 1]->Incref(); |
1047 | 0 | if (IsAnchorEnd(&sub, depth+1)) { |
1048 | 0 | PODArray<Regexp*> subcopy(re->nsub()); |
1049 | 0 | subcopy[re->nsub() - 1] = sub; // already have reference |
1050 | 0 | for (int i = 0; i < re->nsub() - 1; i++) |
1051 | 0 | subcopy[i] = re->sub()[i]->Incref(); |
1052 | 0 | *pre = Regexp::Concat(subcopy.data(), re->nsub(), re->parse_flags()); |
1053 | 0 | re->Decref(); |
1054 | 0 | return true; |
1055 | 0 | } |
1056 | 0 | sub->Decref(); |
1057 | 0 | } |
1058 | 0 | break; |
1059 | 0 | case kRegexpCapture: |
1060 | 0 | sub = re->sub()[0]->Incref(); |
1061 | 0 | if (IsAnchorEnd(&sub, depth+1)) { |
1062 | 0 | *pre = Regexp::Capture(sub, re->parse_flags(), re->cap()); |
1063 | 0 | re->Decref(); |
1064 | 0 | return true; |
1065 | 0 | } |
1066 | 0 | sub->Decref(); |
1067 | 0 | break; |
1068 | 0 | case kRegexpEndText: |
1069 | 0 | *pre = Regexp::LiteralString(NULL, 0, re->parse_flags()); |
1070 | 0 | re->Decref(); |
1071 | 0 | return true; |
1072 | 0 | } |
1073 | 0 | return false; |
1074 | 0 | } |
1075 | | |
1076 | | void Compiler::Setup(Regexp::ParseFlags flags, int64_t max_mem, |
1077 | 0 | RE2::Anchor anchor) { |
1078 | 0 | if (flags & Regexp::Latin1) |
1079 | 0 | encoding_ = kEncodingLatin1; |
1080 | 0 | max_mem_ = max_mem; |
1081 | 0 | if (max_mem <= 0) { |
1082 | 0 | max_ninst_ = 100000; // more than enough |
1083 | 0 | } else if (static_cast<size_t>(max_mem) <= sizeof(Prog)) { |
1084 | | // No room for anything. |
1085 | 0 | max_ninst_ = 0; |
1086 | 0 | } else { |
1087 | 0 | int64_t m = (max_mem - sizeof(Prog)) / sizeof(Prog::Inst); |
1088 | | // Limit instruction count so that inst->id() fits nicely in an int. |
1089 | | // SparseArray also assumes that the indices (inst->id()) are ints. |
1090 | | // The call to WalkExponential uses 2*max_ninst_ below, |
1091 | | // and other places in the code use 2 or 3 * prog->size(). |
1092 | | // Limiting to 2^24 should avoid overflow in those places. |
1093 | | // (The point of allowing more than 32 bits of memory is to |
1094 | | // have plenty of room for the DFA states, not to use it up |
1095 | | // on the program.) |
1096 | 0 | if (m >= 1<<24) |
1097 | 0 | m = 1<<24; |
1098 | | // Inst imposes its own limit (currently bigger than 2^24 but be safe). |
1099 | 0 | if (m > Prog::Inst::kMaxInst) |
1100 | 0 | m = Prog::Inst::kMaxInst; |
1101 | 0 | max_ninst_ = static_cast<int>(m); |
1102 | 0 | } |
1103 | 0 | anchor_ = anchor; |
1104 | 0 | } |
1105 | | |
1106 | | // Compiles re, returning program. |
1107 | | // Caller is responsible for deleting prog_. |
1108 | | // If reversed is true, compiles a program that expects |
1109 | | // to run over the input string backward (reverses all concatenations). |
1110 | | // The reversed flag is also recorded in the returned program. |
1111 | 0 | Prog* Compiler::Compile(Regexp* re, bool reversed, int64_t max_mem) { |
1112 | 0 | Compiler c; |
1113 | 0 | c.Setup(re->parse_flags(), max_mem, RE2::UNANCHORED /* unused */); |
1114 | 0 | c.reversed_ = reversed; |
1115 | | |
1116 | | // Simplify to remove things like counted repetitions |
1117 | | // and character classes like \d. |
1118 | 0 | Regexp* sre = re->Simplify(); |
1119 | 0 | if (sre == NULL) |
1120 | 0 | return NULL; |
1121 | | |
1122 | | // Record whether prog is anchored, removing the anchors. |
1123 | | // (They get in the way of other optimizations.) |
1124 | 0 | bool is_anchor_start = IsAnchorStart(&sre, 0); |
1125 | 0 | bool is_anchor_end = IsAnchorEnd(&sre, 0); |
1126 | | |
1127 | | // Generate fragment for entire regexp. |
1128 | 0 | Frag all = c.WalkExponential(sre, Frag(), 2*c.max_ninst_); |
1129 | 0 | sre->Decref(); |
1130 | 0 | if (c.failed_) |
1131 | 0 | return NULL; |
1132 | | |
1133 | | // Success! Finish by putting Match node at end, and record start. |
1134 | | // Turn off c.reversed_ (if it is set) to force the remaining concatenations |
1135 | | // to behave normally. |
1136 | 0 | c.reversed_ = false; |
1137 | 0 | all = c.Cat(all, c.Match(0)); |
1138 | |
|
1139 | 0 | c.prog_->set_reversed(reversed); |
1140 | 0 | if (c.prog_->reversed()) { |
1141 | 0 | c.prog_->set_anchor_start(is_anchor_end); |
1142 | 0 | c.prog_->set_anchor_end(is_anchor_start); |
1143 | 0 | } else { |
1144 | 0 | c.prog_->set_anchor_start(is_anchor_start); |
1145 | 0 | c.prog_->set_anchor_end(is_anchor_end); |
1146 | 0 | } |
1147 | |
|
1148 | 0 | c.prog_->set_start(all.begin); |
1149 | 0 | if (!c.prog_->anchor_start()) { |
1150 | | // Also create unanchored version, which starts with a .*? loop. |
1151 | 0 | all = c.Cat(c.DotStar(), all); |
1152 | 0 | } |
1153 | 0 | c.prog_->set_start_unanchored(all.begin); |
1154 | | |
1155 | | // Hand ownership of prog_ to caller. |
1156 | 0 | return c.Finish(re); |
1157 | 0 | } |
1158 | | |
1159 | 0 | Prog* Compiler::Finish(Regexp* re) { |
1160 | 0 | if (failed_) |
1161 | 0 | return NULL; |
1162 | | |
1163 | 0 | if (prog_->start() == 0 && prog_->start_unanchored() == 0) { |
1164 | | // No possible matches; keep Fail instruction only. |
1165 | 0 | ninst_ = 1; |
1166 | 0 | } |
1167 | | |
1168 | | // Hand off the array to Prog. |
1169 | 0 | prog_->inst_ = std::move(inst_); |
1170 | 0 | prog_->size_ = ninst_; |
1171 | |
|
1172 | 0 | prog_->Optimize(); |
1173 | 0 | prog_->Flatten(); |
1174 | 0 | prog_->ComputeByteMap(); |
1175 | |
|
1176 | 0 | if (!prog_->reversed()) { |
1177 | 0 | std::string prefix; |
1178 | 0 | bool prefix_foldcase; |
1179 | 0 | if (re->RequiredPrefixForAccel(&prefix, &prefix_foldcase)) |
1180 | 0 | prog_->ConfigurePrefixAccel(prefix, prefix_foldcase); |
1181 | 0 | } |
1182 | | |
1183 | | // Record remaining memory for DFA. |
1184 | 0 | if (max_mem_ <= 0) { |
1185 | 0 | prog_->set_dfa_mem(1<<20); |
1186 | 0 | } else { |
1187 | 0 | int64_t m = max_mem_ - sizeof(Prog); |
1188 | 0 | m -= prog_->size_*sizeof(Prog::Inst); // account for inst_ |
1189 | 0 | if (prog_->CanBitState()) |
1190 | 0 | m -= prog_->size_*sizeof(uint16_t); // account for list_heads_ |
1191 | 0 | if (m < 0) |
1192 | 0 | m = 0; |
1193 | 0 | prog_->set_dfa_mem(m); |
1194 | 0 | } |
1195 | |
|
1196 | 0 | Prog* p = prog_; |
1197 | 0 | prog_ = NULL; |
1198 | 0 | return p; |
1199 | 0 | } |
1200 | | |
1201 | | // Converts Regexp to Prog. |
1202 | 0 | Prog* Regexp::CompileToProg(int64_t max_mem) { |
1203 | 0 | return Compiler::Compile(this, false, max_mem); |
1204 | 0 | } |
1205 | | |
1206 | 0 | Prog* Regexp::CompileToReverseProg(int64_t max_mem) { |
1207 | 0 | return Compiler::Compile(this, true, max_mem); |
1208 | 0 | } |
1209 | | |
1210 | 0 | Frag Compiler::DotStar() { |
1211 | 0 | return Star(ByteRange(0x00, 0xff, false), true); |
1212 | 0 | } |
1213 | | |
1214 | | // Compiles RE set to Prog. |
1215 | 0 | Prog* Compiler::CompileSet(Regexp* re, RE2::Anchor anchor, int64_t max_mem) { |
1216 | 0 | Compiler c; |
1217 | 0 | c.Setup(re->parse_flags(), max_mem, anchor); |
1218 | |
|
1219 | 0 | Regexp* sre = re->Simplify(); |
1220 | 0 | if (sre == NULL) |
1221 | 0 | return NULL; |
1222 | | |
1223 | 0 | Frag all = c.WalkExponential(sre, Frag(), 2*c.max_ninst_); |
1224 | 0 | sre->Decref(); |
1225 | 0 | if (c.failed_) |
1226 | 0 | return NULL; |
1227 | | |
1228 | 0 | c.prog_->set_anchor_start(true); |
1229 | 0 | c.prog_->set_anchor_end(true); |
1230 | |
|
1231 | 0 | if (anchor == RE2::UNANCHORED) { |
1232 | | // Prepend .* or else the expression will effectively be anchored. |
1233 | | // Complemented by the ANCHOR_BOTH case in PostVisit(). |
1234 | 0 | all = c.Cat(c.DotStar(), all); |
1235 | 0 | } |
1236 | 0 | c.prog_->set_start(all.begin); |
1237 | 0 | c.prog_->set_start_unanchored(all.begin); |
1238 | |
|
1239 | 0 | Prog* prog = c.Finish(re); |
1240 | 0 | if (prog == NULL) |
1241 | 0 | return NULL; |
1242 | | |
1243 | | // Make sure DFA has enough memory to operate, |
1244 | | // since we're not going to fall back to the NFA. |
1245 | 0 | bool dfa_failed = false; |
1246 | 0 | StringPiece sp = "hello, world"; |
1247 | 0 | prog->SearchDFA(sp, sp, Prog::kAnchored, Prog::kManyMatch, |
1248 | 0 | NULL, &dfa_failed, NULL); |
1249 | 0 | if (dfa_failed) { |
1250 | 0 | delete prog; |
1251 | 0 | return NULL; |
1252 | 0 | } |
1253 | | |
1254 | 0 | return prog; |
1255 | 0 | } |
1256 | | |
1257 | 0 | Prog* Prog::CompileSet(Regexp* re, RE2::Anchor anchor, int64_t max_mem) { |
1258 | 0 | return Compiler::CompileSet(re, anchor, max_mem); |
1259 | 0 | } |
1260 | | |
1261 | | } // namespace re2 |