Coverage for /pythoncovmergedfiles/medio/medio/usr/local/lib/python3.11/site-packages/pygments/lexers/sql.py: 52%
Shortcuts on this page
r m x toggle line displays
j k next/prev highlighted chunk
0 (zero) top of page
1 (one) first highlighted chunk
Shortcuts on this page
r m x toggle line displays
j k next/prev highlighted chunk
0 (zero) top of page
1 (one) first highlighted chunk
1"""
2 pygments.lexers.sql
3 ~~~~~~~~~~~~~~~~~~~
5 Lexers for various SQL dialects and related interactive sessions.
7 Postgres specific lexers:
9 `PostgresLexer`
10 A SQL lexer for the PostgreSQL dialect. Differences w.r.t. the SQL
11 lexer are:
13 - keywords and data types list parsed from the PG docs (run the
14 `_postgres_builtins` module to update them);
15 - Content of $-strings parsed using a specific lexer, e.g. the content
16 of a PL/Python function is parsed using the Python lexer;
17 - parse PG specific constructs: E-strings, $-strings, U&-strings,
18 different operators and punctuation.
20 `PlPgsqlLexer`
21 A lexer for the PL/pgSQL language. Adds a few specific construct on
22 top of the PG SQL lexer (such as <<label>>).
24 `PostgresConsoleLexer`
25 A lexer to highlight an interactive psql session:
27 - identifies the prompt and does its best to detect the end of command
28 in multiline statement where not all the lines are prefixed by a
29 prompt, telling them apart from the output;
30 - highlights errors in the output and notification levels;
31 - handles psql backslash commands.
33 `PostgresExplainLexer`
34 A lexer to highlight Postgres execution plan.
36 The ``tests/examplefiles`` contains a few test files with data to be
37 parsed by these lexers.
39 :copyright: Copyright 2006-2025 by the Pygments team, see AUTHORS.
40 :license: BSD, see LICENSE for details.
41"""
43import collections
44import re
46from pygments.lexer import Lexer, RegexLexer, do_insertions, bygroups, words
47from pygments.lexers import _googlesql_builtins
48from pygments.lexers import _mysql_builtins
49from pygments.lexers import _postgres_builtins
50from pygments.lexers import _sql_builtins
51from pygments.lexers import _tsql_builtins
52from pygments.lexers import get_lexer_by_name, ClassNotFound
53from pygments.token import Punctuation, Whitespace, Text, Comment, Operator, \
54 Keyword, Name, String, Number, Generic, Literal
57__all__ = ['GoogleSqlLexer', 'PostgresLexer', 'PlPgsqlLexer',
58 'PostgresConsoleLexer', 'PostgresExplainLexer', 'SqlLexer',
59 'TransactSqlLexer', 'MySqlLexer', 'SqliteConsoleLexer', 'RqlLexer']
61line_re = re.compile('.*?\n')
62sqlite_prompt_re = re.compile(r'^(?:sqlite| ...)>(?= )')
64language_re = re.compile(r"\s+LANGUAGE\s+'?(\w+)'?", re.IGNORECASE)
66do_re = re.compile(r'\bDO\b', re.IGNORECASE)
68# Regular expressions for analyse_text()
69name_between_bracket_re = re.compile(r'\[[a-zA-Z_]\w*\]')
70name_between_backtick_re = re.compile(r'`[a-zA-Z_]\w*`')
71tsql_go_re = re.compile(r'\bgo\b', re.IGNORECASE)
72tsql_declare_re = re.compile(r'\bdeclare\s+@', re.IGNORECASE)
73tsql_variable_re = re.compile(r'@[a-zA-Z_]\w*\b')
75# Identifiers for analyse_text()
76googlesql_identifiers = (
77 _googlesql_builtins.functionnames
78 + _googlesql_builtins.keywords
79 + _googlesql_builtins.types)
82def language_callback(lexer, match):
83 """Parse the content of a $-string using a lexer
85 The lexer is chosen looking for a nearby LANGUAGE or assumed as
86 plpgsql if inside a DO statement and no LANGUAGE has been found.
87 """
88 lx = None
89 m = language_re.match(lexer.text[match.end():match.end()+100])
90 if m is not None:
91 lx = lexer._get_lexer(m.group(1))
92 else:
93 m = list(language_re.finditer(
94 lexer.text[max(0, match.start()-100):match.start()]))
95 if m:
96 lx = lexer._get_lexer(m[-1].group(1))
97 else:
98 m = list(do_re.finditer(
99 lexer.text[max(0, match.start()-25):match.start()]))
100 if m:
101 lx = lexer._get_lexer('plpgsql')
103 # 1 = $, 2 = delimiter, 3 = $
104 yield (match.start(1), String, match.group(1))
105 yield (match.start(2), String.Delimiter, match.group(2))
106 yield (match.start(3), String, match.group(3))
107 # 4 = string contents
108 if lx:
109 yield from lx.get_tokens_unprocessed(match.group(4))
110 else:
111 yield (match.start(4), String, match.group(4))
112 # 5 = $, 6 = delimiter, 7 = $
113 yield (match.start(5), String, match.group(5))
114 yield (match.start(6), String.Delimiter, match.group(6))
115 yield (match.start(7), String, match.group(7))
118class PostgresBase:
119 """Base class for Postgres-related lexers.
121 This is implemented as a mixin to avoid the Lexer metaclass kicking in.
122 this way the different lexer don't have a common Lexer ancestor. If they
123 had, _tokens could be created on this ancestor and not updated for the
124 other classes, resulting e.g. in PL/pgSQL parsed as SQL. This shortcoming
125 seem to suggest that regexp lexers are not really subclassable.
126 """
127 def get_tokens_unprocessed(self, text, *args):
128 # Have a copy of the entire text to be used by `language_callback`.
129 self.text = text
130 yield from super().get_tokens_unprocessed(text, *args)
132 def _get_lexer(self, lang):
133 if lang.lower() == 'sql':
134 return get_lexer_by_name('postgresql', **self.options)
136 tries = [lang]
137 if lang.startswith('pl'):
138 tries.append(lang[2:])
139 if lang.endswith('u'):
140 tries.append(lang[:-1])
141 if lang.startswith('pl') and lang.endswith('u'):
142 tries.append(lang[2:-1])
144 for lx in tries:
145 try:
146 return get_lexer_by_name(lx, **self.options)
147 except ClassNotFound:
148 pass
149 else:
150 # TODO: better logging
151 # print >>sys.stderr, "language not found:", lang
152 return None
155class PostgresLexer(PostgresBase, RegexLexer):
156 """
157 Lexer for the PostgreSQL dialect of SQL.
158 """
160 name = 'PostgreSQL SQL dialect'
161 aliases = ['postgresql', 'postgres']
162 mimetypes = ['text/x-postgresql']
163 url = 'https://www.postgresql.org'
164 version_added = '1.5'
166 flags = re.IGNORECASE
167 tokens = {
168 'root': [
169 (r'\s+', Whitespace),
170 (r'--.*\n?', Comment.Single),
171 (r'/\*', Comment.Multiline, 'multiline-comments'),
172 (r'(' + '|'.join(s.replace(" ", r"\s+")
173 for s in _postgres_builtins.DATATYPES +
174 _postgres_builtins.PSEUDO_TYPES) + r')\b',
175 Name.Builtin),
176 (words(_postgres_builtins.KEYWORDS, suffix=r'\b'), Keyword),
177 (r'[+*/<>=~!@#%^&|`?-]+', Operator),
178 (r'::', Operator), # cast
179 (r'\$\d+', Name.Variable),
180 (r'([0-9]*\.[0-9]*|[0-9]+)(e[+-]?[0-9]+)?', Number.Float),
181 (r'[0-9]+', Number.Integer),
182 (r"((?:E|U&)?)(')", bygroups(String.Affix, String.Single), 'string'),
183 # quoted identifier
184 (r'((?:U&)?)(")', bygroups(String.Affix, String.Name), 'quoted-ident'),
185 (r'(?s)(\$)([^$]*)(\$)(.*?)(\$)(\2)(\$)', language_callback),
186 (r'[a-z_]\w*', Name),
188 # psql variable in SQL
189 (r""":(['"]?)[a-z]\w*\b\1""", Name.Variable),
191 (r'[;:()\[\]{},.]', Punctuation),
192 ],
193 'multiline-comments': [
194 (r'/\*', Comment.Multiline, 'multiline-comments'),
195 (r'\*/', Comment.Multiline, '#pop'),
196 (r'[^/*]+', Comment.Multiline),
197 (r'[/*]', Comment.Multiline)
198 ],
199 'string': [
200 (r"[^']+", String.Single),
201 (r"''", String.Single),
202 (r"'", String.Single, '#pop'),
203 ],
204 'quoted-ident': [
205 (r'[^"]+', String.Name),
206 (r'""', String.Name),
207 (r'"', String.Name, '#pop'),
208 ],
209 }
212class PlPgsqlLexer(PostgresBase, RegexLexer):
213 """
214 Handle the extra syntax in Pl/pgSQL language.
215 """
216 name = 'PL/pgSQL'
217 aliases = ['plpgsql']
218 mimetypes = ['text/x-plpgsql']
219 url = 'https://www.postgresql.org/docs/current/plpgsql.html'
220 version_added = '1.5'
222 flags = re.IGNORECASE
223 # FIXME: use inheritance
224 tokens = {name: state[:] for (name, state) in PostgresLexer.tokens.items()}
226 # extend the keywords list
227 for i, pattern in enumerate(tokens['root']):
228 if pattern[1] == Keyword:
229 tokens['root'][i] = (
230 words(_postgres_builtins.KEYWORDS +
231 _postgres_builtins.PLPGSQL_KEYWORDS, suffix=r'\b'),
232 Keyword)
233 del i
234 break
235 else:
236 assert 0, "SQL keywords not found"
238 # Add specific PL/pgSQL rules (before the SQL ones)
239 tokens['root'][:0] = [
240 (r'\%[a-z]\w*\b', Name.Builtin), # actually, a datatype
241 (r':=', Operator),
242 (r'\<\<[a-z]\w*\>\>', Name.Label),
243 (r'\#[a-z]\w*\b', Keyword.Pseudo), # #variable_conflict
244 ]
247class PsqlRegexLexer(PostgresBase, RegexLexer):
248 """
249 Extend the PostgresLexer adding support specific for psql commands.
251 This is not a complete psql lexer yet as it lacks prompt support
252 and output rendering.
253 """
255 name = 'PostgreSQL console - regexp based lexer'
256 aliases = [] # not public
258 flags = re.IGNORECASE
259 tokens = {name: state[:] for (name, state) in PostgresLexer.tokens.items()}
261 tokens['root'].append(
262 (r'\\[^\s]+', Keyword.Pseudo, 'psql-command'))
263 tokens['psql-command'] = [
264 (r'\n', Text, 'root'),
265 (r'\s+', Whitespace),
266 (r'\\[^\s]+', Keyword.Pseudo),
267 (r""":(['"]?)[a-z]\w*\b\1""", Name.Variable),
268 (r"'(''|[^'])*'", String.Single),
269 (r"`([^`])*`", String.Backtick),
270 (r"[^\s]+", String.Symbol),
271 ]
274re_prompt = re.compile(r'^(\S.*?)??[=\-\(\$\'\"][#>]')
275re_psql_command = re.compile(r'\s*\\')
276re_end_command = re.compile(r';\s*(--.*?)?$')
277re_psql_command = re.compile(r'(\s*)(\\.+?)(\s+)$')
278re_error = re.compile(r'(ERROR|FATAL):')
279re_message = re.compile(
280 r'((?:DEBUG|INFO|NOTICE|WARNING|ERROR|'
281 r'FATAL|HINT|DETAIL|CONTEXT|LINE [0-9]+):)(.*?\n)')
284class lookahead:
285 """Wrap an iterator and allow pushing back an item."""
286 def __init__(self, x):
287 self.iter = iter(x)
288 self._nextitem = None
290 def __iter__(self):
291 return self
293 def send(self, i):
294 self._nextitem = i
295 return i
297 def __next__(self):
298 if self._nextitem is not None:
299 ni = self._nextitem
300 self._nextitem = None
301 return ni
302 return next(self.iter)
303 next = __next__
306class PostgresConsoleLexer(Lexer):
307 """
308 Lexer for psql sessions.
309 """
311 name = 'PostgreSQL console (psql)'
312 aliases = ['psql', 'postgresql-console', 'postgres-console']
313 mimetypes = ['text/x-postgresql-psql']
314 url = 'https://www.postgresql.org'
315 version_added = '1.5'
316 _example = "psql/psql_session.txt"
318 def get_tokens_unprocessed(self, data):
319 sql = PsqlRegexLexer(**self.options)
321 lines = lookahead(line_re.findall(data))
323 # prompt-output cycle
324 while 1:
326 # consume the lines of the command: start with an optional prompt
327 # and continue until the end of command is detected
328 curcode = ''
329 insertions = []
330 for line in lines:
331 # Identify a shell prompt in case of psql commandline example
332 if line.startswith('$') and not curcode:
333 lexer = get_lexer_by_name('console', **self.options)
334 yield from lexer.get_tokens_unprocessed(line)
335 break
337 # Identify a psql prompt
338 mprompt = re_prompt.match(line)
339 if mprompt is not None:
340 insertions.append((len(curcode),
341 [(0, Generic.Prompt, mprompt.group())]))
342 curcode += line[len(mprompt.group()):]
343 else:
344 curcode += line
346 # Check if this is the end of the command
347 # TODO: better handle multiline comments at the end with
348 # a lexer with an external state?
349 if re_psql_command.match(curcode) \
350 or re_end_command.search(curcode):
351 break
353 # Emit the combined stream of command and prompt(s)
354 yield from do_insertions(insertions,
355 sql.get_tokens_unprocessed(curcode))
357 # Emit the output lines
358 out_token = Generic.Output
359 for line in lines:
360 mprompt = re_prompt.match(line)
361 if mprompt is not None:
362 # push the line back to have it processed by the prompt
363 lines.send(line)
364 break
366 mmsg = re_message.match(line)
367 if mmsg is not None:
368 if mmsg.group(1).startswith("ERROR") \
369 or mmsg.group(1).startswith("FATAL"):
370 out_token = Generic.Error
371 yield (mmsg.start(1), Generic.Strong, mmsg.group(1))
372 yield (mmsg.start(2), out_token, mmsg.group(2))
373 else:
374 yield (0, out_token, line)
375 else:
376 return
379class PostgresExplainLexer(RegexLexer):
380 """
381 Handle PostgreSQL EXPLAIN output
382 """
384 name = 'PostgreSQL EXPLAIN dialect'
385 aliases = ['postgres-explain']
386 filenames = ['*.explain']
387 mimetypes = ['text/x-postgresql-explain']
388 url = 'https://www.postgresql.org/docs/current/using-explain.html'
389 version_added = '2.15'
391 tokens = {
392 'root': [
393 (r'(:|\(|\)|ms|kB|->|\.\.|\,|\/)', Punctuation),
394 (r'(\s+)', Whitespace),
396 # This match estimated cost and effectively measured counters with ANALYZE
397 # Then, we move to instrumentation state
398 (r'(cost)(=?)', bygroups(Name.Class, Punctuation), 'instrumentation'),
399 (r'(actual)( )(=?)', bygroups(Name.Class, Whitespace, Punctuation), 'instrumentation'),
401 # Misc keywords
402 (words(('actual', 'Memory Usage', 'Disk Usage', 'Memory', 'Buckets', 'Batches',
403 'originally', 'row', 'rows', 'Hits', 'Misses',
404 'Evictions', 'Overflows', 'Planned Partitions'), suffix=r'\b'),
405 Comment.Single),
407 (r'(hit|read|dirtied|written|write|time|calls)(=)', bygroups(Comment.Single, Operator)),
408 (r'(shared|temp|local)', Keyword.Pseudo),
410 # We move to sort state in order to emphasize specific keywords (especially disk access)
411 (r'(Sort Method)(: )', bygroups(Comment.Preproc, Punctuation), 'sort'),
413 # These keywords can be followed by an object, like a table
414 (r'(Sort Key|Group Key|Presorted Key|Hash Key)(:)( )',
415 bygroups(Comment.Preproc, Punctuation, Whitespace), 'object_name'),
416 (r'(Cache Key|Cache Mode)(:)( )', bygroups(Comment, Punctuation, Whitespace), 'object_name'),
418 # These keywords can be followed by a predicate
419 (words(('Join Filter', 'Subplans Removed', 'Filter', 'Merge Cond',
420 'Hash Cond', 'Index Cond', 'Recheck Cond', 'Heap Blocks',
421 'TID Cond', 'Run Condition', 'Order By', 'Function Call',
422 'Table Function Call', 'Inner Unique', 'Params Evaluated',
423 'Single Copy', 'Sampling', 'One-Time Filter', 'Output',
424 'Relations', 'Remote SQL'), suffix=r'\b'),
425 Comment.Preproc, 'predicate'),
427 # Special keyword to handle ON CONFLICT
428 (r'Conflict ', Comment.Preproc, 'conflict'),
430 # Special keyword for InitPlan or SubPlan
431 (r'(InitPlan|SubPlan)( )(\d+)( )',
432 bygroups(Keyword, Whitespace, Number.Integer, Whitespace),
433 'init_plan'),
435 (words(('Sort Method', 'Join Filter', 'Planning time',
436 'Planning Time', 'Execution time', 'Execution Time',
437 'Workers Planned', 'Workers Launched', 'Buffers',
438 'Planning', 'Worker', 'Query Identifier', 'Time',
439 'Full-sort Groups', 'Pre-sorted Groups'), suffix=r'\b'), Comment.Preproc),
441 # Emphasize these keywords
443 (words(('Rows Removed by Join Filter', 'Rows Removed by Filter',
444 'Rows Removed by Index Recheck',
445 'Heap Fetches', 'never executed'),
446 suffix=r'\b'), Name.Exception),
447 (r'(I/O Timings)(:)( )', bygroups(Name.Exception, Punctuation, Whitespace)),
449 (words(_postgres_builtins.EXPLAIN_KEYWORDS, suffix=r'\b'), Keyword),
451 # join keywords
452 (r'((Right|Left|Full|Semi|Anti) Join)', Keyword.Type),
453 (r'(Parallel |Async |Finalize |Partial )', Comment.Preproc),
454 (r'Backward', Comment.Preproc),
455 (r'(Intersect|Except|Hash)', Comment.Preproc),
457 (r'(CTE)( )(\w*)?', bygroups(Comment, Whitespace, Name.Variable)),
460 # Treat "on" and "using" as a punctuation
461 (r'(on|using)', Punctuation, 'object_name'),
464 # strings
465 (r"'(''|[^'])*'", String.Single),
466 # numbers
467 (r'-?\d+\.\d+', Number.Float),
468 (r'(-?\d+)', Number.Integer),
470 # boolean
471 (r'(true|false)', Name.Constant),
472 # explain header
473 (r'\s*QUERY PLAN\s*\n\s*-+', Comment.Single),
474 # Settings
475 (r'(Settings)(:)( )', bygroups(Comment.Preproc, Punctuation, Whitespace), 'setting'),
477 # Handle JIT counters
478 (r'(JIT|Functions|Options|Timing)(:)', bygroups(Comment.Preproc, Punctuation)),
479 (r'(Inlining|Optimization|Expressions|Deforming|Generation|Emission|Total)', Keyword.Pseudo),
481 # Handle Triggers counters
482 (r'(Trigger)( )(\S*)(:)( )',
483 bygroups(Comment.Preproc, Whitespace, Name.Variable, Punctuation, Whitespace)),
485 ],
486 'expression': [
487 # matches any kind of parenthesized expression
488 # the first opening paren is matched by the 'caller'
489 (r'\(', Punctuation, '#push'),
490 (r'\)', Punctuation, '#pop'),
491 (r'(never executed)', Name.Exception),
492 (r'[^)(]+', Comment),
493 ],
494 'object_name': [
496 # This is a cost or analyze measure
497 (r'(\(cost)(=?)', bygroups(Name.Class, Punctuation), 'instrumentation'),
498 (r'(\(actual)( )(=?)', bygroups(Name.Class, Whitespace, Punctuation), 'instrumentation'),
500 # if object_name is parenthesized, mark opening paren as
501 # punctuation, call 'expression', and exit state
502 (r'\(', Punctuation, 'expression'),
503 (r'(on)', Punctuation),
504 # matches possibly schema-qualified table and column names
505 (r'\w+(\.\w+)*( USING \S+| \w+ USING \S+)', Name.Variable),
506 (r'\"?\w+\"?(?:\.\"?\w+\"?)?', Name.Variable),
507 (r'\'\S*\'', Name.Variable),
509 # if we encounter a comma, another object is listed
510 (r',\n', Punctuation, 'object_name'),
511 (r',', Punctuation, 'object_name'),
513 # special case: "*SELECT*"
514 (r'"\*SELECT\*( \d+)?"(.\w+)?', Name.Variable),
515 (r'"\*VALUES\*(_\d+)?"(.\w+)?', Name.Variable),
516 (r'"ANY_subquery"', Name.Variable),
518 # Variable $1 ...
519 (r'\$\d+', Name.Variable),
520 # cast
521 (r'::\w+', Name.Variable),
522 (r' +', Whitespace),
523 (r'"', Punctuation),
524 (r'\[\.\.\.\]', Punctuation),
525 (r'\)', Punctuation, '#pop'),
526 ],
527 'predicate': [
528 # if predicate is parenthesized, mark paren as punctuation
529 (r'(\()([^\n]*)(\))', bygroups(Punctuation, Name.Variable, Punctuation), '#pop'),
530 # otherwise color until newline
531 (r'[^\n]*', Name.Variable, '#pop'),
532 ],
533 'instrumentation': [
534 (r'=|\.\.', Punctuation),
535 (r' +', Whitespace),
536 (r'(rows|width|time|loops)', Name.Class),
537 (r'\d+\.\d+', Number.Float),
538 (r'(\d+)', Number.Integer),
539 (r'\)', Punctuation, '#pop'),
540 ],
541 'conflict': [
542 (r'(Resolution: )(\w+)', bygroups(Comment.Preproc, Name.Variable)),
543 (r'(Arbiter \w+:)', Comment.Preproc, 'object_name'),
544 (r'(Filter: )', Comment.Preproc, 'predicate'),
545 ],
546 'setting': [
547 (r'([a-z_]*?)(\s*)(=)(\s*)(\'.*?\')', bygroups(Name.Attribute, Whitespace, Operator, Whitespace, String)),
548 (r'\, ', Punctuation),
549 ],
550 'init_plan': [
551 (r'\(', Punctuation),
552 (r'returns \$\d+(,\$\d+)?', Name.Variable),
553 (r'\)', Punctuation, '#pop'),
554 ],
555 'sort': [
556 (r':|kB', Punctuation),
557 (r'(quicksort|top-N|heapsort|Average|Memory|Peak)', Comment.Prepoc),
558 (r'(external|merge|Disk|sort)', Name.Exception),
559 (r'(\d+)', Number.Integer),
560 (r' +', Whitespace),
561 ],
562 }
565class SqlLexer(RegexLexer):
566 """
567 Lexer for Structured Query Language. Currently, this lexer does
568 not recognize any special syntax except ANSI SQL.
569 """
571 name = 'SQL'
572 aliases = ['sql']
573 filenames = ['*.sql']
574 mimetypes = ['text/x-sql']
575 url = 'https://en.wikipedia.org/wiki/SQL'
576 version_added = ''
578 flags = re.IGNORECASE
579 tokens = {
580 'root': [
581 (r'\s+', Whitespace),
582 (r'--.*\n?', Comment.Single),
583 (r'/\*', Comment.Multiline, 'multiline-comments'),
584 (words(_sql_builtins.KEYWORDS, suffix=r'\b'), Keyword),
585 (words(_sql_builtins.DATATYPES, suffix=r'\b'), Name.Builtin),
586 (r'[+*/<>=~!@#%^&|`?-]', Operator),
587 (r'[0-9]+', Number.Integer),
588 # TODO: Backslash escapes?
589 (r"'(''|[^'])*'", String.Single),
590 (r'"(""|[^"])*"', String.Symbol), # not a real string literal in ANSI SQL
591 (r'[a-z_][\w$]*', Name), # allow $s in strings for Oracle
592 (r'[;:()\[\],.]', Punctuation)
593 ],
594 'multiline-comments': [
595 (r'/\*', Comment.Multiline, 'multiline-comments'),
596 (r'\*/', Comment.Multiline, '#pop'),
597 (r'[^/*]+', Comment.Multiline),
598 (r'[/*]', Comment.Multiline)
599 ]
600 }
602 def analyse_text(self, text):
603 return
606class TransactSqlLexer(RegexLexer):
607 """
608 Transact-SQL (T-SQL) is Microsoft's and Sybase's proprietary extension to
609 SQL.
611 The list of keywords includes ODBC and keywords reserved for future use.
612 """
614 name = 'Transact-SQL'
615 aliases = ['tsql', 't-sql']
616 filenames = ['*.sql']
617 mimetypes = ['text/x-tsql']
618 url = 'https://www.tsql.info'
619 version_added = ''
621 flags = re.IGNORECASE
623 tokens = {
624 'root': [
625 (r'\s+', Whitespace),
626 (r'--.*[$|\n]?', Comment.Single),
627 (r'/\*', Comment.Multiline, 'multiline-comments'),
628 (words(_tsql_builtins.OPERATORS), Operator),
629 (words(_tsql_builtins.OPERATOR_WORDS, suffix=r'\b'), Operator.Word),
630 (words(_tsql_builtins.TYPES, suffix=r'\b'), Name.Class),
631 (words(_tsql_builtins.FUNCTIONS, suffix=r'\b'), Name.Function),
632 (r'(goto)(\s+)(\w+\b)', bygroups(Keyword, Whitespace, Name.Label)),
633 (words(_tsql_builtins.KEYWORDS, suffix=r'\b'), Keyword),
634 (r'(\[)([^]]+)(\])', bygroups(Operator, Name, Operator)),
635 (r'0x[0-9a-f]+', Number.Hex),
636 # Float variant 1, for example: 1., 1.e2, 1.2e3
637 (r'[0-9]+\.[0-9]*(e[+-]?[0-9]+)?', Number.Float),
638 # Float variant 2, for example: .1, .1e2
639 (r'\.[0-9]+(e[+-]?[0-9]+)?', Number.Float),
640 # Float variant 3, for example: 123e45
641 (r'[0-9]+e[+-]?[0-9]+', Number.Float),
642 (r'[0-9]+', Number.Integer),
643 (r"'(''|[^'])*'", String.Single),
644 (r'"(""|[^"])*"', String.Symbol),
645 (r'[;(),.]', Punctuation),
646 # Below we use \w even for the first "real" character because
647 # tokens starting with a digit have already been recognized
648 # as Number above.
649 (r'@@\w+', Name.Builtin),
650 (r'@\w+', Name.Variable),
651 (r'(\w+)(:)', bygroups(Name.Label, Punctuation)),
652 (r'#?#?\w+', Name), # names for temp tables and anything else
653 (r'\?', Name.Variable.Magic), # parameter for prepared statements
654 ],
655 'multiline-comments': [
656 (r'/\*', Comment.Multiline, 'multiline-comments'),
657 (r'\*/', Comment.Multiline, '#pop'),
658 (r'[^/*]+', Comment.Multiline),
659 (r'[/*]', Comment.Multiline)
660 ]
661 }
663 def analyse_text(text):
664 rating = 0
665 if tsql_declare_re.search(text):
666 # Found T-SQL variable declaration.
667 rating = 1.0
668 else:
669 name_between_backtick_count = len(
670 name_between_backtick_re.findall(text))
671 name_between_bracket_count = len(
672 name_between_bracket_re.findall(text))
673 # We need to check if there are any names using
674 # backticks or brackets, as otherwise both are 0
675 # and 0 >= 2 * 0, so we would always assume it's true
676 dialect_name_count = name_between_backtick_count + name_between_bracket_count
677 if dialect_name_count >= 1 and \
678 name_between_bracket_count >= 2 * name_between_backtick_count:
679 # Found at least twice as many [name] as `name`.
680 rating += 0.5
681 elif name_between_bracket_count > name_between_backtick_count:
682 rating += 0.2
683 elif name_between_bracket_count > 0:
684 rating += 0.1
685 if tsql_variable_re.search(text) is not None:
686 rating += 0.1
687 if tsql_go_re.search(text) is not None:
688 rating += 0.1
689 return rating
692class MySqlLexer(RegexLexer):
693 """The Oracle MySQL lexer.
695 This lexer does not attempt to maintain strict compatibility with
696 MariaDB syntax or keywords. Although MySQL and MariaDB's common code
697 history suggests there may be significant overlap between the two,
698 compatibility between the two is not a target for this lexer.
699 """
701 name = 'MySQL'
702 aliases = ['mysql']
703 mimetypes = ['text/x-mysql']
704 url = 'https://www.mysql.com'
705 version_added = ''
707 flags = re.IGNORECASE
708 tokens = {
709 'root': [
710 (r'\s+', Whitespace),
712 # Comments
713 (r'(?:#|--\s+).*', Comment.Single),
714 (r'/\*\+', Comment.Special, 'optimizer-hints'),
715 (r'/\*', Comment.Multiline, 'multiline-comment'),
717 # Hexadecimal literals
718 (r"x'([0-9a-f]{2})+'", Number.Hex), # MySQL requires paired hex characters in this form.
719 (r'0x[0-9a-f]+', Number.Hex),
721 # Binary literals
722 (r"b'[01]+'", Number.Bin),
723 (r'0b[01]+', Number.Bin),
725 # Numeric literals
726 (r'[0-9]+\.[0-9]*(e[+-]?[0-9]+)?', Number.Float), # Mandatory integer, optional fraction and exponent
727 (r'[0-9]*\.[0-9]+(e[+-]?[0-9]+)?', Number.Float), # Mandatory fraction, optional integer and exponent
728 (r'[0-9]+e[+-]?[0-9]+', Number.Float), # Exponents with integer significands are still floats
729 (r'[0-9]+(?=[^0-9a-z$_\u0080-\uffff])', Number.Integer), # Integers that are not in a schema object name
731 # Date literals
732 (r"\{\s*d\s*(?P<quote>['\"])\s*\d{2}(\d{2})?.?\d{2}.?\d{2}\s*(?P=quote)\s*\}",
733 Literal.Date),
735 # Time literals
736 (r"\{\s*t\s*(?P<quote>['\"])\s*(?:\d+\s+)?\d{1,2}.?\d{1,2}.?\d{1,2}(\.\d*)?\s*(?P=quote)\s*\}",
737 Literal.Date),
739 # Timestamp literals
740 (
741 r"\{\s*ts\s*(?P<quote>['\"])\s*"
742 r"\d{2}(?:\d{2})?.?\d{2}.?\d{2}" # Date part
743 r"\s+" # Whitespace between date and time
744 r"\d{1,2}.?\d{1,2}.?\d{1,2}(\.\d*)?" # Time part
745 r"\s*(?P=quote)\s*\}",
746 Literal.Date
747 ),
749 # String literals
750 (r"'", String.Single, 'single-quoted-string'),
751 (r'"', String.Double, 'double-quoted-string'),
753 # Variables
754 (r'@@(?:global\.|persist\.|persist_only\.|session\.)?[a-z_]+', Name.Variable),
755 (r'@[a-z0-9_$.]+', Name.Variable),
756 (r"@'", Name.Variable, 'single-quoted-variable'),
757 (r'@"', Name.Variable, 'double-quoted-variable'),
758 (r"@`", Name.Variable, 'backtick-quoted-variable'),
759 (r'\?', Name.Variable), # For demonstrating prepared statements
761 # Operators
762 (r'[!%&*+/:<=>^|~-]+', Operator),
764 # Exceptions; these words tokenize differently in different contexts.
765 (r'\b(set)(?!\s*\()', Keyword),
766 (r'\b(character)(\s+)(set)\b', bygroups(Keyword, Whitespace, Keyword)),
767 # In all other known cases, "SET" is tokenized by MYSQL_DATATYPES.
769 (words(_mysql_builtins.MYSQL_CONSTANTS, prefix=r'\b', suffix=r'\b'),
770 Name.Constant),
771 (words(_mysql_builtins.MYSQL_DATATYPES, prefix=r'\b', suffix=r'\b'),
772 Keyword.Type),
773 (words(_mysql_builtins.MYSQL_KEYWORDS, prefix=r'\b', suffix=r'\b'),
774 Keyword),
775 (words(_mysql_builtins.MYSQL_FUNCTIONS, prefix=r'\b', suffix=r'\b(\s*)(\()'),
776 bygroups(Name.Function, Whitespace, Punctuation)),
778 # Schema object names
779 #
780 # Note: Although the first regex supports unquoted all-numeric
781 # identifiers, this will not be a problem in practice because
782 # numeric literals have already been handled above.
783 #
784 ('[0-9a-z$_\u0080-\uffff]+', Name),
785 (r'`', Name.Quoted, 'schema-object-name'),
787 # Punctuation
788 (r'[(),.;]', Punctuation),
789 ],
791 # Multiline comment substates
792 # ---------------------------
794 'optimizer-hints': [
795 (r'[^*a-z]+', Comment.Special),
796 (r'\*/', Comment.Special, '#pop'),
797 (words(_mysql_builtins.MYSQL_OPTIMIZER_HINTS, suffix=r'\b'),
798 Comment.Preproc),
799 ('[a-z]+', Comment.Special),
800 (r'\*', Comment.Special),
801 ],
803 'multiline-comment': [
804 (r'[^*]+', Comment.Multiline),
805 (r'\*/', Comment.Multiline, '#pop'),
806 (r'\*', Comment.Multiline),
807 ],
809 # String substates
810 # ----------------
812 'single-quoted-string': [
813 (r"[^'\\]+", String.Single),
814 (r"''", String.Escape),
815 (r"""\\[0'"bnrtZ\\%_]""", String.Escape),
816 (r"'", String.Single, '#pop'),
817 ],
819 'double-quoted-string': [
820 (r'[^"\\]+', String.Double),
821 (r'""', String.Escape),
822 (r"""\\[0'"bnrtZ\\%_]""", String.Escape),
823 (r'"', String.Double, '#pop'),
824 ],
826 # Variable substates
827 # ------------------
829 'single-quoted-variable': [
830 (r"[^']+", Name.Variable),
831 (r"''", Name.Variable),
832 (r"'", Name.Variable, '#pop'),
833 ],
835 'double-quoted-variable': [
836 (r'[^"]+', Name.Variable),
837 (r'""', Name.Variable),
838 (r'"', Name.Variable, '#pop'),
839 ],
841 'backtick-quoted-variable': [
842 (r'[^`]+', Name.Variable),
843 (r'``', Name.Variable),
844 (r'`', Name.Variable, '#pop'),
845 ],
847 # Schema object name substates
848 # ----------------------------
849 #
850 # "Name.Quoted" and "Name.Quoted.Escape" are non-standard but
851 # formatters will style them as "Name" by default but add
852 # additional styles based on the token name. This gives users
853 # flexibility to add custom styles as desired.
854 #
855 'schema-object-name': [
856 (r'[^`]+', Name.Quoted),
857 (r'``', Name.Quoted.Escape),
858 (r'`', Name.Quoted, '#pop'),
859 ],
860 }
862 def analyse_text(text):
863 rating = 0
864 name_between_backtick_count = len(
865 name_between_backtick_re.findall(text))
866 name_between_bracket_count = len(
867 name_between_bracket_re.findall(text))
868 # Same logic as above in the TSQL analysis
869 dialect_name_count = name_between_backtick_count + name_between_bracket_count
870 if dialect_name_count >= 1 and \
871 name_between_backtick_count >= 2 * name_between_bracket_count:
872 # Found at least twice as many `name` as [name].
873 rating += 0.5
874 elif name_between_backtick_count > name_between_bracket_count:
875 rating += 0.2
876 elif name_between_backtick_count > 0:
877 rating += 0.1
878 return rating
881class GoogleSqlLexer(RegexLexer):
882 """
883 GoogleSQL is Google's standard SQL dialect, formerly known as ZetaSQL.
885 The list of keywords includes reserved words for future use.
886 """
888 name = 'GoogleSQL'
889 aliases = ['googlesql', 'zetasql']
890 filenames = ['*.googlesql', '*.googlesql.sql']
891 mimetypes = ['text/x-google-sql', 'text/x-google-sql-aux']
892 url = 'https://cloud.google.com/bigquery/googlesql'
893 version_added = '2.19'
895 flags = re.IGNORECASE
896 tokens = {
897 'root': [
898 (r'\s+', Whitespace),
900 # Comments
901 (r'(?:#|--\s+).*', Comment.Single),
902 (r'/\*', Comment.Multiline, 'multiline-comment'),
904 # Hexadecimal literals
905 (r"x'([0-9a-f]{2})+'", Number.Hex),
906 (r'0x[0-9a-f]+', Number.Hex),
908 # Binary literals
909 (r"b'[01]+'", Number.Bin),
910 (r'0b[01]+', Number.Bin),
912 # Numeric literals
913 (r'[0-9]+\.[0-9]*(e[+-]?[0-9]+)?', Number.Float), # Mandatory integer, optional fraction and exponent
914 (r'[0-9]*\.[0-9]+(e[+-]?[0-9]+)?', Number.Float), # Mandatory fraction, optional integer and exponent
915 (r'[0-9]+e[+-]?[0-9]+', Number.Float), # Exponents with integer significands are still floats
916 (r'[0-9]+(?=[^0-9a-z$_\u0080-\uffff])', Number.Integer), # Integers that are not in a schema object name
918 # Date literals
919 (r"\{\s*d\s*(?P<quote>['\"])\s*\d{2}(\d{2})?.?\d{2}.?\d{2}\s*(?P=quote)\s*\}",
920 Literal.Date),
922 # Time literals
923 (r"\{\s*t\s*(?P<quote>['\"])\s*(?:\d+\s+)?\d{1,2}.?\d{1,2}.?\d{1,2}(\.\d*)?\s*(?P=quote)\s*\}",
924 Literal.Date),
926 # Timestamp literals
927 (
928 r"\{\s*ts\s*(?P<quote>['\"])\s*"
929 r"\d{2}(?:\d{2})?.?\d{2}.?\d{2}" # Date part
930 r"\s+" # Whitespace between date and time
931 r"\d{1,2}.?\d{1,2}.?\d{1,2}(\.\d*)?" # Time part
932 r"\s*(?P=quote)\s*\}",
933 Literal.Date
934 ),
936 # String literals
937 (r"'", String.Single, 'single-quoted-string'),
938 (r'"', String.Double, 'double-quoted-string'),
940 # Variables
941 (r'@@(?:global\.|persist\.|persist_only\.|session\.)?[a-z_]+', Name.Variable),
942 (r'@[a-z0-9_$.]+', Name.Variable),
943 (r"@'", Name.Variable, 'single-quoted-variable'),
944 (r'@"', Name.Variable, 'double-quoted-variable'),
945 (r"@`", Name.Variable, 'backtick-quoted-variable'),
946 (r'\?', Name.Variable), # For demonstrating prepared statements
948 # Exceptions; these words tokenize differently in different contexts.
949 (r'\b(set)(?!\s*\()', Keyword),
950 (r'\b(character)(\s+)(set)\b', bygroups(Keyword, Whitespace, Keyword)),
952 # Constants, types, keywords, functions, operators
953 (words(_googlesql_builtins.constants, prefix=r'\b', suffix=r'\b'), Name.Constant),
954 (words(_googlesql_builtins.types, prefix=r'\b', suffix=r'\b'), Keyword.Type),
955 (words(_googlesql_builtins.keywords, prefix=r'\b', suffix=r'\b'), Keyword),
956 (words(_googlesql_builtins.functionnames, prefix=r'\b', suffix=r'\b(\s*)(\()'),
957 bygroups(Name.Function, Whitespace, Punctuation)),
958 (words(_googlesql_builtins.operators, prefix=r'\b', suffix=r'\b'), Operator),
960 # Schema object names
961 #
962 # Note: Although the first regex supports unquoted all-numeric
963 # identifiers, this will not be a problem in practice because
964 # numeric literals have already been handled above.
965 #
966 ('[0-9a-z$_\u0080-\uffff]+', Name),
967 (r'`', Name.Quoted, 'schema-object-name'),
969 # Punctuation
970 (r'[(),.;]', Punctuation),
971 ],
973 # Multiline comment substates
974 # ---------------------------
976 'multiline-comment': [
977 (r'[^*]+', Comment.Multiline),
978 (r'\*/', Comment.Multiline, '#pop'),
979 (r'\*', Comment.Multiline),
980 ],
982 # String substates
983 # ----------------
985 'single-quoted-string': [
986 (r"[^'\\]+", String.Single),
987 (r"''", String.Escape),
988 (r"""\\[0'"bnrtZ\\%_]""", String.Escape),
989 (r"'", String.Single, '#pop'),
990 ],
992 'double-quoted-string': [
993 (r'[^"\\]+', String.Double),
994 (r'""', String.Escape),
995 (r"""\\[0'"bnrtZ\\%_]""", String.Escape),
996 (r'"', String.Double, '#pop'),
997 ],
999 # Variable substates
1000 # ------------------
1002 'single-quoted-variable': [
1003 (r"[^']+", Name.Variable),
1004 (r"''", Name.Variable),
1005 (r"'", Name.Variable, '#pop'),
1006 ],
1008 'double-quoted-variable': [
1009 (r'[^"]+', Name.Variable),
1010 (r'""', Name.Variable),
1011 (r'"', Name.Variable, '#pop'),
1012 ],
1014 'backtick-quoted-variable': [
1015 (r'[^`]+', Name.Variable),
1016 (r'``', Name.Variable),
1017 (r'`', Name.Variable, '#pop'),
1018 ],
1020 # Schema object name substates
1021 # ----------------------------
1022 #
1023 # "Name.Quoted" and "Name.Quoted.Escape" are non-standard but
1024 # formatters will style them as "Name" by default but add
1025 # additional styles based on the token name. This gives users
1026 # flexibility to add custom styles as desired.
1027 #
1028 'schema-object-name': [
1029 (r'[^`]+', Name.Quoted),
1030 (r'``', Name.Quoted.Escape),
1031 (r'`', Name.Quoted, '#pop'),
1032 ],
1033 }
1035 def analyse_text(text):
1036 tokens = collections.Counter(text.split())
1037 return 0.001 * sum(count for t, count in tokens.items()
1038 if t in googlesql_identifiers)
1041class SqliteConsoleLexer(Lexer):
1042 """
1043 Lexer for example sessions using sqlite3.
1044 """
1046 name = 'sqlite3con'
1047 aliases = ['sqlite3']
1048 filenames = ['*.sqlite3-console']
1049 mimetypes = ['text/x-sqlite3-console']
1050 url = 'https://www.sqlite.org'
1051 version_added = '0.11'
1052 _example = "sqlite3/sqlite3.sqlite3-console"
1054 def get_tokens_unprocessed(self, data):
1055 sql = SqlLexer(**self.options)
1057 curcode = ''
1058 insertions = []
1059 for match in line_re.finditer(data):
1060 line = match.group()
1061 prompt_match = sqlite_prompt_re.match(line)
1062 if prompt_match is not None:
1063 insertions.append((len(curcode),
1064 [(0, Generic.Prompt, line[:7])]))
1065 insertions.append((len(curcode),
1066 [(7, Whitespace, ' ')]))
1067 curcode += line[8:]
1068 else:
1069 if curcode:
1070 yield from do_insertions(insertions,
1071 sql.get_tokens_unprocessed(curcode))
1072 curcode = ''
1073 insertions = []
1074 if line.startswith('SQL error: '):
1075 yield (match.start(), Generic.Traceback, line)
1076 else:
1077 yield (match.start(), Generic.Output, line)
1078 if curcode:
1079 yield from do_insertions(insertions,
1080 sql.get_tokens_unprocessed(curcode))
1083class RqlLexer(RegexLexer):
1084 """
1085 Lexer for Relation Query Language.
1086 """
1087 name = 'RQL'
1088 url = 'http://www.logilab.org/project/rql'
1089 aliases = ['rql']
1090 filenames = ['*.rql']
1091 mimetypes = ['text/x-rql']
1092 version_added = '2.0'
1094 flags = re.IGNORECASE
1095 tokens = {
1096 'root': [
1097 (r'\s+', Whitespace),
1098 (r'(DELETE|SET|INSERT|UNION|DISTINCT|WITH|WHERE|BEING|OR'
1099 r'|AND|NOT|GROUPBY|HAVING|ORDERBY|ASC|DESC|LIMIT|OFFSET'
1100 r'|TODAY|NOW|TRUE|FALSE|NULL|EXISTS)\b', Keyword),
1101 (r'[+*/<>=%-]', Operator),
1102 (r'(Any|is|instance_of|CWEType|CWRelation)\b', Name.Builtin),
1103 (r'[0-9]+', Number.Integer),
1104 (r'[A-Z_]\w*\??', Name),
1105 (r"'(''|[^'])*'", String.Single),
1106 (r'"(""|[^"])*"', String.Single),
1107 (r'[;:()\[\],.]', Punctuation)
1108 ],
1109 }