You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
307 lines
12 KiB
307 lines
12 KiB
# -*- coding: utf-8 -*-
|
|
"""
|
|
pygments.lexers.prolog
|
|
~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
Lexers for Prolog and Prolog-like languages.
|
|
|
|
:copyright: Copyright 2006-2017 by the Pygments team, see AUTHORS.
|
|
:license: BSD, see LICENSE for details.
|
|
"""
|
|
|
|
import re
|
|
|
|
from pygments.lexer import RegexLexer, bygroups
|
|
from pygments.token import Text, Comment, Operator, Keyword, Name, String, \
|
|
Number, Punctuation
|
|
|
|
__all__ = ['PrologLexer', 'LogtalkLexer']
|
|
|
|
|
|
class PrologLexer(RegexLexer):
|
|
"""
|
|
Lexer for Prolog files.
|
|
"""
|
|
name = 'Prolog'
|
|
aliases = ['prolog']
|
|
filenames = ['*.ecl', '*.prolog', '*.pro', '*.pl']
|
|
mimetypes = ['text/x-prolog']
|
|
|
|
flags = re.UNICODE | re.MULTILINE
|
|
|
|
tokens = {
|
|
'root': [
|
|
(r'^#.*', Comment.Single),
|
|
(r'/\*', Comment.Multiline, 'nested-comment'),
|
|
(r'%.*', Comment.Single),
|
|
# character literal
|
|
(r'0\'.', String.Char),
|
|
(r'0b[01]+', Number.Bin),
|
|
(r'0o[0-7]+', Number.Oct),
|
|
(r'0x[0-9a-fA-F]+', Number.Hex),
|
|
# literal with prepended base
|
|
(r'\d\d?\'[a-zA-Z0-9]+', Number.Integer),
|
|
(r'(\d+\.\d*|\d*\.\d+)([eE][+-]?[0-9]+)?', Number.Float),
|
|
(r'\d+', Number.Integer),
|
|
(r'[\[\](){}|.,;!]', Punctuation),
|
|
(r':-|-->', Punctuation),
|
|
(r'"(?:\\x[0-9a-fA-F]+\\|\\u[0-9a-fA-F]{4}|\\U[0-9a-fA-F]{8}|'
|
|
r'\\[0-7]+\\|\\["\nabcefnrstv]|[^\\"])*"', String.Double),
|
|
(r"'(?:''|[^'])*'", String.Atom), # quoted atom
|
|
# Needs to not be followed by an atom.
|
|
# (r'=(?=\s|[a-zA-Z\[])', Operator),
|
|
(r'is\b', Operator),
|
|
(r'(<|>|=<|>=|==|=:=|=|/|//|\*|\+|-)(?=\s|[a-zA-Z0-9\[])',
|
|
Operator),
|
|
(r'(mod|div|not)\b', Operator),
|
|
(r'_', Keyword), # The don't-care variable
|
|
(r'([a-z]+)(:)', bygroups(Name.Namespace, Punctuation)),
|
|
(u'([a-z\u00c0-\u1fff\u3040-\ud7ff\ue000-\uffef]'
|
|
u'[\\w$\u00c0-\u1fff\u3040-\ud7ff\ue000-\uffef]*)'
|
|
u'(\\s*)(:-|-->)',
|
|
bygroups(Name.Function, Text, Operator)), # function defn
|
|
(u'([a-z\u00c0-\u1fff\u3040-\ud7ff\ue000-\uffef]'
|
|
u'[\\w$\u00c0-\u1fff\u3040-\ud7ff\ue000-\uffef]*)'
|
|
u'(\\s*)(\\()',
|
|
bygroups(Name.Function, Text, Punctuation)),
|
|
(u'[a-z\u00c0-\u1fff\u3040-\ud7ff\ue000-\uffef]'
|
|
u'[\\w$\u00c0-\u1fff\u3040-\ud7ff\ue000-\uffef]*',
|
|
String.Atom), # atom, characters
|
|
# This one includes !
|
|
(u'[#&*+\\-./:<=>?@\\\\^~\u00a1-\u00bf\u2010-\u303f]+',
|
|
String.Atom), # atom, graphics
|
|
(r'[A-Z_]\w*', Name.Variable),
|
|
(u'\\s+|[\u2000-\u200f\ufff0-\ufffe\uffef]', Text),
|
|
],
|
|
'nested-comment': [
|
|
(r'\*/', Comment.Multiline, '#pop'),
|
|
(r'/\*', Comment.Multiline, '#push'),
|
|
(r'[^*/]+', Comment.Multiline),
|
|
(r'[*/]', Comment.Multiline),
|
|
],
|
|
}
|
|
|
|
def analyse_text(text):
|
|
return ':-' in text
|
|
|
|
|
|
class LogtalkLexer(RegexLexer):
|
|
"""
|
|
For `Logtalk <http://logtalk.org/>`_ source code.
|
|
|
|
.. versionadded:: 0.10
|
|
"""
|
|
|
|
name = 'Logtalk'
|
|
aliases = ['logtalk']
|
|
filenames = ['*.lgt', '*.logtalk']
|
|
mimetypes = ['text/x-logtalk']
|
|
|
|
tokens = {
|
|
'root': [
|
|
# Directives
|
|
(r'^\s*:-\s', Punctuation, 'directive'),
|
|
# Comments
|
|
(r'%.*?\n', Comment),
|
|
(r'/\*(.|\n)*?\*/', Comment),
|
|
# Whitespace
|
|
(r'\n', Text),
|
|
(r'\s+', Text),
|
|
# Numbers
|
|
(r"0'.", Number),
|
|
(r'0b[01]+', Number.Bin),
|
|
(r'0o[0-7]+', Number.Oct),
|
|
(r'0x[0-9a-fA-F]+', Number.Hex),
|
|
(r'\d+\.?\d*((e|E)(\+|-)?\d+)?', Number),
|
|
# Variables
|
|
(r'([A-Z_]\w*)', Name.Variable),
|
|
# Event handlers
|
|
(r'(after|before)(?=[(])', Keyword),
|
|
# Message forwarding handler
|
|
(r'forward(?=[(])', Keyword),
|
|
# Execution-context methods
|
|
(r'(parameter|this|se(lf|nder))(?=[(])', Keyword),
|
|
# Reflection
|
|
(r'(current_predicate|predicate_property)(?=[(])', Keyword),
|
|
# DCGs and term expansion
|
|
(r'(expand_(goal|term)|(goal|term)_expansion|phrase)(?=[(])', Keyword),
|
|
# Entity
|
|
(r'(abolish|c(reate|urrent))_(object|protocol|category)(?=[(])', Keyword),
|
|
(r'(object|protocol|category)_property(?=[(])', Keyword),
|
|
# Entity relations
|
|
(r'co(mplements_object|nforms_to_protocol)(?=[(])', Keyword),
|
|
(r'extends_(object|protocol|category)(?=[(])', Keyword),
|
|
(r'imp(lements_protocol|orts_category)(?=[(])', Keyword),
|
|
(r'(instantiat|specializ)es_class(?=[(])', Keyword),
|
|
# Events
|
|
(r'(current_event|(abolish|define)_events)(?=[(])', Keyword),
|
|
# Flags
|
|
(r'(current|set)_logtalk_flag(?=[(])', Keyword),
|
|
# Compiling, loading, and library paths
|
|
(r'logtalk_(compile|l(ibrary_path|oad|oad_context)|make)(?=[(])', Keyword),
|
|
(r'\blogtalk_make\b', Keyword),
|
|
# Database
|
|
(r'(clause|retract(all)?)(?=[(])', Keyword),
|
|
(r'a(bolish|ssert(a|z))(?=[(])', Keyword),
|
|
# Control constructs
|
|
(r'(ca(ll|tch)|throw)(?=[(])', Keyword),
|
|
(r'(fa(il|lse)|true)\b', Keyword),
|
|
# All solutions
|
|
(r'((bag|set)of|f(ind|or)all)(?=[(])', Keyword),
|
|
# Multi-threading meta-predicates
|
|
(r'threaded(_(call|once|ignore|exit|peek|wait|notify))?(?=[(])', Keyword),
|
|
# Term unification
|
|
(r'(subsumes_term|unify_with_occurs_check)(?=[(])', Keyword),
|
|
# Term creation and decomposition
|
|
(r'(functor|arg|copy_term|numbervars|term_variables)(?=[(])', Keyword),
|
|
# Evaluable functors
|
|
(r'(div|rem|m(ax|in|od)|abs|sign)(?=[(])', Keyword),
|
|
(r'float(_(integer|fractional)_part)?(?=[(])', Keyword),
|
|
(r'(floor|t(an|runcate)|round|ceiling)(?=[(])', Keyword),
|
|
# Other arithmetic functors
|
|
(r'(cos|a(cos|sin|tan|tan2)|exp|log|s(in|qrt)|xor)(?=[(])', Keyword),
|
|
# Term testing
|
|
(r'(var|atom(ic)?|integer|float|c(allable|ompound)|n(onvar|umber)|'
|
|
r'ground|acyclic_term)(?=[(])', Keyword),
|
|
# Term comparison
|
|
(r'compare(?=[(])', Keyword),
|
|
# Stream selection and control
|
|
(r'(curren|se)t_(in|out)put(?=[(])', Keyword),
|
|
(r'(open|close)(?=[(])', Keyword),
|
|
(r'flush_output(?=[(])', Keyword),
|
|
(r'(at_end_of_stream|flush_output)\b', Keyword),
|
|
(r'(stream_property|at_end_of_stream|set_stream_position)(?=[(])', Keyword),
|
|
# Character and byte input/output
|
|
(r'(nl|(get|peek|put)_(byte|c(har|ode)))(?=[(])', Keyword),
|
|
(r'\bnl\b', Keyword),
|
|
# Term input/output
|
|
(r'read(_term)?(?=[(])', Keyword),
|
|
(r'write(q|_(canonical|term))?(?=[(])', Keyword),
|
|
(r'(current_)?op(?=[(])', Keyword),
|
|
(r'(current_)?char_conversion(?=[(])', Keyword),
|
|
# Atomic term processing
|
|
(r'atom_(length|c(hars|o(ncat|des)))(?=[(])', Keyword),
|
|
(r'(char_code|sub_atom)(?=[(])', Keyword),
|
|
(r'number_c(har|ode)s(?=[(])', Keyword),
|
|
# Implementation defined hooks functions
|
|
(r'(se|curren)t_prolog_flag(?=[(])', Keyword),
|
|
(r'\bhalt\b', Keyword),
|
|
(r'halt(?=[(])', Keyword),
|
|
# Message sending operators
|
|
(r'(::|:|\^\^)', Operator),
|
|
# External call
|
|
(r'[{}]', Keyword),
|
|
# Logic and control
|
|
(r'(ignore|once)(?=[(])', Keyword),
|
|
(r'\brepeat\b', Keyword),
|
|
# Sorting
|
|
(r'(key)?sort(?=[(])', Keyword),
|
|
# Bitwise functors
|
|
(r'(>>|<<|/\\|\\\\|\\)', Operator),
|
|
# Predicate aliases
|
|
(r'\bas\b', Operator),
|
|
# Arithemtic evaluation
|
|
(r'\bis\b', Keyword),
|
|
# Arithemtic comparison
|
|
(r'(=:=|=\\=|<|=<|>=|>)', Operator),
|
|
# Term creation and decomposition
|
|
(r'=\.\.', Operator),
|
|
# Term unification
|
|
(r'(=|\\=)', Operator),
|
|
# Term comparison
|
|
(r'(==|\\==|@=<|@<|@>=|@>)', Operator),
|
|
# Evaluable functors
|
|
(r'(//|[-+*/])', Operator),
|
|
(r'\b(e|pi|div|mod|rem)\b', Operator),
|
|
# Other arithemtic functors
|
|
(r'\b\*\*\b', Operator),
|
|
# DCG rules
|
|
(r'-->', Operator),
|
|
# Control constructs
|
|
(r'([!;]|->)', Operator),
|
|
# Logic and control
|
|
(r'\\+', Operator),
|
|
# Mode operators
|
|
(r'[?@]', Operator),
|
|
# Existential quantifier
|
|
(r'\^', Operator),
|
|
# Strings
|
|
(r'"(\\\\|\\"|[^"])*"', String),
|
|
# Ponctuation
|
|
(r'[()\[\],.|]', Text),
|
|
# Atoms
|
|
(r"[a-z]\w*", Text),
|
|
(r"'", String, 'quoted_atom'),
|
|
],
|
|
|
|
'quoted_atom': [
|
|
(r"''", String),
|
|
(r"'", String, '#pop'),
|
|
(r'\\([\\abfnrtv"\']|(x[a-fA-F0-9]+|[0-7]+)\\)', String.Escape),
|
|
(r"[^\\'\n]+", String),
|
|
(r'\\', String),
|
|
],
|
|
|
|
'directive': [
|
|
# Conditional compilation directives
|
|
(r'(el)?if(?=[(])', Keyword, 'root'),
|
|
(r'(e(lse|ndif))[.]', Keyword, 'root'),
|
|
# Entity directives
|
|
(r'(category|object|protocol)(?=[(])', Keyword, 'entityrelations'),
|
|
(r'(end_(category|object|protocol))[.]', Keyword, 'root'),
|
|
# Predicate scope directives
|
|
(r'(public|protected|private)(?=[(])', Keyword, 'root'),
|
|
# Other directives
|
|
(r'e(n(coding|sure_loaded)|xport)(?=[(])', Keyword, 'root'),
|
|
(r'in(clude|itialization|fo)(?=[(])', Keyword, 'root'),
|
|
(r'(built_in|dynamic|synchronized|threaded)[.]', Keyword, 'root'),
|
|
(r'(alias|d(ynamic|iscontiguous)|m(eta_(non_terminal|predicate)|ode|ultifile)|'
|
|
r's(et_(logtalk|prolog)_flag|ynchronized))(?=[(])', Keyword, 'root'),
|
|
(r'op(?=[(])', Keyword, 'root'),
|
|
(r'(c(alls|oinductive)|module|reexport|use(s|_module))(?=[(])', Keyword, 'root'),
|
|
(r'[a-z]\w*(?=[(])', Text, 'root'),
|
|
(r'[a-z]\w*[.]', Text, 'root'),
|
|
],
|
|
|
|
'entityrelations': [
|
|
(r'(complements|extends|i(nstantiates|mp(lements|orts))|specializes)(?=[(])', Keyword),
|
|
# Numbers
|
|
(r"0'.", Number),
|
|
(r'0b[01]+', Number.Bin),
|
|
(r'0o[0-7]+', Number.Oct),
|
|
(r'0x[0-9a-fA-F]+', Number.Hex),
|
|
(r'\d+\.?\d*((e|E)(\+|-)?\d+)?', Number),
|
|
# Variables
|
|
(r'([A-Z_]\w*)', Name.Variable),
|
|
# Atoms
|
|
(r"[a-z]\w*", Text),
|
|
(r"'", String, 'quoted_atom'),
|
|
# Strings
|
|
(r'"(\\\\|\\"|[^"])*"', String),
|
|
# End of entity-opening directive
|
|
(r'([)]\.)', Text, 'root'),
|
|
# Scope operator
|
|
(r'(::)', Operator),
|
|
# Ponctuation
|
|
(r'[()\[\],.|]', Text),
|
|
# Comments
|
|
(r'%.*?\n', Comment),
|
|
(r'/\*(.|\n)*?\*/', Comment),
|
|
# Whitespace
|
|
(r'\n', Text),
|
|
(r'\s+', Text),
|
|
]
|
|
}
|
|
|
|
def analyse_text(text):
|
|
if ':- object(' in text:
|
|
return 1.0
|
|
elif ':- protocol(' in text:
|
|
return 1.0
|
|
elif ':- category(' in text:
|
|
return 1.0
|
|
elif re.search(r'^:-\s[a-z]', text, re.M):
|
|
return 0.9
|
|
else:
|
|
return 0.0
|