aboutsummaryrefslogtreecommitdiff
path: root/lua/lexers/erlang.lua
diff options
context:
space:
mode:
Diffstat (limited to 'lua/lexers/erlang.lua')
-rw-r--r--lua/lexers/erlang.lua143
1 files changed, 76 insertions, 67 deletions
diff --git a/lua/lexers/erlang.lua b/lua/lexers/erlang.lua
index d4d48b7..a0a3a7d 100644
--- a/lua/lexers/erlang.lua
+++ b/lua/lexers/erlang.lua
@@ -1,81 +1,90 @@
--- Copyright 2006-2017 Mitchell mitchell.att.foicica.com. See LICENSE.
+-- Copyright 2006-2022 Mitchell. See LICENSE.
-- Erlang LPeg lexer.
-local l = require('lexer')
-local token, word_match = l.token, l.word_match
-local P, R, S = lpeg.P, lpeg.R, lpeg.S
+local lexer = require('lexer')
+local token, word_match = lexer.token, lexer.word_match
+local P, S = lpeg.P, lpeg.S
-local M = {_NAME = 'erlang'}
+local lex = lexer.new('erlang')
-- Whitespace.
-local ws = token(l.WHITESPACE, l.space^1)
+lex:add_rule('whitespace', token(lexer.WHITESPACE, lexer.space^1))
--- Comments.
-local comment = token(l.COMMENT, '%' * l.nonnewline^0)
+-- Keywords.
+lex:add_rule('keyword', token(lexer.KEYWORD, word_match{
+ 'after', 'begin', 'case', 'catch', 'cond', 'end', 'fun', 'if', 'let', 'of', 'query', 'receive',
+ 'try', 'when',
+ -- Operators.
+ 'div', 'rem', 'or', 'xor', 'bor', 'bxor', 'bsl', 'bsr', 'and', 'band', 'not', 'bnot', 'badarg',
+ 'nocookie', 'orelse', 'andalso', 'false', 'true'
+}))
--- Strings.
-local string = token(l.STRING, l.delimited_range('"'))
+-- Functions.
+lex:add_rule('function', token(lexer.FUNCTION, word_match{
+ 'abs', 'alive', 'apply', 'atom_to_list', 'binary_to_list', 'binary_to_term', 'concat_binary',
+ 'date', 'disconnect_node', 'element', 'erase', 'exit', 'float', 'float_to_list', 'get',
+ 'get_keys', 'group_leader', 'halt', 'hd', 'integer_to_list', 'is_alive', 'is_record', 'length',
+ 'link', 'list_to_atom', 'list_to_binary', 'list_to_float', 'list_to_integer', 'list_to_pid',
+ 'list_to_tuple', 'load_module', 'make_ref', 'monitor_node', 'node', 'nodes', 'now', 'open_port',
+ 'pid_to_list', 'process_flag', 'process_info', 'process', 'put', 'register', 'registered',
+ 'round', 'self', 'setelement', 'size', 'spawn', 'spawn_link', 'split_binary', 'statistics',
+ 'term_to_binary', 'throw', 'time', 'tl', 'trunc', 'tuple_to_list', 'unlink', 'unregister',
+ 'whereis',
+ -- Others.
+ 'any', 'atom', 'binary', 'bitstring', 'byte', 'constant', 'function', 'integer', 'list', 'map',
+ 'mfa', 'non_neg_integer', 'number', 'pid', 'ports', 'port_close', 'port_info', 'pos_integer',
+ 'reference', 'record',
+ -- Erlang.
+ 'check_process_code', 'delete_module', 'get_cookie', 'hash', 'math', 'module_loaded', 'preloaded',
+ 'processes', 'purge_module', 'set_cookie', 'set_node',
+ -- Math.
+ 'acos', 'asin', 'atan', 'atan2', 'cos', 'cosh', 'exp', 'log', 'log10', 'min', 'max', 'pi', 'pow',
+ 'power', 'sin', 'sinh', 'sqrt', 'tan', 'tanh'
+}))
--- Numbers.
-local const_char = '$' * (('\\' * l.ascii) + l.any)
-local number = token(l.NUMBER, const_char + l.float + l.integer)
+-- Identifiers.
+lex:add_rule('identifier', token(lexer.IDENTIFIER, lexer.lower * ('_' + lexer.alnum)^0))
--- Atoms.
-local atom_pat = (l.lower * (l.alnum + '_')^0) + l.delimited_range("'")
-local atom = token(l.LABEL, atom_pat)
+-- Variables.
+lex:add_rule('variable', token(lexer.VARIABLE, P('_')^0 * lexer.upper * ('_' + lexer.alnum)^0))
--- Functions.
-local func = token(l.FUNCTION, atom_pat * #l.delimited_range("()", false, false, true))
+-- Directives.
+lex:add_rule('directive', token('directive', '-' * word_match{
+ 'author', 'behaviour', 'behavior', 'compile', 'copyright', 'define', 'doc', 'else', 'endif',
+ 'export', 'file', 'ifdef', 'ifndef', 'import', 'include', 'include_lib', 'module', 'record',
+ 'spec', 'type', 'undef'
+}))
+lex:add_style('directive', lexer.styles.preprocessor)
--- Keywords.
-local keyword = token(l.KEYWORD, word_match{
- 'after', 'begin', 'case', 'catch', 'cond', 'end', 'fun', 'if', 'let', 'of',
- 'query', 'receive', 'when'
-})
+-- Strings.
+local sq_str = lexer.range("'", true)
+local dq_str = lexer.range('"')
+lex:add_rule('string', token(lexer.STRING, sq_str + dq_str + '$' * lexer.any * lexer.alnum^0))
--- Identifiers.
-local identifier = token(l.IDENTIFIER, ((l.upper + '_') * (l.alnum + '_')^0))
+-- Comments.
+lex:add_rule('comment', token(lexer.COMMENT, lexer.to_eol('%')))
--- Operators.
-local named_operator = word_match{
- 'div', 'rem', 'or', 'xor', 'bor', 'bxor', 'bsl', 'bsr', 'and', 'band', 'not',
- 'bnot'
-}
-local operator = token(l.OPERATOR, S('-<>.;=/|#+*:,?!()[]{}') + named_operator)
+-- Numbers.
+lex:add_rule('number', token(lexer.NUMBER, lexer.number))
--- Directives.
-local directive = token('directive', '-' * word_match{
- 'author', 'compile', 'copyright', 'define', 'doc', 'else', 'endif', 'export',
- 'file', 'ifdef', 'ifndef', 'import', 'include_lib', 'include', 'module',
- 'record', 'undef'
-})
-
-M._rules = {
- {'whitespace', ws},
- {'keyword', keyword},
- {'function', func},
- {'operator', operator},
- {'atom', atom},
- {'identifier', identifier},
- {'directive', directive},
- {'string', string},
- {'comment', comment},
- {'number', number}
-}
-
-M._tokenstyles = {
- directive = l.STYLE_PREPROCESSOR
-}
-
-M._foldsymbols = {
- _patterns = {'[a-z]+', '[%(%)%[%]{}]', '%%'},
- [l.KEYWORD] = {
- case = 1, fun = 1, ['if'] = 1, query = 1, receive = 1, ['end'] = -1
- },
- [l.OPERATOR] = {
- ['('] = 1, [')'] = -1, ['['] = 1, [']'] = -1, ['{'] = 1, ['}'] = -1
- },
- [l.COMMENT] = {['%'] = l.fold_line_comments('%')}
-}
-
-return M
+-- Operators.
+lex:add_rule('operator', token(lexer.OPERATOR, S('-<>.;=/|+*:,!()[]{}')))
+
+-- Preprocessor.
+lex:add_rule('preprocessor', token(lexer.TYPE, '?' * lexer.word))
+
+-- Records.
+lex:add_rule('type', token(lexer.TYPE, '#' * lexer.word))
+
+-- Fold points.
+lex:add_fold_point(lexer.KEYWORD, 'case', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'fun', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'if', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'query', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'receive', 'end')
+lex:add_fold_point(lexer.OPERATOR, '(', ')')
+lex:add_fold_point(lexer.OPERATOR, '[', ']')
+lex:add_fold_point(lexer.OPERATOR, '{', '}')
+lex:add_fold_point(lexer.COMMENT, lexer.fold_consecutive_lines('%'))
+
+return lex