aboutsummaryrefslogtreecommitdiff
path: root/lua/lexers/crystal.lua
diff options
context:
space:
mode:
Diffstat (limited to 'lua/lexers/crystal.lua')
-rw-r--r--lua/lexers/crystal.lua185
1 files changed, 73 insertions, 112 deletions
diff --git a/lua/lexers/crystal.lua b/lua/lexers/crystal.lua
index 5195387..f8b2211 100644
--- a/lua/lexers/crystal.lua
+++ b/lua/lexers/crystal.lua
@@ -1,141 +1,102 @@
--- Copyright 2006-2017 Mitchell mitchell.att.foicica.com. See LICENSE.
+-- Copyright 2006-2022 Mitchell. See LICENSE.
-- Copyright 2017 Michel Martens.
-- Crystal LPeg lexer (based on Ruby).
-local l = require('lexer')
-local token, word_match = l.token, l.word_match
-local P, R, S = lpeg.P, lpeg.R, lpeg.S
+local lexer = require('lexer')
+local token, word_match = lexer.token, lexer.word_match
+local P, S = lpeg.P, lpeg.S
-local M = {_NAME = 'crystal'}
+local lex = lexer.new('crystal')
-- Whitespace.
-local ws = token(l.WHITESPACE, l.space^1)
+lex:add_rule('whitespace', token(lexer.WHITESPACE, lexer.space^1))
--- Comments.
-local line_comment = '#' * l.nonnewline_esc^0
-local comment = token(l.COMMENT, line_comment)
+-- Keywords.
+lex:add_rule('keyword', token(lexer.KEYWORD, word_match{
+ 'alias', 'begin', 'break', 'case', 'class', 'def', 'defined?', 'do', 'else', 'elsif', 'end',
+ 'ensure', 'false', 'for', 'if', 'in', 'module', 'next', 'nil', 'not', 'redo', 'rescue', 'retry',
+ 'return', 'self', 'super', 'then', 'true', 'undef', 'unless', 'until', 'when', 'while', 'yield',
+ '__FILE__', '__LINE__'
+}))
-local delimiter_matches = {['('] = ')', ['['] = ']', ['{'] = '}'}
-local literal_delimitted = P(function(input, index)
- local delimiter = input:sub(index, index)
- if not delimiter:find('[%w\r\n\f\t ]') then -- only non alpha-numerics
- local match_pos, patt
- if delimiter_matches[delimiter] then
- -- Handle nested delimiter/matches in strings.
- local s, e = delimiter, delimiter_matches[delimiter]
- patt = l.delimited_range(s..e, false, false, true)
- else
- patt = l.delimited_range(delimiter)
- end
- match_pos = lpeg.match(patt, input, index)
- return match_pos or #input + 1
- end
-end)
+-- Functions.
+lex:add_rule('function', token(lexer.FUNCTION, word_match{
+ 'abort', 'at_exit', 'caller', 'delay', 'exit', 'fork', 'future', 'get_stack_top', 'gets', 'lazy',
+ 'loop', 'main', 'p', 'print', 'printf', 'puts', 'raise', 'rand', 'read_line', 'require', 'sleep',
+ 'spawn', 'sprintf', 'system', 'with_color',
+ -- Macros.
+ 'assert_responds_to', 'debugger', 'parallel', 'pp', 'record', 'redefine_main'
+}) * -S('.:|'))
+
+-- Identifiers.
+local word_char = lexer.alnum + S('_!?')
+local word = (lexer.alpha + '_') * word_char^0
+lex:add_rule('identifier', token(lexer.IDENTIFIER, word))
+
+-- Comments.
+lex:add_rule('comment', token(lexer.COMMENT, lexer.to_eol('#', true)))
-- Strings.
-local cmd_str = l.delimited_range('`')
-local sq_str = l.delimited_range("'")
-local dq_str = l.delimited_range('"')
+local cmd_str = lexer.range('`')
+local sq_str = lexer.range("'")
+local dq_str = lexer.range('"')
local heredoc = '<<' * P(function(input, index)
- local s, e, indented, _, delimiter =
- input:find('(%-?)(["`]?)([%a_][%w_]*)%2[\n\r\f;]+', index)
- if s == index and delimiter then
- local end_heredoc = (#indented > 0 and '[\n\r\f]+ *' or '[\n\r\f]+')
- local _, e = input:find(end_heredoc..delimiter, e)
- return e and e + 1 or #input + 1
- end
+ local _, e, indented, _, delimiter = input:find('^(%-?)(["`]?)([%a_][%w_]*)%2[\n\r\f;]+', index)
+ if not delimiter then return end
+ local end_heredoc = (#indented > 0 and '[\n\r\f]+ *' or '[\n\r\f]+')
+ _, e = input:find(end_heredoc .. delimiter, e)
+ return e and e + 1 or #input + 1
end)
+local string = token(lexer.STRING, (sq_str + dq_str + heredoc + cmd_str) * S('f')^-1)
-- TODO: regex_str fails with `obj.method /patt/` syntax.
-local regex_str = #P('/') * l.last_char_includes('!%^&*([{-=+|:;,?<>~') *
- l.delimited_range('/', true, false) * S('iomx')^0
-local string = token(l.STRING, (sq_str + dq_str + heredoc + cmd_str) *
- S('f')^-1) +
- token(l.REGEX, regex_str)
-
-local word_char = l.alnum + S('_!?')
+local regex_str =
+ #P('/') * lexer.last_char_includes('!%^&*([{-=+|:;,?<>~') * lexer.range('/', true) * S('iomx')^0
+local regex = token(lexer.REGEX, regex_str)
+lex:add_rule('string', string + regex)
-- Numbers.
-local dec = l.digit^1 * ('_' * l.digit^1)^0 * S('ri')^-1
+local dec = lexer.digit^1 * ('_' * lexer.digit^1)^0 * S('ri')^-1
local bin = '0b' * S('01')^1 * ('_' * S('01')^1)^0
-local integer = S('+-')^-1 * (bin + l.hex_num + l.oct_num + dec)
+local integer = S('+-')^-1 * (bin + lexer.hex_num + lexer.oct_num + dec)
-- TODO: meta, control, etc. for numeric_literal.
-local numeric_literal = '?' * (l.any - l.space) * -word_char
-local number = token(l.NUMBER, l.float * S('ri')^-1 + integer + numeric_literal)
-
--- Keywords.
-local keyword = token(l.KEYWORD, word_match({
- 'alias', 'begin', 'break', 'case', 'class', 'def', 'defined?', 'do', 'else',
- 'elsif', 'end', 'ensure', 'false', 'for', 'if', 'in', 'module', 'next', 'nil',
- 'not', 'redo', 'rescue', 'retry', 'return', 'self', 'super', 'then', 'true',
- 'undef', 'unless', 'until', 'when', 'while', 'yield', '__FILE__', '__LINE__'
-}, '?!'))
-
--- Functions.
-local func = token(l.FUNCTION, word_match({
- 'abort', 'at_exit', 'caller', 'delay', 'exit', 'fork', 'future',
- 'get_stack_top', 'gets', 'lazy', 'loop', 'main', 'p', 'print', 'printf',
- 'puts', 'raise', 'rand', 'read_line', 'require', 'sleep', 'spawn', 'sprintf',
- 'system', 'with_color',
- -- Macros
- 'assert_responds_to', 'debugger', 'parallel', 'pp', 'record', 'redefine_main'
-}, '?!')) * -S('.:|')
-
--- Identifiers.
-local word = (l.alpha + '_') * word_char^0
-local identifier = token(l.IDENTIFIER, word)
+local numeric_literal = '?' * (lexer.any - lexer.space) * -word_char
+lex:add_rule('number', token(lexer.NUMBER, lexer.float * S('ri')^-1 + integer + numeric_literal))
-- Variables.
-local global_var = '$' * (word + S('!@L+`\'=~/\\,.;<>_*"$?:') + l.digit + '-' *
- S('0FadiIKlpvw'))
+local global_var = '$' *
+ (word + S('!@L+`\'=~/\\,.;<>_*"$?:') + lexer.digit + '-' * S('0FadiIKlpvw'))
local class_var = '@@' * word
local inst_var = '@' * word
-local variable = token(l.VARIABLE, global_var + class_var + inst_var)
+lex:add_rule('variable', token(lexer.VARIABLE, global_var + class_var + inst_var))
-- Symbols.
-local symbol = token('symbol', ':' * P(function(input, index)
+lex:add_rule('symbol', token('symbol', ':' * P(function(input, index)
if input:sub(index - 2, index - 2) ~= ':' then return index end
-end) * (word_char^1 + sq_str + dq_str))
+end) * (word_char^1 + sq_str + dq_str)))
+lex:add_style('symbol', lexer.styles.constant)
-- Operators.
-local operator = token(l.OPERATOR, S('!%^&*()[]{}-=+/|:;.,?<>~'))
-
-M._rules = {
- {'whitespace', ws},
- {'keyword', keyword},
- {'function', func},
- {'identifier', identifier},
- {'comment', comment},
- {'string', string},
- {'number', number},
- {'variable', variable},
- {'symbol', symbol},
- {'operator', operator},
-}
-
-M._tokenstyles = {
- symbol = l.STYLE_CONSTANT
-}
+lex:add_rule('operator', token(lexer.OPERATOR, S('!%^&*()[]{}-=+/|:;.,?<>~')))
+-- Fold points.
local function disambiguate(text, pos, line, s)
- return line:sub(1, s - 1):match('^%s*$') and
- not text:sub(1, pos - 1):match('\\[ \t]*\r?\n$') and 1 or 0
+ return line:sub(1, s - 1):match('^%s*$') and not text:sub(1, pos - 1):match('\\[ \t]*\r?\n$') and
+ 1 or 0
end
-
-M._foldsymbols = {
- _patterns = {'%l+', '[%(%)%[%]{}]', '#'},
- [l.KEYWORD] = {
- begin = 1, class = 1, def = 1, ['do'] = 1, ['for'] = 1, ['module'] = 1,
- case = 1,
- ['if'] = disambiguate, ['while'] = disambiguate,
- ['unless'] = disambiguate, ['until'] = disambiguate,
- ['end'] = -1
- },
- [l.OPERATOR] = {
- ['('] = 1, [')'] = -1, ['['] = 1, [']'] = -1, ['{'] = 1, ['}'] = -1
- },
- [l.COMMENT] = {
- ['#'] = l.fold_line_comments('#')
- }
-}
-
-return M
+lex:add_fold_point(lexer.KEYWORD, 'begin', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'case', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'class', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'def', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'do', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'for', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'module', 'end')
+lex:add_fold_point(lexer.KEYWORD, 'if', disambiguate)
+lex:add_fold_point(lexer.KEYWORD, 'while', disambiguate)
+lex:add_fold_point(lexer.KEYWORD, 'unless', disambiguate)
+lex:add_fold_point(lexer.KEYWORD, 'until', disambiguate)
+lex:add_fold_point(lexer.OPERATOR, '(', ')')
+lex:add_fold_point(lexer.OPERATOR, '[', ']')
+lex:add_fold_point(lexer.OPERATOR, '{', '}')
+lex:add_fold_point(lexer.COMMENT, lexer.fold_consecutive_lines('#'))
+
+return lex