summaryrefslogtreecommitdiff
path: root/context/data/scite/lexers/scite-context-lexer-web.lua
blob: f59a3205dd8b2afd81153933f7f18330bd9c8f3f (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
local info = {
    version   = 1.002,
    comment   = "scintilla lpeg lexer for w",
    author    = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
    copyright = "PRAGMA ADE / ConTeXt Development Team",
    license   = "see context related readme files",
}

-- this will be extended

if not lexer._CONTEXTEXTENSIONS then require("scite-context-lexer") end

local lexer = lexer
local token, style, colors, exact_match, no_style = lexer.token, lexer.style, lexer.colors, lexer.exact_match, lexer.style_nothing
local P, R, S, C, Cg, Cb, Cs, Cmt, lpegmatch = lpeg.P, lpeg.R, lpeg.S, lpeg.C, lpeg.Cg, lpeg.Cb, lpeg.Cs, lpeg.Cmt, lpeg.match
local setmetatable = setmetatable

local weblexer    = { _NAME = "web", _FILENAME = "scite-context-lexer-web" }
local whitespace  = lexer.WHITESPACE
local context     = lexer.context

local keywords = { -- copied from cpp.lua
    -- c
    'asm', 'auto', 'break', 'case', 'const', 'continue', 'default', 'do', 'else',
    'extern', 'false', 'for', 'goto', 'if', 'inline', 'register', 'return',
    'sizeof', 'static', 'switch', 'true', 'typedef', 'volatile', 'while',
    'restrict',
    -- hm
    '_Bool', '_Complex', '_Pragma', '_Imaginary',
    -- c++.
    'catch', 'class', 'const_cast', 'delete', 'dynamic_cast', 'explicit',
    'export', 'friend', 'mutable', 'namespace', 'new', 'operator', 'private',
    'protected', 'public', 'signals', 'slots', 'reinterpret_cast',
    'static_assert', 'static_cast', 'template', 'this', 'throw', 'try', 'typeid',
    'typename', 'using', 'virtual'
}

local datatypes = { -- copied from cpp.lua
    'bool', 'char', 'double', 'enum', 'float', 'int', 'long', 'short', 'signed',
    'struct', 'union', 'unsigned', 'void'
}

local macros = { -- copied from cpp.lua
    'define', 'elif', 'else', 'endif', 'error', 'if', 'ifdef', 'ifndef', 'import',
    'include', 'line', 'pragma', 'undef', 'using', 'warning'
}

local space         = lexer.space -- S(" \n\r\t\f\v")
local any           = lexer.any
local patterns      = context.patterns
local restofline    = patterns.restofline
local startofline   = patterns.startofline

local squote        = P("'")
local dquote        = P('"')
local escaped       = P("\\") * P(1)
local slashes       = P('//')
local begincomment  = P("/*")
local endcomment    = P("*/")
local percent       = P("%")

local spacing       = token(whitespace, space^1)
local rest          = token("default", any)

local shortcomment  = token("comment", slashes * restofline^0)
local longcomment   = token("comment", begincomment * (1-endcomment)^0 * endcomment^-1)
local texcomment    = token("comment", percent * restofline^0)

local shortstring   = token("quote",  dquote) -- can be shared
                    * token("string", (escaped + (1-dquote))^0)
                    * token("quote",  dquote)
                    + token("quote",  squote)
                    * token("string", (escaped + (1-squote))^0)
                    * token("quote",  squote)

local integer       = P("-")^-1 * (lexer.hex_num + lexer.dec_num)
local number        = token("number", lexer.float + integer)

local validword     = R("AZ","az","__") * R("AZ","az","__","09")^0

local identifier    = token("default",validword)

local operator      = token("special", S('+-*/%^!=<>;:{}[]().&|?~'))

----- optionalspace = spacing^0

local p_keywords    = exact_match(keywords )
local p_datatypes   = exact_match(datatypes)
local p_macros      = exact_match(macros)

local keyword       = token("keyword", p_keywords)
local datatype      = token("keyword", p_datatypes)
local identifier    = token("default", validword)

local macro         = token("data", #P('#') * startofline * P('#') * S('\t ')^0 * p_macros)

local beginweb      = P("@")
local endweb        = P("@c")

local webcomment    = token("comment", #beginweb * startofline * beginweb * (1-endweb)^0 * endweb)

local texlexer      = lexer.load('scite-context-lexer-tex')

lexer.embed_lexer(weblexer, texlexer, #beginweb * startofline * token("comment",beginweb), token("comment",endweb))

weblexer._rules = {
    { 'whitespace',   spacing      },
    { 'keyword',      keyword      },
    { 'type',         datatype     },
    { 'identifier',   identifier   },
    { 'string',       shortstring  },
 -- { 'webcomment',   webcomment   },
    { 'texcomment',   texcomment   },
    { 'longcomment',  longcomment  },
    { 'shortcomment', shortcomment },
    { 'number',       number       },
    { 'macro',        macro        },
    { 'operator',     operator     },
    { 'rest',         rest         },
}

weblexer._tokenstyles = context.styleset

weblexer._foldpattern = P("/*") + P("*/") + S("{}") -- separate entry else interference

weblexer._foldsymbols = {
    _patterns = {
        '[{}]',
        '/%*',
        '%*/',
    },
 -- ["data"] = { -- macro
 --     ['region']    =  1,
 --     ['endregion'] = -1,
 --     ['if']        =  1,
 --     ['ifdef']     =  1,
 --     ['ifndef']    =  1,
 --     ['endif']     = -1,
 -- },
    ["special"] = { -- operator
        ['{'] =  1,
        ['}'] = -1,
    },
    ["comment"] = {
        ['/*'] =  1,
        ['*/'] = -1,
    }
}

-- -- by indentation:
--
weblexer._foldpatterns = nil
weblexer._foldsymbols  = nil

return weblexer