1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
|
local info = {
version = 1.400,
comment = "basics for scintilla lpeg lexer for context/metafun",
author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
copyright = "PRAGMA ADE / ConTeXt Development Team",
license = "see context related readme files",
comment = "contains copyrighted code from mitchell.att.foicica.com",
}
local trace = true -- false
-- GET GOING
--
-- You need to copy this file over lexer.lua. In principle other lexers could
-- work too but not now. Maybe some day. All patterns will move into the patterns
-- name space. I might do the same with styles. If you run an older version of
-- SciTE you can take one of the archives. Pre 3.41 versions can just be copied
-- to the right path, as there we still use part of the normal lexer.
--
-- REMARK
--
-- We started using lpeg lexing as soon as it came available. Because we had
-- rather demanding files an dalso wanted to use nested lexers, we ended up with
-- our own variant (more robust and faster). As a consequence successive versions
-- had to be adapted to changes in the (still unstable) api. In addition to
-- lexing we also have spell checking and such.
--
-- STATUS
--
-- todo: maybe use a special stripped version of the dll (stable api)
-- todo: play with hotspot and other properties
-- wish: access to all scite properties and in fact integrate in scite
-- todo: add proper tracing and so .. not too hard as we can run on mtxrun
-- todo: get rid of these lexers.STYLE_XX and lexers.XX (hide such details)
--
-- HISTORY
--
-- The fold and lex functions are copied and patched from original code by Mitchell
-- (see lexer.lua). All errors are mine. The ability to use lpeg is a real nice
-- adition and a brilliant move. The code is a byproduct of the (mainly Lua based)
-- textadept (still a rapidly moving target) that unfortunately misses a realtime
-- output pane. On the other hand, SciTE is somewhat crippled by the fact that we
-- cannot pop in our own (language dependent) lexer into the output pane (somehow
-- the errorlist lexer is hard coded into the editor). Hopefully that will change
-- some day.
--
-- Starting with SciTE version 3.20 there is an issue with coloring. As we still
-- lack a connection with SciTE itself (properties as well as printing to the log
-- pane) and we cannot trace this (on windows). As far as I can see, there are no
-- fundamental changes in lexer.lua or LexLPeg.cxx so it must be in Scintilla
-- itself. So for the moment I stick to 3.10. Indicators are: no lexing of 'next'
-- and 'goto <label>' in the Lua lexer and no brace highlighting either. Interesting
-- is that it does work ok in the cld lexer (so the Lua code is okay). Also the fact
-- that char-def.lua lexes fast is a signal that the lexer quits somewhere halfway.
-- Maybe there are some hard coded limitations on the amount of styles and/or length
-- if names.
--
-- After checking 3.24 and adapting to the new lexer tables things are okay again.
-- So, this version assumes 3.24 or higher. In 3.24 we have a different token
-- result, i.e. no longer a { tag, pattern } but just two return values. I didn't
-- check other changes but will do that when I run into issues. I had optimized
-- these small tables by hashing which was more efficient but this is no longer
-- needed. For the moment we keep some of that code around as I don't know what
-- happens in future versions.
--
-- In 3.31 another major change took place: some helper constants (maybe they're no
-- longer constants) and functions were moved into the lexer modules namespace but
-- the functions are assigned to the Lua module afterward so we cannot alias them
-- beforehand. We're probably getting close to a stable interface now. I've
-- considered making a whole copy and patch the other functions too as we need an
-- extra nesting model. However, I don't want to maintain too much. An unfortunate
-- change in 3.03 is that no longer a script can be specified. This means that
-- instead of loading the extensions via the properties file, we now need to load
-- them in our own lexers, unless of course we replace lexer.lua completely (which
-- adds another installation issue).
--
-- Another change has been that _LEXERHOME is no longer available. It looks like
-- more and more functionality gets dropped so maybe at some point we need to ship
-- our own dll/so files. For instance, I'd like to have access to the current
-- filename and other scite properties. For instance, we could cache some info with
-- each file, if only we had knowledge of what file we're dealing with.
--
-- For huge files folding can be pretty slow and I do have some large ones that I
-- keep open all the time. Loading is normally no ussue, unless one has remembered
-- the status and the cursor is at the last line of a 200K line file. Optimizing the
-- fold function brought down loading of char-def.lua from 14 sec => 8 sec.
-- Replacing the word_match function and optimizing the lex function gained another
-- 2+ seconds. A 6 second load is quite ok for me. The changed lexer table structure
-- (no subtables) brings loading down to a few seconds.
--
-- When the lexer path is copied to the textadept lexer path, and the theme
-- definition to theme path (as lexer.lua), the lexer works there as well. When I
-- have time and motive I will make a proper setup file to tune the look and feel a
-- bit and associate suffixes with the context lexer. The textadept editor has a
-- nice style tracing option but lacks the tabs for selecting files that scite has.
-- It also has no integrated run that pipes to the log pane. Interesting is that the
-- jit version of textadept crashes on lexing large files (and does not feel faster
-- either; maybe a side effect of known limitations).
--
-- Function load(lexer_name) starts with _lexers.WHITESPACE = lexer_name ..
-- '_whitespace' which means that we need to have it frozen at the moment we load
-- another lexer. Because spacing is used to revert to a parent lexer we need to
-- make sure that we load children as late as possible in order not to get the wrong
-- whitespace trigger. This took me quite a while to figure out (not being that
-- familiar with the internals). The lex and fold functions have been optimized. It
-- is a pitty that there is no proper print available. Another thing needed is a
-- default style in our own theme style definition, as otherwise we get wrong nested
-- lexers, especially if they are larger than a view. This is the hardest part of
-- getting things right.
--
-- It's a pitty that there is no scintillua library for the OSX version of scite.
-- Even better would be to have the scintillua library as integral part of scite as
-- that way I could use OSX alongside windows and linux (depending on needs). Also
-- nice would be to have a proper interface to scite then because currently the
-- lexer is rather isolated and the lua version does not provide all standard
-- libraries. It would also be good to have lpeg support in the regular scite lua
-- extension (currently you need to pick it up from someplace else).
--
-- With 3.41 the interface changed again so it gets time to look into the C++ code
-- and consider compiling and patching myself. Loading is more complicated not as
-- the lexer gets loaded automatically so we have little control over extending the
-- code now. After a few days trying all kind of solutions I decided to follow a
-- different approach: drop in a complete replacement. This of course means that I
-- need to keep track of even more changes (which for sure will happen) but at least
-- I get rid of interferences. The api (lexing and configuration) is simply too
-- unstable across versions. Maybe in a few years things have stabelized. (Or maybe
-- it's not really expected that one writes lexers at all.) A side effect is that I
-- now no longer will use shipped lexers but just the built-in ones. Not that it
-- matters much as the context lexers cover what I need (and I can always write
-- more).
--
-- In fact, the transition to 3.41 was triggered by an unfateful update of Ubuntu
-- which left me with an incompatible SciTE and lexer library and updating was not
-- possible due to the lack of 64 bit libraries. We'll see what the future brings.
--
-- Promissing is that the library now can use another Lua instance so maybe some day
-- it will get properly in SciTE and we can use more clever scripting.
-- TRACING
--
-- The advantage is that we now can check more easily with regular Lua. We can also
-- use wine and print to the console (somehow stdout is intercepted there.) So, I've
-- added a bit of tracing. Interesting is to notice that each document gets its own
-- instance which is pretty inefficient when we are spellchecking (In the past I
-- assumed a shared instance and took some precautions.)
local lpeg = require("lpeg")
local global = _G
local find, gmatch, match, lower, upper, gsub, sub, format = string.find, string.gmatch, string.match, string.lower, string.upper, string.gsub, string.sub, string.format
local concat = table.concat
local type, next, setmetatable, rawset, tonumber, tostring = type, next, setmetatable, rawset, tonumber, tostring
local R, P, S, V, C, Cp, Cs, Ct, Cmt, Cc, Cf, Cg, Carg = lpeg.R, lpeg.P, lpeg.S, lpeg.V, lpeg.C, lpeg.Cp, lpeg.Cs, lpeg.Ct, lpeg.Cmt, lpeg.Cc, lpeg.Cf, lpeg.Cg, lpeg.Carg
local lpegmatch = lpeg.match
local nesting = 0
local function report(fmt,str,...)
if str then
fmt = format(fmt,str,...)
end
print(format("scite lpeg lexer > %s > %s",nesting == 0 and "-" or nesting,fmt))
end
if trace then
report("loading context lexer module (global table: %s)",tostring(global))
end
if not package.searchpath then
-- Unfortunately the io library is only available when we end up
-- in this branch of code.
if trace then
report("using adapted function 'package.searchpath'")
end
function package.searchpath(name,path)
local tried = { }
for part in gmatch(path,"[^;]+") do
local filename = gsub(part,"%?",name)
local f = io.open(filename,"r")
if f then
f:close()
return filename
end
tried[#tried + 1] = format("no file '%s'",filename)
end
-- added: local path .. for testing
local f = io.open(filename,"r")
if f then
f:close()
return filename
end
--
tried[#tried + 1] = format("no file '%s'",filename)
return nil, concat(tried,"\n")
end
end
local lexers = { }
local context = { }
lexers.context = context
local patterns = { }
context.patterns = patterns -- todo: lexers.patterns
lexers._CONTEXTEXTENSIONS = true
lexers.LEXERPATH = package.path -- no need
local usedlexers = { }
local parent_lexer = nil
-- The problem with styles is that there is some nasty interaction with scintilla
-- and each version of lexer dll/so has a different issue. So, from now on we will
-- just add them here. There is also a limit on some 30 styles. Maybe I should
-- hash them in order to reuse.
local default = {
"nothing", "whitespace", "comment", "string", "number", "keyword",
"identifier", "operator", "error", "preprocessor", "constant", "variable",
"function", "type", "label", "embedded",
-- "regex", "class",
"quote", "special", "extra", "reserved", "okay", "warning",
"command", "internal", "preamble", "grouping", "primitive", "plain",
"user",
-- "invisible", "data",
}
local predefined = {
'default', 'linenumber', 'bracelight', 'bracebad', 'controlchar',
'indentguide', 'calltip'
}
-- Bah ... ugly ... nicer would be a proper hash .. we now have properties
-- as well as STYLE_* and some connection between them ... why .. ok, we
-- could delay things but who cares. Anyway, at this moment the properties
-- are still unknown.
local function preparestyles(list)
for i=1,#list do
local k = list[i]
local K = upper(k)
local s = "style." .. k
lexers[K] = k -- is this used
lexers['STYLE_'..K] = '$(' .. k .. ')'
end
end
preparestyles(default)
preparestyles(predefined)
-- These helpers are set afterwards so we delay their initialization ... there
-- is no need to alias each time again and this way we can more easily adapt
-- to updates.
-- These keep changing (values, functions, tables ...) so we nee to check these
-- with each update. Some of them are set in the loader (the require 'lexer' is
-- in fact not a real one as the lexer code is loaded in the dll). It's also not
-- getting more efficient.
-- FOLD_BASE = lexers.FOLD_BASE or SC_FOLDLEVELBASE
-- FOLD_HEADER = lexers.FOLD_HEADER or SC_FOLDLEVELHEADERFLAG
-- FOLD_BLANK = lexers.FOLD_BLANK or SC_FOLDLEVELWHITEFLAG
-- get_style_at = lexers.get_style_at or GetStyleAt
-- get_indent_amount = lexers.get_indent_amount or GetIndentAmount
-- get_property = lexers.get_property or GetProperty
-- get_fold_level = lexers.get_fold_level or GetFoldLevel
-- It needs checking: do we have access to all properties now? I'll clean
-- this up anyway as I want a simple clean and stable model.
local FOLD_BASE = 0
local FOLD_HEADER = 0
local FOLD_BLANK = 0
local style_at = { }
local indent_amount = { }
local fold_level = { }
local function initialize()
FOLD_BASE = lexers.FOLD_BASE
FOLD_HEADER = lexers.FOLD_HEADER
FOLD_BLANK = lexers.FOLD_BLANK
--
style_at = lexers.style_at -- table
indent_amount = lexers.indent_amount -- table
fold_level = lexers.fold_level -- table
--
initialize = nil
--
end
local function get_property(tag,default)
return lexers.property_int[tag] or lexers.property[tag] or default
end
-- Do we really need this?
lexers.property_expanded = setmetatable({ }, {
__index = function(t, key)
return gsub(lexers.property[key],'[$%%]%b()', function(key)
return t[sub(key,3,-2)]
end)
end,
__newindex = function(t,k,v)
report("properties are read-only, '%s' is not changed",k)
end,
})
-- Style handler.
--
-- The property table will be set later (after loading) by the library. The
-- styleset is not needed any more as we predefine all styles as defaults
-- anyway (too bug sensitive otherwise).
local function toproperty(specification)
local serialized = { }
for key, value in next, specification do
if value == true then
serialized[#serialized+1] = key
elseif type(value) == "table" then
serialized[#serialized+1] = key .. ":" .. "#" .. value[1] .. value[2] .. value[3]
else
serialized[#serialized+1] = key .. ":" .. tostring(value)
end
end
return concat(serialized,",")
end
local function tostyles(styles)
local styleset = { }
local property = lexers.property or { }
for k, v in next, styles do
v = toproperty(v)
styleset[k] = v
property["style."..k] = v
end
return styleset
end
context.toproperty = toproperty
context.tostyles = tostyles
-- If we had one instance/state of Lua as well as all regular libraries
-- preloaded we could use the context base libraries. So, let's go poor-
-- mans solution now.
function context.registerstyles(styles)
local styleset = tostyles(styles)
context.styles = styles
context.styleset = styleset
if trace then
local t, n = { }, 0
for k, v in next, styleset do
t[#t+1] = k
if #k > n then
n = #k
end
end
table.sort(t)
local template = " %-" .. n .. "s : %s"
report("initializing styleset:")
for i=1,#t do
local k = t[i]
report(template,k,styleset[k])
end
end
end
-- Some spell checking related stuff. Unfortunately we cannot use a path set
-- by property.
local locations = {
"data", -- optional data directory
"..", -- regular scite directory
"lexers", -- new in 3.41 .. no tracing possible
"lexers/data", -- new in 3.41 .. no tracing possible
"../data", -- new in 3.41 .. no tracing possible
}
local function collect(name)
local okay, definitions = pcall(function () return require(name) end)
if okay then
if type(definitions) == "function" then
definitions = definitions()
end
if type(definitions) == "table" then
return definitions
end
else
end
end
function context.loaddefinitions(name)
for i=1,#locations do
local data = collect(locations[i] .. "/" .. name)
if data then
if trace then
report("definition file '%s' has been loaded",name)
end
return data
end
end
report("unable to load definition file '%s'",name)
end
function context.word_match(words,word_chars,case_insensitive)
local chars = '%w_' -- maybe just "" when word_chars
if word_chars then
chars = '^([' .. chars .. gsub(word_chars,'([%^%]%-])', '%%%1') ..']+)'
else
chars = '^([' .. chars ..']+)'
end
if case_insensitive then
local word_list = { }
for i=1,#words do
word_list[lower(words[i])] = true
end
return P(function(input, index)
local s, e, word = find(input,chars,index)
return word and word_list[lower(word)] and e + 1 or nil
end)
else
local word_list = { }
for i=1,#words do
word_list[words[i]] = true
end
return P(function(input, index)
local s, e, word = find(input,chars,index)
return word and word_list[word] and e + 1 or nil
end)
end
end
-- Patterns are grouped in a separate namespace but the regular lexers expect
-- shortcuts to be present in the lexers library. Maybe I'll incorporate some
-- of l-lpeg later.
do
local anything = P(1)
local idtoken = R("az","AZ","\127\255","__")
local digit = R("09")
local sign = S("+-")
local period = P(".")
local octdigit = R("07")
local hexdigit = R("09","AF","af")
local lower = R('az')
local upper = R('AZ')
local alpha = upper + lower
local space = S(" \n\r\t\f\v")
local eol = S("\r\n")
local backslash = P("\\")
local decimal = digit^1
local octal = P("0")
* octdigit^1
local hexadecimal = P("0") * S("xX")
* (hexdigit^0 * '.' * hexdigit^1 + hexdigit^1 * '.' * hexdigit^0 + hexdigit^1)
* (S("pP") * sign^-1 * hexdigit^1)^-1 -- *
patterns.idtoken = idtoken
patterns.digit = digit
patterns.sign = sign
patterns.period = period
patterns.octdigit = octdigit
patterns.hexdigit = hexdigit
patterns.ascii = R('\000\127') -- useless
patterns.extend = R('\000\255') -- useless
patterns.control = R('\000\031')
patterns.lower = lower
patterns.upper = upper
patterns.alpha = alpha
patterns.decimal = decimal
patterns.octal = octal
patterns.hexadecimal = hexadecimal
patterns.float = sign^-1
* (digit^0 * '.' * digit^1 + digit^1 * '.' * digit^0 + digit^1)
* S("eE") * sign^-1 * digit^1 -- *
patterns.cardinal = decimal
patterns.signeddecimal = sign^-1 * decimal
patterns.signedoctal = sign^-1 * octal
patterns.signedhexadecimal = sign^-1 * hexadecimal
patterns.integer = sign^-1 * (hexadecimal + octal + decimal)
patterns.real =
sign^-1 * ( -- at most one
digit^1 * period * digit^0 -- 10.0 10.
+ digit^0 * period * digit^1 -- 0.10 .10
+ digit^1 -- 10
)
patterns.anything = anything
patterns.any = anything
patterns.restofline = (1-eol)^1
patterns.space = space
patterns.spacing = space^1
patterns.nospacing = (1-space)^1
patterns.eol = eol
patterns.newline = P("\r\n") + eol
local endof = S("\n\r\f")
patterns.startofline = P(function(input,index)
return (index == 1 or lpegmatch(endof,input,index-1)) and index
end)
-- These are the expected ones for other lexers. Maybe all in own namespace
-- and provide compatibility layer.
lexers.any = anything
lexers.ascii = ascii
lexers.extend = extend
lexers.alpha = alpha
lexers.digit = digit
lexers.alnum = alnum
lexers.lower = lower
lexers.upper = upper
lexers.xdigit = hexdigit
lexers.cntrl = control
lexers.graph = R('!~')
lexers.print = R(' ~')
lexers.punct = R('!/', ':@', '[\'', '{~')
lexers.space = space
lexers.newline = S("\r\n\f")^1
lexers.nonnewline = 1 - lexers.newline
lexers.nonnewline_esc = 1 - (lexers.newline + '\\') + backslash * anything
lexers.dec_num = decimal
lexers.oct_num = octal
lexers.hex_num = hexadecimal
lexers.integer = integer
lexers.float = float
lexers.word = (alpha + '_') * (alpha + digit + '_')^0 -- weird, why digits
end
-- end of patterns
function context.exact_match(words,word_chars,case_insensitive)
local characters = concat(words)
local pattern -- the concat catches _ etc
if word_chars == true or word_chars == false or word_chars == nil then
word_chars = ""
end
if type(word_chars) == "string" then
pattern = S(characters) + patterns.idtoken
if case_insensitive then
pattern = pattern + S(upper(characters)) + S(lower(characters))
end
if word_chars ~= "" then
pattern = pattern + S(word_chars)
end
elseif word_chars then
pattern = word_chars
end
if case_insensitive then
local list = { }
if #words == 0 then
for k, v in next, words do
list[lower(k)] = v
end
else
for i=1,#words do
list[lower(words[i])] = true
end
end
return Cmt(pattern^1, function(_,i,s)
return list[lower(s)] -- and i or nil
end)
else
local list = { }
if #words == 0 then
for k, v in next, words do
list[k] = v
end
else
for i=1,#words do
list[words[i]] = true
end
end
return Cmt(pattern^1, function(_,i,s)
return list[s] -- and i or nil
end)
end
end
function context.just_match(words)
local p = P(words[1])
for i=2,#words do
p = p + P(words[i])
end
return p
end
-- spell checking (we can only load lua files)
--
-- return {
-- min = 3,
-- max = 40,
-- n = 12345,
-- words = {
-- ["someword"] = "someword",
-- ["anotherword"] = "Anotherword",
-- },
-- }
local lists = { }
function context.setwordlist(tag,limit) -- returns hash (lowercase keys and original values)
if not tag or tag == "" then
return false, 3
end
local list = lists[tag]
if not list then
list = context.loaddefinitions("spell-" .. tag)
if not list or type(list) ~= "table" then
list = { words = false, min = 3 }
else
list.words = list.words or false
list.min = list.min or 3
end
lists[tag] = list
end
return list.words, list.min
end
patterns.wordtoken = R("az","AZ","\127\255")
patterns.wordpattern = patterns.wordtoken^3 -- todo: if limit and #s < limit then
function context.checkedword(validwords,validminimum,s,i) -- ,limit
if not validwords then -- or #s < validminimum then
return true, "text", i -- true, "default", i
else
-- keys are lower
local word = validwords[s]
if word == s then
return true, "okay", i -- exact match
elseif word then
return true, "warning", i -- case issue
else
local word = validwords[lower(s)]
if word == s then
return true, "okay", i -- exact match
elseif word then
return true, "warning", i -- case issue
elseif upper(s) == s then
return true, "warning", i -- probably a logo or acronym
else
return true, "error", i
end
end
end
end
function context.styleofword(validwords,validminimum,s) -- ,limit
if not validwords or #s < validminimum then
return "text"
else
-- keys are lower
local word = validwords[s]
if word == s then
return "okay" -- exact match
elseif word then
return "warning" -- case issue
else
local word = validwords[lower(s)]
if word == s then
return "okay" -- exact match
elseif word then
return "warning" -- case issue
elseif upper(s) == s then
return "warning" -- probably a logo or acronym
else
return "error"
end
end
end
end
-- overloaded functions
local h_table, b_table, n_table = { }, { }, { } -- from the time small tables were used (optimization)
setmetatable(h_table, { __index = function(t,level) local v = { level, FOLD_HEADER } t[level] = v return v end })
setmetatable(b_table, { __index = function(t,level) local v = { level, FOLD_BLANK } t[level] = v return v end })
setmetatable(n_table, { __index = function(t,level) local v = { level } t[level] = v return v end })
local newline = patterns.newline
local p_yes = Cp() * Cs((1-newline)^1) * newline^-1
local p_nop = newline
local folders = { }
local function fold_by_parsing(text,start_pos,start_line,start_level,lexer)
local folder = folders[lexer]
if not folder then
--
local pattern, folds, text, start_pos, line_num, prev_level, current_level
--
local fold_symbols = lexer._foldsymbols
local fold_pattern = lexer._foldpattern -- use lpeg instead (context extension)
--
if fold_pattern then
-- if no functions are found then we could have a faster one
fold_pattern = Cp() * C(fold_pattern) / function(s,match)
local symbols = fold_symbols[style_at[start_pos + s]]
if symbols then
local l = symbols[match]
if l then
current_level = current_level + l
end
end
end
local action_y = function()
folds[line_num] = prev_level
if current_level > prev_level then
folds[line_num] = prev_level + FOLD_HEADER
end
if current_level < FOLD_BASE then
current_level = FOLD_BASE
end
prev_level = current_level
line_num = line_num + 1
end
local action_n = function()
folds[line_num] = prev_level + FOLD_BLANK
line_num = line_num + 1
end
pattern = ((fold_pattern + (1-newline))^1 * newline / action_y + newline/action_n)^0
else
-- the traditional one but a bit optimized
local fold_symbols_patterns = fold_symbols._patterns
local action_y = function(pos,line)
for j = 1, #fold_symbols_patterns do
for s, match in gmatch(line,fold_symbols_patterns[j]) do -- '()('..patterns[i]..')'
local symbols = fold_symbols[style_at[start_pos + pos + s - 1]]
local l = symbols and symbols[match]
local t = type(l)
if t == 'number' then
current_level = current_level + l
elseif t == 'function' then
current_level = current_level + l(text, pos, line, s, match)
end
end
end
folds[line_num] = prev_level
if current_level > prev_level then
folds[line_num] = prev_level + FOLD_HEADER
end
if current_level < FOLD_BASE then
current_level = FOLD_BASE
end
prev_level = current_level
line_num = line_num + 1
end
local action_n = function()
folds[line_num] = prev_level + FOLD_BLANK
line_num = line_num + 1
end
pattern = (p_yes/action_y + p_nop/action_n)^0
end
--
local reset_parser = lexer._reset_parser
--
folder = function(_text_,_start_pos_,_start_line_,_start_level_)
if reset_parser then
reset_parser()
end
folds = { }
text = _text_
start_pos = _start_pos_
line_num = _start_line_
prev_level = _start_level_
current_level = prev_level
lpegmatch(pattern,text)
-- make folds collectable
local t = folds
folds = nil
return t
end
folders[lexer] = folder
end
return folder(text,start_pos,start_line,start_level,lexer)
end
local folds, current_line, prev_level
local function action_y()
local current_level = FOLD_BASE + indent_amount[current_line]
if current_level > prev_level then -- next level
local i = current_line - 1
local f
while true do
f = folds[i]
if not f then
break
elseif f[2] == FOLD_BLANK then
i = i - 1
else
f[2] = FOLD_HEADER -- low indent
break
end
end
folds[current_line] = { current_level } -- high indent
elseif current_level < prev_level then -- prev level
local f = folds[current_line - 1]
if f then
f[1] = prev_level -- high indent
end
folds[current_line] = { current_level } -- low indent
else -- same level
folds[current_line] = { prev_level }
end
prev_level = current_level
current_line = current_line + 1
end
local function action_n()
folds[current_line] = { prev_level, FOLD_BLANK }
current_line = current_line + 1
end
local pattern = ( S("\t ")^0 * ( (1-patterns.eol)^1 / action_y + P(true) / action_n) * newline )^0
local function fold_by_indentation(text,start_pos,start_line,start_level)
-- initialize
folds = { }
current_line = start_line
prev_level = start_level
-- define
-- -- not here .. pattern binds and local functions are not frozen
-- analyze
lpegmatch(pattern,text)
-- flatten
for line, level in next, folds do
folds[line] = level[1] + (level[2] or 0)
end
-- done, make folds collectable
local t = folds
folds = nil
return t
end
local function fold_by_line(text,start_pos,start_line,start_level)
local folds = { }
-- can also be lpeg'd
for _ in gmatch(text,".-\r?\n") do
folds[start_line] = n_table[start_level] -- { start_level } -- stile tables ? needs checking
start_line = start_line + 1
end
return folds
end
local threshold_by_lexer = 512 * 1024 -- we don't know the filesize yet
local threshold_by_parsing = 512 * 1024 -- we don't know the filesize yet
local threshold_by_indentation = 512 * 1024 -- we don't know the filesize yet
local threshold_by_line = 512 * 1024 -- we don't know the filesize yet
function context.fold(lexer,text,start_pos,start_line,start_level) -- hm, we had size thresholds .. where did they go
if text == '' then
return { }
end
if initialize then
initialize()
end
local fold_by_lexer = lexer._fold
local fold_by_symbols = lexer._foldsymbols
local filesize = 0 -- we don't know that
if fold_by_lexer then
if filesize <= threshold_by_lexer then
return fold_by_lexer(text,start_pos,start_line,start_level,lexer)
end
elseif fold_by_symbols then -- and get_property('fold.by.parsing',1) > 0 then
if filesize <= threshold_by_parsing then
return fold_by_parsing(text,start_pos,start_line,start_level,lexer)
end
elseif get_property('fold.by.indentation',1) > 0 then
if filesize <= threshold_by_indentation then
return fold_by_indentation(text,start_pos,start_line,start_level,lexer)
end
elseif get_property('fold.by.line',1) > 0 then
if filesize <= threshold_by_line then
return fold_by_line(text,start_pos,start_line,start_level,lexer)
end
end
return { }
end
-- function context.fold(lexer,text,start_pos,start_line,start_level) -- hm, we had size thresholds .. where did they go
-- return { }
-- end
-- The following code is mostly unchanged:
local function add_rule(lexer,id,rule) -- unchanged
if not lexer._RULES then
lexer._RULES = { }
lexer._RULEORDER = { }
end
lexer._RULES[id] = rule
lexer._RULEORDER[#lexer._RULEORDER + 1] = id
end
local function add_style(lexer,token_name,style) -- unchanged (well, changed a bit around 3.41)
-- if not lexer._TOKENSTYLES[token_name] then
local num_styles = lexer._numstyles
if num_styles == 32 then
num_styles = num_styles + 8
end
if num_styles >= 255 then
report("there can't be more than %s styles",255)
end
lexer._TOKENSTYLES[token_name] = num_styles
lexer._EXTRASTYLES[token_name] = style
lexer._numstyles = num_styles + 1
-- end
end
-- At some point an 'any' append showed up in the original code ...
-- but I see no need to catch that case ... beter fix the specification.
local function join_tokens(lexer) -- slightly different from the original (no 'any' append)
local patterns = lexer._RULES
local order = lexer._RULEORDER
local token_rule = patterns[order[1]] -- normally whitespace
for i=2, #order do
token_rule = token_rule + patterns[order[i]]
end
if lexer._TYPE ~= "context" then
token_rule = token_rule + lexers.token(lexers.DEFAULT, patterns.any)
end
lexer._TOKENRULE = token_rule
return token_rule
end
local function add_lexer(grammar, lexer, token_rule) -- mostly the same as the original
local token_rule = join_tokens(lexer)
local lexer_name = lexer._NAME
local children = lexer._CHILDREN
for i=1,#children do
local child = children[i]
if child._CHILDREN then
add_lexer(grammar, child)
end
local child_name = child._NAME
local rules = child._EMBEDDEDRULES[lexer_name]
local rules_token_rule = grammar['__'..child_name] or rules.token_rule
grammar[child_name] = (-rules.end_rule * rules_token_rule)^0 * rules.end_rule^-1 * V(lexer_name)
local embedded_child = '_' .. child_name
grammar[embedded_child] = rules.start_rule * (-rules.end_rule * rules_token_rule)^0 * rules.end_rule^-1
token_rule = V(embedded_child) + token_rule
end
grammar['__' .. lexer_name] = token_rule
grammar[lexer_name] = token_rule^0
end
local function build_grammar(lexer,initial_rule) -- same as the original
local children = lexer._CHILDREN
if children then
local lexer_name = lexer._NAME
if not initial_rule then
initial_rule = lexer_name
end
local grammar = { initial_rule }
add_lexer(grammar, lexer)
lexer._INITIALRULE = initial_rule
lexer._GRAMMAR = Ct(P(grammar))
else
lexer._GRAMMAR = Ct(join_tokens(lexer)^0)
end
end
-- So far. We need these local functions in the next one. We have these
-- redefinitions because we memoize the lexers ... it looks like in
-- 3.1.4 something similar now happens with 'lexers'.
local lineparsers = { }
function context.lex(lexer,text,init_style)
-- local lexer = global._LEXER
local grammar = lexer._GRAMMAR
if initialize then
initialize()
end
if not grammar then
return { }
elseif lexer._LEXBYLINE then -- we could keep token
local tokens = { }
local offset = 0
local noftokens = 0
local lineparser = lineparsers[lexer]
if not lineparser then -- probably a cmt is more efficient
lineparser = C((1-newline)^0 * newline) / function(line)
local length = #line
local line_tokens = length > 0 and lpegmatch(grammar,line)
if line_tokens then
for i=1,#line_tokens,2 do
noftokens = noftokens + 1
tokens[noftokens] = line_tokens[i]
noftokens = noftokens + 1
tokens[noftokens] = line_tokens[i + 1] + offset
end
end
offset = offset + length
if noftokens > 0 and tokens[noftokens] ~= offset then
noftokens = noftokens + 1
tokens[noftokens] = 'default'
noftokens = noftokens + 1
tokens[noftokens] = offset + 1
end
end
lineparser = lineparser^0
lineparsers[lexer] = lineparser
end
lpegmatch(lineparser,text)
return tokens
elseif lexer._CHILDREN then
-- as we cannot print, tracing is not possible ... this might change as we can as well
-- generate them all in one go (sharing as much as possible)
local hash = lexer._HASH -- hm, was _hash
if not hash then
hash = { }
lexer._HASH = hash
end
grammar = hash[init_style]
if grammar then
lexer._GRAMMAR = grammar
else
for style, style_num in next, lexer._TOKENSTYLES do
if style_num == init_style then
-- the name of the lexers is filtered from the whitespace
-- specification
local lexer_name = match(style,'^(.+)_whitespace') or lexer._NAME
if lexer._INITIALRULE ~= lexer_name then
grammar = hash[lexer_name]
if not grammar then
build_grammar(lexer,lexer_name)
grammar = lexer._GRAMMAR
hash[lexer_name] = grammar
end
end
break
end
end
grammar = grammar or lexer._GRAMMAR
hash[init_style] = grammar
end
return lpegmatch(grammar,text)
else
return lpegmatch(grammar,text)
end
end
-- so far
-- todo: keywords: one lookup and multiple matches
-- function context.token(name, patt)
-- return Ct(patt * Cc(name) * Cp())
-- end
--
-- -- hm, changed in 3.24 .. no longer a table
function context.token(name, patt)
return patt * Cc(name) * Cp()
end
-- The next ones were mostly unchanged (till now), we moved it here when 3.41
-- became close to impossible to combine with cq. overload and a merge was
-- the only solution. It makes later updates more painful but the update to
-- 3.41 was already a bit of a nightmare anyway.
-- Loading lexers is rather interwoven with what the dll/so sets and
-- it changes over time. So, we need to keep an eye on changes. One
-- problem that we always faced were the limitations in length of
-- lexer names (as they get app/prepended occasionally to strings with
-- a hard coded limit). So, we always used alternative names and now need
-- to make sure this doesn't clash. As I no longer intend to use shipped
-- lexers I could strip away some of the code in the future, but keeping
-- it as reference makes sense.
-- I spend quite some time figuring out why 3.41 didn't work or crashed which
-- is hard when no stdout is available and when the io library is absent. In
-- the end of of the problems was in the _NAME setting. We set _NAME
-- to e.g. 'tex' but load from a file with a longer name, which we do
-- as we don't want to clash with existing files, we end up in
-- lexers not being found.
local function check_properties()
if not lexers.property then
lexers.property = { }
lexers.property_int = setmetatable({ }, {
__index = function(t,k)
return tostring(tonumber(lexers.property[k]) or 0)
end,
__newindex = function(t,k,v)
report("properties are read-only, '%s' is not changed",k)
end,
})
end
end
local function check_styles(lexer)
local numstyles = #default
local tokenstyles = { }
for i=1, #default do
tokenstyles[default[i]] = i - 1
end
for i=1, #predefined do
tokenstyles[predefined[i]] = i + 31
end
lexer._TOKENSTYLES = tokenstyles
lexer._numstyles = numstyles
lexer._EXTRASTYLES = { }
return lexer
end
local whitespaces = { }
local function push_whitespace(name)
table.insert(whitespaces,lexers.WHITESPACE or "whitespace")
lexers.WHITESPACE = name .. "_whitespace"
end
local function pop_whitespace()
lexers.WHITESPACE = table.remove(whitespaces) or "whitespace"
end
local function check_whitespace(lexer,name)
if lexer then
lexer.whitespace = (name or lexer.name or lexer._NAME) .. "_whitespace"
end
end
function context.new(name,filename)
local lexer = {
_TYPE = "context",
--
_NAME = name, -- used for token building
_FILENAME = filename, -- for diagnostic purposed
--
name = name,
filename = filename,
whitespace = whitespace,
}
if trace then
report("initializing lexer tagged '%s' from file '%s'",name,filename or name)
end
check_styles(lexer)
check_whitespace(lexer)
return lexer
end
local function nolexer(name)
local lexer = {
_TYPE = "unset",
_NAME = name,
-- _rules = { },
}
check_styles(lexer)
check_whitespace(lexer)
return lexer
end
local function load_lexer(name)
local lexer, okay = nil, false
-- first locate the file (methods have changed over time)
local lexer_file = package.searchpath(name,lexers.LEXERPATH)
if not lexer_file or lexer_file == "" then
report("lexer file '%s' can't be located",name)
else
if trace then
report("loading lexer file '%s'",lexer_file)
end
push_whitespace(name) -- for traditional lexers .. no alt_name yet
okay, lexer = pcall(dofile, lexer_file or '')
pop_whitespace()
if not okay then
report("invalid lexer file '%s'",lexer_file)
elseif trace then
report("lexer file '%s' has been loaded",lexer_file)
end
end
if type(lexer) ~= "table" then
return nolexer(name)
end
if lexer._TYPE ~= "context" then
lexer._TYPE = "native"
check_styles(lexer)
check_whitespace(lexer,name)
end
if not lexer._NAME then
lexer._NAME = name -- so: filename
end
return lexer
end
-- An optional second argument has been introduced so that one can embed a lexer
-- more than once ... maybe something to look into (as not it's done by remembering
-- the start sequence ... quite okay but maybe suboptimal ... anyway, never change
-- a working solution).
function context.load(filename)
nesting = nesting + 1
local lexer = usedlexers[filename] -- we load by filename but the internal name can be short
if lexer then
if trace then
report("reusing lexer '%s'",filename)
end
nesting = nesting - 1
return lexer
end
if trace then
report("loading lexer '%s'",filename)
end
--
check_properties()
--
parent_lexer = nil
--
lexer = load_lexer(filename) or nolexer(name)
usedlexers[filename] = lexer
--
if not lexer._rules and not lexer._lexer then
lexer._lexer = parent_lexer
end
--
if lexer._lexer then
local _l = lexer._lexer
local _r = lexer._rules
local _s = lexer._tokenstyles
if not _l._tokenstyles then
_l._tokenstyles = { }
end
if _r then
local rules = _l._rules
local name = lexer.name
for i=1,#_r do
local rule = _r[i]
rules[#rules + 1] = {
name .. '_' .. rule[1],
rule[2],
}
end
end
if _s then
local tokenstyles = _l._tokenstyles
for token, style in next, _s do
tokenstyles[token] = style
end
end
lexer = l
end
--
local _r = lexer._rules
if _r then
local _s = lexer._tokenstyles
if _s then
for token, style in next, _s do
add_style(lexer, token, style)
end
end
for i=1,#_r do
local rule = _r[i]
add_rule(lexer, rule[1], rule[2])
end
build_grammar(lexer)
end
--
add_style(lexer, lexer.whitespace, lexers.STYLE_WHITESPACE)
--
local foldsymbols = lexer._foldsymbols
if foldsymbols then
local patterns = foldsymbols._patterns
if patterns then
for i = 1, #patterns do
patterns[i] = '()(' .. patterns[i] .. ')'
end
end
end
--
lexer.lex = lexers.lex
lexer.fold = lexers.fold
--
nesting = nesting - 1
--
return lexer
end
function context.embed_lexer(parent, child, start_rule, end_rule) -- mostly the same as the original
local embeddedrules = child._EMBEDDEDRULES
if not embeddedrules then
embeddedrules = { }
child._EMBEDDEDRULES = embeddedrules
end
if not child._RULES then
local rules = child._rules
if not rules then
report("child lexer '%s' has no rules",chile._NAME or "unknown")
rules = { }
child._rules = rules
end
for i=1,#rules do
local rule = rules[i]
add_rule(child, rule[1], rule[2])
end
end
embeddedrules[parent._NAME] = {
['start_rule'] = start_rule,
['token_rule'] = join_tokens(child),
['end_rule'] = end_rule
}
local children = parent._CHILDREN
if not children then
children = { }
parent._CHILDREN = children
end
children[#children + 1] = child
local tokenstyles = parent._tokenstyles
if not tokenstyles then
tokenstyles = { }
parent._tokenstyles = tokenstyles
end
tokenstyles[child._NAME..'_whitespace'] = lexers.STYLE_WHITESPACE -- check what whitespace
local childstyles = child._tokenstyles
if childstyles then
for token, style in next, childstyles do
tokenstyles[token] = style
end
end
child._lexer = parent
parent_lexer = parent
end
-- we now move the adapted code to the lexers namespace
lexers.new = context.new
lexers.load = context.load
lexers.embed_lexer = context.embed_lexer
lexers.fold = context.fold
lexers.lex = context.lex
lexers.token = context.token
lexers.word_match = context.word_match
lexers.exact_match = context.exact_match
lexers.just_match = context.just_match
-- helper .. alas ... the lexer's lua instance is rather crippled .. not even
-- math is part of it
do
local floor = math and math.floor
local char = string.char
if not floor then
floor = function(n)
return tonumber(format("%d",n))
end
math = math or { }
math.floor = floor
end
local function utfchar(n)
if n < 0x80 then
return char(n)
elseif n < 0x800 then
return char(
0xC0 + floor(n/0x40),
0x80 + (n % 0x40)
)
elseif n < 0x10000 then
return char(
0xE0 + floor(n/0x1000),
0x80 + (floor(n/0x40) % 0x40),
0x80 + (n % 0x40)
)
elseif n < 0x40000 then
return char(
0xF0 + floor(n/0x40000),
0x80 + floor(n/0x1000),
0x80 + (floor(n/0x40) % 0x40),
0x80 + (n % 0x40)
)
else
-- return char(
-- 0xF1 + floor(n/0x1000000),
-- 0x80 + floor(n/0x40000),
-- 0x80 + floor(n/0x1000),
-- 0x80 + (floor(n/0x40) % 0x40),
-- 0x80 + (n % 0x40)
-- )
return "?"
end
end
context.utfchar = utfchar
-- a helper from l-lpeg:
local function make(t)
local p
for k, v in next, t do
if not p then
if next(v) then
p = P(k) * make(v)
else
p = P(k)
end
else
if next(v) then
p = p + P(k) * make(v)
else
p = p + P(k)
end
end
end
return p
end
function lpeg.utfchartabletopattern(list)
local tree = { }
for i=1,#list do
local t = tree
for c in gmatch(list[i],".") do
if not t[c] then
t[c] = { }
end
t = t[c]
end
end
return make(tree)
end
patterns.invisibles = lpeg.utfchartabletopattern {
utfchar(0x00A0), -- nbsp
utfchar(0x2000), -- enquad
utfchar(0x2001), -- emquad
utfchar(0x2002), -- enspace
utfchar(0x2003), -- emspace
utfchar(0x2004), -- threeperemspace
utfchar(0x2005), -- fourperemspace
utfchar(0x2006), -- sixperemspace
utfchar(0x2007), -- figurespace
utfchar(0x2008), -- punctuationspace
utfchar(0x2009), -- breakablethinspace
utfchar(0x200A), -- hairspace
utfchar(0x200B), -- zerowidthspace
utfchar(0x202F), -- narrownobreakspace
utfchar(0x205F), -- math thinspace
}
-- now we can make:
patterns.iwordtoken = patterns.wordtoken - patterns.invisibles
patterns.iwordpattern = patterns.iwordtoken^3
end
-- The following helpers are not used, partyally replace by other mechanism and
-- when needed I'll first optimize them. I only made them somewhat more readable.
function lexers.delimited_range(chars, single_line, no_escape, balanced) -- unchanged
local s = sub(chars,1,1)
local e = #chars == 2 and sub(chars,2,2) or s
local range
local b = balanced and s or ''
local n = single_line and '\n' or ''
if no_escape then
local invalid = S(e .. n .. b)
range = patterns.any - invalid
else
local invalid = S(e .. n .. b) + patterns.backslash
range = patterns.any - invalid + patterns.backslash * patterns.any
end
if balanced and s ~= e then
return P {
s * (range + V(1))^0 * e
}
else
return s * range^0 * P(e)^-1
end
end
function lexers.starts_line(patt) -- unchanged
return P ( function(input, index)
if index == 1 then
return index
end
local char = sub(input,index - 1,index - 1)
if char == '\n' or char == '\r' or char == '\f' then
return index
end
end ) * patt
end
function lexers.last_char_includes(s) -- unchanged
s = '[' .. gsub(s,'[-%%%[]', '%%%1') .. ']'
return P ( function(input, index)
if index == 1 then
return index
end
local i = index
while match(sub(input,i - 1,i - 1),'[ \t\r\n\f]') do
i = i - 1
end
if match(sub(input,i - 1,i - 1),s) then
return index
end
end)
end
function lexers.nested_pair(start_chars, end_chars) -- unchanged
local s = start_chars
local e = P(end_chars)^-1
return P {
s * (patterns.any - s - end_chars + V(1))^0 * e
}
end
local function prev_line_is_comment(prefix, text, pos, line, s) -- unchanged
local start = find(line,'%S')
if start < s and not find(line,prefix,start,true) then
return false
end
local p = pos - 1
if sub(text,p,p) == '\n' then
p = p - 1
if sub(text,p,p) == '\r' then
p = p - 1
end
if sub(text,p,p) ~= '\n' then
while p > 1 and sub(text,p - 1,p - 1) ~= '\n'
do p = p - 1
end
while find(sub(text,p,p),'^[\t ]$') do
p = p + 1
end
return sub(text,p,p + #prefix - 1) == prefix
end
end
return false
end
local function next_line_is_comment(prefix, text, pos, line, s)
local p = find(text,'\n',pos + s)
if p then
p = p + 1
while find(sub(text,p,p),'^[\t ]$') do
p = p + 1
end
return sub(text,p,p + #prefix - 1) == prefix
end
return false
end
function lexers.fold_line_comments(prefix)
local property_int = lexers.property_int
return function(text, pos, line, s)
if property_int['fold.line.comments'] == 0 then
return 0
end
if s > 1 and match(line,'^%s*()') < s then
return 0
end
local prev_line_comment = prev_line_is_comment(prefix, text, pos, line, s)
local next_line_comment = next_line_is_comment(prefix, text, pos, line, s)
if not prev_line_comment and next_line_comment then
return 1
end
if prev_line_comment and not next_line_comment then
return -1
end
return 0
end
end
-- done
return lexers
|