summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMarius <mariausol@gmail.com>2010-11-26 22:40:20 +0200
committerMarius <mariausol@gmail.com>2010-11-26 22:40:20 +0200
commitcc2e8cc985157fdab4e0122b27457d7cd21b2d38 (patch)
tree4cdac08071a209d3af472899db9bbd86b1e10563
parent44d0f4fe01338d0f699e1f09ae13d17088c2db83 (diff)
downloadcontext-cc2e8cc985157fdab4e0122b27457d7cd21b2d38.tar.gz
beta 2010.11.26 21:21
-rw-r--r--context/data/scite/context.properties2
-rw-r--r--scripts/context/lua/mtxrun.lua316
-rw-r--r--scripts/context/stubs/mswin/mtxrun.lua316
-rw-r--r--scripts/context/stubs/unix/mtxrun316
-rw-r--r--tex/context/base/buff-ini.lua565
-rw-r--r--tex/context/base/buff-ini.mkiv345
-rw-r--r--tex/context/base/buff-ver.lua377
-rw-r--r--tex/context/base/buff-ver.mkiv1273
-rw-r--r--tex/context/base/buff-vis.lua74
-rw-r--r--tex/context/base/buff-vis.mkiv24
-rw-r--r--tex/context/base/char-utf.lua8
-rw-r--r--tex/context/base/char-utf.mkiv8
-rw-r--r--tex/context/base/cldf-com.lua30
-rw-r--r--tex/context/base/cldf-ver.lua2
-rw-r--r--tex/context/base/colo-ini.mkiv13
-rw-r--r--tex/context/base/cont-new.mkiv10
-rw-r--r--tex/context/base/cont-new.tex2
-rw-r--r--tex/context/base/context.mkiv12
-rw-r--r--tex/context/base/context.tex2
-rw-r--r--tex/context/base/core-sys.mkiv107
-rw-r--r--tex/context/base/data-env.lua40
-rw-r--r--tex/context/base/data-met.lua7
-rw-r--r--tex/context/base/data-sch.lua21
-rw-r--r--tex/context/base/data-tex.lua296
-rw-r--r--tex/context/base/font-ini.mkiv42
-rw-r--r--tex/context/base/font-otn.lua2
-rw-r--r--tex/context/base/font-syn.lua2
-rw-r--r--tex/context/base/l-lpeg.lua82
-rw-r--r--tex/context/base/l-pdfview.lua3
-rw-r--r--tex/context/base/l-unicode.lua291
-rw-r--r--tex/context/base/lang-ini.lua5
-rw-r--r--tex/context/base/lang-ini.mkiv2
-rw-r--r--tex/context/base/luat-lib.mkiv1
-rw-r--r--tex/context/base/luat-mac.lua162
-rw-r--r--tex/context/base/lxml-dir.lua2
-rw-r--r--tex/context/base/math-ini.lua10
-rw-r--r--tex/context/base/math-ini.mkiv4
-rw-r--r--tex/context/base/mlib-pdf.lua16
-rw-r--r--tex/context/base/mlib-pps.lua18
-rw-r--r--tex/context/base/mult-aux.mkiv48
-rw-r--r--tex/context/base/mult-cld.lua129
-rw-r--r--tex/context/base/mult-ini.mkiv8
-rw-r--r--tex/context/base/node-aux.lua5
-rw-r--r--tex/context/base/node-pro.lua22
-rw-r--r--tex/context/base/node-tra.lua2
-rw-r--r--tex/context/base/node-tsk.lua2
-rw-r--r--tex/context/base/regi-ini.lua42
-rw-r--r--tex/context/base/s-abr-01.tex5
-rw-r--r--tex/context/base/scrn-int.mkiv1
-rw-r--r--tex/context/base/scrp-ini.lua4
-rw-r--r--tex/context/base/strc-sec.mkiv2
-rw-r--r--tex/context/base/tabl-ntb.mkiv43
-rw-r--r--tex/context/base/tabl-tab.mkiv507
-rw-r--r--tex/context/base/tabl-tbl.mkiv26
-rw-r--r--tex/context/base/util-prs.lua4
-rw-r--r--tex/context/base/util-seq.lua137
-rw-r--r--tex/context/base/v-default.lua40
-rw-r--r--tex/context/base/v-default.mkiv43
-rw-r--r--tex/context/base/v-escaped.lua14
-rw-r--r--tex/context/base/v-escaped.mkiv18
-rw-r--r--tex/context/base/v-lua.lua238
-rw-r--r--tex/context/base/v-lua.mkiv61
-rw-r--r--tex/context/base/v-nested.lua80
-rw-r--r--tex/context/base/v-nested.mkiv21
-rw-r--r--tex/context/base/v-tex.lua94
-rw-r--r--tex/context/base/v-tex.mkiv31
-rw-r--r--tex/context/base/v-xml.lua133
-rw-r--r--tex/context/base/v-xml.mkiv71
-rw-r--r--tex/generic/context/luatex-fonts-merged.lua86
69 files changed, 3802 insertions, 2923 deletions
diff --git a/context/data/scite/context.properties b/context/data/scite/context.properties
index 13fa079c0..006cfe640 100644
--- a/context/data/scite/context.properties
+++ b/context/data/scite/context.properties
@@ -39,7 +39,7 @@ textwrapper.length=68
file.patterns.tex=
file.patterns.latex=
-file.patterns.context=*.tex;*.tui;*.tuo;*.sty;*.mkii;*.mkiv;
+file.patterns.context=*.tex;*.tui;*.tuo;*.sty;*.mkii;*.mkiv;*.mkvi;
# ConTeXt : keywords (if not yet loaded already)
diff --git a/scripts/context/lua/mtxrun.lua b/scripts/context/lua/mtxrun.lua
index 7b5348424..c41de7fb8 100644
--- a/scripts/context/lua/mtxrun.lua
+++ b/scripts/context/lua/mtxrun.lua
@@ -177,6 +177,9 @@ local lpeg = require("lpeg")
local type = type
+-- Beware, we predefine a bunch of patterns here and one reason for doing so
+-- is that we get consistent behaviour in some of the visualizers.
+
lpeg.patterns = lpeg.patterns or { } -- so that we can share
local patterns = lpeg.patterns
@@ -193,19 +196,38 @@ local alwaysmatched = P(true)
patterns.anything = anything
patterns.endofstring = endofstring
patterns.beginofstring = alwaysmatched
+patterns.alwaysmatched = alwaysmatched
local digit, sign = R('09'), S('+-')
local cr, lf, crlf = P("\r"), P("\n"), P("\r\n")
+local newline = crlf + cr + lf
local utf8next = R("\128\191")
local escaped = P("\\") * anything
local squote = P("'")
local dquote = P('"')
+local space = P(" ")
+
+patterns.somecontent = (anything - newline - space)^1
+patterns.beginline = #(1-newline)
+
+local utfbom_32_be = P('\000\000\254\255')
+local utfbom_32_le = P('\255\254\000\000')
+local utfbom_16_be = P('\255\254')
+local utfbom_16_le = P('\254\255')
+local utfbom_8 = P('\239\187\191')
+local utfbom = utfbom_32_be + utfbom_32_le
+ + utfbom_16_be + utfbom_16_le
+ + utfbom_8
+local utftype = utfbom_32_be / "utf-32-be" + utfbom_32_le / "utf-32-le"
+ + utfbom_16_be / "utf-16-be" + utfbom_16_le / "utf-16-le"
+ + utfbom_8 / "utf-8" + alwaysmatched / "unknown"
patterns.utf8one = R("\000\127")
patterns.utf8two = R("\194\223") * utf8next
patterns.utf8three = R("\224\239") * utf8next * utf8next
patterns.utf8four = R("\240\244") * utf8next * utf8next * utf8next
-patterns.utfbom = P('\000\000\254\255') + P('\255\254\000\000') + P('\255\254') + P('\254\255') + P('\239\187\191')
+patterns.utfbom = utfbom
+patterns.utftype = utftype
local utf8char = patterns.utf8one + patterns.utf8two + patterns.utf8three + patterns.utf8four
local validutf8char = utf8char^0 * endofstring * Cc(true) + Cc(false)
@@ -231,24 +253,30 @@ patterns.hexadecimal = P("0x") * R("09","AF","af")^1
patterns.lowercase = R("az")
patterns.uppercase = R("AZ")
patterns.letter = patterns.lowercase + patterns.uppercase
-patterns.space = P(" ")
+patterns.space = space
patterns.tab = P("\t")
patterns.spaceortab = patterns.space + patterns.tab
patterns.eol = S("\n\r")
patterns.spacer = S(" \t\f\v") -- + string.char(0xc2, 0xa0) if we want utf (cf mail roberto)
-patterns.newline = crlf + cr + lf
-patterns.nonspace = 1 - patterns.space
+patterns.newline = newline
+patterns.emptyline = newline^1
patterns.nonspacer = 1 - patterns.spacer
patterns.whitespace = patterns.eol + patterns.spacer
patterns.nonwhitespace = 1 - patterns.whitespace
+patterns.equal = P("=")
patterns.comma = P(",")
patterns.commaspacer = P(",") * patterns.spacer^0
patterns.period = P(".")
+patterns.colon = P(":")
+patterns.semicolon = P(";")
+patterns.underscore = P("_")
patterns.escaped = escaped
patterns.squote = squote
patterns.dquote = dquote
-patterns.undouble = (dquote/"") * ((escaped + (1-dquote))^0) * (dquote/"")
-patterns.unsingle = (squote/"") * ((escaped + (1-squote))^0) * (squote/"")
+patterns.nosquote = (escaped + (1-squote))^0
+patterns.nodquote = (escaped + (1-dquote))^0
+patterns.unsingle = (squote/"") * patterns.nosquote * (squote/"")
+patterns.undouble = (dquote/"") * patterns.nodquote * (dquote/"")
patterns.unquoted = patterns.undouble + patterns.unsingle -- more often undouble
patterns.unspacer = ((patterns.spacer^1)/"")^0
@@ -267,19 +295,6 @@ function lpeg.splitter(pattern, action)
return (((1-P(pattern))^1)/action+1)^0
end
-local spacing = patterns.spacer^0 * patterns.newline -- sort of strip
-local empty = spacing * Cc("")
-local nonempty = Cs((1-spacing)^1) * spacing^-1
-local content = (empty + nonempty)^1
-
-local capture = Ct(content^0)
-
-function string.splitlines(str)
- return match(capture,str)
-end
-
-patterns.textline = content
-
local splitters_s, splitters_m = { }, { }
local function splitat(separator,single)
@@ -322,6 +337,30 @@ function string.split(str,separator)
return match(c,str)
end
+local spacing = patterns.spacer^0 * newline -- sort of strip
+local empty = spacing * Cc("")
+local nonempty = Cs((1-spacing)^1) * spacing^-1
+local content = (empty + nonempty)^1
+
+patterns.textline = content
+
+
+local linesplitter = Ct(splitat(newline))
+
+patterns.linesplitter = linesplitter
+
+function string.splitlines(str)
+ return match(linesplitter,str)
+end
+
+local utflinesplitter = utfbom^-1 * Ct(splitat(newline))
+
+patterns.utflinesplitter = utflinesplitter
+
+function string.utfsplitlines(str)
+ return match(utflinesplitter,str)
+end
+
local cache = { }
@@ -3247,6 +3286,8 @@ utf = utf or unicode.utf8
local concat, utfchar, utfgsub = table.concat, utf.char, utf.gsub
local char, byte, find, bytepairs, utfvalues, format = string.char, string.byte, string.find, string.bytepairs, string.utfvalues, string.format
+local utfsplitlines = string.utfsplitlines
+
-- 0 EF BB BF UTF-8
-- 1 FF FE UTF-16-little-endian
-- 2 FE FF UTF-16-big-endian
@@ -3291,111 +3332,131 @@ function unicode.utftype(f)
end
end
-function unicode.utf16_to_utf8(str, endian) -- maybe a gsub is faster or an lpeg
- local result, tmp, n, m, p, r, t = { }, { }, 0, 0, 0, 0, 0 -- we reuse tmp
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+
+
+local function utf16_to_utf8_be(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*left + right
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
+ r = r + 1
+ result[r] = utfchar(now)
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- for l,r in bytepairs(str) do
- if r then
- if endian then
- n = 256*l + r
- else
- n = 256*r + l
- end
- if m > 0 then
- n = (m-0xD800)*0x400 + (n-0xDC00) + 0x10000
- m = 0
- doit()
- elseif n >= 0xD800 and n <= 0xDBFF then
- m = n
- else
- doit()
+ return t
+end
+
+local function utf16_to_utf8_le(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*right + left
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
+ r = r + 1
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
+ r = r + 1
+ result[r] = utfchar(now)
+ end
end
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- end
- return result
+ return t
end
-function unicode.utf32_to_utf8(str, endian)
- local result, tmp, n, m, p, r, t = { }, { }, 0, -1, 0, 0, 0
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+local function utf32_to_utf8_be(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*256*256*a + 256*256*b
+ else
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[t] = utfchar(more + 256*a + b)
+ more = -1
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ else
+ break
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r)
+ end
+ return result
+end
+
+local function utf32_to_utf8_le(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
end
- for a,b in bytepairs(str) do
- if a and b then
- if m < 0 then
- if endian then
- m = 256*256*256*a + 256*256*b
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*b + a
else
- m = 256*b + a
+ r = r + 1
+ result[t] = utfchar(more + 256*256*256*b + 256*256*a)
+ more = -1
end
else
- if endian then
- n = m + 256*a + b
- else
- n = m + 256*256*256*b + 256*256*a
- end
- m = -1
- doit()
+ break
end
- else
- break
end
- end
- if #tmp > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
+ t[i] = concat(result,"",1,r)
end
return result
end
+unicode.utf32_to_utf8_be = utf32_to_utf8_be
+unicode.utf32_to_utf8_le = utf32_to_utf8_le
+unicode.utf16_to_utf8_be = utf16_to_utf8_be
+unicode.utf16_to_utf8_le = utf16_to_utf8_le
+
+function unicode.utf8_to_utf8(t)
+ return type(t) == "string" and utfsplitlines(t) or t
+end
+
+function unicode.utf16_to_utf8(t,endian)
+ return endian and utf16_to_utf8_be(t) or utf16_to_utf8_le(t) or t
+end
+
+function unicode.utf32_to_utf8(t,endian)
+ return endian and utf32_to_utf8_be(t) or utf32_to_utf8_le(t) or t
+end
+
local function little(c)
local b = byte(c)
if b < 0x10000 then
@@ -3436,6 +3497,10 @@ function unicode.utfcodes(str)
end
+function unicode.filetype(data)
+ return data and lpeg.match(lpeg.patterns.utftype,data) or "unknown"
+end
+
end -- of closure
@@ -3905,11 +3970,11 @@ local pattern_b = spaces * comma^0 * spaces * (key * ((spaces * equal * spaces *
local hash = { }
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
@@ -9676,7 +9741,7 @@ local suffixes = allocate() resolvers.suffixes = suffixes
local dangerous = allocate() resolvers.dangerous = dangerous
local suffixmap = allocate() resolvers.suffixmap = suffixmap
-local relations = allocate {
+local relations = allocate { -- todo: handlers also here
core = {
ofm = {
names = { "ofm", "omega font metric", "omega font metrics" },
@@ -9746,7 +9811,7 @@ local relations = allocate {
tex = {
names = { "tex" },
variable = 'TEXINPUTS',
- suffixes = { 'tex', "mkiv", "mkii" },
+ suffixes = { 'tex', "mkiv", "mkiv", "mkii" },
},
icc = {
names = { "icc", "icc profile", "icc profiles" },
@@ -9860,29 +9925,33 @@ resolvers.relations = relations
-- formats: maps a format onto a variable
-for category, categories in next, relations do
- for name, relation in next, categories do
- local rn = relation.names
- local rv = relation.variable
- local rs = relation.suffixes
- if rn and rv then
- for i=1,#rn do
- local rni = lower(gsub(rn[i]," ",""))
- formats[rni] = rv
- if rs then
- suffixes[rni] = rs
- for i=1,#rs do
- local rsi = rs[i]
- suffixmap[rsi] = rni
+function resolvers.updaterelations()
+ for category, categories in next, relations do
+ for name, relation in next, categories do
+ local rn = relation.names
+ local rv = relation.variable
+ local rs = relation.suffixes
+ if rn and rv then
+ for i=1,#rn do
+ local rni = lower(gsub(rn[i]," ",""))
+ formats[rni] = rv
+ if rs then
+ suffixes[rni] = rs
+ for i=1,#rs do
+ local rsi = rs[i]
+ suffixmap[rsi] = rni
+ end
end
end
end
- end
- if rs then
+ if rs then
+ end
end
end
end
+resolvers.updaterelations() -- push this in the metatable -> newindex
+
local function simplified(t,k)
return rawget(t,lower(gsub(k," ","")))
end
@@ -10345,7 +10414,7 @@ resolvers.locators = allocate { notfound = { nil } } -- locate databases
resolvers.hashers = allocate { notfound = { nil } } -- load databases
resolvers.generators = allocate { notfound = { nil } } -- generate databases
-function resolvers.splitmethod(filename)
+function resolvers.splitmethod(filename) -- todo: trigger by suffix
if not filename then
return { } -- safeguard
elseif type(filename) == "table" then
@@ -10364,10 +10433,13 @@ function resolvers.methodhandler(what, filename, filetype) -- ...
local resolver = resolvers[what]
if resolver[scheme] then
if trace_locating then
- report_resolvers("handler '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ report_resolvers("using special handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
end
return resolver[scheme](filename,filetype)
else
+ if trace_locating then
+ report_resolvers("no handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ end
return resolver.tex(filename,filetype) -- todo: specification
end
end
diff --git a/scripts/context/stubs/mswin/mtxrun.lua b/scripts/context/stubs/mswin/mtxrun.lua
index 7b5348424..c41de7fb8 100644
--- a/scripts/context/stubs/mswin/mtxrun.lua
+++ b/scripts/context/stubs/mswin/mtxrun.lua
@@ -177,6 +177,9 @@ local lpeg = require("lpeg")
local type = type
+-- Beware, we predefine a bunch of patterns here and one reason for doing so
+-- is that we get consistent behaviour in some of the visualizers.
+
lpeg.patterns = lpeg.patterns or { } -- so that we can share
local patterns = lpeg.patterns
@@ -193,19 +196,38 @@ local alwaysmatched = P(true)
patterns.anything = anything
patterns.endofstring = endofstring
patterns.beginofstring = alwaysmatched
+patterns.alwaysmatched = alwaysmatched
local digit, sign = R('09'), S('+-')
local cr, lf, crlf = P("\r"), P("\n"), P("\r\n")
+local newline = crlf + cr + lf
local utf8next = R("\128\191")
local escaped = P("\\") * anything
local squote = P("'")
local dquote = P('"')
+local space = P(" ")
+
+patterns.somecontent = (anything - newline - space)^1
+patterns.beginline = #(1-newline)
+
+local utfbom_32_be = P('\000\000\254\255')
+local utfbom_32_le = P('\255\254\000\000')
+local utfbom_16_be = P('\255\254')
+local utfbom_16_le = P('\254\255')
+local utfbom_8 = P('\239\187\191')
+local utfbom = utfbom_32_be + utfbom_32_le
+ + utfbom_16_be + utfbom_16_le
+ + utfbom_8
+local utftype = utfbom_32_be / "utf-32-be" + utfbom_32_le / "utf-32-le"
+ + utfbom_16_be / "utf-16-be" + utfbom_16_le / "utf-16-le"
+ + utfbom_8 / "utf-8" + alwaysmatched / "unknown"
patterns.utf8one = R("\000\127")
patterns.utf8two = R("\194\223") * utf8next
patterns.utf8three = R("\224\239") * utf8next * utf8next
patterns.utf8four = R("\240\244") * utf8next * utf8next * utf8next
-patterns.utfbom = P('\000\000\254\255') + P('\255\254\000\000') + P('\255\254') + P('\254\255') + P('\239\187\191')
+patterns.utfbom = utfbom
+patterns.utftype = utftype
local utf8char = patterns.utf8one + patterns.utf8two + patterns.utf8three + patterns.utf8four
local validutf8char = utf8char^0 * endofstring * Cc(true) + Cc(false)
@@ -231,24 +253,30 @@ patterns.hexadecimal = P("0x") * R("09","AF","af")^1
patterns.lowercase = R("az")
patterns.uppercase = R("AZ")
patterns.letter = patterns.lowercase + patterns.uppercase
-patterns.space = P(" ")
+patterns.space = space
patterns.tab = P("\t")
patterns.spaceortab = patterns.space + patterns.tab
patterns.eol = S("\n\r")
patterns.spacer = S(" \t\f\v") -- + string.char(0xc2, 0xa0) if we want utf (cf mail roberto)
-patterns.newline = crlf + cr + lf
-patterns.nonspace = 1 - patterns.space
+patterns.newline = newline
+patterns.emptyline = newline^1
patterns.nonspacer = 1 - patterns.spacer
patterns.whitespace = patterns.eol + patterns.spacer
patterns.nonwhitespace = 1 - patterns.whitespace
+patterns.equal = P("=")
patterns.comma = P(",")
patterns.commaspacer = P(",") * patterns.spacer^0
patterns.period = P(".")
+patterns.colon = P(":")
+patterns.semicolon = P(";")
+patterns.underscore = P("_")
patterns.escaped = escaped
patterns.squote = squote
patterns.dquote = dquote
-patterns.undouble = (dquote/"") * ((escaped + (1-dquote))^0) * (dquote/"")
-patterns.unsingle = (squote/"") * ((escaped + (1-squote))^0) * (squote/"")
+patterns.nosquote = (escaped + (1-squote))^0
+patterns.nodquote = (escaped + (1-dquote))^0
+patterns.unsingle = (squote/"") * patterns.nosquote * (squote/"")
+patterns.undouble = (dquote/"") * patterns.nodquote * (dquote/"")
patterns.unquoted = patterns.undouble + patterns.unsingle -- more often undouble
patterns.unspacer = ((patterns.spacer^1)/"")^0
@@ -267,19 +295,6 @@ function lpeg.splitter(pattern, action)
return (((1-P(pattern))^1)/action+1)^0
end
-local spacing = patterns.spacer^0 * patterns.newline -- sort of strip
-local empty = spacing * Cc("")
-local nonempty = Cs((1-spacing)^1) * spacing^-1
-local content = (empty + nonempty)^1
-
-local capture = Ct(content^0)
-
-function string.splitlines(str)
- return match(capture,str)
-end
-
-patterns.textline = content
-
local splitters_s, splitters_m = { }, { }
local function splitat(separator,single)
@@ -322,6 +337,30 @@ function string.split(str,separator)
return match(c,str)
end
+local spacing = patterns.spacer^0 * newline -- sort of strip
+local empty = spacing * Cc("")
+local nonempty = Cs((1-spacing)^1) * spacing^-1
+local content = (empty + nonempty)^1
+
+patterns.textline = content
+
+
+local linesplitter = Ct(splitat(newline))
+
+patterns.linesplitter = linesplitter
+
+function string.splitlines(str)
+ return match(linesplitter,str)
+end
+
+local utflinesplitter = utfbom^-1 * Ct(splitat(newline))
+
+patterns.utflinesplitter = utflinesplitter
+
+function string.utfsplitlines(str)
+ return match(utflinesplitter,str)
+end
+
local cache = { }
@@ -3247,6 +3286,8 @@ utf = utf or unicode.utf8
local concat, utfchar, utfgsub = table.concat, utf.char, utf.gsub
local char, byte, find, bytepairs, utfvalues, format = string.char, string.byte, string.find, string.bytepairs, string.utfvalues, string.format
+local utfsplitlines = string.utfsplitlines
+
-- 0 EF BB BF UTF-8
-- 1 FF FE UTF-16-little-endian
-- 2 FE FF UTF-16-big-endian
@@ -3291,111 +3332,131 @@ function unicode.utftype(f)
end
end
-function unicode.utf16_to_utf8(str, endian) -- maybe a gsub is faster or an lpeg
- local result, tmp, n, m, p, r, t = { }, { }, 0, 0, 0, 0, 0 -- we reuse tmp
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+
+
+local function utf16_to_utf8_be(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*left + right
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
+ r = r + 1
+ result[r] = utfchar(now)
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- for l,r in bytepairs(str) do
- if r then
- if endian then
- n = 256*l + r
- else
- n = 256*r + l
- end
- if m > 0 then
- n = (m-0xD800)*0x400 + (n-0xDC00) + 0x10000
- m = 0
- doit()
- elseif n >= 0xD800 and n <= 0xDBFF then
- m = n
- else
- doit()
+ return t
+end
+
+local function utf16_to_utf8_le(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*right + left
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
+ r = r + 1
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
+ r = r + 1
+ result[r] = utfchar(now)
+ end
end
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- end
- return result
+ return t
end
-function unicode.utf32_to_utf8(str, endian)
- local result, tmp, n, m, p, r, t = { }, { }, 0, -1, 0, 0, 0
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+local function utf32_to_utf8_be(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*256*256*a + 256*256*b
+ else
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[t] = utfchar(more + 256*a + b)
+ more = -1
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ else
+ break
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r)
+ end
+ return result
+end
+
+local function utf32_to_utf8_le(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
end
- for a,b in bytepairs(str) do
- if a and b then
- if m < 0 then
- if endian then
- m = 256*256*256*a + 256*256*b
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*b + a
else
- m = 256*b + a
+ r = r + 1
+ result[t] = utfchar(more + 256*256*256*b + 256*256*a)
+ more = -1
end
else
- if endian then
- n = m + 256*a + b
- else
- n = m + 256*256*256*b + 256*256*a
- end
- m = -1
- doit()
+ break
end
- else
- break
end
- end
- if #tmp > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
+ t[i] = concat(result,"",1,r)
end
return result
end
+unicode.utf32_to_utf8_be = utf32_to_utf8_be
+unicode.utf32_to_utf8_le = utf32_to_utf8_le
+unicode.utf16_to_utf8_be = utf16_to_utf8_be
+unicode.utf16_to_utf8_le = utf16_to_utf8_le
+
+function unicode.utf8_to_utf8(t)
+ return type(t) == "string" and utfsplitlines(t) or t
+end
+
+function unicode.utf16_to_utf8(t,endian)
+ return endian and utf16_to_utf8_be(t) or utf16_to_utf8_le(t) or t
+end
+
+function unicode.utf32_to_utf8(t,endian)
+ return endian and utf32_to_utf8_be(t) or utf32_to_utf8_le(t) or t
+end
+
local function little(c)
local b = byte(c)
if b < 0x10000 then
@@ -3436,6 +3497,10 @@ function unicode.utfcodes(str)
end
+function unicode.filetype(data)
+ return data and lpeg.match(lpeg.patterns.utftype,data) or "unknown"
+end
+
end -- of closure
@@ -3905,11 +3970,11 @@ local pattern_b = spaces * comma^0 * spaces * (key * ((spaces * equal * spaces *
local hash = { }
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
@@ -9676,7 +9741,7 @@ local suffixes = allocate() resolvers.suffixes = suffixes
local dangerous = allocate() resolvers.dangerous = dangerous
local suffixmap = allocate() resolvers.suffixmap = suffixmap
-local relations = allocate {
+local relations = allocate { -- todo: handlers also here
core = {
ofm = {
names = { "ofm", "omega font metric", "omega font metrics" },
@@ -9746,7 +9811,7 @@ local relations = allocate {
tex = {
names = { "tex" },
variable = 'TEXINPUTS',
- suffixes = { 'tex', "mkiv", "mkii" },
+ suffixes = { 'tex', "mkiv", "mkiv", "mkii" },
},
icc = {
names = { "icc", "icc profile", "icc profiles" },
@@ -9860,29 +9925,33 @@ resolvers.relations = relations
-- formats: maps a format onto a variable
-for category, categories in next, relations do
- for name, relation in next, categories do
- local rn = relation.names
- local rv = relation.variable
- local rs = relation.suffixes
- if rn and rv then
- for i=1,#rn do
- local rni = lower(gsub(rn[i]," ",""))
- formats[rni] = rv
- if rs then
- suffixes[rni] = rs
- for i=1,#rs do
- local rsi = rs[i]
- suffixmap[rsi] = rni
+function resolvers.updaterelations()
+ for category, categories in next, relations do
+ for name, relation in next, categories do
+ local rn = relation.names
+ local rv = relation.variable
+ local rs = relation.suffixes
+ if rn and rv then
+ for i=1,#rn do
+ local rni = lower(gsub(rn[i]," ",""))
+ formats[rni] = rv
+ if rs then
+ suffixes[rni] = rs
+ for i=1,#rs do
+ local rsi = rs[i]
+ suffixmap[rsi] = rni
+ end
end
end
end
- end
- if rs then
+ if rs then
+ end
end
end
end
+resolvers.updaterelations() -- push this in the metatable -> newindex
+
local function simplified(t,k)
return rawget(t,lower(gsub(k," ","")))
end
@@ -10345,7 +10414,7 @@ resolvers.locators = allocate { notfound = { nil } } -- locate databases
resolvers.hashers = allocate { notfound = { nil } } -- load databases
resolvers.generators = allocate { notfound = { nil } } -- generate databases
-function resolvers.splitmethod(filename)
+function resolvers.splitmethod(filename) -- todo: trigger by suffix
if not filename then
return { } -- safeguard
elseif type(filename) == "table" then
@@ -10364,10 +10433,13 @@ function resolvers.methodhandler(what, filename, filetype) -- ...
local resolver = resolvers[what]
if resolver[scheme] then
if trace_locating then
- report_resolvers("handler '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ report_resolvers("using special handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
end
return resolver[scheme](filename,filetype)
else
+ if trace_locating then
+ report_resolvers("no handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ end
return resolver.tex(filename,filetype) -- todo: specification
end
end
diff --git a/scripts/context/stubs/unix/mtxrun b/scripts/context/stubs/unix/mtxrun
index 7b5348424..c41de7fb8 100644
--- a/scripts/context/stubs/unix/mtxrun
+++ b/scripts/context/stubs/unix/mtxrun
@@ -177,6 +177,9 @@ local lpeg = require("lpeg")
local type = type
+-- Beware, we predefine a bunch of patterns here and one reason for doing so
+-- is that we get consistent behaviour in some of the visualizers.
+
lpeg.patterns = lpeg.patterns or { } -- so that we can share
local patterns = lpeg.patterns
@@ -193,19 +196,38 @@ local alwaysmatched = P(true)
patterns.anything = anything
patterns.endofstring = endofstring
patterns.beginofstring = alwaysmatched
+patterns.alwaysmatched = alwaysmatched
local digit, sign = R('09'), S('+-')
local cr, lf, crlf = P("\r"), P("\n"), P("\r\n")
+local newline = crlf + cr + lf
local utf8next = R("\128\191")
local escaped = P("\\") * anything
local squote = P("'")
local dquote = P('"')
+local space = P(" ")
+
+patterns.somecontent = (anything - newline - space)^1
+patterns.beginline = #(1-newline)
+
+local utfbom_32_be = P('\000\000\254\255')
+local utfbom_32_le = P('\255\254\000\000')
+local utfbom_16_be = P('\255\254')
+local utfbom_16_le = P('\254\255')
+local utfbom_8 = P('\239\187\191')
+local utfbom = utfbom_32_be + utfbom_32_le
+ + utfbom_16_be + utfbom_16_le
+ + utfbom_8
+local utftype = utfbom_32_be / "utf-32-be" + utfbom_32_le / "utf-32-le"
+ + utfbom_16_be / "utf-16-be" + utfbom_16_le / "utf-16-le"
+ + utfbom_8 / "utf-8" + alwaysmatched / "unknown"
patterns.utf8one = R("\000\127")
patterns.utf8two = R("\194\223") * utf8next
patterns.utf8three = R("\224\239") * utf8next * utf8next
patterns.utf8four = R("\240\244") * utf8next * utf8next * utf8next
-patterns.utfbom = P('\000\000\254\255') + P('\255\254\000\000') + P('\255\254') + P('\254\255') + P('\239\187\191')
+patterns.utfbom = utfbom
+patterns.utftype = utftype
local utf8char = patterns.utf8one + patterns.utf8two + patterns.utf8three + patterns.utf8four
local validutf8char = utf8char^0 * endofstring * Cc(true) + Cc(false)
@@ -231,24 +253,30 @@ patterns.hexadecimal = P("0x") * R("09","AF","af")^1
patterns.lowercase = R("az")
patterns.uppercase = R("AZ")
patterns.letter = patterns.lowercase + patterns.uppercase
-patterns.space = P(" ")
+patterns.space = space
patterns.tab = P("\t")
patterns.spaceortab = patterns.space + patterns.tab
patterns.eol = S("\n\r")
patterns.spacer = S(" \t\f\v") -- + string.char(0xc2, 0xa0) if we want utf (cf mail roberto)
-patterns.newline = crlf + cr + lf
-patterns.nonspace = 1 - patterns.space
+patterns.newline = newline
+patterns.emptyline = newline^1
patterns.nonspacer = 1 - patterns.spacer
patterns.whitespace = patterns.eol + patterns.spacer
patterns.nonwhitespace = 1 - patterns.whitespace
+patterns.equal = P("=")
patterns.comma = P(",")
patterns.commaspacer = P(",") * patterns.spacer^0
patterns.period = P(".")
+patterns.colon = P(":")
+patterns.semicolon = P(";")
+patterns.underscore = P("_")
patterns.escaped = escaped
patterns.squote = squote
patterns.dquote = dquote
-patterns.undouble = (dquote/"") * ((escaped + (1-dquote))^0) * (dquote/"")
-patterns.unsingle = (squote/"") * ((escaped + (1-squote))^0) * (squote/"")
+patterns.nosquote = (escaped + (1-squote))^0
+patterns.nodquote = (escaped + (1-dquote))^0
+patterns.unsingle = (squote/"") * patterns.nosquote * (squote/"")
+patterns.undouble = (dquote/"") * patterns.nodquote * (dquote/"")
patterns.unquoted = patterns.undouble + patterns.unsingle -- more often undouble
patterns.unspacer = ((patterns.spacer^1)/"")^0
@@ -267,19 +295,6 @@ function lpeg.splitter(pattern, action)
return (((1-P(pattern))^1)/action+1)^0
end
-local spacing = patterns.spacer^0 * patterns.newline -- sort of strip
-local empty = spacing * Cc("")
-local nonempty = Cs((1-spacing)^1) * spacing^-1
-local content = (empty + nonempty)^1
-
-local capture = Ct(content^0)
-
-function string.splitlines(str)
- return match(capture,str)
-end
-
-patterns.textline = content
-
local splitters_s, splitters_m = { }, { }
local function splitat(separator,single)
@@ -322,6 +337,30 @@ function string.split(str,separator)
return match(c,str)
end
+local spacing = patterns.spacer^0 * newline -- sort of strip
+local empty = spacing * Cc("")
+local nonempty = Cs((1-spacing)^1) * spacing^-1
+local content = (empty + nonempty)^1
+
+patterns.textline = content
+
+
+local linesplitter = Ct(splitat(newline))
+
+patterns.linesplitter = linesplitter
+
+function string.splitlines(str)
+ return match(linesplitter,str)
+end
+
+local utflinesplitter = utfbom^-1 * Ct(splitat(newline))
+
+patterns.utflinesplitter = utflinesplitter
+
+function string.utfsplitlines(str)
+ return match(utflinesplitter,str)
+end
+
local cache = { }
@@ -3247,6 +3286,8 @@ utf = utf or unicode.utf8
local concat, utfchar, utfgsub = table.concat, utf.char, utf.gsub
local char, byte, find, bytepairs, utfvalues, format = string.char, string.byte, string.find, string.bytepairs, string.utfvalues, string.format
+local utfsplitlines = string.utfsplitlines
+
-- 0 EF BB BF UTF-8
-- 1 FF FE UTF-16-little-endian
-- 2 FE FF UTF-16-big-endian
@@ -3291,111 +3332,131 @@ function unicode.utftype(f)
end
end
-function unicode.utf16_to_utf8(str, endian) -- maybe a gsub is faster or an lpeg
- local result, tmp, n, m, p, r, t = { }, { }, 0, 0, 0, 0, 0 -- we reuse tmp
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+
+
+local function utf16_to_utf8_be(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*left + right
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
+ r = r + 1
+ result[r] = utfchar(now)
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- for l,r in bytepairs(str) do
- if r then
- if endian then
- n = 256*l + r
- else
- n = 256*r + l
- end
- if m > 0 then
- n = (m-0xD800)*0x400 + (n-0xDC00) + 0x10000
- m = 0
- doit()
- elseif n >= 0xD800 and n <= 0xDBFF then
- m = n
- else
- doit()
+ return t
+end
+
+local function utf16_to_utf8_le(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*right + left
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
+ r = r + 1
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
+ r = r + 1
+ result[r] = utfchar(now)
+ end
end
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- end
- return result
+ return t
end
-function unicode.utf32_to_utf8(str, endian)
- local result, tmp, n, m, p, r, t = { }, { }, 0, -1, 0, 0, 0
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+local function utf32_to_utf8_be(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*256*256*a + 256*256*b
+ else
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[t] = utfchar(more + 256*a + b)
+ more = -1
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ else
+ break
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r)
+ end
+ return result
+end
+
+local function utf32_to_utf8_le(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
end
- for a,b in bytepairs(str) do
- if a and b then
- if m < 0 then
- if endian then
- m = 256*256*256*a + 256*256*b
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*b + a
else
- m = 256*b + a
+ r = r + 1
+ result[t] = utfchar(more + 256*256*256*b + 256*256*a)
+ more = -1
end
else
- if endian then
- n = m + 256*a + b
- else
- n = m + 256*256*256*b + 256*256*a
- end
- m = -1
- doit()
+ break
end
- else
- break
end
- end
- if #tmp > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
+ t[i] = concat(result,"",1,r)
end
return result
end
+unicode.utf32_to_utf8_be = utf32_to_utf8_be
+unicode.utf32_to_utf8_le = utf32_to_utf8_le
+unicode.utf16_to_utf8_be = utf16_to_utf8_be
+unicode.utf16_to_utf8_le = utf16_to_utf8_le
+
+function unicode.utf8_to_utf8(t)
+ return type(t) == "string" and utfsplitlines(t) or t
+end
+
+function unicode.utf16_to_utf8(t,endian)
+ return endian and utf16_to_utf8_be(t) or utf16_to_utf8_le(t) or t
+end
+
+function unicode.utf32_to_utf8(t,endian)
+ return endian and utf32_to_utf8_be(t) or utf32_to_utf8_le(t) or t
+end
+
local function little(c)
local b = byte(c)
if b < 0x10000 then
@@ -3436,6 +3497,10 @@ function unicode.utfcodes(str)
end
+function unicode.filetype(data)
+ return data and lpeg.match(lpeg.patterns.utftype,data) or "unknown"
+end
+
end -- of closure
@@ -3905,11 +3970,11 @@ local pattern_b = spaces * comma^0 * spaces * (key * ((spaces * equal * spaces *
local hash = { }
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
@@ -9676,7 +9741,7 @@ local suffixes = allocate() resolvers.suffixes = suffixes
local dangerous = allocate() resolvers.dangerous = dangerous
local suffixmap = allocate() resolvers.suffixmap = suffixmap
-local relations = allocate {
+local relations = allocate { -- todo: handlers also here
core = {
ofm = {
names = { "ofm", "omega font metric", "omega font metrics" },
@@ -9746,7 +9811,7 @@ local relations = allocate {
tex = {
names = { "tex" },
variable = 'TEXINPUTS',
- suffixes = { 'tex', "mkiv", "mkii" },
+ suffixes = { 'tex', "mkiv", "mkiv", "mkii" },
},
icc = {
names = { "icc", "icc profile", "icc profiles" },
@@ -9860,29 +9925,33 @@ resolvers.relations = relations
-- formats: maps a format onto a variable
-for category, categories in next, relations do
- for name, relation in next, categories do
- local rn = relation.names
- local rv = relation.variable
- local rs = relation.suffixes
- if rn and rv then
- for i=1,#rn do
- local rni = lower(gsub(rn[i]," ",""))
- formats[rni] = rv
- if rs then
- suffixes[rni] = rs
- for i=1,#rs do
- local rsi = rs[i]
- suffixmap[rsi] = rni
+function resolvers.updaterelations()
+ for category, categories in next, relations do
+ for name, relation in next, categories do
+ local rn = relation.names
+ local rv = relation.variable
+ local rs = relation.suffixes
+ if rn and rv then
+ for i=1,#rn do
+ local rni = lower(gsub(rn[i]," ",""))
+ formats[rni] = rv
+ if rs then
+ suffixes[rni] = rs
+ for i=1,#rs do
+ local rsi = rs[i]
+ suffixmap[rsi] = rni
+ end
end
end
end
- end
- if rs then
+ if rs then
+ end
end
end
end
+resolvers.updaterelations() -- push this in the metatable -> newindex
+
local function simplified(t,k)
return rawget(t,lower(gsub(k," ","")))
end
@@ -10345,7 +10414,7 @@ resolvers.locators = allocate { notfound = { nil } } -- locate databases
resolvers.hashers = allocate { notfound = { nil } } -- load databases
resolvers.generators = allocate { notfound = { nil } } -- generate databases
-function resolvers.splitmethod(filename)
+function resolvers.splitmethod(filename) -- todo: trigger by suffix
if not filename then
return { } -- safeguard
elseif type(filename) == "table" then
@@ -10364,10 +10433,13 @@ function resolvers.methodhandler(what, filename, filetype) -- ...
local resolver = resolvers[what]
if resolver[scheme] then
if trace_locating then
- report_resolvers("handler '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ report_resolvers("using special handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
end
return resolver[scheme](filename,filetype)
else
+ if trace_locating then
+ report_resolvers("no handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ end
return resolver.tex(filename,filetype) -- todo: specification
end
end
diff --git a/tex/context/base/buff-ini.lua b/tex/context/base/buff-ini.lua
index 19e3a142b..593166038 100644
--- a/tex/context/base/buff-ini.lua
+++ b/tex/context/base/buff-ini.lua
@@ -6,15 +6,6 @@ if not modules then modules = { } end modules ['buff-ini'] = {
license = "see context related readme files"
}
--- todo: deal with jobname here, or actually, "" is valid as well
-
--- ctx lua reference model / hooks and such
--- to be optimized
-
--- redefine buffers.get
-
--- if needed we can make 'm local
-
local trace_run = false trackers.register("buffers.run", function(v) trace_run = v end)
local trace_visualize = false trackers.register("buffers.visualize", function(v) trace_visualize = v end)
@@ -22,6 +13,8 @@ local report_buffers = logs.new("buffers")
local utf = unicode.utf8
+-- todo: weed the next list
+
local concat, texprint, texwrite = table.concat, tex.print, tex.write
local utfbyte, utffind, utfgsub = utf.byte, utf.find, utf.gsub
local type, next = type, next
@@ -37,22 +30,15 @@ local allocate = utilities.storage.allocate
local tabtospace = utilities.strings.tabtospace
buffers = {
- data = allocate(),
- hooks = { },
- flags = { },
- commands = { },
- visualizers = { },
+ data = allocate(),
+ flags = { },
}
local buffers = buffers
local context = context
-local data = buffers.data
-local flags = buffers.flags
-local hooks = buffers.hooks
-local visualizers = buffers.visualizers
-
-visualizers.defaultname = variables.typing
+local data = buffers.data
+local flags = buffers.flags
function buffers.raw(name)
return data[name] or { }
@@ -109,40 +95,6 @@ function buffers.doifelsebuffer(name)
commands.testcase(data[name] ~= nil)
end
--- handy helpers
---
--- \sop[color] switch_of_pretty
--- \bop[color] begin_of_pretty
--- \eop end_of_pretty
--- \obs obeyedspace
--- \char <n> special characters
-
-local sop = context.sop
-local bop = context.bop
-local eop = context.eop
-local obs = context.obs
-local par = context.par
-local chr = context.char
-
-local bgroup = context.bgroup
-local egroup = context.egroup
-
-flags.optimizeverbatim = true
-flags.countemptylines = false
-
-local doverbatimnobreak = context.doverbatimnobreak
-local doverbatimgoodbreak = context.doverbatimgoodbreak
-
-function buffers.verbatimbreak(n,m)
- if flags.optimizeverbatim then
- if n == 2 or n == m then
- doverbatimnobreak()
- elseif n > 1 then
- doverbatimgoodbreak()
- end
- end
-end
-
function buffers.strip(lines,first,last)
local first, last = first or 1, last or #lines
for i=first,last do
@@ -209,41 +161,83 @@ function buffers.range(lines,first,last,range) -- 1,3 1,+3 fromhere,tothere
return first, last
end
-function buffers.type(name,realign,range)
- local lines = data[name]
- local action = buffers.typeline
+-- this will go to buff-ver.lua
+
+-- there is some overlap in the following
+
+flags.tablength = 7
+
+local function flush(content,method,settings)
+ local tab = settings.tab
+ tab = tab and (tab == variables.yes and flags.tablength or tonumber(tab))
+ if tab then
+ content = utilities.strings.tabtospace(content,tab)
+ end
+ local visualizer = settings.visualizer
+ if visualizer and visualizer ~= "" then
+ visualizers.visualize(visualizer,method,content,settings)
+ else -- todo:
+ visualizers.visualize("",method,content,settings)
+ end
+end
+
+local function filter(lines,settings) -- todo: inline or display in settings
+ local strip = settings.strip
+ if strip then
+ lines = buffers.realign(lines,strip)
+ end
+ local line, n = 0, 0
+ local first, last, m = buffers.strip(lines)
+ if range then
+ first, last = buffers.range(lines,first,last,range)
+ first, last = buffers.strip(lines,first,last)
+ end
+ local content = concat(lines,(settings.nature == "inline" and " ") or "\n",first,last)
+ return content, m
+end
+
+function buffers.typestring(settings) -- todo: settings.nature = "inline"
+ local content = settings.data
+ if content and content ~= "" then
+ flush(content,"inline",settings)
+ end
+end
+
+function buffers.typebuffer(settings) -- todo: settings.nature = "display"
+ local name = settings.name
+ local lines = name and data[name]
if lines then
if type(lines) == "string" then
lines = splitlines(lines)
data[name] = lines
end
- if realign then
- lines = buffers.realign(lines,realign)
- end
- local line, n = 0, 0
- local first, last, m = buffers.strip(lines)
- if range then
- first, last = buffers.range(lines,first,last,range)
- first, last = buffers.strip(lines,first,last)
+ local content, m = filter(lines,settings)
+ if content and content ~= "" then
+ flush(content,"display",settings)
end
- hooks.begin_of_display()
- for i=first,last do
- n, line = action(lines[i], n, m, line)
- end
- hooks.end_of_display()
end
end
-function buffers.loaddata(filename) -- this one might go away
- -- this will be cleaned up when we have split supp-fil completely
- -- instead of half-half
- local ok, str, n = resolvers.loaders.tex(filename)
- if not str then
- ok, str, n = resolvers.loaders.tex(file.addsuffix(filename,'tex'))
+function buffers.typefile(settings) -- todo: settings.nature = "display"
+ local name = settings.name
+ local str = buffers.loaddata(name)
+ if str and str ~= "" then
+ local regime = settings.regime
+ if regime and regime ~= "" then
+ regimes.load(regime)
+ str = regimes.translate(str,regime)
+ end
+ if str and str~= "" then
+ local lines = splitlines(str)
+ local content, m = filter(lines,settings)
+ if content and content ~= "" then
+ flush(content,"display",settings)
+ end
+ end
end
end
-function buffers.loaddata(filename) -- this one might go away
+function buffers.loaddata(filename) -- this one might go away or become local
local foundname = resolvers.findtexfile(filename) or ""
if foundname == "" then
foundname = resolvers.findtexfile(file.addsuffix(filename,'tex')) or ""
@@ -255,48 +249,6 @@ function buffers.loaddata(filename) -- this one might go away
end
end
-function buffers.typefile(name,realign,range,regime) -- still somewhat messy, since name can be be suffixless
- local str = buffers.loaddata(name)
- if regime and regime ~= "" then
- regimes.load(regime)
- str = regimes.translate(str,regime)
- end
- if str and str~= "" then
- local lines = splitlines(str)
- if realign then
- lines = buffers.realign(lines,realign)
- end
- local line, n, action = 0, 0, buffers.typeline
- local first, last, m = buffers.strip(lines)
- hooks.begin_of_display()
- if range then
- first, last = buffers.range(lines,first,last,range)
- first, last = buffers.strip(lines,first,last)
- end
- for i=first,last do
- n, line = action(lines[i], n, m, line)
- end
- hooks.end_of_display()
- end
-end
-
-function buffers.typeline(str,n,m,line)
- n = n + 1
- buffers.verbatimbreak(n,m)
- if find(str,"%S") then
- line = line + 1
- hooks.begin_of_line(line)
- hooks.flush_line(hooks.line(str))
- hooks.end_of_line()
- else
- if flags.countemptylines then
- line = line + 1
- end
- hooks.empty_line(line)
- end
- return n, line
-end
-
-- The optional prefix hack is there for the typesetbuffer feature and
-- in mkii we needed that (this hidden feature is used in a manual).
@@ -397,7 +349,7 @@ function buffers.collect(names,separator) -- no print
return concat(t,separator or "\r") -- "\n" is safer due to comments and such
end
-function buffers.feedback(names,separator)
+function buffers.feedback(names,separator) -- we can use cld
-- don't change the texprint into texsprint as it fails on mp buffers
-- because (p<nl>enddef) becomes penddef then
texprint(ctxcatcodes,splitlines(buffers.collect(names,separator)))
@@ -422,271 +374,6 @@ function buffers.inspect(name)
end
end
--- maybe just line(n,str) empty(n,str)
-
-visualizers.tablength = 7
-visualizers.enabletab = true -- false
-visualizers.obeyspace = true
-
-function buffers.settablength(tablength)
- visualizers.tablength = tablength and tonumber(tablength) or 7
-end
-
-visualizers.handlers = visualizers.handlers or { }
-
-local handlers = visualizers.handlers
-
-function buffers.newvisualizer(name)
- name = lower(name)
- local handler = { }
- handlers[name] = handler
- return handler
-end
-
-function buffers.getvisualizer(name)
- name = lower(name)
- return handlers[name] or buffers.loadvisualizer(name)
-end
-
-function buffers.loadvisualizer(name)
- name = lower(name)
- local hn = handlers[name]
- if hn then
- return hn
- else
- environment.loadluafile("pret-" .. name)
- local hn = handlers[name]
- if not hn then
- -- hn = buffers.newvisualizer(name)
- hn = handlers[visualizers.defaultname]
- handlers[name] = hn
- if trace_visualize then
- report_buffers("mapping '%s' visualizer onto '%s'",name,visualizers.defaultname)
- end
- elseif trace_visualize then
- report_buffers("loading '%s' visualizer",name)
- end
- return hn
- end
-end
-
--- was "default", should be set at tex end (todo)
-
-local default = buffers.newvisualizer(visualizers.defaultname)
-
-default.begin_of_display = context.doverbatimbeginofdisplay
-default.end_of_display = context.doverbatimendofdisplay
-default.begin_of_inline = context.doverbatimbeginofinline
-default.end_of_inline = context.doverbatimendofinline
-default.begin_of_line = context.doverbatimbeginofline
-default.end_of_line = context.doverbatimendofline
-default.empty_line = context.doverbatimemptyline
-
---~ print(variables.typing) os.exit()
-
--- will become cleaner
-
-local currentvisualizer, currenthandler
-
-function buffers.setvisualizer(str)
- currentvisualizer = lower(str)
- currenthandler = handlers[currentvisualizer]
- if currenthandler then
- -- if trace_visualize then
- -- report_buffers("enabling specific '%s' visualizer",currentvisualizer)
- -- end
- else
- currentvisualizer = visualizers.defaultname
- currenthandler = handlers.default
- -- if trace_visualize then
- -- report_buffers("enabling default visualizer '%s'",currentvisualizer)
- -- end
- end
- if currenthandler.reset then
- currenthandler.reset()
- end
-end
-
-function buffers.resetvisualizer()
- currentvisualizer = visualizers.defaultname
- currenthandler = handlers.default
- if currenthandler.reset then
- currenthandler.reset()
- end
-end
-
-buffers.setvisualizer(visualizers.defaultname)
-
-function visualizers.reset()
-end
-
-function buffers.doifelsevisualizer(str)
- commands.testcase((str ~= "") and (handlers[lower(str)] ~= nil))
-end
-
--- calling routines, don't change
-
-function hooks.begin_of_display()
- (currenthandler.begin_of_display or default.begin_of_display)(currentvisualizer)
-end
-
-function hooks.end_of_display()
- (currenthandler.end_of_display or default.end_of_display)()
-end
-
-function hooks.begin_of_inline()
- (currenthandler.begin_of_inline or default.begin_of_inline)(currentvisualizer)
-end
-
-function hooks.end_of_inline()
- (currenthandler.end_of_inline or default.end_of_inline)()
-end
-
-function hooks.flush_line(str,nesting)
- local fl = currenthandler.flush_line
- if fl then
- str = gsub(str," *[\n\r]+ *"," ") ; -- semi colon needed
- fl(str,nesting)
- else
- -- gsub done later
- default.flush_line(str,nesting)
- end
-end
-
-function hooks.flush_inline(str,nesting)
- hooks.begin_of_inline()
- hooks.flush_line(str,nesting)
- hooks.end_of_inline()
-end
-
-function hooks.begin_of_line(n)
- (currenthandler.begin_of_line or default.begin_of_line)(n)
-end
-
-function hooks.end_of_line()
- (currenthandler.end_of_line or default.end_of_line)()
-end
-
-function hooks.empty_line()
- (currenthandler.empty_line or default.empty_line)()
-end
-
-function hooks.line(str)
- if visualizers.enabletab then
- str = tabtospace(str,visualizers.tablength)
- else
- str = gsub(str,"\t"," ")
- end
- return (currenthandler.line or default.line)(str)
-end
-
-buffers.currentcolors = { } -- todo: registercurrentcolor and preset sop then ... faster or at least precreate tables
-
-function buffers.changestate(n, state)
- if n then
- if state ~= n then
- if state > 0 then
- sop { buffers.currentcolors[n] }
- else
- bop { buffers.currentcolors[n] }
- end
- return n
- end
- elseif state > 0 then
- eop()
- return 0
- end
- return state
-end
-
-function default.line(str)
- return str
-end
-
-function default.flush_line(str)
- str = gsub(str," *[\n\r]+ *"," ")
- if visualizers.obeyspace then
- for c in utfcharacters(str) do
- if c == " " then
- obs()
- else
- texwrite(c)
- end
- end
- else
- texwrite(str)
- end
-end
-
--- special one
-
-buffers.commands.nested = "\\switchslantedtype "
-
-function visualizers.flushnested(str, enable) -- todo: no utf, vrb catcodes, kind of obsolete mess
- str = gsub(str," *[\n\r]+ *"," ")
- local c, nested, i = "", 0, 1
- local commands = buffers.commands -- otherwise wrong commands
- while i < #str do -- slow
- c = sub(str,i,i+1)
- if c == "<<" then
- nested = nested + 1
- bgroup()
- if enable then
- context(commands.nested)
- end
- i = i + 2
- elseif c == ">>" then
- if nested > 0 then
- nested = nested - 1
- egroup()
- end
- i = i + 2
- else
- c = sub(str,i,i)
- if c == " " then
- obs()
- elseif find(c,"%a") then
- context(c)
- else
- chr(byte(c))
- end
- i = i + 1
- end
- end
- chr(byte(sub(str,i,i)))
- for i=1,#nested do
- egroup()
- end
-end
-
-function buffers.finishstate(state)
- if state > 0 then
- eop()
- return 0
- else
- return state
- end
-end
-
-local opennested = rep("\\char"..byte('<').." ",2)
-local closenested = rep("\\char"..byte('>').." ",2)
-
-function buffers.replacenested(result)
- result = gsub(result,opennested, "{")
- result = gsub(result,closenested,"}")
- return result
-end
-
-function buffers.flushresult(result,nested)
- if nested then
- context(buffers.replacenested(concat(result)))
- else
- context(concat(result))
- end
-end
-
--- new
-
function buffers.realign(name,forced_n) -- no, auto, <number>
local n, d
if type(name) == "string" then
@@ -725,111 +412,3 @@ function buffers.realign(name,forced_n) -- no, auto, <number>
end
-- escapes: buffers.setescapepair("tex","/BTEX","/ETEX")
-
-local function flush_escaped_line(str,pattern,flushline)
- while true do
- local a, b, c = match(str,pattern)
- if a and a ~= "" then
- flushline(a)
- end
- if b and b ~= "" then
- context(b)
- end
- if c then
- if c == "" then
- break
- else
- str = c
- end
- else
- flushline(str)
- break
- end
- end
-end
-
-function buffers.setescapepair(name,pair)
- if pair and pair ~= "" then
- local visualizer = buffers.getvisualizer(name)
- visualizer.normal_flush_line = visualizer.normal_flush_line or visualizer.flush_line
- if pair == variables.no then
- visualizer.flush_line = visualizer.normal_flush_line or visualizer.flush_line
- if trace_visualize then
- report_buffers("resetting escape range for visualizer '%s'",name)
- end
- else
- local start, stop
- if pair == variables.yes then
- start, stop = "/BTEX", "/ETEX"
- else
- pair = string.split(pair,",")
- start, stop = escapedpattern(pair[1] or ""), escapedpattern(pair[2] or "")
- end
- if start ~= "" then
- local pattern
- if stop == "" then
- pattern = "^(.-)" .. start .. "(.*)(.*)$"
- else
- pattern = "^(.-)" .. start .. "(.-)" .. stop .. "(.*)$"
- end
- function visualizer.flush_line(str)
- flush_escaped_line(str,pattern,visualizer.normal_flush_line)
- end
- if trace_visualize then
- report_buffers("setting escape range for visualizer '%s' to %s -> %s",name,start,stop)
- end
- elseif trace_visualize then
- report_buffers("problematic escape specification '%s' for visualizer '%s'",pair,name)
- end
- end
- end
-end
-
--- THIS WILL BECOME A FRAMEWORK: the problem with pretty printing is that
--- we deal with snippets and therefore we need tolerant parsing
-
---~ local type = type
-
---~ visualizers = visualizers or { }
-
---~ local function fallback(s) return s end
-
---~ function visualizers.visualize(visualizer,kind,pattern)
---~ if type(visualizer) == "table" and type(kind) == "string" then
---~ kind = visualizer[kind] or visualizer.default or fallback
---~ else
---~ kind = fallback
---~ end
---~ return (lpeg.C(pattern))/kind
---~ end
-
---~ local flusher = texio.write
---~ local format = string.format
-
---~ local visualizer = {
---~ word = function(s) return flusher(format("\\bold{%s}",s)) end,
---~ number = function(s) return flusher(format("\\slanted{%s}",s)) end,
---~ default = function(s) return flusher(s) end,
---~ }
-
---~ local word = lpeg.R("AZ","az")^1
---~ local number = lpeg.R("09")^1
---~ local any = lpeg.P(1)
-
---~ local pattern = lpeg.P { "start",
---~ start = (
---~ visualizers.visualize(visualizer,"word",word) +
---~ visualizers.visualize(visualizer,"number",number) +
---~ visualizers.visualize(visualizer,"default",any)
---~ )^1
---~ }
-
---~ str = [[test 123 test $oeps$]]
-
---~ lpegmatch(pattern,str)
-
-buffers.obsolete = buffers.obsolete or { }
-local obsolete = buffers.obsolete
-
-buffers.finish_state = buffers.finishstate obsolete.finish_state = buffers.finishstate
-buffers.change_state = buffers.finishstate obsolete.change_state = buffers.finishstate
diff --git a/tex/context/base/buff-ini.mkiv b/tex/context/base/buff-ini.mkiv
index 19fabcff6..f195cd8cf 100644
--- a/tex/context/base/buff-ini.mkiv
+++ b/tex/context/base/buff-ini.mkiv
@@ -15,75 +15,32 @@
\registerctxluafile{buff-ini}{1.001}
-% todo: move all to lua, also before and after, just context.beforebuffer()
-% todo: commalist to lua end
-% todo: jobname == "" so no need for testing
-
-% todo:
-%
-% \startluacode
-% local locations = { }
-% function document.set_number(name)
-% locations[name] = {
-% line = status.linenumber,
-% file = status.filename
-% }
-% end
-% function document.add_number(name)
-% local b, l = buffers.raw(name), locations[name]
-% if b and l then
-% for i=1,#b do
-% b[i] = string.gsub(b[i],"# line: <number>","# line: " .. l.line + 2)
-% end
-% end
-% end
-% \stopluacode
-%
-% \starttext
-%
-% \ctxlua{document.set_number("oeps")}
-% \startbuffer[oeps]
-% # line: <number>
-%
-% test
-% test
-% \stopbuffer
-% \ctxlua{document.add_number("oeps")}
-%
-% \typebuffer[oeps]
-%
-% \stoptext
-
-\ifdefined\doinitializeverbatim \else% temp hack
- \def\doinitializeverbatim{\tttf}
-\fi
-
\unprotect
-\setnewconstant\buffernestmode\plusone % 0: not nested, 1: startbuffer nested, 2: all buffers nested
+% number is messy and not needed as we store the number anyway
+% we can get rid of \c!number
-\newconditional\segmentatebuffermode % not implemented in mkiv (yet)
+\newcount\nofdefinedbuffers
-\def\currentbuffer{\jobname}
+\let\currentbuffer\empty
-\def\setcurrentbuffer#1%
- {\doifelsenothing{#1}{\edef\currentbuffer{\jobname}}{\edef\currentbuffer{#1}}}
+\def\doifelsebuffer#1%
+ {\ctxlua{buffers.doifelsebuffer("#1")}}
\def\resetbuffer
{\dosingleempty\doresetbuffer}
\def\doresetbuffer[#1]%
- {\begingroup
- \setcurrentbuffer{#1}%
- \ctxlua{buffers.erase("\currentbuffer")}%
- \endgroup}
+ {\ctxlua{buffers.erase("#1")}}
-\def\dostartbuffer
+\unexpanded\def\dostartdefinedbuffer
{\bgroup
- \obeylines % nodig, anders gaat 't fout als direct \starttable (bv)
+ \obeylines
\doquadrupleempty\dodostartbuffer}
-\def\dodostartbuffer[#1][#2][#3][#4]% upward compatible
+\let\dostartbuffer\dostartdefinedbuffer % used in some modules
+
+\def\dodostartbuffer[#1][#2][#3][#4]% [category] [name] [start] [stop]
{\iffourthargument
\def\next{\dododostartbuffer{#1}{#2}{#3}{#4}}%
\else
@@ -91,65 +48,43 @@
\fi
\next}
-\def\dododostartbuffer#1#2#3#4%
- {%\showmessage\m!systems{15}{#2}%
- \doifelsevalue{\??bu#1\c!paragraph}\v!yes
- {\settrue\segmentatebuffermode} % todo in mkiv
- {\doifnumberelse{\getvalue{\??bu#1\c!paragraph}}
- {\settrue \segmentatebuffermode}
- {\setfalse\segmentatebuffermode}}%
- \doifvalue{\??bu#1\c!local}\v!yes
- {\buffernestmode\plustwo}% permit nesting
- \setcurrentbuffer{#2}%
- \doifelsenothing{#4}
- {\normalexpanded{\noexpand\setbuffercapsules{\e!start\v!buffer}{\e!stop\v!buffer}}%
- \letvalue\bufferstop\relax}
- %{\@EA\setbuffercapsules\@EA{\csname#3\@EA\endcsname\@EA}\@EA{\csname#4\endcsname}}% if we strip later
- {\setbuffercapsules{#3}{#4}}%
- \normalexpanded{\noexpand\dodowithbuffer
- {\currentbuffer}
- {\bufferstart}
- {\bufferstop}
- {\donothing}
- {\egroup
- \noexpand\getvalue{\bufferstop}}}}
-
-\letvalue{\e!start\v!buffer}\dostartbuffer
-
-\let\endbuffer\undefined % to please the dep parser
+\def\dododostartbuffer#1#2#3#4% \donothing needed !
+ {\normalexpanded{\dodowithbuffer{#2}{#3}{#4}{\donothing}{\egroup\noexpand\getvalue{#4}}}}
+
+\setvalue{\e!start\v!buffer}%
+ {\bgroup
+ \obeylines
+ \dosingleempty\redostartbuffer}
+
+\def\redostartbuffer[#1]%
+ {\dododostartbuffer{}{#1}{\e!start\v!buffer}{\e!stop\v!buffer}}
\def\dowithbuffer#1#2#3% name, startsequence, stopsequence, before, after
- {\setbuffercapsules{#2}{#3}%
- \normalexpanded{\noexpand\dodowithbuffer{#1}{\bufferstart}{\bufferstop}}}
+ {\normalexpanded{\dodowithbuffer{#1}{#2}{#3}}}
-\long\def\dodowithbuffer#1#2#3#4#5% name, startsequence, stopsequence, before, after
+\unexpanded\long\def\dodowithbuffer#1#2#3#4#5% name, startsequence, stopsequence, before, after
{#4%
\bgroup
- \setcatcodetable \vrbcatcodes
- \catcode`\\=12
\ctxlua{buffers.erase("#1")}%
+ \setcatcodetable \vrbcatcodes
\long\def\nododowithbuffer
{\egroup
#5}%
\long\def\dododowithbuffer##1#3% is detokenize needed? TEST
- {\ctxlua
- {buffers.grab("#1","#2","#3",\!!bs\detokenize{##1}\!!es)}
- \dododowithbuffer
- \nododowithbuffer}%
+ {\ctxlua{buffers.grab("#1","#2","#3",\!!bs\detokenize{##1}\!!es)} % space ?
+ \dododowithbuffer
+ \nododowithbuffer}%
\dododowithbuffer}
-\def\setbuffercapsules#1#2% \scantextokens not needed (had a reason at some point)
- {\edef\bufferstart{#1}\edef\bufferstart{\scantextokens\expandafter{\bufferstart}}%
- \edef\bufferstop {#2}\edef\bufferstop {\scantextokens\expandafter{\bufferstop }}}
-
\def\setbuffer
{\dosingleempty\dosetbuffer}
+\let\endbuffer\relax
+
\long\def\dosetbuffer[#1]#2\endbuffer % seldom used so we just pass #2
- {\begingroup
- \setcurrentbuffer{#1}%
- \ctxlua{buffers.set("\currentbuffer", \!!bs\detokenize{#2}\!!es)}%
- \endgroup}
+ {\ctxlua{buffers.set("#1", \!!bs\detokenize{#2}\!!es)}}
+
+\def\namedbufferparameter#1#2{\csname\??bu#1#2\endcsname}
\unexpanded\def\setupbuffer
{\dodoubleempty\dosetupbuffer}
@@ -161,191 +96,71 @@
\getparameters[\??bu][#1]%
\fi}
-\def\dodefinebuffer[#1][#2]%
- {\iffirstargument % else problems
- \doglobal\increment\nofdefinedbuffers
- \letvalue{\??bu#1\c!number }\nofdefinedbuffers
- \letvalue{\??bu#1\c!paragraph}\v!no
- \setuevalue{\e!start#1}{\noexpand\dostartbuffer[#1][def-\nofdefinedbuffers][\e!start#1][\e!stop#1]}% maybe also relax stop
- \setuevalue{\e!get #1}{\noexpand\dogetbuffer [#1][def-\nofdefinedbuffers]}%
- \setuevalue{\e!type #1}{\noexpand\dotypebuffer [#1][def-\nofdefinedbuffers]}%
- \getparameters[\??bu#1][#2]%
- \fi}
+\newtoks\everydefinebuffer
\unexpanded\def\definebuffer
{\dodoubleempty\dodefinebuffer}
-\def\thebuffernumber#1%
- {\csname\??bu#1\c!number\endcsname}
-
-\def\thedefinedbuffer#1%
- {def-\csname\??bu#1\c!number\endcsname}
-
-\unexpanded\def\getbuffer
- {\dodoubleempty\dogetbuffer}
-
-\def\dogetbuffer[#1][#2]%
- {\ifsecondargument
- \dodogetbuffer[#1][#2]%
- \else
- \dodogetbuffer[][#1]%
- \fi}
-
-\def\dogetbufferasis{\ctxlua{buffers.get("\currentbuffer")}}
-
-\def\dodogetbuffer[#1][#2]%
- {\getvalue{\??bu#1\c!before}%
- \dobuffer{16}{#2}\dogetbufferasis
- \getvalue{\??bu#1\c!after}}
-
-\unexpanded\def\typebuffer
- {\dodoubleempty\dotypebuffer}
-
-\def\doprocessbufferverbatim
- {\doinitializeverbatim
- \beginofverbatimlines
- \dostarttagged\t!verbatim\currentbuffer
- \ctxlua{buffers.type("\currentbuffer","\typingparameter\c!strip")}%
- \dostoptagged
- \endofverbatimlines}
-
-\def\doprocessbufferlinesverbatim#1#2#3%
- {#2\doprocessbufferverbatim#3}
-
-\def\doifelsebuffer#1%
- {\ctxlua{buffers.doifelsebuffer("#1")}}
-
-\def\dodotypebuffer#1#2#3% see dodotypefile
- {\doifelsebuffer{#3}
- {\dosometyping{#1}{#2}{#3}\doprocessbufferverbatim\doprocessbufferlinesverbatim}
- {\reporttypingerror{#3}}}
-
-\def\dotypefilebuffer{\dodotypebuffer{\v!file}{}{\currentbuffer}}%
-
-\def\dotypebuffer[#1][#2]%
+\def\dodefinebuffer[#1][#2]%
{\iffirstargument
- \dobuffer{17}{#1}\dotypefilebuffer
+ \global\advance\nofdefinedbuffers\plusone
+ \setevalue{\??bu#1\c!number}{\number\nofdefinedbuffers}%
+ \def\currentbuffer{#1}%
+ \getparameters[\??bu#1][#2]%
+ \the\everydefinebuffer
\else
- \dobuffer{17}{#2}\dotypefilebuffer
+ % fatal error
\fi}
-\def\dobuffer#1#2#3%
- {\doifelsenothing{#2}
- {\dodobuffer#3\jobname}
- {\processcommalist[#2]{\dodobuffer#3}}}
+\def\thebuffernumber #1{\csname\??bu#1\c!number\endcsname}
+\def\thedefinedbuffer#1{def-\csname\??bu#1\c!number\endcsname}
+
+\appendtoks
+ \setuevalue{\e!start\currentbuffer}%
+ {\noexpand\dostartdefinedbuffer
+ [\currentbuffer]%
+ [def-\number\nofdefinedbuffers]%
+ [\e!start\currentbuffer]%
+ [\e!stop\currentbuffer]}%
+ \setuevalue{\e!get\currentbuffer}%
+ {\noexpand\dogetdefinedbuffer
+ [\currentbuffer]%
+ [def-\number\nofdefinedbuffers]}%
+\to \everydefinebuffer
+
+\def\doprocessbufferlist#1#2%
+ {\doifelsenothing{#1}
+ {\dododogetbuffer\empty}
+ {\processcommalist[#1]#2}}
+
+\unexpanded\def\getbuffer % no [settings yet]
+ {\dosingleempty\dogetbuffer}
-\def\dodobuffer#1#2% command name
- {\pushmacro\currentbuffer
- \edef\currentbuffer{\ifcsname\??bu#2\c!number\endcsname def-\csname\??bu#2\c!number\endcsname\else#2\fi}%
- #1%
- \popmacro\currentbuffer}
+\unexpanded\def\dogetbuffer[#1]% [name]
+ {\namedbufferparameter\empty\c!before
+ \doprocessbufferlist{#1}\dododogetbuffer
+ \namedbufferparameter\empty\c!after}
-\def\processTEXbuffer{\getbuffer} % handy
+\def\dogetdefinedbuffer[#1][#2]%
+ {\namedbufferparameter{#1}\c!before
+ \dododogetbuffer{#2}%
+ \namedbufferparameter{#1}\c!after}
-% extras:
+\def\dododogetbuffer#1%
+ {\ctxlua{buffers.get("#1")}}
\def\inspectbuffer
{\dosingleempty\doinspectbuffer}
\def\doinspectbuffer[#1]%
- {\setcurrentbuffer{#1}%
- \ctxlua{buffers.inspect("\currentbuffer")}}
-
-% seldom used, only in a few projects that demanded more speed
-
-\let\usememorybuffers\relax
-\let\usefilebuffers \relax
-
-% this features is soldom used (complex examns where we need to fetch
-% special parts of a text
-%
-% this is not yet supported in mkiv (relatively easy to do but there
-% we don't have the par tags but need to grab 'm
-
-\def\skippedbufferparagraphs{0}
-
-\let\startbufferparagraph\relax
-\let\stopbufferparagraph \par % \relax
-
-\newcount\currentbufferparagraph
-
-\def\getbufferparagraphs
- {\dodoubleempty\dogetbufferparagraphs}
-
-\def\dosetbufferoffset#1%
- {\doifnumberelse{\getvalue{\??bu#1\c!paragraph}}
- {\currentbufferparagraph-\getvalue{\??bu#1\c!paragraph}}
- {\currentbufferparagraph \zerocount}%
- \relax}
-
-\def\dogetbufferparagraphs[#1][#2]%
- {\iffirstargument
- \ifsecondargument
- \dosetbufferoffset{#1}%
- \doifelse{#2}\v!all
- {\unexpanded\def\startbufferparagraph{\normalbufferparagraph{#1}}}
- {\unexpanded\def\startbufferparagraph{\filterbufferparagraph{#1}{#2}}}%
- \unexpanded\def\stopbufferparagraph{\dostopbufferparagraph{#1}}%
- \def\next{\getparagraphedbuffer[#1]}%
- \else
- \dosetbufferoffset\empty
- \unexpanded\def\startbufferparagraph{\filterbufferparagraph{}{#1}}%
- \unexpanded\def\stopbufferparagraph{\dostopbufferparagraph{}}%
- \def\next{\getparagraphedbuffer[]}%
- \fi
- \else
- \dosetbufferoffset\empty
- \unexpanded\def\startbufferparagraph{\normalbufferparagraph{}}%
- \unexpanded\def\stopbufferparagraph{\dostopbufferparagraph{}}%
- \def\next{\getparagraphedbuffer[]}%
- \fi
- \next}
-
-\def\dotypeparagraphbuffer{\ctxlua{buffers.get("\currentbuffer")}}
-
-\def\getparagraphedbuffer[#1]%
- {\dobuffer{16}{#1}\dotypeparagraphbuffer}
+ {\ctxlua{buffers.inspect("#1")}}
-\def\dostopbufferparagraph#1%
- {\getvalue{\??bu#1\c!after}\par}
+\definebuffer[\v!hiding] \setupbuffer[\v!hiding][\c!before=,\c!after=]
-\def\dostartbufferparagraph#1%
- {\par\getvalue{\??bu#1\c!before}}
-
-\def\normalbufferparagraph
- {\advance\currentbufferparagraph \plusone
- \ifnum\currentbufferparagraph>\zerocount
- \expandafter\dostartbufferparagraph
- \else
- \expandafter\gobbleoneargument
- \fi}
-
-\def\filterbufferparagraph#1#2%
- {\advance\currentbufferparagraph \plusone
- \ifcase\currentbufferparagraph
- \@EA\gobblebufferparagraph
- \else
- \doifinsetelse{\the\currentbufferparagraph}{#2}
- {\@EA\dostartbufferparagraph}
- {\@EA\fakebufferparagraph}%
- \fi
- {#1}}
-
-\long\def\gobblebufferparagraph#1#2\stopbufferparagraph
- {}
-
-\def\fakebufferparagraph#1%
- {\bgroup
- \unexpanded\def\stopbufferparagraph{\dostopbufferparagraph{#1}\egroup\egroup}%
- \setbox\scratchbox\vbox\bgroup\dostartbufferparagraph{#1}}
-
-% definitions
-
-\definebuffer[\v!hiding] \setupbuffer[\v!hiding][\c!local=\v!yes]
+\let\processTEXbuffer\getbuffer % handy synonym
\setupbuffer
- [\c!paragraph=\v!no,
- \c!before=,
+ [\c!before=,
\c!after=]
% only mkiv:
@@ -375,8 +190,6 @@
{\dosingleempty\doctxluabuffer}
\def\doctxluabuffer[#1]%
- {\doifelsenothing{#1}
- {\ctxlua{buffers.evaluate("\jobname")}}
- {\ctxlua{buffers.evaluate("#1")}}}
+ {\ctxlua{buffers.evaluate("#1")}}
\protect \endinput
diff --git a/tex/context/base/buff-ver.lua b/tex/context/base/buff-ver.lua
new file mode 100644
index 000000000..9574768b5
--- /dev/null
+++ b/tex/context/base/buff-ver.lua
@@ -0,0 +1,377 @@
+if not modules then modules = { } end modules ['buff-ver'] = {
+ version = 1.001,
+ comment = "companion to buff-ver.mkiv",
+ author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
+ copyright = "PRAGMA ADE / ConTeXt Development Team",
+ license = "see context related readme files"
+}
+
+-- The default visualizers have reserved names starting with v-*. Users are
+-- supposed to use different names for their own variants.
+
+local type, rawset, rawget, setmetatable, getmetatable = type, rawset, rawget, setmetatable, getmetatable
+local format, lower, match = string.format, string.lower, string.match
+local C, P, V, Carg = lpeg.C, lpeg.P, lpeg.V, lpeg.Carg
+local patterns, lpegmatch, lpegtype = lpeg.patterns, lpeg.match, lpeg.type
+
+local function is_lpeg(p)
+ return p and lpegtype(p) == "pattern"
+end
+
+visualizers = visualizers or { }
+
+local specifications = { } visualizers.specifications = specifications
+
+local verbatim = context.verbatim
+local variables = interfaces.variables
+local findfile = resolvers.findfile
+local addsuffix = file.addsuffix
+
+local v_yes = variables.yes
+
+-- beware, these all get an argument (like newline)
+
+local doinlineverbatimnewline = context.doinlineverbatimnewline
+local doinlineverbatimbeginline = context.doinlineverbatimbeginline
+local doinlineverbatimemptyline = context.doinlineverbatimemptyline
+local doinlineverbatimstart = context.doinlineverbatimstart
+local doinlineverbatimstop = context.doinlineverbatimstop
+
+local dodisplayverbatimnewline = context.dodisplayverbatimnewline
+local dodisplayverbatimbeginline = context.dodisplayverbatimbeginline
+local dodisplayverbatimemptyline = context.dodisplayverbatimemptyline
+local dodisplayverbatimstart = context.dodisplayverbatimstart
+local dodisplayverbatimstop = context.dodisplayverbatimstop
+
+local doverbatimspace = context.doverbatimspace
+
+local CargOne = Carg(1)
+
+local function f_emptyline(s,settings)
+ if settings and settings.currentnature == "inline" then
+ doinlineverbatimemptyline()
+ else
+ dodisplayverbatimemptyline()
+ end
+end
+
+local function f_beginline(s,settings)
+ if settings and settings.currentnature == "inline" then
+ doinlineverbatimbeginline()
+ else
+ dodisplayverbatimbeginline()
+ end
+end
+
+local function f_newline(s,settings)
+ if settings and settings.currentnature == "inline" then
+ doinlineverbatimnewline()
+ else
+ dodisplayverbatimnewline()
+ end
+end
+
+local function f_start(s,settings)
+ if settings and settings.currentnature == "inline" then
+ doinlineverbatimstart()
+ else
+ dodisplayverbatimstart()
+ end
+end
+
+local function f_stop(s,settings)
+ if settings and settings.currentnature == "inline" then
+ doinlineverbatimstop()
+ else
+ dodisplayverbatimstop()
+ end
+end
+
+local function f_default(s) -- (s,settings)
+ verbatim(s)
+end
+
+local function f_space() -- (s,settings)
+ doverbatimspace()
+end
+
+local functions = { __index = {
+ emptyline = f_emptyline,
+ newline = f_newline,
+ default = f_default,
+ beginline = f_beginline,
+ space = f_space,
+ start = f_start,
+ stop = f_stop,
+ }
+}
+
+local handlers = { }
+
+function visualizers.newhandler(name,data)
+ local tname, tdata = type(name), type(data)
+ if tname == "table" then -- (data)
+ setmetatable(name,getmetatable(name) or functions)
+ return name
+ elseif tname == "string" then
+ if tdata == "string" then -- ("name","parent")
+ local result = { }
+ setmetatable(result,getmetatable(handlers[data]) or functions)
+ handlers[name] = result
+ return result
+ elseif tdata == "table" then -- ("name",data)
+ setmetatable(data,getmetatable(data) or functions)
+ handlers[name] = data
+ return data
+ else -- ("name")
+ local result = { }
+ setmetatable(result,functions)
+ handlers[name] = result
+ return result
+ end
+ else -- ()
+ local result = { }
+ setmetatable(result,functions)
+ return result
+ end
+end
+
+function visualizers.newgrammar(name,t)
+ t = t or { }
+ local g = visualizers.specifications[name]
+ g = g and g.grammar
+ if g then
+ for k,v in next, g do
+ if not t[k] then
+ t[k] = v
+ end
+ if is_lpeg(v) then
+ t[name..":"..k] = v
+ end
+ end
+ end
+ return t
+end
+
+local fallback = context.verbatim
+
+local function makepattern(visualizer,kind,pattern)
+ if not pattern then
+ logs.simple("error in visualizer: %s",kind)
+ return patterns.alwaystrue
+ else
+ if type(visualizer) == "table" and type(kind) == "string" then
+ kind = visualizer[kind] or fallback
+ else
+ kind = fallback
+ end
+ return (C(pattern) * CargOne) / kind
+ end
+end
+
+visualizers.pattern = makepattern
+visualizers.makepattern = makepattern
+
+function visualizers.load(name)
+ if rawget(specifications,name) == nil then
+ name = lower(name)
+ local texname = findfile(format("v-%s.mkiv",name))
+ local luaname = findfile(format("v-%s.lua" ,name))
+ if texname == "" or luaname == "" then
+ -- assume a user specific file
+ luaname = findfile(addsuffix(name,"mkiv"))
+ texname = findfile(addsuffix(name,"lua" ))
+ end
+ if texname == "" or luaname == "" then
+ -- error message
+ else
+ lua.registercode(luaname)
+ context.input(texname)
+ end
+ if rawget(specifications,name) == nil then
+ rawset(specifications,name,false)
+ end
+ end
+end
+
+function commands.doifelsevisualizer(name)
+ commands.testcase(specifications[lower(name)])
+end
+
+function visualizers.register(name,specification)
+ specifications[name] = specification
+ local parser, handler = specification.parser, specification.handler
+ local displayparser = specification.display or parser
+ local inlineparser = specification.inline or parser
+ local isparser = is_lpeg(parser)
+ local start, stop
+ if isparser then
+ start = makepattern(handler,"start",patterns.alwaysmatched)
+ stop = makepattern(handler,"stop",patterns.alwaysmatched)
+ end
+ if handler then
+ if isparser then
+ specification.display = function(content,settings)
+ if handler.startdisplay then handler.startdisplay(settings) end
+ lpegmatch(start * displayparser * stop,content,1,settings)
+ if handler.stopdisplay then handler.stopdisplay(settings) end
+ end
+ specification.inline = function(content,settings)
+ if handler.startinline then handler.startinline(settings) end
+ lpegmatch(start * inlineparser * stop,content,1,settings)
+ if handler.stopinline then handler.stopinline(settings) end
+ end
+ specification.direct = function(content,settings)
+ lpegmatch(parser,content,1,settings)
+ end
+ elseif parser then
+ specification.display = function(content,settings)
+ if handler.startdisplay then handler.startdisplay(settings) end
+ parser(content,settings)
+ if handler.stopdisplay then handler.stopdisplay(settings) end
+ end
+ specification.inline = function(content,settings)
+ if handler.startinline then handler.startinline(settings) end
+ parser(content,settings)
+ if handler.stopinline then handler.stopinline(settings) end
+ end
+ specification.direct = parser
+ end
+ elseif isparser then
+ specification.display = function(content,settings)
+ lpegmatch(start * displayparser * stop,content,1,settings)
+ end
+ specification.inline = function(content,settings)
+ lpegmatch(start * inlineparser * stop,content,1,settings)
+ end
+ specification.direct = function(content,settings)
+ lpegmatch(parser,content,1,settings)
+ end
+ elseif parser then
+ specification.display = parser
+ specification.inline = parser
+ specification.direct = parser
+ end
+ return specification
+end
+
+local function getvisualizer(method,nature)
+ local m = specifications[method] or specifications.default
+ if nature then
+ return m and (m[nature] or m.parser) or nil
+ else
+ return m and m.parser or nil
+ end
+end
+
+local escapepatterns = { } visualizers.escapepatterns = escapepatterns
+
+local function texmethod(s)
+ context.bgroup()
+ context(s)
+ context.egroup()
+end
+
+local function defaultmethod(s,settings)
+ lpegmatch(getvisualizer("default"),s,1,settings)
+end
+
+function visualizers.registerescapepattern(name,before,after,normalmethod,escapemethod)
+ local escapepattern = escapepatterns[name]
+ if not escapepattern then
+ before, after = P(before) * patterns.space^0, patterns.space^0 * P(after)
+ escapepattern = (
+ (before / "")
+ * ((1 - after)^0 / (escapemethod or texmethod))
+ * (after / "")
+ + ((1 - before)^1) / (normalmethod or defaultmethod)
+ )^0
+ escapepatterns[name] = escapepattern
+ end
+ return escapepattern
+end
+
+local escapedvisualizers = { }
+
+local function visualize(method,nature,content,settings) -- maybe also method and nature in settings
+ if content and content ~= "" then
+ local m
+ local e = settings.escape
+ if e and e ~= "" then
+ local newname = format("%s-%s",e,method)
+ local newspec = specifications[newname]
+ if newspec then
+ m = newspec
+ else
+ local start, stop
+ if e == v_yes then
+ start, stop = "/BTEX", "/ETEX"
+ else
+ start,stop = match(e,"^(.-),(.-)$") -- todo: lpeg
+ end
+ if start and stop then
+ local oldvisualizer = specifications[method] or specifications.default
+ local oldparser = oldvisualizer.direct
+ local newparser = visualizers.registerescapepattern(newname,start,stop,oldparser)
+ m = visualizers.register(newname, {
+ parser = newparser,
+ handler = oldvisualizer.handler,
+ })
+ else
+ -- visualizers.register(newname,n)
+ specifications[newname] = m -- old spec so that we have one lookup only
+ end
+ end
+ else
+ m = specifications[method] or specifications.default
+ end
+ local n = m and m[nature]
+ settings.currentnature = nature or "display" -- tricky ... why sometimes no nature
+ if n then
+ n(content,settings)
+ else
+ fallback(content,1,settings)
+ end
+ end
+end
+
+visualizers.visualize = visualize
+visualizers.getvisualizer = getvisualizer
+
+function visualizers.visualizestring(method,content,settings)
+ visualize(method,"inline",content)
+end
+
+function visualizers.visualizefile(method,name,settings)
+ visualize(method,"display",resolvers.loadtexfile(name),settings)
+end
+
+function visualizers.visualizebuffer(method,name,settings)
+ visualize(method,"display",buffers.content(name),settings)
+end
+
+-- --
+
+local space = C(patterns.space) * CargOne / f_space
+local newline = C(patterns.newline) * CargOne / f_newline
+local emptyline = C(patterns.emptyline) * CargOne / f_emptyline
+local beginline = C(patterns.beginline) * CargOne / f_beginline
+local anything = C(patterns.somecontent^1) * CargOne / f_default
+
+local verbosed = (space + newline * (emptyline^0) * beginline + anything)^0
+
+local function write(s,settings) -- bad name
+ lpegmatch(verbosed,s,1,settings or false)
+end
+
+visualizers.write = write
+visualizers.writenewline = f_newline
+visualizers.writeemptyline = f_emptyline
+visualizers.writespace = f_space
+visualizers.writedefault = f_default
+
+function visualizers.writeargument(...)
+ context("{") -- If we didn't have tracing then we could
+ write(...) -- use a faster print to tex variant for the
+ context("}") -- { } tokens as they always have ctxcatcodes.
+end
diff --git a/tex/context/base/buff-ver.mkiv b/tex/context/base/buff-ver.mkiv
index 3be410300..2c478d46d 100644
--- a/tex/context/base/buff-ver.mkiv
+++ b/tex/context/base/buff-ver.mkiv
@@ -13,166 +13,147 @@
\writestatus{loading}{ConTeXt Buffer Macros / Verbatim}
-%D We can optimize esp the initializations a bit.
+\registerctxluafile{buff-ver}{1.001}
\unprotect
-\ifdefined\startlinenumbering\else \let\startlinenumbering \relax \fi
-\ifdefined\stoplinenumbering \else \let\stoplinenumbering \relax \fi
-\ifdefined\setuplinenumbering\else \unexpanded\def\setuplinenumbering[#1]{} \fi
-
\definesystemattribute[verbatimline][public]
\appendtoksonce \attribute\verbatimlineattribute \attributeunsetvalue \to \everyforgetall
-% D \macros
-% D {iflinepar}
-% D
-% D A careful reader will see that \type{\linepar} is reset.
-% D This boolean can be used to determine if the current line is
-% D the first line in a pseudo paragraph and this boolean is set
-% D after each empty line. The \type{\relax} can be used to
-% D determine the end of the line when one implements a scanner
-% D routine.
-%
-% will become obsolete
+%D Initializations.
-\newif\iflinepar
+\newtoks\everyinitializeverbatim
-% \type{ <crlf> char} geeft bagger
+\appendtoks
+ \the\everyresettypesetting
+ \resetfontfeature
+ \resetcharacterspacing
+\to \everyinitializeverbatim
-%D We are going to embed the general verbatim support macros in
-%D a proper environment. First we show the common setup
-%D macro, so we know what features are supported. The options
-%D are hooked into the support macros via the \type{\obey}
-%D macros.
+\def\setverbatimspaceskip % to be checked: must happen after font switch
+ {\spaceskip\fontcharwd\font`x\relax
+ \xspaceskip\spaceskip}
-\newif\ifslantedtypeactivated
-\newif\ifslantedtypepermitted
+\unexpanded\def\specialcontrolspace{\hskip\zeropoint\fastcontrolspace\hskip\zeropoint}
-\def\switchslantedtype
- {\ifslantedtypepermitted
- \ifslantedtypeactivated
- \slantedtypeactivatedfalse\tttf
- \else
- \slantedtypeactivatedtrue\ttsl
- \fi
- \fi}
+\setvalue{\??tp:\c!lines:\v!yes }{\obeybreakpoints}
+\setvalue{\??tp:\c!lines:\v!hyphenated}{\obeyhyphens}
-\def\prettyidentifier {TEX}
-\def\prettypalet {}
+\setvalue{\??tp:\c!space:\v!on }{\let\obeyedspace\specialcontrolspace}
+\setvalue{\??tp:\c!space:\v!stretch }{\let\obeyedspace\specialstretchedspace}
+\setvalue{\??tp:\c!space:\v!normal }{}
+
+\setvalue{\??tp:\c!blank:\v!standard }{\ctxparskip}
+\setvalue{\??tp:\c!blank:\v!small }{\smallskipamount}
+\setvalue{\??tp:\c!blank:\v!medium }{\medskipamount}
+\setvalue{\??tp:\c!blank:\v!big }{\bigskipamount}
+\setvalue{\??tp:\c!blank:\v!halfline }{.5\baselineskip}
+\setvalue{\??tp:\c!blank:\v!line }{\baselineskip}
+\setvalue{\??tp:\c!blank:\v!none }{\zeropoint}
+
+\def\secondstageinitializetype
+ {\let\obeyedspace\specialobeyedspace
+ \csname\??tp:\c!space:\typeparameter\c!space\endcsname
+ \relax\the\everyinitializeverbatim\relax}
-\def\installprettytype
- {\dodoubleargument\doinstallprettytype}
+\def\secondstageinitializetyping
+ {\let\obeyedspace\specialobeyedspace
+ \csname\??tp:\c!space:\typingparameter\c!space\endcsname
+ \csname\??tp:\c!lines:\typingparameter\c!lines\endcsname
+ \relax\the\everyinitializeverbatim\relax}
-\def\doinstallprettytype[#1][#2]% map #1 onto #2
- {\uppercasestring#1\to\asciia
- \uppercasestring#2\to\asciib
- \setevalue{\??ty\??ty\asciia}{\asciib}}
+\def\firststageinitializetype
+ {\let\obeylines\ignorelines
+ \ignorehyphens % default
+ \dosettypeattributes\c!style\c!color
+ \setcatcodetable\vrbcatcodes}
-\unexpanded\def\setupprettiesintype#1%
- {\uppercasestring#1\to\ascii
- \edef\prettyidentifier{\executeifdefined{\??ty\??ty\ascii}{TEX}}%
- \begingroup
- \ctxlua{buffers.loadvisualizer("\ascii")}%
- \endgroup}
+\let\doverbatimsetlinemargin\relax
-\unexpanded\def\setupprettytype
- {\processingverbatimtrue % will move
- \ctxlua{buffers.visualizers.reset()}}
+\def\dodoverbatimsetlinemargin
+ {\getpagestatus
+ \ifrightpage
+ \hskip\typingparameter\c!oddmargin\relax
+ \else
+ \hskip\typingparameter\c!evenmargin\relax
+ \fi}
-\def\setverbatimspaceskip
- {\setbox\scratchbox\hbox{x}%
- \spaceskip\wd\scratchbox
- \xspaceskip\spaceskip}
+\def\docheckverbatimmargins
+ {\scratchskip\typingparameter\c!oddmargin\relax
+ \ifzeropt\scratchskip
+ \else
+ \let\doverbatimsetlinemargin\dodoverbatimsetlinemargin
+ \fi
+ \scratchskip\typingparameter\c!evenmargin\relax
+ \ifzeropt\scratchskip
+ \else
+ \let\doverbatimsetlinemargin\dodoverbatimsetlinemargin
+ \fi
+ \ifx\doverbatimsetlinemargin\relax
+ \doadaptleftskip{\typingparameter\c!margin}%
+ \fi}
-\let\obeycharacters\relax % tab mess can go
+\def\firststageinitializetyping
+ {\switchtobodyfont[\typingparameter\c!bodyfont]%
+ \docheckverbatimmargins
+ \dosettypingattributes\c!style\c!color
+ \doifsomething{\typingparameter\c!align}{\setupalign[\typingparameter\c!align]}%
+ \ignorehyphens} % default
-\setvalue{\??tp:\c!lines:\v!yes }{\obeybreakpoints}
-\setvalue{\??tp:\c!lines:\v!hyphenated}{\obeyhyphens}
+%D \macros
+%D {definetype,setuptype}
+%D
+%D Specific inline verbatim commands can be defined with the
+%D following command.
-%setvalue{\??tp:\c!empty:\v!yes }{\obeyemptylines}
-%setvalue{\??tp:\c!empty:\v!all }{\obeyallemptylines}
-
-\setvalue{\??tp:\c!option:\v!none }{\let\obeycharacters\relax}
-\setvalue{\??tp:\c!option:\empty }{\let\obeycharacters\relax}
-\setvalue{\??tp:\c!option:\v!color }{\setupprettiesintype{\typingparameter\c!option}%
- \let\obeycharacters\setupprettytype}
-\setvalue{\??tp:\c!option:\v!normal }{\let\obeycharacters\setupgroupedtype}
-\setvalue{\??tp:\c!option:\v!slanted }{\let\obeycharacters\setupslantedtype}
-\setvalue{\??tp:\c!option:\s!unknown }{\setupprettiesintype{\typingparameter\c!option}%
- \let\obeycharacters\setupprettytype}
-%setvalue{\??tp:\c!option:\v!commands }{\def\obeycharacters{\setupcommandsintype}}
-
-\def\dosetverbatimfont
- {\redoconvertfont\dosetfontattribute{\currenttypingclass\currenttyping}\c!style
- \normalnoligatures\font}
-
-\unexpanded\def\setupcommonverbatim
- {\let\prettyidentifier\s!default
- %
- \def\prettyidentifierfont{\typingparameter\c!icommand}%
- \def\prettyvariablefont {\typingparameter\c!vcommand}%
- \def\prettynaturalfont {\typingparameter\c!ccommand}%
- %
- \ignorehyphens % default
- \getvalue{\??tp:\c!lines:\typingparameter\c!lines}%
- \getvalue{\??tp:\c!empty:\typingparameter\c!empty}%
- \getvalue{\??tp:\c!option:\ifcsname\??tp:\c!option:\typingparameter\c!option\endcsname\typingparameter\c!option\else\s!unknown\fi}%
- \setupverbatimcolor}
+\installcommandhandler{\??ty}{type}{\??ty}
-\newtoks \everyinitializeverbatim
+\appendtoks
+ \normalexpanded{\dodefinetype{\currenttype}}%
+\to \everydefinetype
+
+\unexpanded\def\dodefinetype#1%
+ {\setuvalue{#1}{\dotype{#1}}}
\appendtoks
- \the\everyresettypesetting
-\to \everyinitializeverbatim
+ \doinitializevisualizer{\typeparameter\c!option}%
+\to \everysetuptype
-\def\ignorebeginofpretty [#1]{}
-\def\ignoreendofpretty {}
-
-\def\doverbatimbop{\bgroup\beginofpretty}
-\def\doverbatimeop{\endofpretty\egroup}
-\def\doverbatimsop{\endofpretty\egroup\bgroup\beginofpretty}
-
-\let\noverbatimbop\ignorebeginofpretty
-\let\noverbatimeop\ignoreendofpretty
-\let\noverbatimsop\ignorebeginofpretty
-
-\setvalue{\??tp:\c!space:\v!on }{\let\obs\fastcontrolspace}
-\setvalue{\??tp:\c!space:\v!stretch}{\let\obs\specialstretchedspace
- \let\specialobeyedspace\specialstretchedspace % I need to clean
- \let\obeyedspace\specialstretchedspace} % up this mess.
-\setvalue{\??tp:\c!space:\v!normal }{\let\obs\specialobeyedspace}
-
-\def\doinitializeverbatim % todo: combine all in one call is faster
- {\ctxlua{buffers.visualizers.reset()}%
- \executeifdefined{\??tp:\c!space:\typingparameter\c!space}{\let\obs\specialobeyedspace}%
- \edef\askedverbatimtab{\typingparameter\c!tab}%
- \doifelse\askedverbatimtab\v!no
- {\ctxlua{buffers.settablength(1)}}
- {\doifnumberelse{\askedverbatimtab}
- {\ctxlua{buffers.settablength(\askedverbatimtab)}}
- {\ctxlua{buffers.settablength()}}}%
- \ctxlua{buffers.doifelsevisualizer("\prettyidentifier")}
- {\ctxlua{buffers.setvisualizer("\prettyidentifier")}%
- \let\bop\doverbatimbop
- \let\eop\doverbatimeop
- \let\sop\doverbatimsop}%
- {\ctxlua{buffers.setvisualizer("\v!typing")}% or resetdefaultvisualizer
- \let\bop\noverbatimbop
- \let\eop\noverbatimeop
- \let\sop\noverbatimsop}%
- \relax\the\everyinitializeverbatim\relax}
+%D \macros
+%D {definetyping,setuptyping}
+%D
+%D For most users the standard \type{\start}||\type{\stop}||pair
+%D will suffice, but for documentation purposes the next
+%D definition command can be of use:
+%D
+%D \starttyping
+%D \definetyping[extratyping][margin=3em]
+%D
+%D \startextratyping
+%D these extra ones are indented by 1 em
+%D \stopextratyping
+%D \stoptyping
+%D
+%D The definitions default to the standard typing values.
+
+\installcommandhandler{\??tp}{typing}{\??tp}
\appendtoks
- \resetfontfeature
- \resetcharacterspacing
-\to \everyinitializeverbatim
+ \normalexpanded{\dodefinetyping{\currenttyping}}%
+\to \everydefinetyping
+
+\unexpanded\def\dodefinetyping#1%
+ {\setvalue{\e!start#1}{\dostarttyping{#1}}%
+ \setvalue{\e!stop #1}{\dostoptyping {#1}}}
+
+\appendtoks
+ \doinitializevisualizer{\typingparameter\c!option}% we can check at the tex end
+\to \everysetuptyping
-% BEWARE: the noligatures will globally change the verbatim font's behaviour
+\def\doinitializevisualizer#1%
+ {\ifproductionrun\ctxlua{visualizers.load("#1")}\fi}
-% test case:
-%
% \definetype[typeTEX][option=tex]
%
% \typeTEX|\example---oeps|. this---ligates---again.
@@ -219,33 +200,9 @@
%D
%D \typebuffer \start \getbuffer \stop
-\unexpanded\def\setupcommandsintype
- {\ctxlua{buffers.setescapepair("\currenttyping",\!!bs\typingparameter\c!escape\!!es)}}
-
-\appendtoks
- \setupcommandsintype
-\to \everyinitializeverbatim
-
-\unexpanded\def\setupslantedtype
- {\slantedtypepermittedtrue}
-
-\ifx\setupprettytype \undefined \let\setupprettytype \relax \fi
-\ifx\setupslantedtype \undefined \let\setupslantedtype \relax \fi
-\ifx\setupgroupedtype \undefined \let\setupgroupedtype \relax \fi
-\ifx\normalnoligatures\undefined \let\normalnoligatures\gobbleoneargument \fi
-
-%D The verbatim commands have a rather long and turbulent
-%D history. Most users of \CONTEXT\ probably will never use
-%D some of the features, but I've kept in mind that when one is
-%D writing a users manual, about everything can and undoubtly
-%D will be subject to a verbatim treatment.
-%D
%D Verbatim command are very sensitive to argument processing,
%D which is a direct result of the \CATCODES\ being fixed at
-%D reading time. With our growing understanding of \TEX,
-%D especially of the mechanism that can be used for looking
-%D ahead and manipulating \CATCODES, the verbatim support
-%D became more and more advanced and natural.
+%D reading time.
%D
%D Typesetting inline verbatim can be accomplished by
%D \type{\type}, which in this sentence was typeset by saying
@@ -262,14 +219,11 @@
%D \type{<<} and \type{>>} as delimiters. This alternative can
%D be used in situations where slanted typeseting is needed.
-% todo: we can use \letter... here:
-
\def\lesscharacter {<}
\def\morecharacter {>}
-
-\chardef\texescape = `\\
-\chardef\leftargument = `\{
-\chardef\rightargument = `\}
+\let\texescape \textbackslash
+\let\leftargument \textbraceleft
+\let\rightargument \textbraceright
%D \macros
%D {type}
@@ -277,17 +231,17 @@
%D We define \type{\type} as a protected command. This command
%D has several invocations: grouped, wirt boundary characters,
%D and with font switches.
-
-% \starttyping
-% normal: \par \type{xx<<..xx..<<xx <<xx>> >>..>>xx} \par \type<<....>> \par \type<<..<<xx>>..>> \par
-% normal: \par \type{xx<..xx..<xx <slanted> >..>xx} \par \type{<....>} \par \type{<..<xx>..>}
-% \setuptype[option=slanted]
-% slanted: \par \type{xx<<..sl..<<xx <<sl>> xx>>..sl..>>xx} \par \type<<..xx..>> \par \type<<..<<sl>>..>> \par
-% slanted: \par \type{xx<<..sl..<xx <sl> xx>..sl..>>xx} \par \type<<..xx..>> \par \type<<..<sl>..>> \par
-% \setuptype[option=none]
-% none: \par \type{xx<<..xx..<<xx <<xx>> >>..>>xx} \par \type<<....>> \par \type<<..<<xx>>..>> \par
-% \stoptyping
-
+%D
+%D \starttyping
+%D normal: \par \type{xx<<..xx..<<xx <<xx>> >>..>>xx} \par \type<<....>> \par \type<<..<<xx>>..>> \par
+%D normal: \par \type{xx<..xx..<xx <slanted> >..>xx} \par \type{<....>} \par \type{<..<xx>..>}
+%D \setuptype[option=slanted]
+%D slanted: \par \type{xx<<..sl..<<xx <<sl>> xx>>..sl..>>xx} \par \type<<..xx..>> \par \type<<..<<sl>>..>> \par
+%D slanted: \par \type{xx<<..sl..<xx <sl> xx>..sl..>>xx} \par \type<<..xx..>> \par \type<<..<sl>..>> \par
+%D \setuptype[option=none]
+%D none: \par \type{xx<<..xx..<<xx <<xx>> >>..>>xx} \par \type<<....>> \par \type<<..<<xx>>..>> \par
+%D \stoptyping
+%D
%D When writing the manual to \CONTEXT\ and documenting this
%D source we needed to typeset \type{<<} and \type{>>}. Because
%D we wanted to do this in the natural way, we've adapted the
@@ -296,7 +250,7 @@
%D further and treats the lone \type{<<} and \type{>>} a bit
%D different. The \type {\null} prevents ligatures, which
%D unfortunately turn up in Lucida fonts.
-
+%D
%D The following lines show what happens when we set
%D \type {option=commands}.
%D
@@ -316,127 +270,90 @@
%D \typebuffer
\unexpanded\def\type{\dotype\empty}
+\unexpanded\def\typ {\dotyp \empty}
-\def\dotype#1% was \dotype
- {\dontleavehmode \bgroup
- % new, \strut enables leading space in \type { abc } at par start / begstrut
- % else no hyphenation (replaced by \dontleavehmode which saves unboxing)
- % \begstrut
- \let\currenttypingclass\??ty
- \edef\currenttyping{#1}%
- \catcode`\<=\othercatcode
- \catcode`\>=\othercatcode
- \futurelet\next\dodotype}
+\def\dotype#1%
+ {\dontleavehmode
+ \bgroup
+ \edef\currenttype{#1}%
+ \doifnextoptionalelse\redotype\dodotype}
+
+\def\redotype[#1]%
+ {\getparameters[\??ty\currenttype][#1]%
+ \dodotype}
\def\dodotype
+ {\futurelet\next\dodotypeone}
+
+\def\dotyp#1%
+ {\dontleavehmode
+ \bgroup
+ \edef\currenttype{#1}%
+ \dolettypeparameter\v!lines\v!hyphenated
+ \let\specialobeyedspace\specialstretchedspace
+ \doifnextoptionalelse\redotype\dodotype}
+
+\def\dodotypeone
{\ifx\next\bgroup
\@EA\dodotypeA
\else
- \@EA\dodotypeAD
+ \@EA\dodotypetwo
\fi}
-\def\dodotypeAD
+\def\dodotypetwo
+ {\catcode`<=\othercatcode % old precaution
+ \catcode`>=\othercatcode % old precaution
+ \futurelet\next\dodotypethree}
+
+\def\dodotypethree
{\if\next<%
- \doifelse{\typingparameter\c!option}\v!none{\@EA\dodotypeB}{\@EA\dodotypeC}%
+ \@EA\dodotypeB
\else
- \@EA\dodotypeD
+ \@EA\dodotypeC
\fi}
\def\dodotypeA
- {\initializetype % probably too much
- \verbatimcolor
- \dosetverbatimfont
- \setcatcodetable \typcatcodesa
- \dodotypeAA}
-
-\def\dodotypeAA#1%
- {\doinitializeverbatim
- \def\obs{\obeyedspace}%
- \dostarttagged\t!verbatim\currenttyping
- \ctxlua{buffers.hooks.flush_inline(\!!bs\detokenize{#1}\!!es)}%
- \dostoptagged
- \egroup}
+ {\firststageinitializetype
+ \setcatcodetable\typcatcodesa
+ \dodotypenormal}
\def\dodotypeB#1%
- {\initializetype
- \verbatimcolor
- \dosetverbatimfont
- \setcatcodetable \typcatcodesb
- \dodotypeBB}
-
-\def\dodotypeBB#1%
- {\doinitializeverbatim
- \dostarttagged\t!verbatim\currenttyping
- \ctxlua{buffers.visualizers.flushnested(\!!bs\detokenize{#1}\!!es,false)}%
- \dostoptagged
- \egroup
- \gobbleoneargument} % grab last >
+ {\firststageinitializetype
+ \setcatcodetable\typcatcodesb
+ \dodotypenested}
\def\dodotypeC#1%
- {\initializetype
- \verbatimcolor
- \dosetverbatimfont
- \setcatcodetable \typcatcodesb
- \dodotypeCC}
-
-\def\dodotypeCC#1%
- {\doinitializeverbatim
- \ifx\obeycharacters\setupprettytype % temp hack, we need a proper signal
- \dostarttagged\t!verbatim\currenttyping
- \ctxlua{buffers.hooks.flush_inline([\!!bs\detokenize{#1}\!!es,true)}%
- \dostoptagged
- \else
- \def\obs{\obeyedspace}%
- \ctxlua{buffers.visualizers.flushnested(\!!bs\detokenize{#1}\!!es,true)}%
- \fi
- \egroup
- \gobbleoneargument} % grab last >
+ {\firststageinitializetype
+ \setcatcodetable\typcatcodesc
+ \def\next##1#1{\dodotypenormal{##1}}%
+ \next}
+
+\def\dodotypenormal#1%
+ {\secondstageinitializetype
+ \dostarttagged\t!verbatim\currenttype
+ \ctxlua{buffers.typestring {
+ data = \!!bs\detokenize{#1}\!!es,
+ tab = "\typeparameter\c!tab",
+ visualizer = "\typeparameter\c!option",
+ }}%
+ \dostoptagged
+ \egroup}
-\def\dodotypeD#1%
- {\initializetype
- \verbatimcolor
- \dosetverbatimfont
- \setcatcodetable \typcatcodesc % was a
- \def\dodotypeDD##1#1{\dodotypeAA{##1}}%
- \dodotypeDD}
-
-\def\dodotypeDD#1%
- {\doinitializeverbatim
- \dostarttagged\t!verbatim\currenttyping
- \ctxlua{buffers.hooks.flush_inline(\!!bs\detokenize{#1}\!!es,true)}%
+\def\dodotypenested#1%
+ {\secondstageinitializetype
+ \dostarttagged\t!verbatim\currenttype
+ \ctxlua{buffers.typestring {
+ data = \!!bs\detokenize{#1}\!!es,
+ tab = "\typeparameter\c!tab",
+ visualizer = "nested", % we force a special visualizer
+ option = "\typeparameter\c!option", % extra visualizer (maybe: nested,\typeparameter\c!option)
+ }}%
\dostoptagged
\egroup
\gobbleoneargument} % grab last >
-%D The neccessary initializations are done by calling
-%D \type{\initializetype} which in return calls for the support
-%D macro \type{\setupinlineverbatim}.
-
-\def\initializetype
- {\let\obeylines\ignorelines
- \setupcommonverbatim
- %\dosetverbatimfont
- %\setverbatimspaceskip
- %\setupcopyverbatim % not needed
- \setcatcodetable\vrbcatcodes}
-
%D \macros
-%D {setuptype}
-%D
-%D Some characteristics of \type{\type} can be set up by:
-
-\unexpanded\def\setuptype
- {\dodoubleempty\dosetuptype}
-
-\def\dosetuptype[#1][#2]%
- {\ifsecondargument
- \getparameters[\??ty#1][#2]%
- \else
- \getparameters[\??ty][#1]%
- \fi}
-
-%D \macros
-%D {typ,obeyhyphens,obeybreakpoints}
+%D {obeyhyphens,obeybreakpoints}
%D
%D Although it's not clear from the macros, one character
%D trait of this macros, which are build on top of the support
@@ -464,12 +381,6 @@
\let\controlspace\specialcontrolspace
\spaceskip.5em\relax}
-\unexpanded\def\typ
- {\bgroup
- \let\@@tylines\v!hyphenated
- \let\specialobeyedspace\specialstretchedspace
- \futurelet\next\dodotype}
-
%D \macros
%D {tex,arg,mat,dis}
%D
@@ -480,36 +391,9 @@
%D but we've decided not to use that slow and sometimes
%D troublesome solution. Instead we stick to some 'old'
%D \CONTEXT\ macros for typesetting typical \TEX\ characters.
-%D
-%D The next implementation is more clear but less versatile,
-%D so we treated it for a beter one.
-%D
-%D \starttyping
-%D \def\dospecialtype#1#2%
-%D {\bgroup
-%D \initializetype
-%D \catcode`\{=\begingroupcatcode
-%D \catcode`\}=\endgroupcatcode
-%D \def\dospecialtype%
-%D {\def\dospecialtype{#2\egroup}%
-%D \bgroup
-%D \aftergroup\dospecialtype
-%D #1}%
-%D \afterassignment\dospecialtype
-%D \let\next=}
-%D
-%D \unexpanded\def\tex{\dospecialtype\texescape\relax}
-%D \unexpanded\def\arg{\dospecialtype\leftargument\rightargument}
-%D \unexpanded\def\mat{\dospecialtype\$\$}
-%D \unexpanded\def\dis{\dospecialtype{\$\$}{\$\$}}
-%D \stoptyping
\def\setgroupedtype
- {\let\currenttypingclass\??ty
- \initializetype
- \verbatimcolor
- \dosetverbatimfont
- %\setcatcodetable \typcatcodesa
+ {\firststageinitializetype
\catcode`\{=\begingroupcatcode
\catcode`\}=\endgroupcatcode}
@@ -523,105 +407,17 @@
\let\normaltextmat\mat
\let\normaltextdis\dis
-\def\astype
- {\groupedcommand\dorawtype\relax}
-
-\def\dorawtype
- {\let\currenttypingclass\??ty
- \normalverbatimcolor % \verbatimcolor
- \dosetverbatimfont}
-
-%D \macros
-%D {starttyping}
-%D
-%D Display verbatim is realized far more easy, which is mostly
-%D due to the fact that we use \type{\stop...} as delimiter.
-%D The implementation inherits some features, for instance the
-%D support of linenumbering, which can best be studied in the
-%D documented support module.
-
-\let\currenttyping \empty
-\let\currenttypingclass\??ty % saveguard
-
-\def\typingparameter#1%
- {\ifcsname\currenttypingclass\currenttyping#1\endcsname
- \csname\currenttypingclass\currenttyping#1\endcsname
- \else\ifcsname\currenttypingclass#1\endcsname
- \csname\currenttypingclass#1\endcsname
- \fi\fi}
-
-\def\settypingparameter#1#2%
- {\setvalue{\currenttypingclass\currenttyping#1}{#2}}
-
-\def\setxtypingparameter#1#2%
- {\setxvalue{\currenttypingclass\currenttyping#1}{#2}}
-
-\setvalue{\??tp:\c!blank:\v!standard}{\ctxparskip}
-\setvalue{\??tp:\c!blank:\v!small }{\smallskipamount}
-\setvalue{\??tp:\c!blank:\v!medium }{\medskipamount}
-\setvalue{\??tp:\c!blank:\v!big }{\bigskipamount}
-\setvalue{\??tp:\c!blank:\v!halfline}{.5\baselineskip}
-\setvalue{\??tp:\c!blank:\v!line }{\baselineskip}
-\setvalue{\??tp:\c!blank:\v!none }{\zeropoint}
-
-\def\doopenupverbatimlineindeed
- {\getpagestatus
- \ifrightpage
- \hskip\typingparameter\c!oddmargin\relax
- \else
- \hskip\typingparameter\c!evenmargin\relax
- \fi}
-
-\def\initializetyping
- {%\donefalse
- \switchtobodyfont[\typingparameter\c!bodyfont]%
- \donefalse
- \scratchskip\typingparameter\c!oddmargin\relax
- \ifzeropt\scratchskip\else\donetrue\fi
- \scratchskip\typingparameter\c!evenmargin\relax
- \ifzeropt\scratchskip\else\donetrue\fi
- \ifdone
- \let\doopenupverbatimline\doopenupverbatimlineindeed
- \else
- \doadaptleftskip{\typingparameter\c!margin}%
- \fi
- % no symbolic blanks !
- \edef\!!stringa{\executeifdefined{\??bo\typingparameter\c!blank}{\typingparameter\c!blank}}%
- \scratchskip\executeifdefined{\??tp:\c!blank:\!!stringa}\!!stringa\relax
- \ifgridsnapping
- % this will be adapted
- \ifdim\scratchskip=.5\baselineskip\relax
- \edef\verbatimbaselineskip{\the\scratchskip}% new
- \else
- \edef\verbatimbaselineskip{\the\baselineskip}%
- \fi
- \else
- \edef\verbatimbaselineskip{\the\scratchskip}%
- \fi
-\doifsomething{\typingparameter\c!align}{\setupalign[\typingparameter\c!align]}%
- \setupcommonverbatim}
+\def\astype{\groupedcommand{\dosettypeattributes\c!style\c!color}{}}
%D The basic display verbatim commands are defined in an
%D indirect way. As we will see, they are a specific case of a
%D more general mechanism.
-\newif\ifoptimizeverbatim \optimizeverbatimtrue
-
-\let \beginofverbatimlines \relax
-\let \endofverbatimlines \relax
-
-\def\doverbatimnobreak
- {\ifoptimizeverbatim\penalty500 \fi}
-
-\def\doverbatimgoodbreak
- {\ifoptimizeverbatim\penalty\linepenalty\fi}
-
-% we need this hack because otherwise verbatim skips
-% the first line (everything after the initial command)
+\let\beginofverbatimlines\relax
+\let\endofverbatimlines \relax
\def\dostarttyping#1% tricky non standard lookahead
{\bgroup
- \let\currenttypingclass\??tp
\edef\currenttyping{#1}%
\obeylines
\futurelet\nexttoken\dodostarttyping}
@@ -636,40 +432,37 @@
\def\nododostarttyping
{\dododostarttyping[]}
-\def\dotypefileverbatim
- {\doinitializeverbatim
- \beginofverbatimlines
- \dostarttagged\t!verbatimblock\currenttyping
- \ctxlua{buffers.typefile("\readfilename","\typingparameter\c!strip","\typingparameter\c!range","\currentregime")}%
- \dostoptagged
- \endofverbatimlines}
-
-\def\dotypefilelinesverbatim#1#2%
- {#1\dotypefileverbatim#2}
+\def\dododostarttyping[#1]%
+ {\typingparameter\c!before
+ \startpacked[\v!blank]
+ \doifassignmentelse{#1}
+ {\setuptyping[\currenttyping][#1]}
+ {\doifinset\v!continue{#1}{\setuptyping[\currenttyping][\c!continue=\v!yes]}}%
+ \dosetuptypelinenumbering
+ \firststageinitializetyping
+ \normalexpanded{\dotypeblockverbatim{\e!start\currenttyping}{\e!stop\currenttyping}}}
\unexpanded\def\dotypeblockverbatim#1#2%
- {\dowithbuffer{_typing_}{#1}{#2}
- {}
- {\doinitializeverbatim
- \beginofverbatimlines
- \dostarttagged\t!verbatimblock\currenttyping
- \ctxlua{buffers.type("_typing_","\typingparameter\c!strip","\typingparameter\c!range")}%
- \dostoptagged
- \endofverbatimlines
- \csname#2\endcsname}}
+ {\dowithbuffer{_typing_}{#1}{#2}{}{\dodotypeblockverbatim{#1}{#2}}}
-\def\dododostarttyping[#1]%
- {\typingparameter\c!before
- \startpacked % includes \bgroup
- \dosetuptypelinenumbering{#1}%
- \initializetyping
- \dosetverbatimfont
- \startverbatimcolor
- \normalexpanded{\dotypeblockverbatim{\e!start\currenttyping}{\e!stop\currenttyping}}} % was s!start
+\def\dodotypeblockverbatim#1#2%
+ {\secondstageinitializetyping
+ \beginofverbatimlines
+ \dostarttagged\t!verbatimblock\currenttyping
+ \ctxlua{buffers.typebuffer {
+ name = "_typing_",
+ strip = "\typingparameter\c!strip",
+ range = "\typingparameter\c!range",
+ tab = "\typingparameter\c!tab",
+ visualizer = "\typingparameter\c!option",
+ escape = \!!bs\typingparameter\c!escape\!!es,
+ }}%
+ \dostoptagged
+ \endofverbatimlines
+ \csname#2\endcsname}
\def\dostoptyping#1% hm, currenttyping
- {\stopverbatimcolor
- \stoppacked % includes \egroup
+ {\stoppacked
\typingparameter\c!after
\normalexpanded{\egroup\checknextindentation[\typingparameter\c!indentnext]}%
\dorechecknextindentation}
@@ -710,8 +503,7 @@
%D
%D \starttext
%D
-%D % \typefile[file][range={3,6}]{whatever.tex}
-%D
+%D % \typefile[file][range={3,6}] {whatever.tex}
%D % \typefile[file][range={3,+2}]{whatever.tex}
%D
%D \typefile[file][range={label:start:one,label:stop:one}]{whatever.tex}
@@ -720,128 +512,6 @@
%D \stoptyping
%D \macros
-%D {setuptyping}
-%D
-%D The setup of typing accepts two arguments. The optional
-%D first one identifies the user defined ones. If only one
-%D argument is given, the values apply to both the standard
-%D command \type{\starttyping} and \type{\typefile}.
-
-\def\dosetuptyping[#1][#2]%
- {\ifsecondargument
- \getparameters[\??tp#1][#2]%
- \else
- \getparameters[\??tp][#1]%
- \fi}
-
-\unexpanded\def\setuptyping
- {\dodoubleempty\dosetuptyping}
-
-%D \macros
-%D {definetype}
-%D
-%D Specific inline verbatim commands can be defined with the
-%D following command.
-
-\unexpanded\def\definetype
- {\dodoubleempty\dodefinetype}
-
-\def\dodefinetype[#1][#2]%
- {\setuvalue{#1}{\dotype{#1}}%
- \getparameters[\??ty#1][#2]}
-
-%D \macros
-%D {definetyping}
-%D
-%D For most users the standard \type{\start}||\type{\stop}||pair
-%D will suffice, but for documentation purposes the next
-%D definition command can be of use:
-%D
-%D \starttyping
-%D \definetyping[extratyping][margin=3em]
-%D
-%D \startextratyping
-%D these extra ones are indented by 1 em
-%D \stopextratyping
-%D \stoptyping
-%D
-%D The definitions default to the standard typing values.
-
-% TODO: parent etc !
-
-\def\presettyping[#1][#2]% brrr also use parent here
- {\copyparameters[\??tp#1][\??tp][\c!color,\c!style]%
- \getparameters [\??tp#1][#2]}
-
-\def\dodefinetyping[#1][#2]%
- {\setvalue{\e!start#1}{\dostarttyping{#1}}%
- \setvalue{\e!stop #1}{\dostoptyping {#1}}%
- \presettyping[#1][#2]}
-
-\unexpanded\def\definetyping
- {\dodoubleempty\dodefinetyping}
-
-%D We can use some core color commands. These are faster than
-%D the standard color switching ones and work ok on a line by
-%D line basis.
-%D
-%D \starttyping
-%D \unexpanded\def\setupverbatimcolor%
-%D {\edef\prettypalet{\prettyidentifier\typingparameter\c!palet}%
-%D \def\beginofpretty[##1]{\startcolormode{\prettypalet:##1}}%
-%D \def\endofpretty {\stopcolormode}}
-%D \stoptyping
-%D
-%D Since we support a global color too, the folowing
-%D definition is better:
-
-\def\normalbeginofpretty [#1]{\startcolor[\prettypalet:#1]}%
-\def\normalendofpretty {\stopcolor}
-\def\normalstartverbatimcolor{\startcolor[\typingparameter\c!color]}%
-\def\normalstopverbatimcolor {\stopcolor}
-\def\normalverbatimcolor {\getvalue{\typingparameter\c!color}}% command !
-
-\unexpanded\def\setupnormalprettyverbatim
- {\edef\prettypalet{\prettyidentifier\prettypalet}%
- \let\beginofpretty \normalbeginofpretty
- \let\endofpretty \normalendofpretty
- \let\startverbatimcolor\normalstartverbatimcolor
- \let\stopverbatimcolor \normalstopverbatimcolor
- \let\verbatimcolor \normalverbatimcolor}
-
-\unexpanded\def\setupignoreprettyverbatim
- {\let\prettypalet \empty
- \let\beginofpretty \ignorebeginofpretty
- \let\endofpretty \ignoreendofpretty
- \let\startverbatimcolor\normalstartverbatimcolor
- \let\stopverbatimcolor \normalstopverbatimcolor
- \let\verbatimcolor \normalverbatimcolor}
-
-\unexpanded\def\setupverbatimcolor
- {\edef\prettypalet{\typingparameter\c!palet}%
- \ifx\prettypalet\empty
- \setupignoreprettyverbatim
- \else
- \setupnormalprettyverbatim
- \fi}
-
-\let\beginofpretty \ignorebeginofpretty
-\let\endofpretty \ignoreendofpretty
-\let\prettypalet \empty
-\let\startverbatimcolor\relax
-\let\stopverbatimcolor \relax
-\let\verbatimcolor \relax
-
-%D \macros
-%D {EveryPar, EveryLine, iflinepar}
-%D
-%D One of the features of these commands is the support of
-%D \type{\EveryPar}, \type{\EveryLine} and \type{\iflinepar}.
-%D In the documentation of the verbatim support module we give
-%D some examples of line- and paragraph numbering using these
-%D macros.
-
-%D \macros
%D {typefile}
%D
%D Typesetting files verbatim (for the moment) only supports
@@ -849,118 +519,92 @@
%D setup values are inherited from display verbatim.
%D The implementation of \type{\typefile} is straightforward:
-% new feature (not yet 100\% ok)
-%
-% \setuptyping[file][numbering=file]
-%
-% \typefile[start=2,nlines=3]{zapf}
-% \typefile[start=continue,nlines=13]{zapf}
-% \typefile{zapf}
-%
-% \setuptyping[file][numbering=line]
-%
-% \typefile[start=4,step=3]{zapf}
-% \typefile{zapf}
+% [category] [settings] {name} % for historic reasons, all filenames are {}
\def\typefile
{\dodoubleempty\dotypefile}
\def\dotypefile[#1][#2]#3%
- {\ifsecondargument
- \dodotypefile[#1][#2]{#3}%
+ {\bgroup
+ \ifsecondargument
+ \setuptyping[#1][#2]%
+ \dododotypefile{#1}{#3}%
\else\iffirstargument
\doifassignmentelse{#1}
- {\dodotypefile[\v!file][#1]{#3}}
- {\dodotypefile[#1][]{#3}}%
+ {\setuptyping[\v!file][#1]
+ \dododotypefile\v!file{#3}}
+ {\dododotypefile{#1}{#3}}%
\else
- \dodotypefile[\v!file][]{#3}%
- \fi\fi}
-
-\def\dosetuptypelinenumbering#1% fuzzy
- {%\ifcsname\currenttypingclass\currenttyping\c!start\endcsname \else
- % \setuptyping[\currenttyping][\c!start=1,\c!stop=,\c!step=1,\c!continue=\v!no,\c!nlines=]%
- %\fi
- \doifassignmentelse{#1}{\setuptyping[\currenttyping][#1]}\donothing
- \doifelse{\typingparameter\c!numbering}\v!file
- {% kind of special: filters lines !
- \setuplinenumbering[\currenttyping][\c!method=\v!file]%
- \donetrue}
- {\doifelse{\typingparameter\c!numbering}\v!line
- {\doifinset\v!continue{#1}{\setuptyping[\currenttyping][\c!continue=\v!yes]}% fails: \settypingparameter\c!continue{\v!yes}
- \donetrue}
- {\donefalse}}%
- \ifdone
- \edef\beginofverbatimlines{\noexpand\startlinenumbering
- [\currenttyping]%
- [\c!continue=\typingparameter\c!continue,
- \c!method=\v!type,
- \c!start=\typingparameter\c!start,
- \c!stop=\typingparameter\c!stop, % ?
- \c!step=\typingparameter\c!step]%
- }%
- \def\endofverbatimlines{\stoplinenumbering}%
- \fi}
-
-\def\reporttypingerror#1% temp hack
- {\blank
- \dontleavehmode\hbox\bgroup
- \expanded{\defconvertedargument\noexpand\ascii{#1}}%
- \tttf[\makemessage\m!verbatims1\ascii]%
- \showmessage\m!verbatims1\ascii
- \egroup
- \blank}
-
-\def\dosometyping#1#2#3#4#5%
- {\bgroup
- \let\currenttypingclass\??tp
- \edef\currenttyping{#1}%
- \typingparameter\c!before
- \startpacked % includes \bgroup
- \dosetuptypelinenumbering{#2}%
- \doifinset{\typingparameter\c!option}{\v!commands,\v!slanted,\v!normal}
- {\setuptyping[#1][\c!option=\v!none]}%
- \doif{\typingparameter\c!option}\v!color
- {\expandafter\aftersplitstring#3\at.\to\prettyidentifier
- \settypingparameter\c!option{\prettyidentifier}}%
- \initializetyping
- \dosetverbatimfont
- \startverbatimcolor
- \scratchcounter
- \ifcsname\currenttypingclass#3\v!global\c!start\endcsname
- \numexpr\csname\currenttypingclass#3\v!global\c!start\endcsname+\plusone\relax
- \else
- \plusone
- \fi
- \setxvalue{\currenttypingclass#3\v!global\c!start}{\the\scratchcounter}% no direct setxvalue as it defines beforehand
- \doifelsenothing{\typingparameter\c!start}
- {#4}
- {\doif{\typingparameter\c!start}\v!continue
- {\setevalue{\currenttypingclass#1\c!start}%
- {\getvalue{\currenttypingclass#3\v!global\c!start}}}%
- \doifelsenothing{\typingparameter\c!stop}
- {\doifelsenothing{\typingparameter\c!nlines}
- {#4}
- {\setxvalue{\currenttypingclass#3\v!global\c!start}%
- {\the\numexpr\typingparameter\c!start+\typingparameter\c!nlines+\minusone\relax}%
- #5{\typingparameter\c!start}{\getvalue{\currenttypingclass#3\v!global\c!start}}}}%
- {#5{\typingparameter\c!start}{\typingparameter\c!stop}}}%
- \stopverbatimcolor
- \stoppacked
- \typingparameter\c!after
+ \dododotypefile\v!file{#3}%
+ \fi\fi
\egroup}
+\def\dododotypefile#1#2%
+ {\doifelsetypingfile{#2}
+ {\dodododotypefile{#1}\readfilename}
+ {\reporttypingerror{#2}}}
+
\def\doifelsetypingfile#1% sets \readfilename (we will make this proper mkiv i.e. less messy)
{\doiflocfileelse{#1}
{\firstoftwoarguments}
{\doifinputfileelse{#1}
-% {\def\readfilename{\pathplusfile\filepath{#1}}\firstoftwoarguments} % messy, looks wrong too
{\def\readfilename{#1}\firstoftwoarguments} % messy, looks wrong too
{\secondoftwoarguments}}}
-\def\dodotypefile[#1][#2]#3%
- {\doifelsetypingfile{#3}
- {\dosometyping{#1}{#2}{#3}\dotypefileverbatim\dotypefilelinesverbatim}
- {\reporttypingerror{#3}}}
+\def\dodododotypefile#1#2% category name
+ {\edef\currenttyping{#1}%
+ \typingparameter\c!before
+ \startpacked[\v!blank]
+ \dosetuptypelinenumbering
+ \firststageinitializetyping
+ \secondstageinitializetyping
+ \beginofverbatimlines
+ \dostarttagged\t!verbatimblock\currenttyping
+ \ctxlua{buffers.typefile {
+ name = "#2",
+ strip = "\typingparameter\c!strip",
+ range = "\typingparameter\c!range",
+ regime = "\currentregime",
+ tab = "\typingparameter\c!tab",
+ visualizer = "\typingparameter\c!option",
+ }}%
+ \dostoptagged
+ \endofverbatimlines
+ \stoppacked
+ \typingparameter\c!after}
+
+%D Line numbering:
+
+\newconditional\verbatimnumberinglines
+
+\def\beginofverbatimlinenumbering
+ {\startlinenumbering
+ [\currenttyping]%
+ [\c!continue=\typingparameter\c!continue,
+ \c!method=\v!type,
+ \c!start=\typingparameter\c!start,
+ \c!stop=\typingparameter\c!stop, % ?
+ \c!step=\typingparameter\c!step]}
+
+\def\endofverbatimlinenumbering
+ {\stoplinenumbering}
+
+\def\dosetuptypelinenumbering
+ {\ifdefined\startlinenumbering
+ \edef\currenttypingnumbering{\typingparameter\c!numbering}%
+ \ifx\currenttypingnumbering\v!file
+ \setuplinenumbering[\currenttyping][\c!method=\v!file]%
+ \settrue\verbatimnumberinglines
+ \else\ifx\currenttypingnumbering\v!line
+ \settrue\verbatimnumberinglines
+ \else
+ \setfalse\verbatimnumberinglines
+ \fi\fi
+ \ifconditional\verbatimnumberinglines
+ \let\beginofverbatimlines\beginofverbatimlinenumbering
+ \let\endofverbatimlines \endofverbatimlinenumbering
+ \fi
+ \fi}
%D \macros
%D {filename}
@@ -975,9 +619,12 @@
\unexpanded\def\filename#1{{\tttf\hyphenatedfilename{#1}}}
-%D And a bonus macro:
+%D \macros
+%D {verbatim}
+%d
+%D And a bonus macro, an unexpanded detokenize:
-\def\verbatim#1{\defconvertedargument\ascii{#1}\ascii}
+\unexpanded\def\verbatim#1{\detokenize{#1}}
%D The setups for display verbatim and file verbatim are
%D shared. One can adapt the extra defined typing environments,
@@ -987,224 +634,182 @@
\setuptyping
[ \c!before=\blank,
\c!after=\blank,
- \c!bodyfont=,
- \c!color=,
+ %\c!bodyfont=,
+ %\c!color=,
\c!space=\v!off,
\c!page=\v!no,
\c!tab=\s!ascii,
\c!option=\v!none,
- \c!palet=colorpretty,
\c!text=\v!no,
\c!style=\tttf,
- \c!icommand=\ttsl,
- \c!vcommand=,
- \c!ccommand=\tttf,
\c!indentnext=\v!yes,
\c!margin=\!!zeropoint,
\c!evenmargin=\!!zeropoint,
\c!oddmargin=\!!zeropoint,
\c!blank=\v!line,
- \c!escape=, % yes | no | START,STOP BTEX,ETEX
+ %\c!escape=, % yes | no | {START,STOP} | default when yes: {BTEX,ETEX}
\c!numbering=\v!no,
- \c!lines=,
- \c!range=,
- \c!empty=,
+ %\c!lines=,
+ %\c!range=,
\c!start=1,
- \c!stop=,
+ %\c!stop=,
\c!step=1,
- \c!continue=,
- \c!strip=\v!no, % auto or number
- \c!nlines=]
+ %\c!continue=,
+ \c!strip=\v!no] %auto or number
\definetyping[\v!typing]
-\presettyping[\v!file][]
-
-% \setuptyping % not needed
-% [\v!file]
-% [\c!start=1,
-% \c!stop=,
-% \c!step=1,
-% \c!continue=,
-% \c!nlines=]
+\setuptyping[\v!file] [\s!parent=\??tp\v!typing] % we don't want \start..\stop overload
+\setuptyping[\v!buffer][\s!parent=\??tp\v!file] % we don't want \start..\stop overload
%D The setups for inline verbatim default to:
\setuptype
[ \c!space=\v!off,
- \c!color=,
+ %\c!color=,
\c!style=\tt\tf, % \tttf gives problems with {\tx \type...}
+ %\c!option=\v!normal,
\c!page=\v!no,
- \c!tab=\v!yes,
- \c!palet=colorpretty,
- \c!option=\v!normal]
-
-%D Beware: only a few are currently (re)implemented in \MKIV.
-
-\definetyping[RAW] [\c!option=RAW]
-\definetyping[MP] [\c!option=MP] % done
-\definetyping[PL] [\c!option=PL]
-\definetyping[PM] [\c!option=PL]
-\definetyping[JS] [\c!option=JS]
-\definetyping[JV] [\c!option=JV]
-\definetyping[SQL] [\c!option=SQL]
-\definetyping[TEX] [\c!option=TEX] % done
-\definetyping[PAS] [\c!option=PAS]
-\definetyping[PASCAL][\c!option=PAS]
-\definetyping[MOD] [\c!option=PAS]
-\definetyping[MODULA][\c!option=PAS]
-\definetyping[DELPHI][\c!option=PAS]
-\definetyping[EIFFEL][\c!option=EIF]
-\definetyping[XML] [\c!option=XML]
-\definetyping[LUA] [\c!option=LUA] % done
-
-\installprettytype [RAW] [RAW]
-
-\installprettytype [TEX] [TEX]
-
-\installprettytype [PERL] [PL]
-\installprettytype [PL] [PL]
-\installprettytype [PM] [PL]
-
-\installprettytype [METAPOST] [MP]
-\installprettytype [METAFONT] [MP]
-\installprettytype [MP] [MP]
-\installprettytype [MF] [MP]
-
-\installprettytype [JAVASCRIPT] [JS]
-\installprettytype [JAVA] [JV]
-\installprettytype [JS] [JS]
-\installprettytype [JV] [JV]
-
-\installprettytype [SQL] [SQL]
-
-\installprettytype [PASCAL] [PAS]
-\installprettytype [PAS] [PAS]
-\installprettytype [MODULA] [PAS]
-\installprettytype [MOD] [PAS]
-
-\installprettytype [EIFFEL] [EIF]
-\installprettytype [EIF] [EIF]
-\installprettytype [E] [EIF]
-
-\installprettytype [XML] [XML]
-
-\installprettytype [LUA] [LUA]
-
-%D We use the \CONTEXT\ color system for switching to and from
-%D color mode. We can always redefine these colors afterwards.
-
-\definecolor [colorprettyone] [r=.9, g=.0, b=.0] % red
-\definecolor [colorprettytwo] [r=.0, g=.8, b=.0] % green
-\definecolor [colorprettythree] [r=.0, g=.0, b=.9] % blue
-\definecolor [colorprettyfour] [r=.8, g=.8, b=.6] % yellow
-
-\definecolor [grayprettyone] [s=.30]
-\definecolor [grayprettytwo] [s=.45]
-\definecolor [grayprettythree] [s=.60]
-\definecolor [grayprettyfour] [s=.75]
-
-\definepalet
- [colorpretty]
- [ prettyone=colorprettyone,
- prettytwo=colorprettytwo,
- prettythree=colorprettythree,
- prettyfour=colorprettyfour]
-
-\definepalet
- [graypretty]
- [ prettyone=grayprettyone,
- prettytwo=grayprettytwo,
- prettythree=grayprettythree,
- prettyfour=grayprettyfour]
-
-\definepalet [TEXcolorpretty] [colorpretty]
-\definepalet [TEXgraypretty] [graypretty]
-\definepalet [PLcolorpretty] [colorpretty]
-\definepalet [PLgraypretty] [graypretty]
-\definepalet [PMcolorpretty] [colorpretty]
-\definepalet [PMgraypretty] [graypretty]
-\definepalet [MPcolorpretty] [colorpretty]
-\definepalet [MPgraypretty] [graypretty]
-\definepalet [JVcolorpretty] [colorpretty]
-\definepalet [JVgraypretty] [graypretty]
-\definepalet [JScolorpretty] [colorpretty]
-\definepalet [JSgraypretty] [graypretty]
-\definepalet [SQLcolorpretty] [colorpretty]
-\definepalet [SQLgraypretty] [graypretty]
-\definepalet [PAScolorpretty] [colorpretty]
-\definepalet [PASgraypretty] [graypretty]
-\definepalet [EIFcolorpretty] [colorpretty]
-\definepalet [EIFgraypretty] [graypretty]
-\definepalet [XMLcolorpretty] [colorpretty]
-\definepalet [XMLgraypretty] [graypretty]
-\definepalet [LUAcolorpretty] [colorpretty]
-\definepalet [LUAgraypretty] [graypretty]
-
-% patched from verb-ini (todo)
-
-% \let\beginverbatimline\relax
-% \let\endverbatimline \relax
-
-% \appendtoks whatever\par\to\everyverbatimbeginofdisplay
-% \appendtoks whatever\to\everyverbatimendofdisplay
-% \appendtoks [\to\everyverbatimbeginofinline
-% \appendtoks ]\to\everyverbatimendofinline
-
-\let\doopenupverbatimline\empty
-
-\newtoks\everyverbatimbeginofdisplay
-\newtoks\everyverbatimendofdisplay
-\newtoks\everyverbatimbeginofinline
-\newtoks\everyverbatimendofinline
-
-\let\currentverbatimpretty\empty
-
-\def\doverbatimbeginofdisplay#1%
- {\edef\currentverbatimpretty{#1}%
- \the\everyverbatimbeginofdisplay}
-
-\def\doverbatimendofdisplay
- {\the\everyverbatimendofdisplay}
-
-\def\doverbatimbeginofinline#1%
- {\edef\currentverbatimpretty{#1}%
- \the\everyverbatimbeginofinline}
-
-\def\doverbatimendofinline
- {\the\everyverbatimendofinline}
+ \c!tab=\v!yes]
+
+%D Buffers
+
+% [name] [settings] | [name] | [settings]
+
+\unexpanded\def\typebuffer
+ {\dodoubleempty\dotypebuffer}
+
+\unexpanded\def\dotypedefinedbuffer
+ {\dotripleempty\dodotypedefinedbuffer}
+
+\appendtoks
+ \setuevalue{\e!type\currentbuffer}%
+ {\noexpand\dotypedefinedbuffer
+ [\v!buffer]%
+ [def-\number\nofdefinedbuffers]}%
+\to \everydefinebuffer
+
+\def\dotypebuffer[#1][#2]%
+ {\bgroup
+ \ifsecondargument
+ \setuptyping[\v!buffer][#2]%
+ \processcommalist[#1]{\dododotypebuffer\v!buffer}% [name] [settings]
+ \else\iffirstargument
+ \doifassignmentelse{#1}
+ {\setuptyping[\v!buffer][#1]%
+ \dododotypebuffer\v!buffer\empty}% [settings]
+ {\processcommalist[#1]{\dododotypebuffer\v!buffer}}% [name]
+ \else
+ \dododotypebuffer\v!buffer\empty% []
+ \fi\fi
+ \egroup}
+
+\def\dodotypedefinedbuffer[#1][#2][#3]% category name settings
+ {\bgroup
+ \ifthirdargument
+ \setuptyping[#1][#3]%
+ \fi
+ \dododotypebuffer{#1}{#2}%
+ \egroup}
+
+\def\dododotypebuffer#1#2% category name
+ {\edef\currenttyping{#1}%
+ \typingparameter\c!before
+ \startpacked[\v!blank]
+ \dosetuptypelinenumbering
+ \firststageinitializetyping
+ \secondstageinitializetyping
+ \beginofverbatimlines
+ \dostarttagged\t!verbatim{#1}%
+ \ctxlua{buffers.typebuffer {
+ name = "#2",
+ strip = "\typingparameter\c!strip",
+ range = "\typingparameter\c!range",
+ regime = "\currentregime",
+ tab = "\typingparameter\c!tab",
+ visualizer = "\typingparameter\c!option",
+ }}%
+ \dostoptagged
+ \endofverbatimlines
+ \stoppacked
+ \typingparameter\c!after}
+
+% process buffers .. settings
+
+\unexpanded\def\processbuffer
+ {\dodoubleempty\doprocessbuffer}
+
+\def\doprocessbuffer[#1][#2]%
+ {\bgroup
+ \ifsecondargument
+ \setuptyping[\v!buffer][#2]%
+ \processcommalist[#1]{\dodoprocessbuffer\v!buffer}% [name] [settings]
+ \else\iffirstargument
+ \doifassignmentelse{#1}
+ {\setuptyping[\v!buffer][#1]%
+ \dodoprocessbuffer\v!buffer\empty}% [settings]
+ {\processcommalist[#1]{\dodoprocessbuffer\v!buffer}}% [name]
+ \else
+ \dodoprocessbuffer\v!buffer\empty% []
+ \fi\fi
+ \egroup}
+
+% get : before/after of buffer
+% typing : before/after of typing
+% process : no before/after (special case anyway)
+
+\def\dodoprocessbuffer#1#2%
+ {\edef\currenttyping{#1}%
+ \ctxlua{buffers.typebuffer {
+ name = "#2",
+ strip = "\typingparameter\c!strip",
+ tab = "\typingparameter\c!tab",
+ visualizer = "\typingparameter\c!option",
+ }}}
+
+% line numbering, keep broken lines together
\newcount\nofverbatimlines
-\def\doverbatimbeginofline#1% linenumber
- {\global\advance\nofverbatimlines\plusone
- \attribute\verbatimlineattribute\nofverbatimlines
- \bgroup % due to pretty status
- \iflinepar\else\EveryPar{}\fi
- \noindent % was wrong: \dontleavehmode
- \xdef\dokeepverbatimlinedata % hm, still needed?
- {\parindent \the\parindent
- \hangindent\the\hangindent
- \hangafter \the\hangafter
- \leftskip \the\leftskip
- \rightskip \the\rightskip}%
- \egroup
- \dokeepverbatimlinedata
- \doopenupverbatimline
+\def\doverbatimbeginofline
+ {\ifconditional\verbatimnumberinglines
+ \global\advance\nofverbatimlines\plusone
+ \attribute\verbatimlineattribute\nofverbatimlines
+ \fi
+ \noindent
+ \doverbatimsetlinemargin
\the\everyline\strut
- \dostarttagged\t!verbatimline\empty
- }%\beginverbatimline}
+ \dostarttagged\t!verbatimline\empty}
\def\doverbatimendofline
- {%\endverbatimline
- \dostoptagged
- \global\lineparfalse
- \obeyedline\par
- \attribute\verbatimlineattribute\attributeunsetvalue}
+ {\dostoptagged
+ \obeyedline % still needed?
+ \par
+ \ifconditional\verbatimnumberinglines
+ \attribute\verbatimlineattribute\attributeunsetvalue
+ \fi}
\def\doverbatimemptyline
- {\strut
- \par
- \global\linepartrue}
+ {\ifconditional\verbatimnumberinglines
+ \par\strut\par % this will be an option where we use a signal instead of a strut
+ \else
+ \blank[\typingparameter\c!blank]%
+ \fi}
+
+% hooks:
+
+\def\doinlineverbatimstart {}
+\def\doinlineverbatimstop {}
+\def\doinlineverbatimbeginline {}
+\def\doinlineverbatimnewline {\obeyedspace}
+\def\doinlineverbatimemptyline {\obeyedspace}
+
+\def\dodisplayverbatimstart {\doverbatimbeginofline}
+\def\dodisplayverbatimstop {\doverbatimendofline}
+\def\dodisplayverbatimbeginline {\doverbatimbeginofline}
+\def\dodisplayverbatimnewline {\doverbatimendofline\par}
+\def\dodisplayverbatimemptyline {\doverbatimemptyline}
+
+\def\doverbatimspace {\obeyedspace}
\protect \endinput
diff --git a/tex/context/base/buff-vis.lua b/tex/context/base/buff-vis.lua
deleted file mode 100644
index 1a97f2591..000000000
--- a/tex/context/base/buff-vis.lua
+++ /dev/null
@@ -1,74 +0,0 @@
-if not modules then modules = { } end modules ['buff-vis'] = {
- version = 1.001,
- comment = "companion to buff-vis.mkiv",
- author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
- copyright = "PRAGMA ADE / ConTeXt Development Team",
- license = "see context related readme files"
-}
-
-local type = type
-local format = string.format
-local C, P, V, patterns, lpegmatch = lpeg.C, lpeg.P, lpeg.V, lpeg.patterns, lpeg.match
-
-visualizers = visualizers or { }
-
-local patterns = { } visualizers.patterns = patterns
-
-local fallback = context.verbatim
-
-function visualizers.pattern(visualizer,kind,pattern)
- if type(visualizer) == "table" and type(kind) == "string" then
- kind = visualizer[kind] or visualizer.default or fallback
- else
- kind = fallback
- end
- return C(pattern)/kind
-end
-
-setmetatable(patterns, {
- __index = function(t,k)
- local v = require(format("v-%s.lua",k)) or false
- context.input(format("v-%s.mkiv",k))
- t[k] = v
- return v
- end
-} )
-
-local function visualizestring(method,content)
- if content and content ~= "" then
- lpegmatch(patterns[method],content)
- end
-end
-
-visualizers.visualizestring = visualizestring
-
-function visualizers.visualizefile(method,name)
- visualizestring(method,resolvers.loadtexfile(name))
-end
-
-function visualizers.visualizebuffer(method,name)
- lpegmatch(method,buffers.content(name))
-end
-
-local visualizer = {
- start = function() context.startSnippet() end,
- stop = function() context.stopSnippet() end ,
- default = context.verbatim,
-}
-
-local patterns = lpeg.patterns
-local pattern = visualizers.pattern
-
-local texvisualizer = P { "process",
- process =
- V("start") * V("content") * V("stop"),
- start =
- pattern(visualizer,"start",patterns.beginofstring),
- stop =
- pattern(visualizer,"stop",patterns.endofstring),
- content = (
- pattern(visualizer,"default",patterns.anything)
- )^1
-}
-
-return texvisualizer
diff --git a/tex/context/base/buff-vis.mkiv b/tex/context/base/buff-vis.mkiv
deleted file mode 100644
index 7739dc21f..000000000
--- a/tex/context/base/buff-vis.mkiv
+++ /dev/null
@@ -1,24 +0,0 @@
-%D \module
-%D [ file=buff-vis,
-%D version=2010.10.19, % replaces old visualizers as we now have lpeg
-%D title=\CONTEXT\ Buffer Macros,
-%D subtitle=Visualizers,
-%D author=Hans Hagen,
-%D date=\currentdate,
-%D copyright=PRAGMA ADE]
-%C
-%C This module is part of the \CONTEXT\ macro||package and is
-%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
-%C details.
-
-\writestatus{loading}{ConTeXt Buffer Macros / Visualizers}
-
-\registerctxluafile{buff-vis}{1.001}
-
-\unprotect
-
-\definestartstop
- [Snippet]
- [\c!style=\tt]
-
-\protect \endinput
diff --git a/tex/context/base/char-utf.lua b/tex/context/base/char-utf.lua
index 87ed861ed..27e87d29c 100644
--- a/tex/context/base/char-utf.lua
+++ b/tex/context/base/char-utf.lua
@@ -43,9 +43,6 @@ local filters = characters.filters
filters.utf = filters.utf or { }
local utffilters = characters.filters.utf
-utffilters.collapsing = true
-utffilters.expanding = true
-
--[[ldx--
<p>It only makes sense to collapse at runtime, since we don't expect
source code to depend on collapsing.</p>
@@ -169,7 +166,7 @@ not collecting tokens is not only faster but also saves garbage collecting.
-- lpeg variant is not faster
function utffilters.collapse(str) -- not really tested (we could preallocate a table)
- if utffilters.collapsing and str then
+ if str and str ~= "" then
local nstr = #str
if nstr > 1 then
if initialize then -- saves a call
@@ -248,6 +245,9 @@ function utffilters.collapse(str) -- not really tested (we could preallocate a t
return str
end
+utilities.sequencers.appendaction (resolvers.openers.textfileactions,"system","characters.filters.utf.collapse")
+utilities.sequencers.disableaction(resolvers.openers.textfileactions,"characters.filters.utf.collapse")
+
--[[ldx--
<p>Next we implement some commands that are used in the user interface.</p>
--ldx]]--
diff --git a/tex/context/base/char-utf.mkiv b/tex/context/base/char-utf.mkiv
index 8992b098e..b59d2f569 100644
--- a/tex/context/base/char-utf.mkiv
+++ b/tex/context/base/char-utf.mkiv
@@ -28,11 +28,11 @@
%D since the source files are rather simple, we postpone the
%D initialization till runtime.
+% resolvers.filters.install('utf',characters.filters.utf.collapse)
+
\appendtoks
- \ctxlua {
- characters.filters.utf.collapsing = true
- resolvers.filters.install('utf',characters.filters.utf.collapse)
- }%
+ \ctxlua{utilities.sequencers.enableaction
+ (resolvers.openers.textfileactions,"characters.filters.utf.collapse")}%
\to \everyjob
%D The next one influences input parsing.
diff --git a/tex/context/base/cldf-com.lua b/tex/context/base/cldf-com.lua
index 5ab1d8c8d..9d03a450a 100644
--- a/tex/context/base/cldf-com.lua
+++ b/tex/context/base/cldf-com.lua
@@ -48,3 +48,33 @@ function context.concat(t,separator)
end
end
end
+
+function context.char(k) -- todo: if catcode == letter or other then just the utf
+ if type(k) == "table" then
+ for i=1,#k do
+ context(format([[\char%s\relax]],k[i]))
+ end
+ elseif k then
+ context(format([[\char%s\relax]],k))
+ end
+end
+
+function context.utfchar(k)
+ context(utfchar(k))
+end
+
+function context.chardef(cs,u)
+ context(format([[\chardef\%s=%s\relax]],k))
+end
+
+function context.par()
+ context([[\par]]) -- no need to add {} there
+end
+
+function context.bgroup()
+ context("{")
+end
+
+function context.egroup()
+ context("}")
+end
diff --git a/tex/context/base/cldf-ver.lua b/tex/context/base/cldf-ver.lua
index 8607cb578..120bc468a 100644
--- a/tex/context/base/cldf-ver.lua
+++ b/tex/context/base/cldf-ver.lua
@@ -9,7 +9,7 @@ if not modules then modules = { } end modules ['cldf-ver'] = {
-- tex.print == newline
local concat, tohandle = table.concat, table.tohandle
-local splitlines = string.splitlines
+local find, splitlines = string.find, string.splitlines
local tostring, type = tostring, type
local context = context
diff --git a/tex/context/base/colo-ini.mkiv b/tex/context/base/colo-ini.mkiv
index 406cc619d..af3c1ebf4 100644
--- a/tex/context/base/colo-ini.mkiv
+++ b/tex/context/base/colo-ini.mkiv
@@ -140,7 +140,7 @@
%D \showsetup{color}
%D \showsetup{graycolor}
- \def\switchtocolor [#1]{\getvalue{#1}}
+ \def\switchtocolor [#1]{\csname#1\endcsname}
\unexpanded\def\color [#1]{\groupedcommand{\doactivatecolor{#1}}{}}
\unexpanded\def\startcolor [#1]{\begingroup\doactivatecolor{#1}}
\unexpanded\def\stopcolor {\endgroup}
@@ -149,10 +149,13 @@
\unexpanded\def\fastcolored [#1]#2{\begingroup\dodefinefastcolor[@colored@][#1]\doactivatecolor{@colored@}#2\endgroup}
\def\predefinecolor [#1]{\flushatshipout{\hbox{\color[#1]{}}}}
\def\predefineindexcolor[#1]{\flushatshipout{\hbox{\color[#1]{}}}}
- \unexpanded\def\startcolorpage {\startcolor[\ifx\maintextcolor\empty\defaulttextcolor\else\maintextcolor\fi]}
- \unexpanded\def\stopcolorpage {\stopcolor}
- \unexpanded\def\startraster [#1]{\dosetrastercolor{#1}}
- \unexpanded\def\stopraster {}
+
+% some of this will go away
+
+\unexpanded\def\startcolorpage {\startcolor[\ifx\maintextcolor\empty\defaulttextcolor\else\maintextcolor\fi]}
+\unexpanded\def\stopcolorpage {\stopcolor}
+\unexpanded\def\startraster [#1]{\dosetrastercolor{#1}}
+\unexpanded\def\stopraster {}
\def\raster [#1]{\groupedcommand{\dosetrastercolor{#1}}{}}
\def\faststartcolor [#1]{\doactivatecolor{#1}}
\def\faststopcolor {}
diff --git a/tex/context/base/cont-new.mkiv b/tex/context/base/cont-new.mkiv
index e400dfaa0..9e97001b7 100644
--- a/tex/context/base/cont-new.mkiv
+++ b/tex/context/base/cont-new.mkiv
@@ -39,10 +39,10 @@
% for a while as these can be used in user modules
-\startluacode
- jobregisters = logs.obsolete("jobregisters", "structures.registers")
- buffers.finish_state = logs.obsolete("buffers.finish_state","buffers.finishstate")
- buffers.change_state = logs.obsolete("buffers.change_state","buffers.finishstate")
-\stopluacode
+% \startluacode
+% jobregisters = logs.obsolete("jobregisters", "structures.registers")
+% buffers.finish_state = logs.obsolete("buffers.finish_state","buffers.finishstate")
+% buffers.change_state = logs.obsolete("buffers.change_state","buffers.finishstate")
+% \stopluacode
\protect \endinput
diff --git a/tex/context/base/cont-new.tex b/tex/context/base/cont-new.tex
index f024ae024..42d506e11 100644
--- a/tex/context/base/cont-new.tex
+++ b/tex/context/base/cont-new.tex
@@ -11,7 +11,7 @@
%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
%C details.
-\newcontextversion{2010.11.20 12:51}
+\newcontextversion{2010.11.26 21:21}
%D This file is loaded at runtime, thereby providing an
%D excellent place for hacks, patches, extensions and new
diff --git a/tex/context/base/context.mkiv b/tex/context/base/context.mkiv
index c888140b0..4e790adad 100644
--- a/tex/context/base/context.mkiv
+++ b/tex/context/base/context.mkiv
@@ -232,7 +232,17 @@
\loadmarkfile{buff-ini}
\loadmarkfile{buff-ver}
-\loadmarkfile{buff-vis}
+
+\loadmarkfile{v-default} % preloaded as otherwise spurious spaces inline due to loading
+\loadmarkfile{v-nested} % preloaded as otherwise spurious spaces inline due to loading
+\loadmarkfile{v-escaped} % for completeness
+\loadmarkfile{v-tex} % optional as also runtime if not loaded
+\loadmarkfile{v-mp} % optional as also runtime if not loaded
+\loadmarkfile{v-lua} % optional as also runtime if not loaded
+\loadmarkfile{v-xml} % optional as also runtime if not loaded
+
+\loadmarkfile{v-parsed-xml} % optional
+\loadmarkfile{v-parsed-lua} % optional
\loadmarkfile{strc-blk}
diff --git a/tex/context/base/context.tex b/tex/context/base/context.tex
index e88fbee62..11e0e02e1 100644
--- a/tex/context/base/context.tex
+++ b/tex/context/base/context.tex
@@ -20,7 +20,7 @@
%D your styles an modules.
\edef\contextformat {\jobname}
-\edef\contextversion{2010.11.20 12:51}
+\edef\contextversion{2010.11.26 21:21}
%D For those who want to use this:
diff --git a/tex/context/base/core-sys.mkiv b/tex/context/base/core-sys.mkiv
index f5bb45e35..cdca352bd 100644
--- a/tex/context/base/core-sys.mkiv
+++ b/tex/context/base/core-sys.mkiv
@@ -194,48 +194,79 @@
\definecomplexorsimple\start
\definecomplexorsimple\stop
-\def\dododefinestartstop[#1][#2]% todo: use indirect commands
- {\getparameters
- [\??be#1]
- [\c!before=,
- \c!after=,
- \c!inbetween=,
- \c!commands=,
- \c!style=,
- #2]%
- \setuvalue{#1}%
- {\groupedcommand
- {\getvalue{\??be#1\c!commands}%
- \dostarttagged\t!construct{#1}%
- \dostartattributes{\??be#1}\c!style\c!color}
- {\dostopattributes
- \dostoptagged
- \getvalue{\??be#1\c!inbetween}}}%
- \setvalue{\e!start#1}%
- {\getvalue{\??be#1\c!before}%
- \bgroup
- \getvalue{\??be#1\c!commands}%
- \dostarttagged\t!construct{#1}%
- \dostartattributes{\??be#1}\c!style\c!color\empty}%
- \setvalue{\e!stop#1}%
- {\dostopattributes
- \dostoptagged
- \egroup
- \getvalue{\??be#1\c!after}}}
+% \def\dododefinestartstop[#1][#2]% todo: use indirect commands
+% {\getparameters
+% [\??be#1]
+% [\c!before=,
+% \c!after=,
+% \c!inbetween=,
+% \c!commands=,
+% \c!style=,
+% #2]%
+% \setuvalue{#1}%
+% {\groupedcommand
+% {\getvalue{\??be#1\c!commands}%
+% \dostarttagged\t!construct{#1}%
+% \dostartattributes{\??be#1}\c!style\c!color}
+% {\dostopattributes
+% \dostoptagged
+% \getvalue{\??be#1\c!inbetween}}}%
+% \setuvalue{\e!start#1}%
+% {\getvalue{\??be#1\c!before}%
+% \bgroup
+% \getvalue{\??be#1\c!commands}%
+% \dostarttagged\t!construct{#1}%
+% \dostartattributes{\??be#1}\c!style\c!color\empty}%
+% \setuvalue{\e!stop#1}%
+% {\dostopattributes
+% \dostoptagged
+% \egroup
+% \getvalue{\??be#1\c!after}}}
+%
+% \def\dodefinestartstop[#1][#2]%
+% {\def\docommand##1{\dododefinestartstop[##1][#2]}%
+% \processcommalist[#1]\docommand}
+%
+% \unexpanded\def\definestartstop
+% {\dodoubleargument\dodefinestartstop}
+%
+% \def\dosetupstartstop[#1][#2]%
+% {\def\docommand##1{\getparameters[\??be##1][#2]}%
+% \processcommalist[#1]\docommand}
+%
+% \unexpanded\def\setupstartstop
+% {\dodoubleargument\dosetupstartstop}
+
-\def\dodefinestartstop[#1][#2]%
- {\def\docommand##1{\dododefinestartstop[##1][#2]}%
- \processcommalist[#1]\docommand}
+% \c!before \c!after \c!inbetween \c!commands \c!style \c!color
-\unexpanded\def\definestartstop
- {\dodoubleargument\dodefinestartstop}
+\installcommandhandler{\??be}{startstop}{\??be}
-\def\dosetupstartstop[#1][#2]%
- {\def\docommand##1{\getparameters[\??be##1][#2]}%
- \processcommalist[#1]\docommand}
+\appendtoks
+ \normalexpanded{\dodefinestartstop{\currentstartstop}}%
+\to \everydefinestartstop
-\unexpanded\def\setupstartstop
- {\dodoubleargument\dosetupstartstop}
+\unexpanded\def\dodefinestartstop#1%
+ {\setuvalue{#1}%
+ {\groupedcommand
+ {\def\currentstartstop{#1}%
+ \startstopparameter\c!commands
+ \dostarttagged\t!construct\currentstartstop
+ \dosetstartstopattributes\c!style\c!color}
+ {\def\currentstartstop{#1}%
+ \dostoptagged
+ \startstopparameter\c!inbetween}}%
+ \setuvalue{\e!start#1}%
+ {\namedstartstopparameter{#1}\c!before
+ \bgroup
+ \def\currentstartstop{#1}%
+ \startstopparameter\c!commands
+ \dostarttagged\t!construct\currentstartstop
+ \dosetstartstopattributes\c!style\c!color}%
+ \setuvalue{\e!stop#1}%
+ {\dostoptagged
+ \egroup
+ \namedstartstopparameter{#1}\c!after}}
% \docommand kan niet worden gebruikt omdat deze macro
% soms lokaal wordt gebruikt
diff --git a/tex/context/base/data-env.lua b/tex/context/base/data-env.lua
index c6035d799..d2c38262d 100644
--- a/tex/context/base/data-env.lua
+++ b/tex/context/base/data-env.lua
@@ -18,7 +18,7 @@ local suffixes = allocate() resolvers.suffixes = suffixes
local dangerous = allocate() resolvers.dangerous = dangerous
local suffixmap = allocate() resolvers.suffixmap = suffixmap
-local relations = allocate {
+local relations = allocate { -- todo: handlers also here
core = {
ofm = {
names = { "ofm", "omega font metric", "omega font metrics" },
@@ -88,7 +88,7 @@ local relations = allocate {
tex = {
names = { "tex" },
variable = 'TEXINPUTS',
- suffixes = { 'tex', "mkiv", "mkii" },
+ suffixes = { 'tex', "mkiv", "mkiv", "mkii" },
},
icc = {
names = { "icc", "icc profile", "icc profiles" },
@@ -202,29 +202,33 @@ resolvers.relations = relations
-- formats: maps a format onto a variable
-for category, categories in next, relations do
- for name, relation in next, categories do
- local rn = relation.names
- local rv = relation.variable
- local rs = relation.suffixes
- if rn and rv then
- for i=1,#rn do
- local rni = lower(gsub(rn[i]," ",""))
- formats[rni] = rv
- if rs then
- suffixes[rni] = rs
- for i=1,#rs do
- local rsi = rs[i]
- suffixmap[rsi] = rni
+function resolvers.updaterelations()
+ for category, categories in next, relations do
+ for name, relation in next, categories do
+ local rn = relation.names
+ local rv = relation.variable
+ local rs = relation.suffixes
+ if rn and rv then
+ for i=1,#rn do
+ local rni = lower(gsub(rn[i]," ",""))
+ formats[rni] = rv
+ if rs then
+ suffixes[rni] = rs
+ for i=1,#rs do
+ local rsi = rs[i]
+ suffixmap[rsi] = rni
+ end
end
end
end
- end
- if rs then
+ if rs then
+ end
end
end
end
+resolvers.updaterelations() -- push this in the metatable -> newindex
+
local function simplified(t,k)
return rawget(t,lower(gsub(k," ","")))
end
diff --git a/tex/context/base/data-met.lua b/tex/context/base/data-met.lua
index b51c8b57b..c9c3cde5e 100644
--- a/tex/context/base/data-met.lua
+++ b/tex/context/base/data-met.lua
@@ -21,7 +21,7 @@ resolvers.locators = allocate { notfound = { nil } } -- locate databases
resolvers.hashers = allocate { notfound = { nil } } -- load databases
resolvers.generators = allocate { notfound = { nil } } -- generate databases
-function resolvers.splitmethod(filename)
+function resolvers.splitmethod(filename) -- todo: trigger by suffix
if not filename then
return { } -- safeguard
elseif type(filename) == "table" then
@@ -40,10 +40,13 @@ function resolvers.methodhandler(what, filename, filetype) -- ...
local resolver = resolvers[what]
if resolver[scheme] then
if trace_locating then
- report_resolvers("handler '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ report_resolvers("using special handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
end
return resolver[scheme](filename,filetype)
else
+ if trace_locating then
+ report_resolvers("no handler for '%s' -> '%s' -> '%s'",specification.original,what,table.sequenced(specification))
+ end
return resolver.tex(filename,filetype) -- todo: specification
end
end
diff --git a/tex/context/base/data-sch.lua b/tex/context/base/data-sch.lua
index c8d209798..655cd8209 100644
--- a/tex/context/base/data-sch.lua
+++ b/tex/context/base/data-sch.lua
@@ -23,7 +23,7 @@ schemes.threshold = 24 * 60 * 60
directives.register("schemes.threshold", function(v) schemes.threshold = tonumber(v) or schemes.threshold end)
-local cached, loaded, reused = { }, { }, { }
+local cached, loaded, reused, thresholds = { }, { }, { }, { }
function schemes.curl(name,cachename) -- will use sockets instead or the curl library
local command = "curl --silent --create-dirs --output " .. cachename .. " " .. name -- no protocol .. "://"
@@ -35,7 +35,8 @@ function schemes.fetch(protocol,name,handler)
local cachename = caches.setfirstwritablefile(cleanname,"schemes")
if not cached[name] then
statistics.starttiming(schemes)
- if not io.exists(cachename) or (os.difftime(os.time(),lfs.attributes(cachename).modification) > schemes.threshold) then
+ if not io.exists(cachename) or (os.difftime(os.time(),lfs.attributes(cachename).modification) >
+ (thresholds[protocol] or schemes.threshold)) then
cached[name] = cachename
if handler then
if trace_schemes then
@@ -86,14 +87,15 @@ function loaders.schemes(protocol,filename)
return loaders.generic(protocol,filename)
end
--- could be metatable
+-- could be metatable and proper subtables
-function schemes.install(protocol,handler)
- loaded [protocol] = 0
- reused [protocol] = 0
- finders[protocol] = function (filename,filetype) return finders.schemes(protocol,filename,handler) end
- openers[protocol] = function (filename) return openers.schemes(protocol,filename) end
- loaders[protocol] = function (filename) return loaders.schemes(protocol,filename) end
+function schemes.install(protocol,handler,threshold)
+ loaded [protocol] = 0
+ reused [protocol] = 0
+ finders [protocol] = function (filename,filetype) return finders.schemes(protocol,filename,handler) end
+ openers [protocol] = function (filename) return openers.schemes(protocol,filename) end
+ loaders [protocol] = function (filename) return loaders.schemes(protocol,filename) end
+ thresholds[protocol] = threshold or schemes.threshold
end
local function http_handler(protocol,name,cachename)
@@ -109,6 +111,7 @@ local function http_handler(protocol,name,cachename)
os.remove(cachename)
os.rename(tempname,cachename)
end
+ return cachename
end
schemes.install('http',http_handler)
diff --git a/tex/context/base/data-tex.lua b/tex/context/base/data-tex.lua
index bd6eb0599..3c42882fd 100644
--- a/tex/context/base/data-tex.lua
+++ b/tex/context/base/data-tex.lua
@@ -36,125 +36,216 @@ function finders.generic(tag,filename,filetype)
end
end
---~ local lpegmatch = lpeg.match
---~ local getlines = lpeg.Ct(lpeg.patterns.textline)
+-- -- keep this one as reference as it's the first version
+--
+-- resolvers.filters = resolvers.filters or { }
+--
+-- local input_translator, utf_translator, user_translator = nil, nil, nil
+--
+-- function resolvers.filters.install(name,func)
+-- if name == "input" then input_translator = func
+-- elseif name == "utf" then utf_translator = func
+-- elseif name == "user" then user_translator = func end
+-- end
+--
+-- function openers.textopener(filename,file_handle,tag)
+-- local u = unicode.utftype(file_handle)
+-- local t = { }
+-- if u > 0 then
+-- if trace_locating then
+-- report_resolvers("%s opener, file '%s' opened using method '%s'",tag,filename,unicode.utfname[u])
+-- end
+-- local l
+-- local data = file_handle:read("*a")
+-- if u > 2 then
+-- l = unicode.utf32_to_utf8(data,u==4)
+-- elseif u > 1 then
+-- l = unicode.utf16_to_utf8(data,u==2)
+-- else
+-- l = string.splitlines(data)
+-- end
+-- file_handle:close()
+-- t = {
+-- utftype = u, -- may go away
+-- lines = l,
+-- current = 0, -- line number, not really needed
+-- handle = nil,
+-- noflines = #l,
+-- close = function()
+-- if trace_locating then
+-- report_resolvers("%s closer, file '%s' closed",tag,filename)
+-- end
+-- logs.show_close(filename)
+-- t = nil
+-- end,
+-- reader = function(self)
+-- self = self or t
+-- local current, lines = self.current, self.lines
+-- if current >= #lines then
+-- return nil
+-- else
+-- current = current + 1
+-- self.current = current
+-- local line = lines[current]
+-- if not line then
+-- return nil
+-- elseif line == "" then
+-- return ""
+-- else
+-- if input_translator then
+-- line = input_translator(line)
+-- end
+-- if utf_translator then
+-- line = utf_translator(line)
+-- end
+-- if user_translator then
+-- line = user_translator(line)
+-- end
+-- return line
+-- end
+-- end
+-- end
+-- }
+-- else
+-- if trace_locating then
+-- report_resolvers("%s opener, file '%s' opened",tag,filename)
+-- end
+-- -- todo: file;name -> freeze / eerste regel scannen -> freeze
+-- --~ local data = lpegmatch(getlines,file_handle:read("*a"))
+-- --~ local n = 0
+-- t = {
+-- reader = function() -- self
+-- local line = file_handle:read()
+-- --~ n = n + 1
+-- --~ local line = data[n]
+-- --~ print(line)
+-- if not line then
+-- return nil
+-- elseif line == "" then
+-- return ""
+-- else
+-- if input_translator then
+-- line = input_translator(line)
+-- end
+-- if utf_translator then
+-- line = utf_translator(line)
+-- end
+-- if user_translator then
+-- line = user_translator(line)
+-- end
+-- return line
+-- end
+-- end,
+-- close = function()
+-- if trace_locating then
+-- report_resolvers("%s closer, file '%s' closed",tag,filename)
+-- end
+-- logs.show_close(filename)
+-- file_handle:close()
+-- t = nil
+-- collectgarbage("step") -- saves some memory, maybe checkgarbage but no #
+-- end,
+-- handle = function()
+-- return file_handle
+-- end,
+-- noflines = function()
+-- t.noflines = io.noflines(file_handle)
+-- return t.noflines
+-- end
+-- }
+-- end
+-- return t
+-- end
-resolvers.filters = resolvers.filters or { }
-local input_translator, utf_translator, user_translator = nil, nil, nil
+-- the main text reader --
-function resolvers.filters.install(name,func)
- if name == "input" then input_translator = func
- elseif name == "utf" then utf_translator = func
- elseif name == "user" then user_translator = func end
-end
+local sequencers = utilities.sequencers
+
+local fileprocessor = nil
+local lineprocessor = nil
+
+local textfileactions = sequencers.reset {
+ arguments = "str,filename",
+ returnvalues = "str",
+ results = "str",
+}
+
+local textlineactions = sequencers.reset {
+ arguments = "str,filename,linenumber",
+ returnvalues = "str",
+ results = "str",
+}
+
+openers.textfileactions = textfileactions
+openers.textlineactions = textlineactions
+
+sequencers.appendgroup(textfileactions,"system")
+sequencers.appendgroup(textfileactions,"user")
+
+sequencers.appendgroup(textlineactions,"system")
+sequencers.appendgroup(textlineactions,"user")
function openers.textopener(filename,file_handle,tag)
- local u = unicode.utftype(file_handle)
- local t = { }
- if u > 0 then
- if trace_locating then
- report_resolvers("%s opener, file '%s' opened using method '%s'",tag,filename,unicode.utfname[u])
- end
- local l
- if u > 2 then
- l = unicode.utf32_to_utf8(file_handle:read("*a"),u==4)
- else
- l = unicode.utf16_to_utf8(file_handle:read("*a"),u==2)
- end
- file_handle:close()
- t = {
- utftype = u, -- may go away
- lines = l,
- current = 0, -- line number, not really needed
- handle = nil,
- noflines = #l,
- close = function()
- if trace_locating then
- report_resolvers("%s closer, file '%s' closed",tag,filename)
- end
- logs.show_close(filename)
- t = nil
- end,
- reader = function(self)
- self = self or t
- local current, lines = self.current, self.lines
- if current >= #lines then
- return nil
- else
- current = current + 1
- self.current = current
- local line = lines[current]
- if not line then
- return nil
- elseif line == "" then
- return ""
- else
- if input_translator then
- line = input_translator(line)
- end
- if utf_translator then
- line = utf_translator(line)
- end
- if user_translator then
- line = user_translator(line)
- end
- return line
- end
- end
+ if trace_locating then
+ report_resolvers("%s opener, file '%s' opened using method '%s'",tag,filename,unicode.utfname[u])
+ end
+ if textfileactions.dirty then
+ fileprocessor = sequencers.compile(textfileactions)
+ end
+ local lines = io.loaddata(filename)
+ local kind = unicode.filetype(lines)
+ if kind == "utf-16-be" then
+ lines = unicode.utf16_to_utf8_be(lines)
+ elseif kind == "utf-16-le" then
+ lines = unicode.utf16_to_utf8_le(lines)
+ elseif kind == "utf-32-be" then
+ lines = unicode.utf32_to_utf8_be(lines)
+ elseif kind == "utf-32-le" then
+ lines = unicode.utf32_to_utf8_le(lines)
+ else -- utf8 or unknown
+ lines = fileprocessor(lines,filename) or lines
+ lines = string.splitlines(lines)
+ end
+ local t = {
+ lines = lines,
+ current = 0,
+ handle = nil,
+ noflines = #lines,
+ close = function()
+ if trace_locating then
+ report_resolvers("%s closer, file '%s' closed",tag,filename)
end
- }
- else
- if trace_locating then
- report_resolvers("%s opener, file '%s' opened",tag,filename)
- end
- -- todo: file;name -> freeze / eerste regel scannen -> freeze
- --~ local data = lpegmatch(getlines,file_handle:read("*a"))
- --~ local n = 0
- t = {
- reader = function() -- self
- local line = file_handle:read()
- --~ n = n + 1
- --~ local line = data[n]
- --~ print(line)
+ logs.show_close(filename)
+ t = nil
+ end,
+ reader = function(self)
+ self = self or t
+ local current, noflines = self.current, self.noflines
+ if current >= noflines then
+ return nil
+ else
+ current = current + 1
+ self.current = current
+ local line = lines[current]
if not line then
return nil
elseif line == "" then
return ""
else
- if input_translator then
- line = input_translator(line)
- end
- if utf_translator then
- line = utf_translator(line)
+ if textlineactions.dirty then
+ lineprocessor = sequencers.compile(textlineactions)
end
- if user_translator then
- line = user_translator(line)
- end
- return line
- end
- end,
- close = function()
- if trace_locating then
- report_resolvers("%s closer, file '%s' closed",tag,filename)
+ return lineprocessor(line,filename,current) or line
end
- logs.show_close(filename)
- file_handle:close()
- t = nil
- collectgarbage("step") -- saves some memory, maybe checkgarbage but no #
- end,
- handle = function()
- return file_handle
- end,
- noflines = function()
- t.noflines = io.noflines(file_handle)
- return t.noflines
end
- }
- end
+ end
+ }
return t
end
+-- -- --
+
function openers.generic(tag,filename)
if filename and filename ~= "" then
local f = io.open(filename,"r")
@@ -226,6 +317,7 @@ function resolvers.openfile(filename)
end
function resolvers.loadtexfile(filename, filetype)
+ -- todo: apply filters
local ok, data, size = resolvers.loadbinfile(filename, filetype)
return data or ""
end
diff --git a/tex/context/base/font-ini.mkiv b/tex/context/base/font-ini.mkiv
index b362b35a0..3a479497b 100644
--- a/tex/context/base/font-ini.mkiv
+++ b/tex/context/base/font-ini.mkiv
@@ -3206,12 +3206,37 @@
\getvalue\emphasizedtypeface
\fi\fi\fi}
-\def\emphasistypeface{\doemphasistypeface\sl\it}
-\def\emphasisboldface{\doemphasistypeface\bs\bi}
+\unexpanded\def\emphasistypeface{\doemphasistypeface\sl\it}
+\unexpanded\def\emphasisboldface{\doemphasistypeface\bs\bi}
+
+\unexpanded\def\normalboldface
+ {\relax\ifx\fontalternative\c!it \bi
+ \else \ifx\fontalternative\c!sl \bs
+ \else \bf \fi\fi}
+
+\unexpanded\def\normaltypeface
+ {\relax
+ \ifx\fontalternative\c!bi \it \else
+ \ifx\fontalternative\c!bs \sl \else
+ \tf \fi\fi}
+
+\let\typeface\normaltypeface
+\let\boldface\normalboldface
+
+\unexpanded\def\swaptypeface
+ {\relax
+ \ifx\fontalternative\c!it \tf \else
+ \ifx\fontalternative\c!sl \tf \else
+ \ifx\fontalternative\c!bf \emphasisboldface \else
+ \ifx\fontalternative\c!bs \bf \else
+ \ifx\fontalternative\c!bi \bf \else
+ \emphasistypeface \fi\fi\fi\fi\fi}
%D To be set with the default body font environment: \type
%D {em} being \type {slanted} or \type {italic}.
+% maybe a \csname...\fontalternative\endcsname
+
\newconditional\emneeded
\newtoks\everyemphasized
@@ -3238,8 +3263,7 @@
\emphasistypeface
\fi\fi\fi\fi\fi
\the\everyemphasized
- \ifconditional\emneeded\relax
- \else
+ \ifconditional\emneeded\else
\expandafter\aftergroup
\fi
\emphasiscorrection}
@@ -3250,7 +3274,7 @@
% \setupbodyfontenvironment [default] [em={\italic\color[red]}]
%D The next feature was not present in previous versions. It
-%D takes care of \type {\em \bf ...} sitiations.
+%D takes care of \type {\em \bf ...} situations.
\def\setemphasisboldface
{\let\savedemphasisboldface\bf
@@ -3274,10 +3298,10 @@
\let\italiccorrection=\/ % tex primitive
-\def\emphasiscorrection
- {\ifhmode
- \expandafter\emphasislook
- \fi}
+\def\emphasiscorrection % not in raw alignment groups, else omit problem
+ {\ifhmode\ifnum\currentgrouptype=\@@aligngroup\else
+ \expandafter\expandafter\expandafter\emphasislook
+ \fi\fi}
\def\emphasislook
{\begingroup
diff --git a/tex/context/base/font-otn.lua b/tex/context/base/font-otn.lua
index 41169618f..29fed2944 100644
--- a/tex/context/base/font-otn.lua
+++ b/tex/context/base/font-otn.lua
@@ -239,7 +239,7 @@ local contextsetups = specifiers.contextsetups
local contextnumbers = specifiers.contextnumbers
local contextmerged = specifiers.contextmerged
--- we cannot optimize with "start = first_character(head)" because then we don't
+-- we cannot optimize with "start = first_glyph(head)" because then we don't
-- know which rlmode we're in which messes up cursive handling later on
--
-- head is always a whatsit so we can safely assume that head is not changed
diff --git a/tex/context/base/font-syn.lua b/tex/context/base/font-syn.lua
index a0dbd4f7e..c6f7b0393 100644
--- a/tex/context/base/font-syn.lua
+++ b/tex/context/base/font-syn.lua
@@ -57,7 +57,7 @@ names.cache = containers.define("fonts","data",names.version,true)
<p>A few helpers.</p>
--ldx]]--
-local P, C, Cc, Cs, Carg = lpeg.P, lpeg.C, lpeg.Cc, lpeg.Cs, lpeg.Carg
+local P, C, Cc, Cs = lpeg.P, lpeg.C, lpeg.Cc, lpeg.Cs
-- what to do with 'thin'
diff --git a/tex/context/base/l-lpeg.lua b/tex/context/base/l-lpeg.lua
index ff49101eb..940652ee3 100644
--- a/tex/context/base/l-lpeg.lua
+++ b/tex/context/base/l-lpeg.lua
@@ -10,6 +10,9 @@ local lpeg = require("lpeg")
local type = type
+-- Beware, we predefine a bunch of patterns here and one reason for doing so
+-- is that we get consistent behaviour in some of the visualizers.
+
lpeg.patterns = lpeg.patterns or { } -- so that we can share
local patterns = lpeg.patterns
@@ -26,19 +29,38 @@ local alwaysmatched = P(true)
patterns.anything = anything
patterns.endofstring = endofstring
patterns.beginofstring = alwaysmatched
+patterns.alwaysmatched = alwaysmatched
local digit, sign = R('09'), S('+-')
local cr, lf, crlf = P("\r"), P("\n"), P("\r\n")
+local newline = crlf + cr + lf
local utf8next = R("\128\191")
local escaped = P("\\") * anything
local squote = P("'")
local dquote = P('"')
+local space = P(" ")
+
+patterns.somecontent = (anything - newline - space)^1
+patterns.beginline = #(1-newline)
+
+local utfbom_32_be = P('\000\000\254\255')
+local utfbom_32_le = P('\255\254\000\000')
+local utfbom_16_be = P('\255\254')
+local utfbom_16_le = P('\254\255')
+local utfbom_8 = P('\239\187\191')
+local utfbom = utfbom_32_be + utfbom_32_le
+ + utfbom_16_be + utfbom_16_le
+ + utfbom_8
+local utftype = utfbom_32_be / "utf-32-be" + utfbom_32_le / "utf-32-le"
+ + utfbom_16_be / "utf-16-be" + utfbom_16_le / "utf-16-le"
+ + utfbom_8 / "utf-8" + alwaysmatched / "unknown"
patterns.utf8one = R("\000\127")
patterns.utf8two = R("\194\223") * utf8next
patterns.utf8three = R("\224\239") * utf8next * utf8next
patterns.utf8four = R("\240\244") * utf8next * utf8next * utf8next
-patterns.utfbom = P('\000\000\254\255') + P('\255\254\000\000') + P('\255\254') + P('\254\255') + P('\239\187\191')
+patterns.utfbom = utfbom
+patterns.utftype = utftype
local utf8char = patterns.utf8one + patterns.utf8two + patterns.utf8three + patterns.utf8four
local validutf8char = utf8char^0 * endofstring * Cc(true) + Cc(false)
@@ -64,24 +86,30 @@ patterns.hexadecimal = P("0x") * R("09","AF","af")^1
patterns.lowercase = R("az")
patterns.uppercase = R("AZ")
patterns.letter = patterns.lowercase + patterns.uppercase
-patterns.space = P(" ")
+patterns.space = space
patterns.tab = P("\t")
patterns.spaceortab = patterns.space + patterns.tab
patterns.eol = S("\n\r")
patterns.spacer = S(" \t\f\v") -- + string.char(0xc2, 0xa0) if we want utf (cf mail roberto)
-patterns.newline = crlf + cr + lf
-patterns.nonspace = 1 - patterns.space
+patterns.newline = newline
+patterns.emptyline = newline^1
patterns.nonspacer = 1 - patterns.spacer
patterns.whitespace = patterns.eol + patterns.spacer
patterns.nonwhitespace = 1 - patterns.whitespace
+patterns.equal = P("=")
patterns.comma = P(",")
patterns.commaspacer = P(",") * patterns.spacer^0
patterns.period = P(".")
+patterns.colon = P(":")
+patterns.semicolon = P(";")
+patterns.underscore = P("_")
patterns.escaped = escaped
patterns.squote = squote
patterns.dquote = dquote
-patterns.undouble = (dquote/"") * ((escaped + (1-dquote))^0) * (dquote/"")
-patterns.unsingle = (squote/"") * ((escaped + (1-squote))^0) * (squote/"")
+patterns.nosquote = (escaped + (1-squote))^0
+patterns.nodquote = (escaped + (1-dquote))^0
+patterns.unsingle = (squote/"") * patterns.nosquote * (squote/"")
+patterns.undouble = (dquote/"") * patterns.nodquote * (dquote/"")
patterns.unquoted = patterns.undouble + patterns.unsingle -- more often undouble
patterns.unspacer = ((patterns.spacer^1)/"")^0
@@ -104,19 +132,6 @@ function lpeg.splitter(pattern, action)
return (((1-P(pattern))^1)/action+1)^0
end
-local spacing = patterns.spacer^0 * patterns.newline -- sort of strip
-local empty = spacing * Cc("")
-local nonempty = Cs((1-spacing)^1) * spacing^-1
-local content = (empty + nonempty)^1
-
-local capture = Ct(content^0)
-
-function string.splitlines(str)
- return match(capture,str)
-end
-
-patterns.textline = content
-
local splitters_s, splitters_m = { }, { }
local function splitat(separator,single)
@@ -163,6 +178,35 @@ function string.split(str,separator)
return match(c,str)
end
+local spacing = patterns.spacer^0 * newline -- sort of strip
+local empty = spacing * Cc("")
+local nonempty = Cs((1-spacing)^1) * spacing^-1
+local content = (empty + nonempty)^1
+
+patterns.textline = content
+
+--~ local linesplitter = Ct(content^0)
+--~
+--~ function string.splitlines(str)
+--~ return match(linesplitter,str)
+--~ end
+
+local linesplitter = Ct(splitat(newline))
+
+patterns.linesplitter = linesplitter
+
+function string.splitlines(str)
+ return match(linesplitter,str)
+end
+
+local utflinesplitter = utfbom^-1 * Ct(splitat(newline))
+
+patterns.utflinesplitter = utflinesplitter
+
+function string.utfsplitlines(str)
+ return match(utflinesplitter,str)
+end
+
--~ lpeg.splitters = cache -- no longer public
local cache = { }
diff --git a/tex/context/base/l-pdfview.lua b/tex/context/base/l-pdfview.lua
index c16df036c..85545cd10 100644
--- a/tex/context/base/l-pdfview.lua
+++ b/tex/context/base/l-pdfview.lua
@@ -26,7 +26,8 @@ local allcalls = {
}
if os.type == "windows" then
- opencalls['okular'] = 'start "test" "c:/program files/kde/bin/okular.exe" --unique' -- todo: get focus
+--~ opencalls['okular'] = 'start "test" "c:/program files/kde/bin/okular.exe" --unique' -- todo: get focus
+ opencalls['okular'] = 'start "test" "c:/data/system/kde/bin/okular.exe" --unique' -- todo: get focus
else
opencalls['okular'] = 'okular --unique'
end
diff --git a/tex/context/base/l-unicode.lua b/tex/context/base/l-unicode.lua
index a97f01d1e..445909d88 100644
--- a/tex/context/base/l-unicode.lua
+++ b/tex/context/base/l-unicode.lua
@@ -36,6 +36,8 @@ utf = utf or unicode.utf8
local concat, utfchar, utfgsub = table.concat, utf.char, utf.gsub
local char, byte, find, bytepairs, utfvalues, format = string.char, string.byte, string.find, string.bytepairs, string.utfvalues, string.format
+local utfsplitlines = string.utfsplitlines
+
-- 0 EF BB BF UTF-8
-- 1 FF FE UTF-16-little-endian
-- 2 FE FF UTF-16-big-endian
@@ -80,111 +82,234 @@ function unicode.utftype(f)
end
end
-function unicode.utf16_to_utf8(str, endian) -- maybe a gsub is faster or an lpeg
- local result, tmp, n, m, p, r, t = { }, { }, 0, 0, 0, 0, 0 -- we reuse tmp
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+--~ function unicode.utf16_to_utf8(str, endian) -- maybe a gsub is faster or an lpeg
+--~ local result, tmp, n, m, p, r, t = { }, { }, 0, 0, 0, 0, 0 -- we reuse tmp
+--~ -- lf | cr | crlf / (cr:13, lf:10)
+--~ local function doit() -- inline this
+--~ if n == 10 then
+--~ if p ~= 13 then
+--~ if t > 0 then
+--~ r = r + 1
+--~ result[r] = concat(tmp,"",1,t)
+--~ t = 0
+--~ end
+--~ p = 0
+--~ end
+--~ elseif n == 13 then
+--~ if t > 0 then
+--~ r = r + 1
+--~ result[r] = concat(tmp,"",1,t)
+--~ t = 0
+--~ end
+--~ p = n
+--~ else
+--~ t = t + 1
+--~ tmp[t] = utfchar(n)
+--~ p = 0
+--~ end
+--~ end
+--~ for l,r in bytepairs(str) do
+--~ if r then
+--~ if endian then -- maybe make two loops
+--~ n = 256*l + r
+--~ else
+--~ n = 256*r + l
+--~ end
+--~ if m > 0 then
+--~ n = (m-0xD800)*0x400 + (n-0xDC00) + 0x10000
+--~ m = 0
+--~ doit()
+--~ elseif n >= 0xD800 and n <= 0xDBFF then
+--~ m = n
+--~ else
+--~ doit()
+--~ end
+--~ end
+--~ end
+--~ if t > 0 then
+--~ r = r + 1
+--~ result[r] = concat(tmp,"",1,t) -- we reused tmp, hence t
+--~ end
+--~ return result
+--~ end
+
+--~ function unicode.utf32_to_utf8(str, endian)
+--~ local result, tmp, n, m, p, r, t = { }, { }, 0, -1, 0, 0, 0
+--~ -- lf | cr | crlf / (cr:13, lf:10)
+--~ local function doit() -- inline this
+--~ if n == 10 then
+--~ if p ~= 13 then
+--~ if t > 0 then
+--~ r = r + 1
+--~ result[r] = concat(tmp,"",1,t)
+--~ t = 0
+--~ end
+--~ p = 0
+--~ end
+--~ elseif n == 13 then
+--~ if t > 0 then
+--~ r = r + 1
+--~ result[r] = concat(tmp,"",1,t)
+--~ t = 0
+--~ end
+--~ p = n
+--~ else
+--~ t = t + 1
+--~ tmp[t] = utfchar(n)
+--~ p = 0
+--~ end
+--~ end
+--~ for a,b in bytepairs(str) do
+--~ if a and b then
+--~ if m < 0 then
+--~ if endian then -- maybe make two loops
+--~ m = 256*256*256*a + 256*256*b
+--~ else
+--~ m = 256*b + a
+--~ end
+--~ else
+--~ if endian then -- maybe make two loops
+--~ n = m + 256*a + b
+--~ else
+--~ n = m + 256*256*256*b + 256*256*a
+--~ end
+--~ m = -1
+--~ doit()
+--~ end
+--~ else
+--~ break
+--~ end
+--~ end
+--~ if #tmp > 0 then
+--~ r = r + 1
+--~ result[r] = concat(tmp,"",1,t) -- we reused tmp, hence t
+--~ end
+--~ return result
+--~ end
+
+local function utf16_to_utf8_be(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*left + right
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
+ r = r + 1
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[r] = utfchar(now)
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- for l,r in bytepairs(str) do
- if r then
- if endian then
- n = 256*l + r
- else
- n = 256*r + l
- end
- if m > 0 then
- n = (m-0xD800)*0x400 + (n-0xDC00) + 0x10000
- m = 0
- doit()
- elseif n >= 0xD800 and n <= 0xDBFF then
- m = n
- else
- doit()
+ return t
+end
+
+local function utf16_to_utf8_le(t)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, 0
+ for left, right in bytepairs(t[i]) do
+ if right then
+ local now = 256*right + left
+ if more > 0 then
+ now = (more-0xD800)*0x400 + (now-0xDC00) + 0x10000
+ more = 0
+ r = r + 1
+ result[r] = utfchar(now)
+ elseif now >= 0xD800 and now <= 0xDBFF then
+ more = now
+ else
+ r = r + 1
+ result[r] = utfchar(now)
+ end
end
end
+ t[i] = concat(result,"",1,r) -- we reused tmp, hence t
end
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- end
- return result
+ return t
end
-function unicode.utf32_to_utf8(str, endian)
- local result, tmp, n, m, p, r, t = { }, { }, 0, -1, 0, 0, 0
- -- lf | cr | crlf / (cr:13, lf:10)
- local function doit()
- if n == 10 then
- if p ~= 13 then
- if t > 0 then
+local function utf32_to_utf8_be(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
+ end
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*256*256*a + 256*256*b
+ else
r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ result[t] = utfchar(more + 256*a + b)
+ more = -1
end
- p = 0
- end
- elseif n == 13 then
- if t > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
- t = 0
+ else
+ break
end
- p = n
- else
- t = t + 1
- tmp[t] = utfchar(n)
- p = 0
end
+ t[i] = concat(result,"",1,r)
+ end
+ return result
+end
+
+local function utf32_to_utf8_le(str)
+ if type(t) == "string" then
+ t = utfsplitlines(str)
end
- for a,b in bytepairs(str) do
- if a and b then
- if m < 0 then
- if endian then
- m = 256*256*256*a + 256*256*b
+ local result = { } -- we reuse result
+ for i=1,#t do
+ local r, more = 0, -1
+ for a,b in bytepairs(str) do
+ if a and b then
+ if more < 0 then
+ more = 256*b + a
else
- m = 256*b + a
+ r = r + 1
+ result[t] = utfchar(more + 256*256*256*b + 256*256*a)
+ more = -1
end
else
- if endian then
- n = m + 256*a + b
- else
- n = m + 256*256*256*b + 256*256*a
- end
- m = -1
- doit()
+ break
end
- else
- break
end
- end
- if #tmp > 0 then
- r = r + 1
- result[r] = concat(tmp,"",1,t)
+ t[i] = concat(result,"",1,r)
end
return result
end
+unicode.utf32_to_utf8_be = utf32_to_utf8_be
+unicode.utf32_to_utf8_le = utf32_to_utf8_le
+unicode.utf16_to_utf8_be = utf16_to_utf8_be
+unicode.utf16_to_utf8_le = utf16_to_utf8_le
+
+function unicode.utf8_to_utf8(t)
+ return type(t) == "string" and utfsplitlines(t) or t
+end
+
+function unicode.utf16_to_utf8(t,endian)
+ return endian and utf16_to_utf8_be(t) or utf16_to_utf8_le(t) or t
+end
+
+function unicode.utf32_to_utf8(t,endian)
+ return endian and utf32_to_utf8_be(t) or utf32_to_utf8_le(t) or t
+end
+
local function little(c)
local b = byte(c)
if b < 0x10000 then
@@ -225,3 +350,7 @@ function unicode.utfcodes(str)
end
--~ print(unicode.utfcodes(str))
+
+function unicode.filetype(data)
+ return data and lpeg.match(lpeg.patterns.utftype,data) or "unknown"
+end
diff --git a/tex/context/base/lang-ini.lua b/tex/context/base/lang-ini.lua
index 507d59a9a..8744bff95 100644
--- a/tex/context/base/lang-ini.lua
+++ b/tex/context/base/lang-ini.lua
@@ -16,6 +16,7 @@ if not modules then modules = { } end modules ['lang-ini'] = {
--~ lang:hyphenation(string) string = lang:hyphenation() lang:clear_hyphenation()
+local type, tonumber = type, tonumber
local utf = unicode.utf8
local utfbyte = utf.byte
local format, gsub = string.format, string.gsub
@@ -160,6 +161,10 @@ function languages.installed(separator)
context(concat(table.sortedkeys(registered),separator or ","))
end
+function languages.current(n)
+ return numbers[n and tonumber(n) or tex.language]
+end
+
function languages.associate(tag,script,language) -- not yet used
associated[tag] = { script, language }
end
diff --git a/tex/context/base/lang-ini.mkiv b/tex/context/base/lang-ini.mkiv
index 0a0cbf016..a7371e298 100644
--- a/tex/context/base/lang-ini.mkiv
+++ b/tex/context/base/lang-ini.mkiv
@@ -13,6 +13,8 @@
%D This module needs a further cleanup (real split between ii/iv).
+% \ctxlua{tex.sprint(languages.numbers[tex.count.mainlanguagenumber])}
+
%D This module implements the (for the moment still simple)
%D multi||language support of \CONTEXT, which should not be
%D confused with the multi||lingual interface. This support
diff --git a/tex/context/base/luat-lib.mkiv b/tex/context/base/luat-lib.mkiv
index 2f73fae3d..b02a88b51 100644
--- a/tex/context/base/luat-lib.mkiv
+++ b/tex/context/base/luat-lib.mkiv
@@ -62,6 +62,7 @@
\registerctxluafile{luat-exe}{1.001}
\registerctxluafile{luat-iop}{1.001}
\registerctxluafile{luat-bwc}{1.001}
+\registerctxluafile{luat-mac}{1.001}
\registerctxluafile{lxml-tab}{1.001}
\registerctxluafile{lxml-lpt}{1.001}
diff --git a/tex/context/base/luat-mac.lua b/tex/context/base/luat-mac.lua
new file mode 100644
index 000000000..f4f7779da
--- /dev/null
+++ b/tex/context/base/luat-mac.lua
@@ -0,0 +1,162 @@
+if not modules then modules = { } end modules ['luat-mac'] = {
+ version = 1.001,
+ comment = "companion to luat-lib.mkiv",
+ author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
+ copyright = "PRAGMA ADE / ConTeXt Development Team",
+ license = "see context related readme files"
+}
+
+local P, V, S, R, C, Cs = lpeg.P, lpeg.V, lpeg.S, lpeg.R, lpeg.C, lpeg.Cs
+local lpegmatch, patterns = lpeg.match, lpeg.patterns
+
+local insert, remove = table.insert, table.remove
+local rep = string.rep
+local setmetatable = setmetatable
+
+local report_macros = logs.new("macros")
+
+local stack, top, n, hashes = { }, nil, 0, { }
+
+local function set(s)
+ if top then
+ n = n + 1
+ if n > 9 then
+ report_macros("number of arguments > 9, ignoring %s",s)
+ else
+ local ns = #stack
+ local h = hashes[ns]
+ if not h then
+ h = rep("#",ns)
+ hashes[ns] = h
+ end
+ m = h .. n
+ top[s] = m
+ return m
+ end
+ end
+end
+
+local function get(s)
+ local m = top and top[s] or s
+ return m
+end
+
+local function push()
+ top = { }
+ n = 0
+ local s = stack[#stack]
+ if s then
+ setmetatable(top,{ __index = s })
+ end
+ insert(stack,top)
+end
+
+local function pop()
+ top = remove(stack)
+end
+
+local leftbrace = P("{")
+local rightbrace = P("}")
+local escape = P("\\")
+
+local space = patterns.space
+local spaces = space^1
+local newline = patterns.newline
+local nobrace = 1 - leftbrace - rightbrace
+
+local name = R("AZ","az")^1
+local variable = P("#") * name
+local escapedname = escape * name
+local definer = escape * (P("def") + P("egdx") * P("def"))
+local startcode = P("\\starttexdefinition")
+local stopcode = P("\\stoptexdefinition")
+local anything = patterns.anything
+local always = patterns.alwaysmatched
+
+local pushlocal = always / push
+local poplocal = always / pop
+local declaration = variable / set
+local identifier = variable / get
+
+local grammar = { "converter",
+ texcode = pushlocal
+ * startcode
+ * spaces
+ * name
+ * spaces
+ * (declaration + (1 - newline - space))^0
+ * V("texbody")
+ * stopcode
+ * poplocal,
+ texbody = ( V("definition")
+ + V("braced")
+ + identifier
+ + (1 - stopcode)
+ )^0,
+ definition = pushlocal
+ * definer
+ * escapedname
+ * (declaration + (1-leftbrace))^0
+ * V("braced")
+ * poplocal,
+ braced = leftbrace
+ * ( V("definition")
+ + V("texcode")
+ + V("braced")
+ + identifier
+ + nobrace
+ )^0
+ * rightbrace,
+
+ pattern = V("definition") + V("texcode") + anything,
+
+ converter = V("pattern")^1,
+}
+
+local parser = Cs(grammar)
+
+local checker = P("%") * (1 - newline - P("macros"))^0
+ * P("macros") * space^0 * P("=") * space^0 * C(patterns.letter^1)
+
+-- maybe namespace
+
+local macros = { } resolvers.macros = macros
+
+function macros.preprocessed(data)
+ return lpegmatch(parser,data)
+end
+
+function macros.convertfile(oldname,newname)
+ local data = resolvers.loadtexfile(oldname)
+ data = interfaces.preprocessed(data) or ""
+ io.savedata(newname,data)
+end
+
+function macros.version(data)
+ return lpegmatch(checker,data)
+end
+
+local function handler(protocol,name,cachename)
+ local hashed = url.hashed(name)
+ local path = hashed.path
+ if path and path ~= "" then
+ local data = resolvers.loadtexfile(path)
+ data = lpegmatch(parser,data) or ""
+ io.savedata(cachename,data)
+ end
+ return cachename
+end
+
+resolvers.schemes.install('mkvi',handler,1) -- this will cache !
+
+function macros.processmkvi(str,filename)
+ if file.suffix(filename) == "mkvi" or lpegmatch(checker,str) == "mkvi" then
+ return lpegmatch(parser,str) or str
+ else
+ return str
+ end
+end
+
+utilities.sequencers.appendaction(resolvers.openers.textfileactions,"system","resolvers.macros.processmkvi")
+-- utilities.sequencers.disableaction(resolvers.openers.textfileactions,"resolvers.macros.processmkvi")
+
diff --git a/tex/context/base/lxml-dir.lua b/tex/context/base/lxml-dir.lua
index bcb846bba..ecf7e3eb7 100644
--- a/tex/context/base/lxml-dir.lua
+++ b/tex/context/base/lxml-dir.lua
@@ -41,7 +41,7 @@ local function load_setup(filename)
if fullname ~= "" then
filename = fullname
end
- local collection = xmlparseapply({ getid(xml.load(filename)) },"directive")
+ local collection = xml.applylpath({ getid(xml.load(filename)) },"directive") -- is { } needed ?
if collection then
local valid = 0
for i=1,#collection do
diff --git a/tex/context/base/math-ini.lua b/tex/context/base/math-ini.lua
index b4716d06a..d84e30189 100644
--- a/tex/context/base/math-ini.lua
+++ b/tex/context/base/math-ini.lua
@@ -100,8 +100,14 @@ end
local function mathtopaccent(class,family,slot)
return format('\\Umathaccent "%X "%X "%X ',0,family,slot) -- no class
end
-local function mathbotaccent(class,family,slot)
- return format('\\Umathbotaccent "%X "%X "%X ',0,family,slot) -- no class
+if tex.luatexversion < 65 then -- this will disappear at 0.70
+ local function mathbotaccent(class,family,slot)
+ return format('\\Umathbotaccent "%X "%X "%X ',0,family,slot) -- no class
+ end
+else
+ local function mathbotaccent(class,family,slot)
+ return format('\\Umathaccent bottom "%X "%X "%X ',0,family,slot) -- no class
+ end
end
local function mathtopdelimiter(class,family,slot)
return format('\\Udelimiterover "%X "%X ',family,slot) -- no class
diff --git a/tex/context/base/math-ini.mkiv b/tex/context/base/math-ini.mkiv
index e057e1cc6..bab8792f7 100644
--- a/tex/context/base/math-ini.mkiv
+++ b/tex/context/base/math-ini.mkiv
@@ -681,4 +681,8 @@
\def\mathoptext#1{\mathop{\text{#1}}}
+% for a while:
+
+\def\Umathbotaccent{\Umathaccent bottom }
+
\protect \endinput
diff --git a/tex/context/base/mlib-pdf.lua b/tex/context/base/mlib-pdf.lua
index 5ca554292..80ab9b1d0 100644
--- a/tex/context/base/mlib-pdf.lua
+++ b/tex/context/base/mlib-pdf.lua
@@ -315,7 +315,7 @@ function metapost.flush(result,flusher,askedfig) -- pdf flusher, table en dan co
postscript = object.postscript,
}
--
- local before, inbetween, after = nil, nil, nil
+ local before, inbetween, after, grouped = nil, nil, nil, false
--
local cs, cr = currentobject.color, nil
-- todo document why ...
@@ -329,7 +329,7 @@ function metapost.flush(result,flusher,askedfig) -- pdf flusher, table en dan co
-- move test to function
local special = metapost.specials[prescript]
if special then
- currentobject, before, inbetween, after = special(currentobject.postscript,currentobject,t,flusher)
+ currentobject, before, inbetween, after, grouped = special(currentobject.postscript,currentobject,t,flusher)
objecttype = currentobject.type
end
end
@@ -339,7 +339,9 @@ function metapost.flush(result,flusher,askedfig) -- pdf flusher, table en dan co
t[#t+1], cr = colorconverter(cs)
end
--
- if before then currentobject, t = before() end
+ if before then
+ currentobject, t = before()
+ end
local ml = currentobject.miterlimit
if ml and ml ~= miterlimit then
miterlimit = ml
@@ -426,7 +428,13 @@ function metapost.flush(result,flusher,askedfig) -- pdf flusher, table en dan co
if cr then
t[#t+1] = cr
end
- if after then currentobject, t = after() end
+ if after then
+ currentobject, t = after()
+ end
+ if grouped then
+ -- can be qQ'd so changes can end up in groups
+ miterlimit, linecap, linejoin, dashed = -1, -1, -1, false
+ end
end
end
end
diff --git a/tex/context/base/mlib-pps.lua b/tex/context/base/mlib-pps.lua
index 6fae400fc..396853d3e 100644
--- a/tex/context/base/mlib-pps.lua
+++ b/tex/context/base/mlib-pps.lua
@@ -245,7 +245,7 @@ function specials.fg(specification,object,result,flusher) -- graphics
object.path = nil
return object, { }
end
- return { } , before, nil, nil -- replace { } by object for tracing
+ return { } , before, nil, nil, true -- replace { } by object for tracing
end
function specials.ps(specification,object,result) -- positions
@@ -257,7 +257,7 @@ function specials.ps(specification,object,result) -- positions
x = x - metapost.llx
y = metapost.ury - y
context.MPLIBpositionwhd(label,x,y,w,h)
- return { }, nil, nil, nil
+ return { }, nil, nil, nil, true
end
local nofshades = 0 -- todo: hash resources, start at 1000 in order not to clash with older
@@ -347,7 +347,7 @@ local function resources(object,name,flusher,result)
return object, result
end
object.color, object.type = nil, nil
- return object, before, nil, after
+ return object, before, nil, after, true
end
-- todo: we need a way to move/scale
@@ -361,7 +361,7 @@ function specials.cs(specification,object,result,flusher) -- spot colors?
local coordinates = { tonumber(t[5]), tonumber(t[6]), tonumber(t[7]), tonumber(t[9]), tonumber(t[10]), tonumber(t[11]) }
local ca, cb, colorspace, name = checkandconvert(ca,cb)
lpdf.circularshade(name,domain,ca,cb,1,colorspace,coordinates) -- backend specific (will be renamed)
- return resources(object,name,flusher,result) -- object, before, nil, after
+ return resources(object,name,flusher,result) -- object, before, nil, after, grouped
end
function specials.ls(specification,object,result,flusher)
@@ -373,7 +373,7 @@ function specials.ls(specification,object,result,flusher)
local coordinates = { tonumber(t[5]), tonumber(t[6]), tonumber(t[8]), tonumber(t[9]) }
local ca, cb, colorspace, name = checkandconvert(ca,cb)
lpdf.linearshade(name,domain,ca,cb,1,colorspace,coordinates) -- backend specific (will be renamed)
- return resources(object,name,flusher,result) -- object, before, nil, after
+ return resources(object,name,flusher,result) -- object, before, nil, after, grouped
end
-- no need for a before here
@@ -428,7 +428,7 @@ function specials.tf(specification,object)
context.MPLIBsettext(n,str)
metapost.multipass = true
end
- return { }, nil, nil, nil
+ return { }, nil, nil, nil, true
end
local factor = 65536*(7227/7200)
@@ -473,9 +473,9 @@ function specials.ts(specification,object,result,flusher)
result = { "Q" }
return object, result
end
- return { }, before, nil, nil -- replace { } by object for tracing
+ return { }, before, nil, nil, true -- replace { } by object for tracing
else
- return { }, nil, nil, nil -- replace { } by object for tracing
+ return { }, nil, nil, nil, true -- replace { } by object for tracing
end
end
@@ -804,7 +804,7 @@ function specials.gt(specification,object) -- number, so that we can reorder
graphics[#graphics+1] = format("\\MPLIBgraphictext{%s}",specification)
metapost.intermediate.needed = true
metapost.multipass = true
- return { }, nil, nil, nil
+ return { }, nil, nil, nil, true
end
function metapost.intermediate.actions.makempy()
diff --git a/tex/context/base/mult-aux.mkiv b/tex/context/base/mult-aux.mkiv
index 628914b6d..c86c522cb 100644
--- a/tex/context/base/mult-aux.mkiv
+++ b/tex/context/base/mult-aux.mkiv
@@ -57,12 +57,14 @@
% todo: \def\detokenized...parameter#1{\detokenize\expandafter\expandafter\expandafter{\csname#1#2\endcsname}} % always root
-\unexpanded\def\doinstallparameterhandler#1#2#3#4#5#6#7%
+\unexpanded\def\doinstallparameterhandler#1#2#3#4#5#6#7#8#9%
{\def#3##1{\csname#4{#1#2}{##1}\endcsname}%
\def#4##1##2{\ifcsname##1##2\endcsname##1##2\else\expandafter#5\csname##1\s!parent\endcsname{##2}\fi}%
\def#5##1##2{\ifx##1\relax\s!empty\else#4{##1}{##2}\fi}%
\def#6##1##2{\csname#4{#1##1}{##2}\endcsname}%
- \def#7##1{\detokenize\expandafter\expandafter\expandafter{\csname#1##1\endcsname}}} % always root
+ \def#7##1{\detokenize\expandafter\expandafter\expandafter{\csname#1##1\endcsname}}% always root
+ \def#8{\dosetvalue{#1}}% ##1 {##2} (braces are mandate)
+ \def#9{\doletvalue{#1}}}% ##1 ##2
\unexpanded\def\installparameterhandler#1#2%
{%\message{\detokenize{#1}/\detokenize{#2}}%
@@ -74,7 +76,9 @@
\expandafter\noexpand\csname do#2parameter\endcsname
\expandafter\noexpand\csname do#2parentparameter\endcsname
\expandafter\noexpand\csname named#2parameter\endcsname
- \expandafter\noexpand\csname detokenized#2parameter\endcsname}}
+ \expandafter\noexpand\csname detokenized#2parameter\endcsname
+ \expandafter\noexpand\csname doset#2parameter\endcsname
+ \expandafter\noexpand\csname dolet#2parameter\endcsname}}
\unexpanded\def\doinstallparameterhashhandler#1#2#3#4#5%
{\def#3##1{#4{#1#2}{##1}}%
@@ -104,18 +108,38 @@
\expandafter\noexpand\csname doset#2attributes\endcsname
\expandafter\noexpand\csname #2parameterhash\endcsname}}
+% \unexpanded\def\doinstalldefinehandler#1#2#3#4#5#6#7%
+% {\unexpanded\def#2{\dotripleempty#5}%
+% \newtoks#6%
+% \newtoks#7%
+% \def#5[##1][##2][##3]% [child][parent][settings]
+% {\edef#4{##1}% [child][settings]
+% \the#6% predefine [child]
+% \ifsecondargument
+% \doifassignmentelse{##2}
+% {\getparameters[#1#4][\s!parent=#3,##2]}
+% {\getparameters[#1#4][\s!parent=#1##2,##3]}%
+% \else
+% \getparameters[#1#4][\s!parent=#3]%
+% \fi
+% \the#7}}
+
\unexpanded\def\doinstalldefinehandler#1#2#3#4#5#6#7%
{\unexpanded\def#2{\dotripleempty#5}%
\newtoks#6%
\newtoks#7%
- \def#5[##1][##2][##3]%
- {\edef#4{##1}%
- \the#6% predefine
- \ifsecondargument
+ \def#5[##1][##2][##3]% [child][parent][settings]
+ {\edef#4{##1}% % [child] [settings]
+ \the#6% predefine % [child][parent]
+ \ifthirdargument % [child]
+ \getparameters[#1#4][\s!parent=#1##2,##3]%
+ \else\ifsecondargument
\doifassignmentelse{##2}
{\getparameters[#1#4][\s!parent=#3,##2]}
- {\getparameters[#1#4][\s!parent=#1##2,##3]}%
- \fi
+ {\getparameters[#1#4][\s!parent=#1##2]}%
+ \else
+ \getparameters[#1#4][\s!parent=#3]%
+ \fi\fi
\the#7}}
\unexpanded\def\installdefinehandler#1#2#3%
@@ -125,14 +149,14 @@
\expandafter\noexpand\csname define#2\endcsname
{\noexpand#3}% root
\expandafter\noexpand\csname current#2\endcsname
- \expandafter\noexpand\csname dodefine#2\endcsname
+ \expandafter\noexpand\csname d@define#2\endcsname
\expandafter\noexpand\csname everypreset#2\endcsname
\expandafter\noexpand\csname everydefine#2\endcsname}}
\unexpanded\def\doinstallsetuphandler#1#2#3#4#5%
{\unexpanded\def#2{\dodoubleempty#4}%
\newtoks#5%
- \def#4[##1][##2]%
+ \def#4[##1][##2]% maybe helper
{\ifsecondargument
\def\docommand####1% we will have a simple one as well
{\edef#3{####1}%
@@ -151,7 +175,7 @@
{\noexpand#1}% \??aa
\expandafter\noexpand\csname setup#2\endcsname
\expandafter\noexpand\csname current#2\endcsname
- \expandafter\noexpand\csname dosetup#2\endcsname
+ \expandafter\noexpand\csname d@setup#2\endcsname
\expandafter\noexpand\csname everysetup#2\endcsname}}
\unexpanded\def\installcommandhandler#1#2#3% \??self name \??parent (can be \??self)
diff --git a/tex/context/base/mult-cld.lua b/tex/context/base/mult-cld.lua
index 0542c5fcf..1154eefe3 100644
--- a/tex/context/base/mult-cld.lua
+++ b/tex/context/base/mult-cld.lua
@@ -19,6 +19,8 @@ if not modules then modules = { } end modules ['mult-cld'] = {
-- tflush needs checking ... sort of weird that it's not a table
+-- __flushlines is an experiment and rather ugly so it will go away
+
context = context or { }
local context = context
@@ -130,40 +132,41 @@ local emptyline = space^0 * newline^2
local endofline = space^0 * newline * space^0
local simpleline = endofline * lpeg.P(-1)
-function lpeg.texlinesplitter(f_content,f_endofline,f_emptyline,f_simpleline)
- local splitlines =
- simpleline / (f_simpleline or f_endofline)
- + (
- emptyline / f_emptyline
- + endofline / f_endofline
- + content / f_content
- )^0
- return function(str) return lpegmatch(splitlines,str) end
-end
-
-local function f_content(s)
+local function n_content(s)
flush(contentcatcodes,s)
end
-local function f_endofline()
+local function n_endofline()
texsprint(" ")
end
-local function f_emptyline()
+local function n_emptyline()
texprint("")
end
-local function f_simpleline()
+local function n_simpleline()
texprint("")
end
-local flushlines = lpeg.texlinesplitter(f_content,f_endofline,f_emptyline,f_simpleline)
+function lpeg.texlinesplitter(f_content,f_endofline,f_emptyline,f_simpleline)
+ local splitlines =
+ simpleline / (f_simpleline or n_simpleline)
+ + (
+ emptyline / (f_emptyline or n_emptyline)
+ + endofline / (f_endofline or n_emptyline)
+ + content / (f_content or n_content)
+ )^0
+ return function(str) return lpegmatch(splitlines,str) end
+end
-context.flushlines = flushlines -- maybe context.helpers.flushtexlines
+local flushlines = lpeg.texlinesplitter(n_content,n_endofline,n_emptyline,n_simpleline)
+
+context.__flushlines = flushlines -- maybe context.helpers.flushtexlines
+context.__flush = flush
-- -- --
-local function writer(command,first,...)
+local function writer(parent,command,first,...)
local t = { first, ... }
flush(currentcatcodes,command) -- todo: ctx|prt|texcatcodes
local direct = false
@@ -184,6 +187,7 @@ local function writer(command,first,...)
elseif typ == "string" then
if processlines and find(ti,"\n") then -- we can check for ti == "\n"
flush(currentcatcodes,"{")
+ local flushlines = parent.__flushlines or flushlines
flushlines(ti)
flush(currentcatcodes,"}")
elseif currentcatcodes == contentcatcodes then
@@ -250,28 +254,25 @@ local function writer(command,first,...)
trace_context("error: '%s' gets a weird argument '%s'",command,tostring(ti))
end
end
- if direct then
- trace_context("error: direct flushing used in '%s' without following argument",command)
- end
end
local generics = { } context.generics = generics
-local function indexer(t,k)
+local function indexer(parent,k)
local c = "\\" .. (generics[k] or k)
local f = function(first,...)
if first == nil then
flush(currentcatcodes,c)
else
- return writer(c,first,...)
+ return writer(parent,c,first,...)
end
end
- t[k] = f
+ parent[k] = f
return f
end
-local function caller(t,f,a,...)
- if not t then
+local function caller(parent,f,a,...)
+ if not parent then
-- so we don't need to test in the calling (slower but often no issue) (will go)
elseif f ~= nil then
local typ = type(f)
@@ -279,6 +280,7 @@ local function caller(t,f,a,...)
if a then
flush(contentcatcodes,format(f,a,...)) -- was currentcatcodes
elseif processlines and find(f,"\n") then
+ local flushlines = parent.__flushlines or flushlines
flushlines(f)
else
flush(contentcatcodes,f)
@@ -295,6 +297,7 @@ local function caller(t,f,a,...)
elseif typ == "boolean" then
if f then
if a ~= nil then
+ local flushlines = parent.__flushlines or flushlines
flushlines(f)
-- ignore ... maybe some day
else
@@ -304,7 +307,7 @@ local function caller(t,f,a,...)
else
if a ~= nil then
-- no command, same as context(a,...)
- writer("",a,...)
+ writer(parent,"",a,...)
else
-- ignored
end
@@ -339,7 +342,7 @@ statistics.register("traced context", function()
end
end)
-local tracedwriter = function(...)
+local tracedwriter = function(parent,...)
nofwriters = nofwriters + 1
local t, f, n = { "w : " }, flush, 0
flush = function(...)
@@ -347,7 +350,7 @@ local tracedwriter = function(...)
t[n] = concat({...},"",2)
normalflush(...)
end
- normalwriter(...)
+ normalwriter(parent,...)
flush = f
currenttrace(concat(t))
end
@@ -376,12 +379,14 @@ local function pushlogger(trace)
insert(trace_stack,currenttrace)
currenttrace = trace
flush, writer = tracedflush, tracedwriter
+ context.__flush = flush
end
local function poplogger()
currenttrace = remove(trace_stack)
if not currenttrace then
flush, writer = normalflush, normalwriter
+ context.__flush = flush
end
end
@@ -393,6 +398,8 @@ local function settracing(v)
end
end
+-- todo: share flushers so that we can define in other files
+
trackers.register("context.trace",settracing)
context.pushlogger = pushlogger
@@ -427,58 +434,26 @@ end
function context.direct(first,...)
if first ~= nil then
- return writer("",first,...)
- end
-end
-
--- todo: use flush directly
-
-function context.char(k) -- todo: if catcode == letter or other then just the utf
- if type(k) == "table" then
- for i=1,#k do
- context(format([[\char%s\relax]],k[i]))
- end
- elseif k then
- context(format([[\char%s\relax]],k))
+ return writer(context,"",first,...)
end
end
-function context.utfchar(k)
- context(utfchar(k))
-end
-
-function context.chardef(cs,u)
- context(format([[\chardef\%s=%s\relax]],k))
-end
-
-function context.par()
- context([[\par]]) -- no need to add {} there
-end
-
-function context.bgroup()
- context("{")
-end
-
-function context.egroup()
- context("}")
-end
-
-- context.delayed (todo: lines)
local delayed = { } context.delayed = delayed -- maybe also store them
-local function indexer(t,k)
+local function indexer(parent,k)
local f = function(...)
local a = { ... }
return function()
return context[k](unpack(a))
end
end
- t[k] = f
+ parent[k] = f
return f
end
-local function caller(t,...)
+local function caller(parent,...)
local a = { ... }
return function()
return context(unpack(a))
@@ -491,7 +466,7 @@ setmetatable(delayed, { __index = indexer, __call = caller } )
local nested = { } context.nested = nested
-local function indexer(t,k)
+local function indexer(parent,k)
local f = function(...)
local t, savedflush, n = { }, flush, 0
flush = function(c,f,s,...) -- catcodes are ignored
@@ -502,11 +477,11 @@ local function indexer(t,k)
flush = savedflush
return concat(t)
end
- t[k] = f
+ parent[k] = f
return f
end
-local function caller(t,...)
+local function caller(parent,...)
local t, savedflush, n = { }, flush, 0
flush = function(c,f,s,...) -- catcodes are ignored
n = n + 1
@@ -523,7 +498,7 @@ setmetatable(nested, { __index = indexer, __call = caller } )
local verbatim = { } context.verbatim = verbatim
-local function indexer(t,k)
+local function indexer(parent,k)
local command = context[k]
local f = function(...)
local savedcatcodes = contentcatcodes
@@ -531,14 +506,14 @@ local function indexer(t,k)
command(...)
contentcatcodes = savedcatcodes
end
- t[k] = f
+ parent[k] = f
return f
end
-local function caller(t,...)
+local function caller(parent,...)
local savedcatcodes = contentcatcodes
contentcatcodes = vrbcatcodes
- defaultcaller(t,...)
+ defaultcaller(parent,...)
contentcatcodes = savedcatcodes
end
@@ -550,8 +525,8 @@ local metafun = { } context.metafun = metafun
local mpdrawing = "\\MPdrawing"
-local function caller(t,f,a,...)
- if not t then
+local function caller(parent,f,a,...)
+ if not parent then
-- skip
elseif f then
local typ = type(f)
@@ -602,19 +577,19 @@ end
local delayed = { } metafun.delayed = delayed
-local function indexer(t,k)
+local function indexer(parent,k)
local f = function(...)
local a = { ... }
return function()
return metafun[k](unpack(a))
end
end
- t[k] = f
+ parent[k] = f
return f
end
-local function caller(t,...)
+local function caller(parent,...)
local a = { ... }
return function()
return metafun(unpack(a))
diff --git a/tex/context/base/mult-ini.mkiv b/tex/context/base/mult-ini.mkiv
index b8b1c5387..eddee98a4 100644
--- a/tex/context/base/mult-ini.mkiv
+++ b/tex/context/base/mult-ini.mkiv
@@ -433,11 +433,8 @@
%D interface. (We no longer need the link back to the main
%D internal interface.)
-\def\doresetvalue#1#2%
- {\dosetvalue{#1}{#2}{}}
-
-\def\doignorevalue#1#2#3%
- {\dosetvalue{#1}{#2}{}}
+\def\doresetvalue #1#2{\dosetvalue{#1}{#2}{}}
+\def\doignorevalue#1#2#3{\dosetvalue{#1}{#2}{}}
% \def\dosetvalue#1#2%
% {\let\c!internal!\c!internal!n
@@ -502,6 +499,7 @@
% \@EA\def\csname#1#3\@EA\endcsname\@EA{\csname#2#3\endcsname}%
% \fi}
+\def\doletvalue #1#2{\@EA \let\csname#1#2\endcsname}
\def\dosetvalue #1#2{\@EA \def\csname#1#2\endcsname}
\def\dosetevalue #1#2{\@EA\edef\csname#1#2\endcsname}
\def\dosetgvalue #1#2{\@EA\gdef\csname#1#2\endcsname}
diff --git a/tex/context/base/node-aux.lua b/tex/context/base/node-aux.lua
index 1eb42d95e..2fd6fc005 100644
--- a/tex/context/base/node-aux.lua
+++ b/tex/context/base/node-aux.lua
@@ -22,7 +22,7 @@ local has_attribute = node.has_attribute
local set_attribute = node.set_attribute
local get_attribute = node.get_attribute
local unset_attribute = node.unset_attribute
-local first_character = node.first_character
+local first_glyph = node.first_glyph or node.first_character
local texbox = tex.box
@@ -69,6 +69,7 @@ end
nodes.set_attribute = set_attribute
nodes.unset_attribute = unset_attribute
nodes.has_attribute = has_attribute
+nodes.first_glyph = first_glyph
nodes.set_attributes = set_attributes
nodes.set_unset_attributes = set_unset_attributes
@@ -169,7 +170,7 @@ nodes.unsetattributes = unset_attributes
function nodes.firstcharacter(n,untagged) -- tagged == subtype > 255
if untagged then
- return first_character(n)
+ return first_glyph(n)
else
for g in traverse_id(glyph_code,n) do
return g
diff --git a/tex/context/base/node-pro.lua b/tex/context/base/node-pro.lua
index 9eb431d6b..0ed510cd2 100644
--- a/tex/context/base/node-pro.lua
+++ b/tex/context/base/node-pro.lua
@@ -16,15 +16,15 @@ local report_nodes = logs.new("nodes")
local nodes, node = nodes, node
-local nodecodes = nodes.nodecodes
-local glyph_code = nodecodes.glyph
-local tasks = nodes.tasks
+local nodecodes = nodes.nodecodes
+local glyph_code = nodecodes.glyph
+local tasks = nodes.tasks
-local free_node = node.free
-local first_character = node.first_character
+local free_node = node.free
+local first_glyph = node.first_glyph or node.first_character
-nodes.processors = nodes.processors or { }
-local processors = nodes.processors
+nodes.processors = nodes.processors or { }
+local processors = nodes.processors
-- vbox: grouptype: vbox vtop output split_off split_keep | box_type: exactly|aditional
-- hbox: grouptype: hbox adjusted_hbox(=hbox_in_vmode) | box_type: exactly|aditional
@@ -67,7 +67,7 @@ processors.tracer = tracer
processors.enabled = true -- this will become a proper state (like trackers)
function processors.pre_linebreak_filter(head,groupcode,size,packtype,direction)
- local first, found = first_character(head)
+ local first, found = first_glyph(head)
if found then
if trace_callbacks then
local before = nodes.count(head,true)
@@ -91,7 +91,7 @@ function processors.pre_linebreak_filter(head,groupcode,size,packtype,direction)
end
function processors.hpack_filter(head,groupcode,size,packtype,direction)
- local first, found = first_character(head)
+ local first, found = first_glyph(head)
if found then
if trace_callbacks then
local before = nodes.count(head,true)
@@ -119,13 +119,13 @@ callbacks.register('hpack_filter' , processors.hpack_filter,"all kind of
local actions = tasks.actions("finalizers",1) -- head, where
--- beware, these are packaged boxes so no first_character test
+-- beware, these are packaged boxes so no first_glyph test
-- maybe some day a hash with valid groupcodes
--
-- beware, much can pass twice, for instance vadjust passes two times
function processors.post_linebreak_filter(head,groupcode)
---~ local first, found = first_character(head)
+--~ local first, found = first_glyph(head)
--~ if found then
if trace_callbacks then
local before = nodes.count(head,true)
diff --git a/tex/context/base/node-tra.lua b/tex/context/base/node-tra.lua
index ff038a816..27665d60b 100644
--- a/tex/context/base/node-tra.lua
+++ b/tex/context/base/node-tra.lua
@@ -235,7 +235,7 @@ function step_tracers.glyphs(n,i)
end
function step_tracers.features()
- -- we cannot use first_character here as it only finds characters with subtype < 256
+ -- we cannot use first_glyph here as it only finds characters with subtype < 256
local f = collection[1]
while f do
if f.id == glyph_code then
diff --git a/tex/context/base/node-tsk.lua b/tex/context/base/node-tsk.lua
index a6890c2d5..d19d1c6ae 100644
--- a/tex/context/base/node-tsk.lua
+++ b/tex/context/base/node-tsk.lua
@@ -6,7 +6,7 @@ if not modules then modules = { } end modules ['node-tsk'] = {
license = "see context related readme files"
}
--- this might move to task-*
+-- this might move to task-* .. we already have dirty flags there
local trace_tasks = false trackers.register("tasks.creation", function(v) trace_tasks = v end)
diff --git a/tex/context/base/regi-ini.lua b/tex/context/base/regi-ini.lua
index 561b8d8f4..eb9a58593 100644
--- a/tex/context/base/regi-ini.lua
+++ b/tex/context/base/regi-ini.lua
@@ -26,7 +26,11 @@ regimes.utf = regimes.utf or { }
regimes.synonyms = regimes.synonyms or { }
local synonyms = regimes.synonyms
-storage.register("regimes/synonyms", synonyms, "regimes.synonyms")
+if storage then
+ storage.register("regimes/synonyms", synonyms, "regimes.synonyms")
+else
+ regimes.synonyms = { }
+end
-- setmetatable(regimes.data,_empty_table_)
@@ -72,20 +76,44 @@ function regimes.translate(line,regime)
return line
end
+-- function regimes.enable(regime)
+-- regime = synonyms[regime] or regime
+-- if data[regime] then
+-- regimes.currentregime = regime
+-- local translate = regimes.translate
+-- resolvers.filters.install('input',function(s)
+-- return translate(s,regime)
+-- end)
+-- else
+-- regimes.disable()
+-- end
+-- end
+--
+-- function regimes.disable()
+-- regimes.currentregime = "utf"
+-- resolvers.filters.install('input',nil)
+-- end
+
+local sequencers = utilities.sequencers
+
+function regimes.process(s)
+ return translate(s,regimes.currentregime)
+end
+
function regimes.enable(regime)
regime = synonyms[regime] or regime
if data[regime] then
regimes.currentregime = regime
- local translate = regimes.translate
- resolvers.filters.install('input',function(s)
- return translate(s,regime)
- end)
+ sequencers.enableaction(resolvers.openers.textfileactions,"regimes.process")
else
- regimes.disable()
+ sequencers.disableaction(resolvers.openers.textfileactions,"regimes.process")
end
end
function regimes.disable()
regimes.currentregime = "utf"
- resolvers.filters.install('input',nil)
+ sequencers.disableaction(resolvers.openers.textfileactions,"regimes.process")
end
+
+utilities.sequencers.prependaction(resolvers.openers.textfileactions,"system","regimes.process")
+utilities.sequencers.disableaction(resolvers.openers.textfileactions,"regimes.process")
diff --git a/tex/context/base/s-abr-01.tex b/tex/context/base/s-abr-01.tex
index b233eeee2..85b263468 100644
--- a/tex/context/base/s-abr-01.tex
+++ b/tex/context/base/s-abr-01.tex
@@ -19,10 +19,11 @@
\protect
-\logo [MKI] {MkI}
+\logo [MKI] {MkI} % joke
\logo [MKII] {MkII}
-\logo [MKIII] {MkIII}
+\logo [MKIII] {MkIII} % joke
\logo [MKIV] {MkIV}
+\logo [MKVI] {MkVI}
%logo [FGA] {fga}
%logo [FGBBS] {fgbbs}
diff --git a/tex/context/base/scrn-int.mkiv b/tex/context/base/scrn-int.mkiv
index dc919a8aa..7f143f7b4 100644
--- a/tex/context/base/scrn-int.mkiv
+++ b/tex/context/base/scrn-int.mkiv
@@ -274,7 +274,6 @@
\def\dostartcomment[#1][#2][#3]%
{\bgroup
\doifassignmentelse{#1}{\getparameters[\??cc][#1]}{\getparameters[\??cc][\c!title=#1,#2]}%
- \setcurrentbuffer{\v!comment\v!buffer}%
\dostartbuffer[\v!comment\v!buffer][\v!comment\v!buffer][\e!start\v!comment][\e!stop\v!comment]}
\unexpanded\def\stopcomment
diff --git a/tex/context/base/scrp-ini.lua b/tex/context/base/scrp-ini.lua
index 11ffd02f8..5b012890d 100644
--- a/tex/context/base/scrp-ini.lua
+++ b/tex/context/base/scrp-ini.lua
@@ -19,7 +19,7 @@ local allocate = utilities.storage.allocate
local set_attribute = node.set_attribute
local has_attribute = node.has_attribute
-local first_character = node.first_character
+local first_glyph = node.first_glyph or node.first_character
local traverse_id = node.traverse_id
local nodecodes = nodes.nodecodes
@@ -284,7 +284,7 @@ end
-- one of the time consuming functiions:
function scripts.preprocess(head)
- local start = first_character(head)
+ local start = first_glyph(head)
if not start then
return head, false
else
diff --git a/tex/context/base/strc-sec.mkiv b/tex/context/base/strc-sec.mkiv
index 0d1cd17a6..6fa9348b2 100644
--- a/tex/context/base/strc-sec.mkiv
+++ b/tex/context/base/strc-sec.mkiv
@@ -729,7 +729,7 @@
\setvalue{\??nh:\??mk:n:\v!page }{}
\setvalue{\??nh:\??mk:n:\v!reset}{\resetcurrentstructuremarks}
-\setvalue{\??nh:\??mk:y:\v!page }{\resetcurrentstructuremarks}
+\setvalue{\??nh:\??mk:y:\v!page }{} % to be checked: {\resetcurrentstructuremarks}
\setvalue{\??nh:\??mk:y:\v!reset}{\resetcurrentstructuremarks}
\def\docheckstructureheadlayout
diff --git a/tex/context/base/tabl-ntb.mkiv b/tex/context/base/tabl-ntb.mkiv
index e2420850d..d5eee61a0 100644
--- a/tex/context/base/tabl-ntb.mkiv
+++ b/tex/context/base/tabl-ntb.mkiv
@@ -1574,49 +1574,6 @@
\fi
\fi}
-%D Spacing:
-%
-% \starttabulate
-% \NC text \NC text \NC \NR
-% \TB[small]
-% \NC text \NC text \NC \NR
-% \TB[4*big]
-% \NC text \NC text \NC \NR
-% \stoptabulate
-%
-% \starttable[|||]
-% \VL text \VL text \VL \AR
-% \TB[small]
-% \VL text \VL text \VL \AR
-% \TB[4*big]
-% \VL text \VL text \VL \AR
-% \stoptable
-
-\def\complexTableTB[#1]{\TABLEnoalign{\blank[#1]}}
-\def\simpleTableTB {\TABLEnoalign{\blank}}
-
-\def\TabulateTB
- {\complexorsimpleTable{TB}}
-
-\def\doTableinterline% #1
- {\ifnum\currentTABLEcolumn>\maxTABLEcolumn
- \chuckTABLEautorow
- \else\ifnum\currentTABLEcolumn=\zerocount
- \TABLEnoalign
- {\globalletempty\checkTABLEautorow
- \globalletempty\chuckTABLEautorow}%
- \else
- \setTABLEerror\TABLEmissingcolumn
- \handleTABLEerror
- \fi\fi
- \complexorsimpleTable} % {#1}
-
-\def\TableHL{\doTableinterline{HL}}
-\def\TableTB{\doTableinterline{TB}}
-
-\appendtoks\let\TB\TableTB \to\everytable
-\appendtoks\let\TB\TabulateTB\to\everytabulate % strange place
-
% new (for Olivier Turlier)
%
% \defineTABLEsetup [xx] [foregroundcolor=red]
diff --git a/tex/context/base/tabl-tab.mkiv b/tex/context/base/tabl-tab.mkiv
index d6007372d..3c9515e60 100644
--- a/tex/context/base/tabl-tab.mkiv
+++ b/tex/context/base/tabl-tab.mkiv
@@ -13,6 +13,8 @@
\writestatus{loading}{ConTeXt Table Macros / TaBlE Embedding}
+% Todo: consistent namespace and get rid of not used code
+
% In \MKIV\ the old table macros are sort of obsolete. The
% color extensions have been removed and some code is stripped.
% For practical reasons the \TABLE\ macros that are used are
@@ -36,7 +38,6 @@
\newdimen\tablecolumnwidthunit
\newdimen\tablekernunit
-\def\tablelinethicknessfactor {2}
\def\tablestrutheightfactor {8}
\def\tablestrutdepthfactor {3}
\def\tableintercolumnspacefactor {3}
@@ -44,27 +45,6 @@
\def\tablevspacefactor {2}
\def\tablekernfactor {1}
-\newtoks\NormalTLTU
-\newtoks\NormalTSU
-\newtoks\NormalTICSU
-\newtoks\NormalTCWU
-\newtoks\NormalTKU
-
-\NormalTLTU ={1in \divide \tablelinethicknessunit by 300 }
-\NormalTSU ={\normalbaselineskip \divide \tablestrutunit by 11 }
-\NormalTICSU={.5em plus 1fil minus .25em}
-\NormalTCWU ={.5em}
-\NormalTKU ={.5em}
-
-\def\NormalTableUnits
- {\tablelinethicknessunit \the\NormalTLTU
- \tablestrutunit \the\NormalTSU
- \tableintercolumnspaceunit\the\NormalTICSU
- \tablecolumnwidthunit \the\NormalTCWU
- \tablekernunit \the\NormalTKU}
-
-\NormalTableUnits
-
\newtoks\everytable
\newtoks\everytableparbox
@@ -219,7 +199,7 @@
\noexpand\!!width
\ifnum\!tgCode=\plusone
\ifx\!tgValue\empty
- \tablelinethicknessfactor
+ \tablevrulethicknessfactor
\else
\!tgValue
\fi
@@ -770,7 +750,7 @@
\def\tableStandardTableStrut
{\tableMakeStrut
{\tablestrutheightfactor\tablestrutunit}
- {\tablestrutdepthfactor\tablestrutunit }}
+ {\tablestrutdepthfactor \tablestrutunit }}
\def\tableAugmentedTableStrut#1#2%
{\tableMakeStrut
@@ -996,11 +976,13 @@
\def\@VLn{1}
\def\@VLd{.125em}
+\let\tablecurrentvrulecolor\empty
+
\def\do!ttInsertVrule % will be merged in 2005
{\vrule\!!width
\ifnum\!tgCode=\plusone
\ifx\!tgValue\empty
- \tablelinethicknessfactor
+ \tablevrulethicknessfactor
\else
\!tgValue
\fi
@@ -1012,6 +994,9 @@
\def\!ttInsertVrule
{\hfil
+ \ifx\tablecurrentvrulecolor\empty\else
+ \switchtocolor[\tablecurrentvrulecolor]%
+ \fi
\ifcase\@VLn\or
\do!ttInsertVrule
\unskip
@@ -1020,6 +1005,7 @@
\gdef\@VLn{1}%
\unskip
\fi
+ \global\let\tablecurrentvrulecolor\empty
\hfil
&}
@@ -1043,26 +1029,23 @@
{\ifnum#1>\plusone
\omit
\global\TABLEdivisionfalse
- \scratchcounter\numexpr\currentTABLEcolumn+#1-\plusone\relax % added
- \ifnum\scratchcounter>\maxTABLEcolumn % added
- \def\next % added
- {\setTABLEerror\TABLEspanoverflow % added
- \handleTABLEerror}% % added
- \else % added
- \def\next % added
- {\global\advance\currentTABLEcolumn#1-\plusone\relax % added
- \scratchcounter2\numexpr#1-\plusone\relax % changed
- \!thLoop
- \ifnum\scratchcounter>\plusone
- \span
- \omit
- \advance\scratchcounter\minusone
- \repeat
- \span}%
- \fi % added
+ \scratchcounter\currentTABLEcolumn % added
+ \advance\scratchcounter #1% % added
+ \advance\scratchcounter \minusone % added
+ \def\next % added
+ {\global\advance\currentTABLEcolumn #1% % added
+ \global\advance\currentTABLEcolumn \minusone % added
+ \scratchcounter#1% \mscount is in Plain
+ \advance\scratchcounter \minusone
+ \advance\scratchcounter \scratchcounter
+ \!thLoop
+ \ifnum\scratchcounter>\plusone
+ \spanomit \advance\scratchcounter\minusone
+ \repeat
+ \span}%
\else % added
\def\next % conflicts with possible next \omit % added
- {\global\advance\currentTABLEcolumn\plusone}% % added
+ {\global\advance\currentTABLEcolumn \plusone}% % added
\fi
\next} % added
@@ -1084,7 +1067,11 @@
\def\!ttFullHruleA
{\!ttGetHalfRuleThickness
+ \ifx\tablecurrenthrulecolor\empty\else
+ \switchtocolor[\tablecurrenthrulecolor]%
+ \fi
\hrule\!!height\dimen0\!!depth\dimen0
+ \global\let\tablecurrenthrulecolor\empty
%\penalty0 % removed
\egroup}
@@ -1094,18 +1081,22 @@
\def\!ttShortHruleA
{\!ttGetHalfRuleThickness
+ \ifx\tablecurrenthrulecolor\empty\else
+ \switchtocolor[\tablecurrenthrulecolor]%
+ \fi
\leaders\hrule\!!height\dimen0\!!depth\dimen0\hfill
\null
+ \global\let\tablecurrenthrulecolor\empty
\ignorespaces}
\def\!ttLongHrule
{\omit\span\omit\span\!ttShortHrule}
\def\!ttGetHalfRuleThickness
- {\dimen0=
+ {\dimen0=\dimexpr
\ifnum\!tgCode=\plusone
\ifx\!tgValue\empty
- \tablelinethicknessfactor
+ \tablehrulethicknessfactor
\else
\!tgValue % user-specified integer
\fi
@@ -1122,10 +1113,8 @@
\def\tableRight #1{\hfill#1}
\def\tableOpenUp#1#2%
- {\scratchdimen\tablestrutheightfactor \points \advance\scratchdimen #1\points
- \edef\tablestrutheightfactor{\withoutpt\the\scratchdimen}%
- \scratchdimen\tablestrutdepthfactor \points \advance\scratchdimen #2\points
- \edef\tablestrutdepthfactor{\withoutpt\the\scratchdimen}}
+ {\edef\tablestrutheightfactor{\withoutpt\the\dimexpr\tablestrutheightfactor\points+#1\points}%
+ \edef\tablestrutdepthfactor {\withoutpt\the\dimexpr\tablestrutdepthfactor \points+#2\points}}
\def\tableSetTableToWidth#1%
{\doifelsenothing{#1}{\!taTableSpread\emptytoks}{\!taTableSpread{to #1}}}
@@ -1595,8 +1584,21 @@
%D such hacks become a second nature. However, redefining \type
%D {\omit} and \type{\span} is not that easy.}
+\def\doTABLEnoalign
+ {\let\next=}
+
\def\TABLEnoalign
- {\noalign\bgroup\let\noalign\relax\let\next=}
+ {\noalign
+ \bgroup
+ \let\noalign\relax
+ \let\TABLEnoalign\doTABLEnoalign
+ \doTABLEnoalign}
+
+\def\startTABLEnoalign
+ {\TABLEnoalign\bgroup}
+
+\def\stopTABLEnoalign
+ {\iffalse{\else}\fi}
%D \macros
%D {starttable}
@@ -1780,19 +1782,16 @@
%D Now we can start the table.
+\newtoks \localtabledefinitions
+
\def\thirdstagestartTABLE#1%
{\global\setTABLEactiontrue
\setTABLEaction\TABLEunknown
\setTABLEforce\TABLEunknown
- \resetVLvalues
+ \tableresetVLvalues
\appendtoks\localTABLEsetup\to\everytable
\tablestandardbegin[\ifsplittables u\else b\fi]%
- \defineTABLEunits
- \defineTABLEsteps
- \defineTABLErules
- \defineTABLEdivisions
- \defineTABLEshorthands
- \defineTABLEendings
+ \the\localtabledefinitions
\forgetall % added
\doifsomething{#1}
{\def\TABLEformat{#1}%
@@ -1907,10 +1906,10 @@
\let\@@TABLEhead\empty \def\TABLEhead{\@@TABLEhead}
\let\@@TABLEtail\empty \def\TABLEtail{\@@TABLEtail}
-\letvalue{\e!start\v!tablehead}=\undefined
-\letvalue{\e!stop \v!tablehead}=\undefined
-\letvalue{\e!start\v!tabletail}=\undefined
-\letvalue{\e!stop \v!tabletail}=\undefined
+\letbeundefined{\e!start\v!tablehead}
+\letbeundefined{\e!stop \v!tablehead}
+\letbeundefined{\e!start\v!tabletail}
+\letbeundefined{\e!stop \v!tabletail}
\expanded
{\long\def\csname\e!start\v!tablehead\endcsname##1\csname\e!stop\v!tablehead\endcsname%
@@ -1935,23 +1934,6 @@
%D Redudant \type{\HL}'s are removed automatically, so
%D mid||lines can be used without problems.
-%D We need an alternative for the normal complex or simple
-%D commands, because assignments in these system commands
-%D conflict with \type{\noalign}. This alternative is about
-%D as efficient as possible.
-
-\def\complexorsimpleTable#1#2%
- {\csname\if[\noexpand#2\s!complex\else\s!simple\fi\c!Table#1\endcsname#2}
-
-%D The next one is used in \type{\VL} cum suis and honours
-%D the next grouping.
-
-\def\docomplexorsimpleTable#1#2%
- {\ifx\next\bgroup\@EA#2\else\@EA\dodocomplexorsimpleTable\@EA#1\@EA#2\fi}
-
-\def\dodocomplexorsimpleTable#1#2#3%
- {\if[\noexpand#3\@EA#1\else\@EA#2\fi#3}
-
%D The order of the next macros is more or less random. First
%D we implement error recovery. Errors are reported to the
%D screen and log file as well as visualized in the table in
@@ -2011,12 +1993,13 @@
\setnewconstant\TABLErowzero\zerocount
-\unexpanded\def\defineTABLEendings
- {\let\SR\TableSR
- \let\FR\TableFR
- \let\MR\TableMR
- \let\LR\TableLR
- \let\AR\TableAR}
+\appendtoks
+ \let\SR\TableSR
+ \let\FR\TableFR
+ \let\MR\TableMR
+ \let\LR\TableLR
+ \let\AR\TableAR
+\to \localtabledefinitions
\unexpanded\def\TableSR
{\ifnum\TABLEaction=\TABLEfirstrow
@@ -2144,7 +2127,6 @@
%D one more than the number of columns.
\newcount\currentTABLEcolumn
-\newcount\maxTABLEcolumn % needed for reset vl properties
%D While defining this macro we change the \CATCODE\ of
%D \type{|}. When counting the bars, we use a non active
@@ -2161,7 +2143,7 @@
\bgroup \catcode`\|=\othercatcode
-\gdef\getTABLEnofcolumns#1%
+\gdef\getTABLEnofcolumns#1% todo: also divert this to lua as with tabulate
{\bgroup
\cleanupfeatures % needed !
\@@useotherbar
@@ -2177,34 +2159,46 @@
%D \sym{\type{\HC}} a horizontal colored line
%D \stopitemize
-\unexpanded\def\defineTABLErules
- {\let\VL\TableVL
- \let\VC\TableVC
- \let\HL\TableHL
- \let\HC\TableHC
- \let\VS\TableVS
- \let\VD\TableVD
- \let\VT\TableVT
- \let\VN\TableVN}
+\newcount\tablevrulethicknessfactor
+\newcount\tablehrulethicknessfactor
+\newcount\tabledrulespan
+\let \tablecurrentvrulecolor \empty
+\let \tablecurrenthrulecolor \empty
-\unexpanded\def\TableVL
- {\checkTABLEautorow
- \global\advance\currentTABLEcolumn \plusone
- \doTableVL}
-
-\def\doTableVL
- {\futurelet\next\dodoTableVL}
+\appendtoks
+ \let\VL\TableVL
+ \let\VC\TableVC
+ \let\HL\TableHL
+ \let\HC\TableHC
+ \let\VS\TableVS
+ \let\VD\TableVD
+ \let\VT\TableVT
+ \let\VN\TableVN
+\to \localtabledefinitions
+
+\def\tableresetVLvalues
+ {\global\currentTABLEcolumn\zerocount}
-\def\dodoTableVL
- {\docomplexorsimpleTable\complexTableVL\simpleTableVL}
+\def\dotablevrulecommand#1% global assignments
+ {\doifnumberelse{#1}
+ {\global\tablevrulethicknessfactor#1\relax
+ \global\multiply\tablevrulethicknessfactor\@@tiVLwidth\relax}
+ {\xdef\tablecurrentvrulecolor{#1}}}
-\def\complexTableVL[#1]%
- {\scratchcounter=0#1%
- \multiply\scratchcounter \@@tiVLwidth
- \simpleTableVL}
+\unexpanded\def\TableVL
+ {\checkTABLEautorow
+ \global\advance\currentTABLEcolumn\plusone
+ \dosingleempty\doTableVL}
+
+\def\doTableVL[#1]%
+ {\global\tablecurrentvrulecolor\empty
+ \global\tablevrulethicknessfactor\@@tiVLwidth\relax
+ \iffirstargument
+ \rawprocesscommalist[#1]\dotablevrulecommand
+ \fi
+ \normalexpanded{\noexpand\normalTABLEcomplexbar\the\tablevrulethicknessfactor} }% \relax breaks \use
-\def\simpleTableVL
- {\expanded{\normalTABLEcomplexbar\@@tiVLwidth\space}}% \relax breaks \use
+\let\TableVC\TableVL % for mojca
% \starttable[|||]
% \HL
@@ -2218,79 +2212,46 @@
\unexpanded\def\TableVD {\VN2}
\unexpanded\def\TableVT {\VN3}
\unexpanded\def\TableVN#1{\gdef\@VLn{#1}\VL}
-
-\def\resetVLvalues
- {\global\currentTABLEcolumn\zerocount}
-
-\unexpanded\def\TableVC
- {\checkTABLEautorow
- \doTableVC}
-
-\def\doTableVC
- {\futurelet\next\dodoTableVC}
-
-\def\dodoTableVC
- {\docomplexorsimpleTable\complexTableVC\simpleTableVC}
-\def\complexTableVC[#1]%
- {\setgvalue{bVC\the\currentTABLEcolumn}{\startcolor[#1]}%
- \setgvalue{eVC\the\currentTABLEcolumn}{\stopcolor}%
- \simpleTableVC}
-
-\def\simpleTableVC
- {\setgvalue{bVL\the\currentTABLEcolumn}{\getvalue{bVC\the\currentTABLEcolumn}}%
- \setgvalue{eVL\the\currentTABLEcolumn}{\getvalue{eVC\the\currentTABLEcolumn}}%
- \doTableVL}
+\def\dotablehrulecommand#1% global assignments
+ {\doifnumberelse{#1}
+ {\global\tablehrulethicknessfactor#1\relax
+ \global\multiply\tablehrulethicknessfactor\@@tiHLheight\relax}
+ {\xdef\tablecurrenthrulecolor{#1}}}
\unexpanded\def\TableHL
{\finishTABLErow
- \complexorsimpleTable{HL}}
-
-\def\complexTableHL[#1]%
- {\TABLEnoalign
- {\scratchcounter0#1%
- \multiply\scratchcounter \@@tiHLheight
- \edef\@@tiHLheight{\the\scratchcounter}}%
- \simpleTableHL}
-
-\def\simpleTableHL
- {\TABLEnoalign
- {\nobreak
- \ifnum\TABLEaction=\TABLErule
- \writestatus\m!TABLE{skipping \string\HL}% \statusmessage
- \else
- \ifnum\TABLEaction=\TABLEmidrow
- \writestatus\m!TABLE{change \string\MR\space into \string\LR/\string\SR}%
- \else\ifnum\TABLEaction=\TABLEfirstrow
- \writestatus\m!TABLE{change \string\MR\space into \string\SR}%
- \fi\fi
- \startHLcommand
- \expandafter\normalTABLEfullrule\@@tiHLheight
- \stopHLcommand
- \globalletempty\startHLcommand
- \globalletempty\stopHLcommand
- \accountTABLElinewidth
- \fi
- \setTABLEaction\TABLErule
- \nobreak}}
-
-\let\startHLcommand\empty
-\let\stopHLcommand \empty
-
-\unexpanded\def\TableHC
- {\complexorsimpleTable{HC}}
+ \startTABLEnoalign
+ \dosingleempty\doTableHL}
-\def\complexTableHC[#1]%
- {\TABLEnoalign
- {\gdef\startHCcommand{\startcolor[#1]}%
- \gdef\stopHCcommand {\stopcolor}}%
- \simpleTableHC}
+\def\doTableHL[#1]%
+ {\nobreak
+ \ifnum\TABLEaction=\TABLErule
+ \writestatus\m!TABLE{skipping \string\HL}% \statusmessage
+ \else
+ \ifnum\TABLEaction=\TABLEmidrow
+ \writestatus\m!TABLE{change \string\MR\space into \string\LR/\string\SR}%
+ \else\ifnum\TABLEaction=\TABLEfirstrow
+ \writestatus\m!TABLE{change \string\MR\space into \string\SR}%
+ \fi\fi
+ \bgroup
+ \global\tablehrulethicknessfactor\@@tiHLheight\relax
+ \iffirstargument
+ \global\let\tablecurrenthrulecolor\empty
+ \rawprocesscommalist[#1]\dotablehrulecommand
+ \ifx\tablecurrenthrulecolor\empty\else
+ \switchtocolor[\tablecurrenthrulecolor]%
+ \fi
+ \fi
+ \normalexpanded{\noexpand\normalTABLEfullrule\the\tablehrulethicknessfactor} %
+ \egroup
+ \accountTABLElinewidth
+ \fi
+ \setTABLEaction\TABLErule
+ \nobreak
+ \stopTABLEnoalign}
-\def\simpleTableHC
- {\TABLEnoalign
- {\globallet\startHLcommand\startHCcommand
- \globallet\stopHLcommand \stopHCcommand}%
- \HL}
+\let\TableHC\TableHL % for mojca
%D \startitemize[3*ruim]
%D \sym{\type{\NL}} a vertical skip
@@ -2301,30 +2262,37 @@
%D \sym{\type{\LC}} a last column
%D \stopitemize
-% n+1 uitleggen
+% \starttable[|||]
+% \VL text \VL text \VL \AR
+% \TB[small]
+% \VL text \VL text \VL \AR
+% \TB[4*big]
+% \VL text \VL text \VL \AR
+% \stoptable
-\unexpanded\def\defineTABLEsteps
- {\let\NL\TableNL
- \let\NR\TableNR
- \let\NC\TableNC
- \let\FC\TableNC
- \let\MC\TableNC
- \let\LC\TableNC}
+% n+1 uitleggen
-\unexpanded\def\TableNL
- {\complexorsimpleTable{NL}}
+\appendtoks
+ \let\TB\TableTB
+ \let\NL\TableNL % old
+ \let\NR\TableNR
+ \let\NC\TableNC
+ \let\FC\TableNC
+ \let\MC\TableNC
+ \let\LC\TableNC
+\to \localtabledefinitions
+
+\unexpanded\def\TableTB
+ {\finishTABLErow
+ \startTABLEnoalign
+ \dosingleempty\doTableTB}
-\def\complexTableNL[#1]%
- {\TABLEnoalign
- {\edef\@@tiNL{#1}%
- \simpleTableNL}}%
+\def\doTableTB[#1]%
+ {\blank[\iffirstargument#1\else\@@tiNL\fi]%
+ \nobreak
+ \stopTABLEnoalign}
-\def\simpleTableNL
- {\TABLEnoalign
- {\nobreak
- \setbox0\vbox{\blank[\@@tiNL]}%
- \vskip\ht0
- \nobreak}}
+\let\TableNL\TableTB
\unexpanded\def\TableNR
{\global\currentTABLEcolumn\zerocount
@@ -2347,12 +2315,13 @@
\newif\ifTABLEdivision
-\unexpanded\def\defineTABLEdivisions
- {\global\TABLEdivisionfalse % in start
- \let\DL\TableDL
- \let\DC\TableDC
- \let\DV\TableDV
- \let\DR\TableDR}
+\appendtoks
+ \global\TABLEdivisionfalse % in start
+ \let\DL\TableDL
+ \let\DC\TableDC
+ \let\DV\TableDV
+ \let\DR\TableDR
+\to \localtabledefinitions
\def\checkTABLEdivision
{\ifTABLEdivision \else
@@ -2361,14 +2330,21 @@
\global\TABLEdivisiontrue
\fi}
+\def\dotabledrulecommand#1% global assignments
+ {\doifnumberelse{#1}
+ {\ifcase\tabledrulespan
+ \global\tabledrulespan#1\relax
+ \else
+ \global\tablehrulethicknessfactor#1\relax
+ \global\multiply\tablehrulethicknessfactor\@@tiVLwidth\relax
+ \fi}
+ {\xdef\tablecurrenthrulecolor{#1}}}
+
\unexpanded\def\TableDL
{\checkTABLEdivision
- \complexorsimpleTable{DL}}
+ \dosingleempty\doTableDL}
-\def\simpleTableDL
- {\complexTableDL[1]}
-
-\def\complexTableDL[#1]%
+\def\doTableDL[#1]%
{\ifnum\TABLEaction=\TABLErule
\writestatus\m!TABLE{skipping \string\DL}%
\else
@@ -2378,12 +2354,24 @@
\writestatus\m!TABLE{change \string\MR\space into \string\SR}%
\fi\fi
\setTABLEaction\TABLEunknown
- \ifnum#1=\plusone
+ \global\tablehrulethicknessfactor\@@tiHLheight\relax
+ \global\tabledrulespan\zerocount
+ \iffirstargument
+ \global\let\tablecurrenthrulecolor\empty
+ \rawprocesscommalist[#1]\dotabledrulecommand
+ \ifx\tablecurrenthrulecolor\empty\else
+ \switchtocolor[\tablecurrenthrulecolor]%
+ \fi
+ \fi
+ \ifcase\tabledrulespan
+ \global\advance\currentTABLEcolumn \plusone
+ \let\next\normalTABLEsinglerule
+ \or
\global\advance\currentTABLEcolumn \plustwo
\let\next\normalTABLEsinglerule
\else
\global\advance\currentTABLEcolumn \plusone
- \def\next{\normalTABLEmultirule{#1}}%
+ \edef\next{\noexpand\normalTABLEmultirule{\tabledrulespan} }%
\fi
\next
\fi}
@@ -2412,25 +2400,39 @@
\def\accountTABLElinewidth
{\scratchdimen\tablelinethicknessunit}
-\unexpanded\def\defineTABLEshorthands
- {\def\SPAN##1{\use{##1}}%
- \def\TWO {\use2}%
- \def\THREE {\use3}%
- \def\FOUR {\use4}%
- \def\FIVE {\use5}%
- \def\SIX {\use6}%
- \def\REF {\ReFormat}}
-
-\unexpanded\def\defineTABLEunits
- {\processaction
- [\@@tidistance]
- [ \v!none=>\OpenUp00\def\LOW{\Lower6 },
- \v!small=>\OpenUp00\def\LOW{\Lower6 }, % == baseline
- \v!medium=>\OpenUp11\def\LOW{\Lower7 },
- \v!big=>\OpenUp22\def\LOW{\Lower8 }]%
- \doifelse\@@tidistance\v!none
- {\TABLErowfactor\zerocount}
- {\TABLErowfactor\plustwo }}
+\def\doTableSPAN#1{\use{#1}}
+\def\doTableTWO {\use2}
+\def\doTableTHREE {\use3}
+\def\doTableFOUR {\use4}
+\def\doTableFIVE {\use5}
+\def\doTableSIX {\use6}
+\def\doTableREF {\ReFormat}
+
+\appendtoks
+ \let\SPAN \doTableSPAN
+ \let\TWO \doTableTWO
+ \let\THREE\doTableTHREE
+ \let\FOUR \doTableFOUR
+ \let\FIVE \doTableFIVE
+ \let\SIX \doTableSIX
+ \let\REF \doTableREF
+\to \localtabledefinitions
+
+\setvalue{\??ti:\c!distance:\v!none }{\OpenUp00\def\LOW{\Lower6 }}
+\setvalue{\??ti:\c!distance:\v!small }{\OpenUp00\def\LOW{\Lower6 }} % == baseline
+\setvalue{\??ti:\c!distance:\v!medium}{\OpenUp11\def\LOW{\Lower7 }}
+\setvalue{\??ti:\c!distance:\v!big }{\OpenUp22\def\LOW{\Lower8 }}
+
+\appendtoks
+ \getvalue{\??ti:\c!distance:\@@tidistance}%
+\to \localtabledefinitions
+
+\appendtoks
+ \doifelse\@@tidistance\v!none
+ {\TABLErowfactor\zerocount}
+ {\TABLErowfactor\plustwo }%
+\to \localtabledefinitions
+
\def\dohandlebar % here ?
{\ifmmode
@@ -2446,7 +2448,7 @@
\def\dosetuptables[#1]%
{\getparameters[\??ti][#1]%
- \processaction
+ \processaction % we have a command for this
[\@@tialign]
[ \v!right=>\def\TABLEparalignment{\raggedright},
\v!left=>\def\TABLEparalignment{\raggedleft},
@@ -2456,43 +2458,24 @@
\assignalfadimension\@@tiVL\@@tiVLwidth 246%
\assignalfadimension\@@tiHL\@@tiHLheight246}
-\def\localTABLEsetup
- {\@@ticommands\relax
- \expanded{\switchtobodyfont[\@@tibodyfont]}%
- \def\tablestrutheightfactor {8}%
- \def\tablestrutdepthfactor {4}%
- \def\tablelinethicknessfactor{4}%
- \NormalTLTU {.1pt}%
- \NormalTSU {\normalbaselineskip\divide\tablestrutunit 12 }%
- \NormalTableUnits}
-
-%D And then I wrote the tabulate environment. That
-%D alternative supports setting the rule thickness and color,
-%D so here is the table alternative.
-
-% \let\startglobalTABLEcolor\empty
-% \let\stopglobalTABLEcolor \empty
+\let\tablelinethicknessfactor\plusfour
\def\localTABLEsetup
{\@@ticommands\relax
- \expanded{\switchtobodyfont[\@@tibodyfont]}%
- \def\tablelinethicknessfactor{4}%
- \scratchdimen\@@tirulethickness
- \divide\scratchdimen \tablelinethicknessfactor
- \expanded{\NormalTLTU{\the\scratchdimen}}%
+ \expanded{\switchtobodyfont[\@@tibodyfont]}% expanded ?
+ \tablelinethicknessunit\dimexpr\@@tirulethickness/\tablelinethicknessfactor\relax
\doifelse\@@tiheight\v!strut
- {\edef\tablestrutheightfactor{\tablestrutheightfactor}}
- {\edef\tablestrutheightfactor{\@@tiheight}}%
+ {\let\tablestrutheightfactor\tablestrutheightfactor}
+ {\let\tablestrutheightfactor\@@tiheight}%
\doifelse\@@tidepth\v!strut
- {\edef\tablestrutdepthfactor{\tablestrutdepthfactor}}
- {\edef\tablestrutdepthfactor{\@@tidepth}}%
- \scratchdimen\tablestrutheightfactor\points \multiply\scratchdimen 10%
- \edef\tablestrutheightfactor{\withoutpt\the\scratchdimen}%
- \scratchdimen\tablestrutdepthfactor \points \multiply\scratchdimen 10%
- \edef\tablestrutdepthfactor{\withoutpt\the\scratchdimen}%
- % units
- \NormalTSU{\normalbaselineskip\divide\tablestrutunit 12 }%
- \NormalTableUnits}
+ {\let\tablestrutdepthfactor\tablestrutdepthfactor}
+ {\let\tablestrutdepthfactor\@@tidepth}%
+ \edef\tablestrutheightfactor{\withoutpt\the\dimexpr10\dimexpr\tablestrutheightfactor\points}%
+ \edef\tablestrutdepthfactor {\withoutpt\the\dimexpr10\dimexpr\tablestrutdepthfactor \points}%
+ \tablestrutunit\dimexpr\normalbaselineskip/12\relax % 12 is default bodyfont
+ \tableintercolumnspaceunit.5em plus 1fil minus .25em\relax
+ \tablecolumnwidthunit .5em\relax
+ \tablekernunit .5em\relax}
%D As one can see, we didn't only add color, but also more
%D control over spacing.
diff --git a/tex/context/base/tabl-tbl.mkiv b/tex/context/base/tabl-tbl.mkiv
index 3cfda9ad7..9428ea3dc 100644
--- a/tex/context/base/tabl-tbl.mkiv
+++ b/tex/context/base/tabl-tbl.mkiv
@@ -849,7 +849,7 @@
% The much needed hook:
-\appendtoks \optimizeverbatimfalse \to \everytabulate
+% \appendtoks \optimizeverbatimfalse \to \everytabulate
% Todo: proper inheritance
@@ -1661,6 +1661,30 @@
%D \stoptabulate
%D \stoptyping
+%D Spacing:
+%
+% \starttabulate
+% \NC text \NC text \NC \NR
+% \TB[small]
+% \NC text \NC text \NC \NR
+% \TB[4*big]
+% \NC text \NC text \NC \NR
+% \stoptabulate
+
+\def\TabulateTB
+ {\startTABLEnoalign
+ \dosingleempty\doTabulateTB}
+
+\def\doTabulateTB[#1]%
+ {\iffirstargument
+ \blank[#1]
+ \else
+ \blank
+ \fi
+ \stopTABLEnoalign}
+
+\appendtoks\let\TB\TabulateTB\to\everytabulate
+
% \starttabulatie[|mc|]
% \NC \digits{100.000,00} \NC\NR
% \NC \digits{@10.000,00} \NC\NR
diff --git a/tex/context/base/util-prs.lua b/tex/context/base/util-prs.lua
index 6ca7c3396..a369552dc 100644
--- a/tex/context/base/util-prs.lua
+++ b/tex/context/base/util-prs.lua
@@ -50,11 +50,11 @@ local pattern_b = spaces * comma^0 * spaces * (key * ((spaces * equal * spaces *
local hash = { }
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
-local function set(key,value) -- using Carg is slower here
+local function set(key,value)
hash[key] = value
end
diff --git a/tex/context/base/util-seq.lua b/tex/context/base/util-seq.lua
index 028d0cf8d..7ce2cd345 100644
--- a/tex/context/base/util-seq.lua
+++ b/tex/context/base/util-seq.lua
@@ -37,88 +37,95 @@ local function validaction(action)
return true
end
-function sequencers.reset()
- return {
+function sequencers.reset(t)
+ local s = {
list = { },
order = { },
kind = { },
askip = { },
gskip = { },
}
+ if t then
+ s.arguments = t.arguments
+ s.returnvalues = t.returnvalues
+ s.results = t.results
+ end
+ s.dirty = true
+ return s
end
function sequencers.prependgroup(t,group,where)
- local list, order = t.list, t.order
- removevalue(order,group)
- insertbeforevalue(order,where,group)
- list[group] = { }
+ if t then
+ local list, order = t.list, t.order
+ removevalue(order,group)
+ insertbeforevalue(order,where,group)
+ list[group] = { }
+ t.dirty = true
+ end
end
function sequencers.appendgroup(t,group,where)
- local list, order = t.list, t.order
- removevalue(order,group)
- insertaftervalue(order,where,group)
- list[group] = { }
+ if t then
+ local list, order = t.list, t.order
+ removevalue(order,group)
+ insertaftervalue(order,where,group)
+ list[group] = { }
+ t.dirty = true
+ end
end
function sequencers.prependaction(t,group,action,where,kind,force)
- local g = t.list[group]
- if g and (force or validaction(action)) then
- removevalue(g,action)
- insertbeforevalue(g,where,action)
- t.kind[action] = kind
+ if t then
+ local g = t.list[group]
+ if g and (force or validaction(action)) then
+ removevalue(g,action)
+ insertbeforevalue(g,where,action)
+ t.kind[action] = kind
+ t.dirty = true
+ end
end
end
function sequencers.appendaction(t,group,action,where,kind,force)
- local g = t.list[group]
- if g and (force or validaction(action)) then
- removevalue(g,action)
- insertaftervalue(g,where,action)
- t.kind[action] = kind
+ if t then
+ local g = t.list[group]
+ if g and (force or validaction(action)) then
+ removevalue(g,action)
+ insertaftervalue(g,where,action)
+ t.kind[action] = kind
+ t.dirty = true
+ end
end
end
-function sequencers.enableaction (t,action) t.askip[action] = false end
-function sequencers.disableaction(t,action) t.askip[action] = true end
-function sequencers.enablegroup (t,group) t.gskip[group] = false end
-function sequencers.disablegroup (t,group) t.gskip[group] = true end
+function sequencers.enableaction (t,action) if t then t.dirty = true t.askip[action] = false end end
+function sequencers.disableaction(t,action) if t then t.dirty = true t.askip[action] = true end end
+function sequencers.enablegroup (t,group) if t then t.dirty = true t.gskip[group] = false end end
+function sequencers.disablegroup (t,group) if t then t.dirty = true t.gskip[group] = true end end
function sequencers.setkind(t,action,kind)
- t.kind[action] = kind
+ if t then
+ t.kind[action] = kind
+ t.dirty = true
+ end
end
function sequencers.removeaction(t,group,action,force)
- local g = t.list[group]
+ local g = t and t.list[group]
if g and (force or validaction(action)) then
removevalue(g,action)
+ t.dirty = true
end
end
-function sequencers.compile(t,compiler,n)
- if type(t) == "string" then
- -- already compiled
- elseif compiler then
- t = compiler(t,n)
- else
- t = sequencers.tostring(t)
- end
- return loadstring(t)()
-end
-
local function localize(str)
return (gsub(str,"%.","_"))
end
-local template = [[
-%s
-return function(...)
-%s
-end]]
-
-function sequencers.tostring(t)
+local function construct(t,nodummy)
local list, order, kind, gskip, askip = t.list, t.order, t.kind, t.gskip, t.askip
- local vars, calls, args, n = { }, { }, nil, 0
+ local arguments, returnvalues, results = t.arguments or "...", t.returnvalues, t.results
+ local variables, calls, n = { }, { }, 0
for i=1,#order do
local group = order[i]
if not gskip[group] then
@@ -128,13 +135,44 @@ function sequencers.tostring(t)
if not askip[action] then
local localized = localize(action)
n = n + 1
- vars [n] = format("local %s = %s", localized, action)
- calls[n] = format(" %s(...) -- %s %i", localized, group, i)
+ variables[n] = format("local %s = %s",localized,action)
+ if not returnvalues then
+ calls[n] = format("%s(%s)",localized,arguments)
+ elseif n == 1 then
+ calls[n] = format("local %s = %s(%s)",returnvalues,localized,arguments)
+ else
+ calls[n] = format("%s = %s(%s)",returnvalues,localized,arguments)
+ end
end
end
end
end
- return format(template,concat(vars,"\n"),concat(calls,"\n"))
+ t.dirty = false
+ if nodummy and #calls == 0 then
+ return nil
+ else
+ variables = concat(variables,"\n")
+ calls = concat(calls,"\n")
+ if results then
+ return format("%s\nreturn function(%s)\n%s\nreturn %s\nend",variables,arguments,calls,results)
+ else
+ return format("%s\nreturn function(%s)\n%s\nend",variables,arguments,calls)
+ end
+ end
+end
+
+sequencers.tostring = construct
+sequencers.localize = localize
+
+function sequencers.compile(t,compiler,n)
+ if not t or type(t) == "string" then
+ -- already compiled
+ elseif compiler then
+ t = compiler(t,n)
+ else
+ t = construct(t)
+ end
+ return loadstring(t)()
end
-- we used to deal with tail as well but now that the lists are always
@@ -151,7 +189,7 @@ return function(head%s)
return head, done
end]]
-function sequencers.nodeprocessor(t,nofarguments)
+function sequencers.nodeprocessor(t,nofarguments) -- todo: handle 'kind' in plug into tostring
local list, order, kind, gskip, askip = t.list, t.order, t.kind, t.gskip, t.askip
local vars, calls, args, n = { }, { }, nil, 0
if nofarguments == 0 then
@@ -179,6 +217,7 @@ function sequencers.nodeprocessor(t,nofarguments)
local localized = localize(action)
n = n + 1
vars[n] = format("local %s = %s",localized,action)
+ -- only difference with tostring is kind and rets (why no return)
if kind[action] == "nohead" then
calls[n] = format(" ok = %s(head%s) done = done or ok -- %s %i",localized,args,group,i)
else
diff --git a/tex/context/base/v-default.lua b/tex/context/base/v-default.lua
new file mode 100644
index 000000000..107c4c4ca
--- /dev/null
+++ b/tex/context/base/v-default.lua
@@ -0,0 +1,40 @@
+if not modules then modules = { } end modules ['v-default'] = {
+ version = 1.001,
+ comment = "companion to v-default.mkiv",
+ author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
+ copyright = "PRAGMA ADE / ConTeXt Development Team",
+ license = "see context related readme files"
+}
+
+local patterns, P, V = lpeg.patterns, lpeg.P, lpeg.V
+local makepattern = visualizers.makepattern
+
+local handler = visualizers.newhandler()
+
+local grammar = { "visualizer",
+
+ -- basic
+
+ emptyline = makepattern(handler,"emptyline",patterns.emptyline),
+ beginline = makepattern(handler,"beginline",patterns.beginline),
+ newline = makepattern(handler,"newline", patterns.newline),
+ space = makepattern(handler,"space", patterns.space),
+ default = makepattern(handler,"default", patterns.anything),
+ content = makepattern(handler,"default", patterns.somecontent),
+
+ -- handy
+
+ line = V("newline") * V("emptyline")^0 * V("beginline"),
+ whitespace = (V("space") + V("line"))^1,
+ optionalwhitespace = (V("space") + V("line"))^0,
+
+ -- used
+
+ pattern = V("line") + V("space") + V("content"),
+ visualizer = V("pattern")^1
+
+}
+
+local parser = P(grammar)
+
+visualizers.register("default", { parser = parser, handler = handler, grammar = grammar })
diff --git a/tex/context/base/v-default.mkiv b/tex/context/base/v-default.mkiv
new file mode 100644
index 000000000..f9a821f0f
--- /dev/null
+++ b/tex/context/base/v-default.mkiv
@@ -0,0 +1,43 @@
+%D \module
+%D [ file=v-default,
+%D version=2010.10.19,
+%D title=\CONTEXT\ Visualizer Macros,
+%D subtitle=Default,
+%D author=Hans Hagen,
+%D date=\currentdate,
+%D copyright=PRAGMA-ADE]
+%C
+%C This module is part of the \CONTEXT\ macro||package and is
+%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
+%C details.
+
+\registerctxluafile{v-default.lua}{1.001}
+
+\unprotect
+
+\definestartstop
+ [DefaultSnippet]
+ [\c!before=\blank,
+ \c!after=\blank,
+ \c!style=\tt]
+
+% Name
+% NamePrimitive
+% NamePlain
+% NameMetafun
+% Group
+% Boundary
+% Special
+% Comment
+% Constructor
+% Key
+% Entity
+% String
+% Equal
+% Cdata
+% Instruction
+
+% Value
+% Quote
+
+\protect \endinput
diff --git a/tex/context/base/v-escaped.lua b/tex/context/base/v-escaped.lua
new file mode 100644
index 000000000..060f48ea3
--- /dev/null
+++ b/tex/context/base/v-escaped.lua
@@ -0,0 +1,14 @@
+if not modules then modules = { } end modules ['v-escaped'] = {
+ version = 1.001,
+ comment = "companion to v-escaped.mkiv",
+ author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
+ copyright = "PRAGMA ADE / ConTeXt Development Team",
+ license = "see context related readme files"
+}
+
+visualizers.registerescapepattern("/BTEX/ETEX","/BTEX","/ETEX")
+
+visualizers.register("escaped", {
+ parser = visualizers.escapepatterns["/BTEX/ETEX"],
+ handler = visualizers.newhandler(),
+})
diff --git a/tex/context/base/v-escaped.mkiv b/tex/context/base/v-escaped.mkiv
new file mode 100644
index 000000000..780d0b518
--- /dev/null
+++ b/tex/context/base/v-escaped.mkiv
@@ -0,0 +1,18 @@
+%D \module
+%D [ file=v-escaped,
+%D version=2010.10.19,
+%D title=\CONTEXT\ Visualizer Macros,
+%D subtitle=Escaped,
+%D author=Hans Hagen,
+%D date=\currentdate,
+%D copyright=PRAGMA-ADE]
+%C
+%C This module is part of the \CONTEXT\ macro||package and is
+%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
+%C details.
+
+\registerctxluafile{v-escaped.lua}{1.001}
+
+\unprotect
+
+\protect \endinput
diff --git a/tex/context/base/v-lua.lua b/tex/context/base/v-lua.lua
new file mode 100644
index 000000000..af1db27a3
--- /dev/null
+++ b/tex/context/base/v-lua.lua
@@ -0,0 +1,238 @@
+if not modules then modules = { } end modules ['v-lua'] = {
+ version = 1.001,
+ comment = "companion to v-lua.mkiv",
+ author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
+ copyright = "PRAGMA ADE / ConTeXt Development Team",
+ license = "see context related readme files"
+}
+
+-- borrowed from scite
+--
+-- depricated:
+--
+-- gcinfo unpack getfenv setfenv loadlib
+-- table.maxn table.getn table.setn
+-- math.log10 math.mod math.modf math.fmod
+
+local format, tohash = string.format, table.tohash
+local P, S, V, patterns = lpeg.P, lpeg.S, lpeg.V, lpeg.patterns
+local C, Cs, Cg, Cb, Cmt, Carg = lpeg.C, lpeg.Cs, lpeg.Cg, lpeg.Cb, lpeg.Cmt, lpeg.Carg
+
+local core = tohash {
+ "and", "break", "do", "else", "elseif", "end", "false", "for", "function",
+ "if", "in", "local", "nil", "not", "or", "repeat", "return", "then",
+ "true", "until", "while"
+}
+
+local base = tohash {
+ "assert", "collectgarbage", "dofile", "error", "loadfile",
+ "loadstring", "print", "rawget", "rawset", "require", "tonumber",
+ "tostring", "type", "_G", "getmetatable", "ipairs", "next", "pairs",
+ "pcall", "rawequal", "setmetatable", "xpcall", "module", "select",
+}
+
+local libraries = {
+ coroutine = tohash {
+ "create", "resume", "status", "wrap", "yield", "running",
+ },
+ package = tohash{
+ "cpath", "loaded", "loadlib", "path", "config", "preload", "seeall",
+ },
+ io = tohash{
+ "close", "flush", "input", "lines", "open", "output", "read", "tmpfile",
+ "type", "write", "stdin", "stdout", "stderr", "popen",
+ },
+ math = tohash{
+ "abs", "acos", "asin", "atan", "atan2", "ceil", "cos", "deg", "exp",
+ "floor ", "ldexp", "log", "max", "min", "pi", "pow", "rad", "random",
+ "randomseed", "sin", "sqrt", "tan", "cosh", "sinh", "tanh", "huge",
+ },
+ string = tohash{
+ "byte", "char", "dump", "find", "len", "lower", "rep", "sub", "upper",
+ "format", "gfind", "gsub", "gmatch", "match", "reverse",
+ },
+ table = tohash{
+ "concat", "foreach", "foreachi", "sort", "insert", "remove", "pack",
+ "unpack",
+ },
+ os = tohash{
+ "clock", "date", "difftime", "execute", "exit", "getenv", "remove",
+ "rename", "setlocale", "time", "tmpname",
+ },
+ lpeg = tohash{
+ "print", "match", "locale", "type", "version", "setmaxstack",
+ "P", "R", "S", "C", "V", "Cs", "Ct", "Cs", "Cp", "Carg",
+ "Cg", "Cb", "Cmt", "Cf", "B",
+ },
+ -- bit
+ -- debug
+}
+
+local context = context
+local verbatim = context.verbatim
+local makepattern = visualizers.makepattern
+
+local LuaSnippet = context.LuaSnippet
+local startLuaSnippet = context.startLuaSnippet
+local stopLuaSnippet = context.stopLuaSnippet
+
+local LuaSnippetBoundary = verbatim.LuaSnippetBoundary
+local LuaSnippetSpecial = verbatim.LuaSnippetSpecial
+local LuaSnippetComment = verbatim.LuaSnippetComment
+local LuaSnippetNameCore = verbatim.LuaSnippetNameCore
+local LuaSnippetNameBase = verbatim.LuaSnippetNameBase
+local LuaSnippetNameLibraries = verbatim.LuaSnippetNameLibraries
+local LuaSnippetName = verbatim.LuaSnippetName
+
+local namespace
+
+local function visualizename_a(s)
+ if core[s] then
+ namespace = nil
+ LuaSnippetNameCore(s)
+ elseif base[s] then
+ namespace = nil
+ LuaSnippetNameBase(s)
+ else
+ namespace = libraries[s]
+ if namespace then
+ LuaSnippetNameLibraries(s)
+ else
+ LuaSnippetName(s)
+ end
+ end
+end
+
+local function visualizename_b(s)
+ if namespace and namespace[s] then
+ namespace = nil
+ LuaSnippetNameLibraries(s)
+ else
+ LuaSnippetName(s)
+ end
+end
+
+local function visualizename_c(s)
+ LuaSnippetName(s)
+end
+
+local handler = visualizers.newhandler {
+ startinline = function() LuaSnippet(false,"{") end,
+ stopinline = function() context("}") end,
+ startdisplay = function() startLuaSnippet() end,
+ stopdisplay = function() stopLuaSnippet() end ,
+ boundary = function(s) LuaSnippetBoundary(s) end,
+ special = function(s) LuaSnippetSpecial (s) end,
+ comment = function(s) LuaSnippetComment (s) end,
+ period = function(s) verbatim(s) end,
+ name_a = visualizename_a,
+ name_b = visualizename_b,
+ name_c = visualizename_c,
+}
+
+local space = patterns.space
+local anything = patterns.anything
+local newline = patterns.newline
+local emptyline = patterns.emptyline
+local beginline = patterns.beginline
+local somecontent = patterns.somecontent
+
+local comment = P("--")
+local name = (patterns.letter + patterns.underscore)
+ * (patterns.letter + patterns.underscore + patterns.digit)^0
+local boundary = S('()[]{}')
+local special = S("-+/*^%=#") + P("..")
+
+-- The following longstring parser is taken from Roberto's documentation
+-- that can be found at http://www.inf.puc-rio.br/~roberto/lpeg/lpeg.html.
+
+local equals = P("=")^0
+local open = P("[") * Cg(equals, "init") * P("[") * P("\n")^-1
+local close = P("]") * C(equals) * P("]")
+local closeeq = Cmt(close * Cb("init"), function(s,i,a,b) return a == b end)
+local longstring = open * Cs((P(1) - closeeq)^0) * close * Carg(1)
+
+--~ local simple = P ( -- here we hook into the handler but it is default so we could use that
+--~ makepattern(handler,"space",space)
+--~ + makepattern(handler,"newline",newline)
+--~ * makepattern(handler,"emptyline",emptyline)
+--~ * makepattern(handler,"beginline",beginline)
+--~ + makepattern(handler,"default",anything)
+--~ )^0
+
+local function long(content,equals,settings)
+ handler.boundary(format("[%s[",equals or ""))
+ visualizers.write(content,settings) -- unhandled
+ handler.boundary(format("]%s]",equals or ""))
+end
+
+local grammar = visualizers.newgrammar("default", { "visualizer",
+--~ emptyline =
+--~ makepattern(handler,"emptyline",emptyline),
+--~ beginline =
+--~ makepattern(handler,"beginline",beginline),
+--~ newline =
+--~ makepattern(handler,"newline",newline),
+--~ space =
+--~ makepattern(handler,"space",space),
+--~ default =
+--~ makepattern(handler,"default",anything),
+--~ line =
+--~ V("newline") * V("emptyline")^0 * V("beginline"),
+--~ whitespace =
+--~ (V("space") + V("line"))^1,
+--~ optionalwhitespace =
+--~ (V("space") + V("line"))^0,
+--~ content =
+--~ makepattern(handler,"default",somecontent),
+
+ sstring =
+ makepattern(handler,"string",patterns.dquote)
+ * (V("whitespace") + makepattern(handler,"default",1-patterns.dquote))^0
+ * makepattern(handler,"string",patterns.dquote),
+ dstring =
+ makepattern(handler,"string",patterns.squote)
+ * (V("whitespace") + makepattern(handler,"default",1-patterns.squote))^0
+ * makepattern(handler,"string",patterns.squote),
+ longstring =
+ longstring / long,
+ comment =
+ makepattern(handler,"comment",comment)
+ * (V("space") + V("content"))^0,
+ longcomment =
+ makepattern(handler,"comment",comment)
+ * longstring / long,
+ name =
+ makepattern(handler,"name_a",name)
+ * ( V("optionalwhitespace")
+ * makepattern(handler,"default",patterns.period)
+ * V("optionalwhitespace")
+ * makepattern(handler,"name_b",name)
+ )^-1
+ * ( V("optionalwhitespace")
+ * makepattern(handler,"default",patterns.period)
+ * V("optionalwhitespace")
+ * makepattern(handler,"name_c",name)
+ )^0,
+
+ pattern =
+ V("longcomment")
+ + V("comment")
+ + V("longstring")
+ + V("dstring")
+ + V("sstring")
+ + V("name")
+ + makepattern(handler,"boundary",boundary)
+ + makepattern(handler,"special",special)
+
+ + V("space")
+ + V("line")
+ + V("default"),
+
+ visualizer =
+ V("pattern")^1
+} )
+
+local parser = P(grammar)
+
+visualizers.register("lua", { parser = parser, handler = handler, grammar = grammar } )
diff --git a/tex/context/base/v-lua.mkiv b/tex/context/base/v-lua.mkiv
new file mode 100644
index 000000000..6bfe4a963
--- /dev/null
+++ b/tex/context/base/v-lua.mkiv
@@ -0,0 +1,61 @@
+%D \module
+%D [ file=v-lua,
+%D version=2010.10.19,
+%D title=\CONTEXT\ Visualizer Macros,
+%D subtitle=\LUA,
+%D author=Hans Hagen,
+%D date=\currentdate,
+%D copyright=PRAGMA-ADE]
+%C
+%C This module is part of the \CONTEXT\ macro||package and is
+%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
+%C details.
+
+\registerctxluafile{v-lua.lua}{1.001}
+
+\unprotect
+
+\definestartstop
+ [LuaSnippet]
+ [DefaultSnippet]
+
+\definestartstop
+ [LuaSnippetName]
+ [\c!color=,
+ \c!style=boldface]
+
+\definestartstop
+ [LuaSnippetNameCore]
+ [\c!color=darkgreen,
+ \c!style=boldface]
+
+\definestartstop
+ [LuaSnippetNameBase]
+ [\c!color=darkgreen,
+ \c!style=boldface]
+
+\definestartstop
+ [LuaSnippetNameLibraries]
+ [\c!color=darkgreen,
+ \c!style=boldface]
+
+\definestartstop
+ [LuaSnippetBoundary]
+ [\c!color=darkblue,
+ \c!style=boldface]
+
+\definestartstop
+ [LuaSnippetSpecial]
+ [\c!color=darkred,
+ \c!style=boldface]
+
+\definestartstop
+ [LuaSnippetComment]
+ [\c!color=darkyellow,
+ \c!style=boldface]
+
+\definetyping
+ [LUA]
+ [\c!option=lua]
+
+\protect \endinput
diff --git a/tex/context/base/v-nested.lua b/tex/context/base/v-nested.lua
new file mode 100644
index 000000000..e37e1bb58
--- /dev/null
+++ b/tex/context/base/v-nested.lua
@@ -0,0 +1,80 @@
+if not modules then modules = { } end modules ['v-nested'] = {
+ version = 1.001,
+ comment = "companion to v-nested.mkiv",
+ author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
+ copyright = "PRAGMA ADE / ConTeXt Development Team",
+ license = "see context related readme files"
+}
+
+local lpegmatch, patterns = lpeg.match, lpeg.patterns
+local P, V, Carg = lpeg.P, lpeg.V, lpeg.Carg
+
+local context = context
+local verbatim = context.verbatim
+local variables = interfaces.variables
+
+local makepattern = visualizers.makepattern
+local getvisualizer = visualizers.getvisualizer
+
+local nested = nil
+
+local donestedtypingstart = context.donestedtypingstart
+local donestedtypingstop = context.donestedtypingstop
+
+local v_none = variables.none
+local v_slanted = variables.slanted
+
+local handler = visualizers.newhandler {
+ initialize = function(settings)
+ local option = settings and settings.option
+ if not option or option == "" then
+ nested = nil
+ elseif option == v_slanted then
+ nested = nil
+ elseif option == v_none then
+ nested = nil
+ else
+ nested = getvisualizer(option,"direct")
+ end
+ end,
+ open = function()
+ donestedtypingstart()
+ end,
+ close = function()
+ donestedtypingstop()
+ end,
+ content = function(s)
+ if nested then
+ nested(s)
+ else
+ verbatim(s)
+ end
+ end,
+}
+
+local open = P("<<")
+local close = P(">>")
+local rest = (1 - open - close - patterns.space - patterns.newline)^1
+
+local grammar = visualizers.newgrammar("default", {
+
+ initialize = patterns.beginofstring * Carg(1) / handler.initialize,
+
+ open = makepattern(handler,"open",open),
+ close = makepattern(handler,"close",close),
+ rest = makepattern(handler,"content",rest),
+
+ nested = V("open") * (V("pattern")^0) * V("close"),
+ pattern = V("line") + V("space") + V("nested") + V("rest"),
+
+ visualizer = V("initialize") * (V("pattern")^1)
+
+} )
+
+local parser = P(grammar)
+
+visualizers.register("nested", { parser = parser, handler = handler, grammar = grammar } )
+
+-- lpeg.match(parser,[[<<tf<<sl>>tf<<sl>>tf>>]]) context.par()
+-- lpeg.match(parser,[[<<tf<<sl<<tf>>sl>>tf>>]]) context.par()
+-- lpeg.match(parser,[[sl<<tf<<sl>>tf>>sl]]) context.par()
diff --git a/tex/context/base/v-nested.mkiv b/tex/context/base/v-nested.mkiv
new file mode 100644
index 000000000..aa68483a3
--- /dev/null
+++ b/tex/context/base/v-nested.mkiv
@@ -0,0 +1,21 @@
+%D \module
+%D [ file=v-nested,
+%D version=2010.10.19,
+%D title=\CONTEXT\ Visualizer Macros,
+%D subtitle=Nested,
+%D author=Hans Hagen,
+%D date=\currentdate,
+%D copyright=PRAGMA-ADE]
+%C
+%C This module is part of the \CONTEXT\ macro||package and is
+%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
+%C details.
+
+\registerctxluafile{v-nested.lua}{1.001}
+
+\unprotect
+
+\unexpanded\def\donestedtypingstart{\bgroup\swaptypeface{}}
+\unexpanded\def\donestedtypingstop {\egroup{}}
+
+\unprotect \endinput
diff --git a/tex/context/base/v-tex.lua b/tex/context/base/v-tex.lua
index 0a97d28d6..0977ca571 100644
--- a/tex/context/base/v-tex.lua
+++ b/tex/context/base/v-tex.lua
@@ -1,55 +1,71 @@
if not modules then modules = { } end modules ['v-tex'] = {
version = 1.001,
- comment = "companion to buff-vis.mkiv",
+ comment = "companion to v-tex.mkiv",
author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
copyright = "PRAGMA ADE / ConTeXt Development Team",
license = "see context related readme files"
}
-local verbatim = context.verbatim
local P, S, V, patterns = lpeg.P, lpeg.S, lpeg.V, lpeg.patterns
-local visualizer = {
- start = function() context.startTexSnippet() end,
- stop = function() context.stopTexSnippet() end ,
- name = function(s) verbatim.TexSnippetName(s) end,
- group = function(s) verbatim.TexSnippetGroup(s) end,
- boundary = function(s) verbatim.TexSnippetBoundary(s) end,
- special = function(s) verbatim.TexSnippetSpecial(s) end,
- comment = function(s) verbatim.TexSnippetComment(s) end,
- default = function(s) verbatim(s) end,
+local context = context
+local verbatim = context.verbatim
+local makepattern = visualizers.makepattern
+
+local TexSnippet = context.TexSnippet
+local startTexSnippet = context.startTexSnippet
+local stopTexSnippet = context.stopTexSnippet
+
+local TexSnippetName = verbatim.TexSnippetName
+local TexSnippetGroup = verbatim.TexSnippetGroup
+local TexSnippetBoundary = verbatim.TexSnippetBoundary
+local TexSnippetSpecial = verbatim.TexSnippetSpecial
+local TexSnippetComment = verbatim.TexSnippetComment
+
+local handler = visualizers.newhandler {
+ startinline = function() TexSnippet(false,"{") end,
+ stopinline = function() context("}") end,
+ startdisplay = function() startTexSnippet() end,
+ stopdisplay = function() stopTexSnippet() end ,
+ name = function(s) TexSnippetName(s) end,
+ group = function(s) TexSnippetGroup(s) end,
+ boundary = function(s) TexSnippetBoundary(s) end,
+ special = function(s) TexSnippetSpecial(s) end,
+ comment = function(s) TexSnippetComment(s) end,
}
--- todo: unicode letters
+-- todo: unicode letters in control sequences (slow as we need to test the nature)
-local comment = S("%")
-local restofline = (1-patterns.newline)^0
-local anything = patterns.anything
-local name = P("\\") * (patterns.letter + S("@!?"))^1
-local escape = P("\\") * (anything - patterns.newline)^-1 -- else we get \n
-local group = S("${}")
-local boundary = S('[]()<>#="')
-local special = S("/^_-&+'`|")
+local comment = S("%")
+local name = P("\\") * (patterns.letter + S("@!?"))^1
+local escape = P("\\") * (patterns.anything - patterns.newline)^-1 -- else we get \n
+local group = S("${}")
+local boundary = S('[]()<>#="')
+local special = S("/^_-&+'`|")
local pattern = visualizers.pattern
-local texvisualizer = P { "process",
- process =
- V("start") * V("content") * V("stop"),
- start =
- pattern(visualizer,"start",patterns.beginofstring),
- stop =
- pattern(visualizer,"stop",patterns.endofstring),
- content = (
- pattern(visualizer,"comment",comment)
- * pattern(visualizer,"default",restofline)
- + pattern(visualizer,"name",name)
- + pattern(visualizer,"name",escape)
- + pattern(visualizer,"group",group)
- + pattern(visualizer,"boundary",boundary)
- + pattern(visualizer,"special",special)
- + pattern(visualizer,"default",anything)
- )^1
-}
+local grammar = visualizers.newgrammar("default", { "visualizer",
+
+ comment = makepattern(handler,"comment",comment)
+ * (V("space") + V("content"))^0,
+ name = makepattern(handler,"name",name),
+ escape = makepattern(handler,"name",escape),
+ group = makepattern(handler,"group",group),
+ boundary = makepattern(handler,"boundary",boundary),
+ special = makepattern(handler,"special",special),
+
+ pattern =
+ V("comment") + V("name") + V("escape") + V("group") + V("boundary") + V("special")
+ + V("newline") * V("emptyline")^0 * V("beginline")
+ + V("space")
+ + V("default"),
+
+ visualizer =
+ V("pattern")^1
+
+} )
+
+local parser = P(grammar)
-return texvisualizer
+visualizers.register("tex", { parser = parser, handler = handler, grammar = grammar } )
diff --git a/tex/context/base/v-tex.mkiv b/tex/context/base/v-tex.mkiv
index 57aab2272..5cd5753bb 100644
--- a/tex/context/base/v-tex.mkiv
+++ b/tex/context/base/v-tex.mkiv
@@ -1,34 +1,51 @@
-% todo: global
+%D \module
+%D [ file=v-tex,
+%D version=2010.10.19,
+%D title=\CONTEXT\ Visualizer Macros,
+%D subtitle=\TEX,
+%D author=Hans Hagen,
+%D date=\currentdate,
+%D copyright=PRAGMA-ADE]
+%C
+%C This module is part of the \CONTEXT\ macro||package and is
+%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
+%C details.
+
+\registerctxluafile{v-tex.lua}{1.001}
\unprotect
\definestartstop
[TexSnippet]
- [\c!style=\tt]
+ [DefaultSnippet]
\definestartstop
[TexSnippetName]
[\c!color=darkgreen,
- \c!style=bold]
+ \c!style=boldface]
\definestartstop
[TexSnippetGroup]
[\c!color=darkred,
- \c!style=bold]
+ \c!style=boldface]
\definestartstop
[TexSnippetBoundary]
[\c!color=darkblue,
- \c!style=bold]
+ \c!style=boldface]
\definestartstop
[TexSnippetSpecial]
[\c!color=darkyellow,
- \c!style=bold]
+ \c!style=boldface]
\definestartstop
[TexSnippetComment]
[\c!color=darkyellow,
- \c!style=bold]
+ \c!style=boldface]
+
+\definetyping
+ [TEX]
+ [\c!option=tex]
\protect \endinput
diff --git a/tex/context/base/v-xml.lua b/tex/context/base/v-xml.lua
new file mode 100644
index 000000000..bf6945e26
--- /dev/null
+++ b/tex/context/base/v-xml.lua
@@ -0,0 +1,133 @@
+if not modules then modules = { } end modules ['v-xml'] = {
+ version = 1.001,
+ comment = "companion to v-xml.mkiv",
+ author = "Hans Hagen, PRAGMA-ADE, Hasselt NL",
+ copyright = "PRAGMA ADE / ConTeXt Development Team",
+ license = "see context related readme files"
+}
+
+local P, S, V, patterns = lpeg.P, lpeg.S, lpeg.V, lpeg.patterns
+
+local context = context
+local verbatim = context.verbatim
+local makepattern = visualizers.makepattern
+
+local XmlSnippet = context.XmlSnippet
+local startXmlSnippet = context.startXmlSnippet
+local stopXmlSnippet = context.stopXmlSnippet
+
+local XmlSnippetName = verbatim.XmlSnippetName
+local XmlSnippetKey = verbatim.XmlSnippetKey
+local XmlSnippetBoundary = verbatim.XmlSnippetBoundary
+local XmlSnippetString = verbatim.XmlSnippetString
+local XmlSnippetEqual = verbatim.XmlSnippetEqual
+local XmlSnippetEntity = verbatim.XmlSnippetEntity
+local XmlSnippetComment = verbatim.XmlSnippetComment
+local XmlSnippetCdata = verbatim.XmlSnippetCdata
+
+local handler = visualizers.newhandler {
+ startinline = function() XmlSnippet(false,"{") end,
+ stopinline = function() context("}") end,
+ startdisplay = function() startXmlSnippet() end,
+ stopdisplay = function() stopXmlSnippet () end,
+ name = function(s) XmlSnippetName(s) end,
+ key = function(s) XmlSnippetKey(s) end,
+ boundary = function(s) XmlSnippetBoundary(s) end,
+ string = function(s) XmlSnippetString(s) end,
+ equal = function(s) XmlSnippetEqual(s) end,
+ entity = function(s) XmlSnippetEntity(s) end,
+ comment = function(s) XmlSnippetComment(s) end,
+ cdata = function(s) XmlSnippetCdata(s) end,
+}
+
+local comment = P("--")
+local name = (patterns.letter + patterns.digit + S('_-.'))^1
+local entity = P("&") * (1-P(";"))^1 * P(";")
+local openbegin = P("<")
+local openend = P("</")
+local closebegin = P("/>") + P(">")
+local closeend = P(">")
+local opencomment = P("<!--")
+local closecomment = P("-->")
+local openinstruction = P("<?")
+local closeinstruction = P("?>")
+local opencdata = P("<![CDATA[")
+local closecdata = P("]]>")
+
+local grammar = visualizers.newgrammar("default", { "visualizer",
+ sstring =
+ makepattern(handler,"string",patterns.dquote)
+ * (V("whitespace") + makepattern(handler,"default",1-patterns.dquote))^0
+ * makepattern(handler,"string",patterns.dquote),
+ dstring =
+ makepattern(handler,"string",patterns.squote)
+ * (V("whitespace") + makepattern(handler,"default",1-patterns.squote))^0
+ * makepattern(handler,"string",patterns.squote),
+ entity =
+ makepattern(handler,"entity",entity),
+ name =
+ makepattern(handler,"name",name)
+ * (
+ makepattern(handler,"default",patterns.colon)
+ * makepattern(handler,"name",name)
+ )^0,
+ key =
+ makepattern(handler,"key",name)
+ * (
+ makepattern(handler,"default",patterns.colon)
+ * makepattern(handler,"key",name)
+ )^0,
+ attributes = (
+ V("optionalwhitespace")
+ * V("key")
+ * V("optionalwhitespace")
+ * makepattern(handler,"equal",patterns.equal)
+ * V("optionalwhitespace")
+ * (V("dstring") + V("sstring"))
+ * V("optionalwhitespace")
+ )^0,
+ open =
+ makepattern(handler,"boundary",openbegin)
+ * V("name")
+ * V("optionalwhitespace")
+ * V("attributes")
+ * makepattern(handler,"boundary",closebegin),
+ close =
+ makepattern(handler,"boundary",openend)
+ * V("name")
+ * V("optionalwhitespace")
+ * makepattern(handler,"boundary",closeend),
+ comment =
+ makepattern(handler,"boundary",opencomment)
+ * (V("whitespace") + makepattern(handler,"comment",(1-closecomment)))^0 -- slow
+ * makepattern(handler,"boundary",closecomment),
+ cdata =
+ makepattern(handler,"boundary",opencdata)
+ * (V("whitespace") + makepattern(handler,"comment",(1-closecdata)))^0 -- slow
+ * makepattern(handler,"boundary",closecdata),
+ instruction =
+ makepattern(handler,"boundary",openinstruction)
+ * V("name")
+ * V("optionalwhitespace")
+ * V("attributes")
+ * V("optionalwhitespace")
+ * makepattern(handler,"boundary",closeinstruction),
+
+ pattern =
+ V("comment")
+ + V("instruction")
+ + V("cdata")
+ + V("close")
+ + V("open")
+ + V("entity")
+ + V("space")
+ + V("line")
+ + V("default"),
+
+ visualizer =
+ V("pattern")^1
+} )
+
+local parser = P(grammar)
+
+visualizers.register("xml", { parser = parser, handler = handler, grammar = grammar } )
diff --git a/tex/context/base/v-xml.mkiv b/tex/context/base/v-xml.mkiv
new file mode 100644
index 000000000..3cc9244f2
--- /dev/null
+++ b/tex/context/base/v-xml.mkiv
@@ -0,0 +1,71 @@
+%D \module
+%D [ file=v-xml,
+%D version=2010.10.19,
+%D title=\CONTEXT\ Visualizer Macros,
+%D subtitle=\XML,
+%D author=Hans Hagen,
+%D date=\currentdate,
+%D copyright=PRAGMA-ADE]
+%C
+%C This module is part of the \CONTEXT\ macro||package and is
+%C therefore copyrighted by \PRAGMA. See mreadme.pdf for
+%C details.
+
+\registerctxluafile{v-xml.lua}{1.001}
+
+\unprotect
+
+\definestartstop
+ [XmlSnippet]
+ [DefaultSnippet]
+
+\definestartstop
+ [XmlSnippetName]
+ [\c!color=darkblue,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetKey]
+ [\c!color=darkgreen,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetBoundary]
+ [\c!color=darkblue,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetEntity]
+ [\c!color=darkred,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetString]
+ [\c!color=darkblue,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetEqual]
+ [\c!color=darkblue,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetComment]
+ [\c!color=darkyellow,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetCdata]
+ [\c!color=darkyellow,
+ \c!style=boldface]
+
+\definestartstop
+ [XmlSnippetInstruction]
+ [\c!color=darkyellow,
+ \c!style=boldface]
+
+\definetyping
+ [XML]
+ [\c!option=xml]
+
+\protect \endinput
diff --git a/tex/generic/context/luatex-fonts-merged.lua b/tex/generic/context/luatex-fonts-merged.lua
index 0209d16bf..b50799488 100644
--- a/tex/generic/context/luatex-fonts-merged.lua
+++ b/tex/generic/context/luatex-fonts-merged.lua
@@ -1,6 +1,6 @@
-- merged file : luatex-fonts-merged.lua
-- parent file : luatex-fonts.lua
--- merge date : 11/20/10 12:51:46
+-- merge date : 11/26/10 21:21:25
do -- begin closure to overcome local limits and interference
@@ -143,6 +143,9 @@ local lpeg = require("lpeg")
local type = type
+-- Beware, we predefine a bunch of patterns here and one reason for doing so
+-- is that we get consistent behaviour in some of the visualizers.
+
lpeg.patterns = lpeg.patterns or { } -- so that we can share
local patterns = lpeg.patterns
@@ -159,19 +162,38 @@ local alwaysmatched = P(true)
patterns.anything = anything
patterns.endofstring = endofstring
patterns.beginofstring = alwaysmatched
+patterns.alwaysmatched = alwaysmatched
local digit, sign = R('09'), S('+-')
local cr, lf, crlf = P("\r"), P("\n"), P("\r\n")
+local newline = crlf + cr + lf
local utf8next = R("\128\191")
local escaped = P("\\") * anything
local squote = P("'")
local dquote = P('"')
+local space = P(" ")
+
+patterns.somecontent = (anything - newline - space)^1
+patterns.beginline = #(1-newline)
+
+local utfbom_32_be = P('\000\000\254\255')
+local utfbom_32_le = P('\255\254\000\000')
+local utfbom_16_be = P('\255\254')
+local utfbom_16_le = P('\254\255')
+local utfbom_8 = P('\239\187\191')
+local utfbom = utfbom_32_be + utfbom_32_le
+ + utfbom_16_be + utfbom_16_le
+ + utfbom_8
+local utftype = utfbom_32_be / "utf-32-be" + utfbom_32_le / "utf-32-le"
+ + utfbom_16_be / "utf-16-be" + utfbom_16_le / "utf-16-le"
+ + utfbom_8 / "utf-8" + alwaysmatched / "unknown"
patterns.utf8one = R("\000\127")
patterns.utf8two = R("\194\223") * utf8next
patterns.utf8three = R("\224\239") * utf8next * utf8next
patterns.utf8four = R("\240\244") * utf8next * utf8next * utf8next
-patterns.utfbom = P('\000\000\254\255') + P('\255\254\000\000') + P('\255\254') + P('\254\255') + P('\239\187\191')
+patterns.utfbom = utfbom
+patterns.utftype = utftype
local utf8char = patterns.utf8one + patterns.utf8two + patterns.utf8three + patterns.utf8four
local validutf8char = utf8char^0 * endofstring * Cc(true) + Cc(false)
@@ -197,24 +219,30 @@ patterns.hexadecimal = P("0x") * R("09","AF","af")^1
patterns.lowercase = R("az")
patterns.uppercase = R("AZ")
patterns.letter = patterns.lowercase + patterns.uppercase
-patterns.space = P(" ")
+patterns.space = space
patterns.tab = P("\t")
patterns.spaceortab = patterns.space + patterns.tab
patterns.eol = S("\n\r")
patterns.spacer = S(" \t\f\v") -- + string.char(0xc2, 0xa0) if we want utf (cf mail roberto)
-patterns.newline = crlf + cr + lf
-patterns.nonspace = 1 - patterns.space
+patterns.newline = newline
+patterns.emptyline = newline^1
patterns.nonspacer = 1 - patterns.spacer
patterns.whitespace = patterns.eol + patterns.spacer
patterns.nonwhitespace = 1 - patterns.whitespace
+patterns.equal = P("=")
patterns.comma = P(",")
patterns.commaspacer = P(",") * patterns.spacer^0
patterns.period = P(".")
+patterns.colon = P(":")
+patterns.semicolon = P(";")
+patterns.underscore = P("_")
patterns.escaped = escaped
patterns.squote = squote
patterns.dquote = dquote
-patterns.undouble = (dquote/"") * ((escaped + (1-dquote))^0) * (dquote/"")
-patterns.unsingle = (squote/"") * ((escaped + (1-squote))^0) * (squote/"")
+patterns.nosquote = (escaped + (1-squote))^0
+patterns.nodquote = (escaped + (1-dquote))^0
+patterns.unsingle = (squote/"") * patterns.nosquote * (squote/"")
+patterns.undouble = (dquote/"") * patterns.nodquote * (dquote/"")
patterns.unquoted = patterns.undouble + patterns.unsingle -- more often undouble
patterns.unspacer = ((patterns.spacer^1)/"")^0
@@ -237,19 +265,6 @@ function lpeg.splitter(pattern, action)
return (((1-P(pattern))^1)/action+1)^0
end
-local spacing = patterns.spacer^0 * patterns.newline -- sort of strip
-local empty = spacing * Cc("")
-local nonempty = Cs((1-spacing)^1) * spacing^-1
-local content = (empty + nonempty)^1
-
-local capture = Ct(content^0)
-
-function string.splitlines(str)
- return match(capture,str)
-end
-
-patterns.textline = content
-
local splitters_s, splitters_m = { }, { }
local function splitat(separator,single)
@@ -296,6 +311,35 @@ function string.split(str,separator)
return match(c,str)
end
+local spacing = patterns.spacer^0 * newline -- sort of strip
+local empty = spacing * Cc("")
+local nonempty = Cs((1-spacing)^1) * spacing^-1
+local content = (empty + nonempty)^1
+
+patterns.textline = content
+
+--~ local linesplitter = Ct(content^0)
+--~
+--~ function string.splitlines(str)
+--~ return match(linesplitter,str)
+--~ end
+
+local linesplitter = Ct(splitat(newline))
+
+patterns.linesplitter = linesplitter
+
+function string.splitlines(str)
+ return match(linesplitter,str)
+end
+
+local utflinesplitter = utfbom^-1 * Ct(splitat(newline))
+
+patterns.utflinesplitter = utflinesplitter
+
+function string.utfsplitlines(str)
+ return match(utflinesplitter,str)
+end
+
--~ lpeg.splitters = cache -- no longer public
local cache = { }
@@ -8235,7 +8279,7 @@ local contextsetups = specifiers.contextsetups
local contextnumbers = specifiers.contextnumbers
local contextmerged = specifiers.contextmerged
--- we cannot optimize with "start = first_character(head)" because then we don't
+-- we cannot optimize with "start = first_glyph(head)" because then we don't
-- know which rlmode we're in which messes up cursive handling later on
--
-- head is always a whatsit so we can safely assume that head is not changed