mediawiki-extensions-Visual.../modules/parser/pegParser.pegjs.txt

1092 lines
28 KiB
Plaintext
Raw Normal View History

2011-11-02 21:09:19 +00:00
/* Produces output more or less compatible with FakeParser; plug it into FP's output and see */
{
/* Fixme: use static functions to separate module! Unfortunately, this
* does not work:
* var tu = require('./mediawiki.tokenizer.utils.js');
* console.log(tu.flatten([]));
* Using exports in the module gets a bit further, but accesses to
* tu.flatten in productions still fail. Thus, I just moved the functions
* here until a solution is found:
*/
/* Static utilities */
// Flatten a list of lists.
var flatten = function ( e ) {
var es = [];
// flatten sub-arrays
for(var i = 0, length = e.length; i < length; i++) {
var ei = e[i];
if ($.isArray(ei))
es = es.concat(flatten(ei));
else
es.push(ei);
};
return es;
};
// Remove escaped quotes from attributes etc
var unquote = function (quotec, text) {
return text.replace('\\' + quotec, quotec);
};
// Debug print with global switch
2011-11-02 21:09:19 +00:00
var dp = function ( msg ) {
if ( false ) {
console.log(msg);
}
};
var pp = function ( s ) { return JSON.stringify(s, null, 2); }
/*
* Annotate a token stream with list items with appropriate list tokens
*
* @static
* @method
* @param {[tokens]} Token stream with li tokens
* @returns {[tokens]} Token stream, possibly with additional list tokens
* */
var annotateList = function ( tokens ) {
var out = [], // List of tokens
bstack = [], // Bullet stack, previous element's listStyle
bnext = [], // Next element's listStyle
endtags = []; // Stack of end tags
var commonPrefixLength = function (x, y) {
var minLength = Math.min(x.length, y.length);
for(var i = 0; i < minLength; i++) {
if (x[i] != y[i])
break;
}
return i;
};
var pushList = function ( listName, itemName ) {
out.push({type: 'TAG', name: listName});
out.push({type: 'TAG', name: itemName});
endtags.push({type: 'ENDTAG', name: listName});
endtags.push({type: 'ENDTAG', name: itemName});
};
var popTags = function ( n ) {
for(;n > 0; n--) {
// push list item..
out.push(endtags.pop());
// and the list end tag
out.push(endtags.pop());
}
};
var isDlDd = function (a, b) {
var ab = [a,b].sort();
return (ab[0] === ':' && ab[1] === ';');
};
var doListItem = function ( bs, bn ) {
var prefixLen = commonPrefixLength (bs, bn);
var changeLen = bn.length - prefixLen;
var prefix = bn.slice(0, prefixLen);
// emit close tag tokens for closed lists
if (changeLen === 0) {
var itemToken = endtags.pop();
out.push(itemToken);
out.push({type: 'TAG', name: itemToken.name});
endtags.push({type: 'ENDTAG', name: itemToken.name});
} else if ( bs.length == bn.length
&& changeLen == 1
&& isDlDd( bs[prefixLen], bn[prefixLen] ) ) {
// handle dd/dt transitions
out.push(endtags.pop());
if( bn[prefixLen] == ';') {
var newName = 'dt';
} else {
var newName = 'dd';
}
out.push({type: 'TAG', name: newName});
endtags.push({type: 'ENDTAG', name: newName});
} else {
for(var i = prefixLen; i < bn.length; i++) {
switch (bn[i]) {
case '*':
pushList('ul', 'li');
break;
case '#':
pushList('ol', 'li');
break;
case ';':
pushList('dl', 'dt');
break;
case ':':
pushList('dl', 'dd');
break;
default:
throw("Unknown node prefix " + prefix[i]);
}
}
}
};
for (var i = 0, length = tokens.length; i < length; i++) {
var token = tokens[i];
switch ( token.type ) {
case 'TAG':
switch (token.name) {
case 'list':
// ignore token
break;
case 'listItem':
// convert listItem to list and list item tokens
bnext = token.bullets;
doListItem( bstack, bnext );
bstack = bnext;
break;
default:
// pass through all remaining start tags
out.push(token);
break;
}
break;
case 'ENDTAG':
if ( token.name == 'list' ) {
// pop all open list item tokens
popTags(bstack.length);
bstack = "";
} else {
out.push(token);
}
break;
default:
out.push(token);
break;
}
}
return out;
};
/*
* Italic/Bold handling.
*
* - list of tokens
* - NEWLINE
* - ticks (2+) -> list with link in line token list?
* - process on newline
* - need access to text nodes before/after for conversion back to text
*/
var doQuotes = function ( tokens ) {
var italics = [],
bolds = [],
out = [],
inserted = 0;
var convertBold = function ( i ) {
var index = bolds[i];
var txt = out[index - 1];
txt.value += "'";
bolds = bolds.slice(0, i)
.concat(bolds.slice(i + 1, bolds.length - i - 1));
italics.push(index);
italics.sort();
};
// convert italics/bolds into tags
var quotesToTags = function ( offsets, name ) {
var toggle = true;
for (var j = 0; j < offsets.length; j++) {
var t = out[offsets[j]];
if(toggle) {
t.type = 'TAG';
} else {
t.type = 'ENDTAG';
}
t.name = name;
delete t.value;
toggle = !toggle;
}
if (!toggle) {
// add end tag
out.push({type: 'ENDTAG', name: name});
inserted++;
}
toggle = true;
};
for (var i = 0, length = tokens.length; i < length; i++) {
var token = tokens[i];
switch (token.type) {
case 'QUOTE':
// depending on length, add starting 's to preceding text node
// (if any)
// add token index to italic/bold lists
// add placeholder for token
var qlen = token.value.length;
switch (qlen) {
case 2: italics.push(i + inserted); out.push(token); break;
case 3: bolds.push(i + inserted); out.push(token); break;
case 4:
token.value = "'''";
if (i > 0 && tokens[i-1].type === 'TEXT') {
tokens[i-1].value += "'";
} else {
out.push({type: 'TEXT', value: "'"});
inserted++;
}
bolds.push(i + inserted);
out.push(token);
break;
case 5:
// order does not matter here, will be fixed
// by HTML parser backend
bolds.push(i + inserted);
out.push({type: 'QUOTE', value: "'''"});
inserted++;
italics.push(i + inserted);
out.push({type: 'QUOTE', value: "''"});
break;
default: // longer than 5, only use the last 5 ticks
token.value = "'''''";
var newvalue = token.value.substr(0, qlen - 5 );
if (i > 0 && tokens[i-1].type === 'TEXT') {
tokens[i-1].value += newvalue;
} else {
out.push({type: 'TEXT', value: newvalue});
inserted++;
}
bolds.push(i + inserted);
out.push({type: 'QUOTE', value: "'''"});
inserted++;
italics.push(i + inserted);
out.push({type: 'QUOTE', value: "''"});
break;
}
break;
case 'NEWLINE':
// balance out tokens, convert placeholders into tags
if (italics.length % 2 && bolds.length % 2) {
dp("balancing!");
var firstsingleletterword = -1,
firstmultiletterword = -1,
firstspace = -1;
for (var j = 0; j < bolds.length; j++) {
var ticki = bolds[j];
if (ticki > 0 && out[ticki - 1].type === 'TEXT') {
var txt = out[ticki - 1],
lastchar = txt.value[txt.value.length - 1],
secondtolastchar = txt.value[txt.value.length - 2];
dp('txt: ' + pp(txt));
if (lastchar === ' ' && firstspace === -1) {
firstspace = j;
} else if (lastchar !== ' ') {
if ( secondtolastchar === ' ' &&
firstsingleletterword === -1) {
firstsingleletterword = j;
} else if ( secondtolastchar &&
secondtolastchar !== ' ') {
firstmultiletterword = j;
}
}
}
}
// now see if we can convert a bold to an italic and
// an apostrophe
if (firstsingleletterword > -1) {
convertBold(firstsingleletterword);
} else if (firstmultiletterword > -1) {
convertBold(firstmultiletterword);
} else if (firstspace > -1) {
convertBold(firstspace);
}
}
quotesToTags(bolds, 'b');
quotesToTags(italics, 'i');
bolds = [];
italics = [];
out.push(token);
break;
default:
out.push(token);
}
}
return out;
};
/* End static utilities */
/*
* Flags for specific parse environments (inside tables, links etc). Flags
* trigger syntactic stops in the inline_breaks production, which
* terminates inline and attribute matches. Flags merely reduce the number
* of productions needed: The grammar is still context-free as the
* productions can just be unrolled for all combinations of environments
* at the cost of a much larger grammar.
*/
var syntaxFlags = {};
var setFlag = function(flag) {
if (syntaxFlags[flag] !== undefined) {
syntaxFlags[flag]++;
} else {
syntaxFlags[flag] = 1;
2011-11-02 21:09:19 +00:00
}
return true;
};
var clearFlag = function(flag) {
syntaxFlags[flag]--;
};
// Start position of top-level block
// Could also provide positions for lower-level blocks using a stack.
var blockStart = 0;
// cache the input length
var inputLength = input.length;
// pseudo-production that matches at end of input
var isEOF = function (pos) {
return pos === inputLength;
};
2011-11-02 21:09:19 +00:00
}
start
= e:toplevelblock* newline* {
return flatten(e);
}
2011-11-02 21:09:19 +00:00
2011-11-02 21:09:19 +00:00
anyblock = block / inline
anyblockline = block / inlineline
// All chars that cannot start syntactic structures
text = t:[A-Za-z0-9,._ -]+ { return t.join('') }
space
= s:[ \t]+ { return s.join(''); }
// Start of line
sol = (newline / & { return pos === 0; } { return true; })
cn:(c:comment n:newline? { return [c, {type: 'TEXT', value: n}] })* {
return [{type: 'NEWLINE'}].concat(cn);
}
eof = & { return isEOF(pos); } { return true; }
2011-11-02 21:09:19 +00:00
newline
= '\n' / '\r\n'
2011-11-02 21:09:19 +00:00
eolf = newline / eof
toplevelblock
= & { blockStart = pos; return true; } b:block {
b = flatten(b);
var bs = b[0];
dp('toplevelblock:' + pp(b));
if (bs.attribs === undefined) {
bs.attribs = [];
}
bs.attribs.push(['data-sourcePos', blockStart + ':' + pos]);
// XXX: only run this for lines that actually need it!
b.push({type: 'NEWLINE'});
b = doQuotes(b);
return b;
}
2011-11-02 21:09:19 +00:00
block
= block_lines
/ pre
/ comment &eolf
/ para
/ (s:sol {
if (s) {
return [s, {type: 'NEWLINE'}];
} else {
return [{type: 'NEWLINE'}];
}
}
)
block_lines
= s:sol (space* newline)?
bl:block_line { return s.concat(bl); }
// Block structures with start-of-line wiki syntax
block_line
= h
2011-11-02 21:09:19 +00:00
/ table
/ lists
/ pre_indent
/ pre
/ space* !nowiki generic_tag space* &eolf // XXX: only match block elements!
2011-11-02 21:09:19 +00:00
/* Headings */
2011-11-02 21:09:19 +00:00
h = h1 / h2 / h3 / h4 / h5 / h6
h1 = sol '='
(
& { setFlag('h'); return setFlag('h1') }
c:inlineline '=' comment? &newline {
clearFlag('h');
clearFlag('h1');
return [{type: 'TAG', name: 'h1'}]
.concat(c, [{type: 'ENDTAG', name: 'h1'}]);
}
/ { clearFlag('h'); clearFlag('h1'); return null }
)
h2 = sol '=='
(
& { setFlag('h'); return setFlag('h2') }
c:inlineline '==' comment? &newline {
clearFlag('h');
clearFlag('h2');
return [{type: 'TAG', name: 'h2'}]
.concat(c, [{type: 'ENDTAG', name: 'h2'}]);
}
/ { clearFlag('h'); clearFlag('h2'); return null }
)
h3 = sol '==='
(
& { setFlag('h'); return setFlag('h3') }
c:inlineline '===' comment? &newline {
clearFlag('h');
clearFlag('h3');
return [{type: 'TAG', name: 'h3'}]
.concat(c, [{type: 'ENDTAG', name: 'h3'}]);
}
/ { clearFlag('h'); clearFlag('h3'); return null }
)
h4 = sol '===='
(
& { setFlag('h'); return setFlag('h4') }
c:inlineline '====' comment? &newline {
clearFlag('h');
clearFlag('h4');
return [{type: 'TAG', name: 'h4'}]
.concat(c, [{type: 'ENDTAG', name: 'h4'}]);
}
/ { clearFlag('h'); clearFlag('h4'); return null }
)
h5 = sol '====='
(& { setFlag('h'); return setFlag('h5') }
c:inlineline '=====' comment? &newline {
clearFlag('h');
clearFlag('h5');
return [{type: 'TAG', name: 'h5'}]
.concat(c, [{type: 'ENDTAG', name: 'h5'}]);
}
/ { clearFlag('h'); clearFlag('h5'); return null }
)
h6 = sol '======'
(& { setFlag('h'); return setFlag('h6') }
c:inlineline '======' comment? &newline {
clearFlag('h');
clearFlag('h6');
return [{type: 'TAG', name: 'h6'}]
.concat(c, [{type: 'ENDTAG', name: 'h6'}]);
}
/ { clearFlag('h'); clearFlag('h6'); return null }
)
2011-11-02 21:09:19 +00:00
heading_marker
= '=' '='*
heading_text
= h:( !(heading_marker newline) x:inlineline { return x } )* { return h.join(''); }
2011-11-02 21:09:19 +00:00
// TODO: convert inline content to annotations!
2011-11-02 21:09:19 +00:00
para
= (sol br)? pl:para_line pls:(!block_lines para_line)* {
return [{type: 'TAG', name: 'p'}]
.concat([pl], pls, [{type: 'ENDTAG', name: 'p'}]);
}
para_line
= s:sol? c:inlineline {
var res = [];
if (s !== '') {
return s.concat(c);
} else {
return c;
}
}
br = space* &newline { return {type: 'SELFCLOSINGTAG', name: 'br'} }
pre_indent
= l:pre_indent_line ls:(sol pre_indent_line)* {
return [{type: 'TAG', name: 'pre'}]
.concat( [l], ls
, [{type: 'ENDTAG', name: 'pre'}]);
}
pre_indent_line = space l:inlineline {
return [{type: 'TEXT', value: '\n'}].concat(l);
}
// Syntax that stops inline expansion
inline_breaks
= //& { console.log(pp(syntaxFlags)); return true; }
& { return syntaxFlags['table']; }
a:(newline [!|] / '||' / '!!' / '|}') { dp("table break" + pp(a)); return true; }
/ & { return syntaxFlags['linkdesc']; } link_end { return true; }
/ & { return syntaxFlags['h']; }
( & { return syntaxFlags['h1'] } '=' newline { return true; }
/ & { return syntaxFlags['h2'] } '==' newline { return true; }
/ & { return syntaxFlags['h3'] } '===' newline { return true; }
/ & { return syntaxFlags['h4'] } '====' newline { return true; }
/ & { return syntaxFlags['h5'] } '=====' newline { return true; }
/ & { return syntaxFlags['h6'] } '======' newline { return true; }
)
2011-11-02 21:09:19 +00:00
inline
= c:(text / inline_element / (!inline_breaks ch:. { return ch; }))+ {
var out = [];
var text = [];
c = flatten(c);
for (var i = 0; i < c.length; i++) {
if (typeof c[i] == 'string') {
text.push(c[i]);
} else {
if (text.length) {
out.push({ type: "TEXT", value: text.join('') });
text = [];
}
out.concat(c[i]);
}
}
if (text.length) {
out.push({ type: 'TEXT', value: text.join('') });
2011-11-02 21:09:19 +00:00
}
return out;
2011-11-02 21:09:19 +00:00
}
inlineline
= c:(text / !inline_breaks (inline_element / [^\n]))+ {
var out = [];
var text = [];
c = flatten(c);
for (var i = 0; i < c.length; i++) {
if (typeof c[i] == 'string') {
text.push(c[i]);
} else {
if (text.length) {
out.push({type: 'TEXT', value: text.join('')});
text = [];
}
out.push(c[i]);
}
}
if (text.length) {
out.push({type: 'TEXT', value: text.join('')});
2011-11-02 21:09:19 +00:00
}
//dp('inlineline out:', pp(out));
return out;
2011-11-02 21:09:19 +00:00
}
/* TODO: convert all these to annotations!
* -> need (start, end) offsets within block
*/
2011-11-02 21:09:19 +00:00
inline_element
= comment
/ xmlish_tag
/ extlink
/ template
/ link
/ quote
2011-11-02 21:09:19 +00:00
comment
= '<!--' c:comment_chars* ('-->' / eof)
cs:(space* newline space* cn:comment { return cn })* {
return [{ type: 'COMMENT', value: c.join('') }].concat(cs);
}
2011-11-02 21:09:19 +00:00
comment_chars
= c:[^-] { return c; }
/ c:'-' !'->' { return c; }
2011-11-02 21:09:19 +00:00
extlink
= "[" target:url " " text:extlink_text "]" {
return [ { type: 'TAG',
name: 'a',
attribs: [['href', target]] }
, {type: 'TEXT', value: text}
, {type: 'ENDTAG', name: 'a'}];
}
2011-11-02 21:09:19 +00:00
// = "[" target:url text:extlink_text "]" { return { type: 'extlink', target: target, text: text } }
url
= proto:"http:" rest:([^ \]]+) { return proto + rest.join(''); }
extlink_text
= c:[^\]]+ { return c.join(''); }
template
= "{{" target:link_target params:("|" p:template_param { return p })* "}}" {
var obj = { type: 'SELFCLOSINGTAG', name: 'template', attribs: [['target', target]] }
if (params && params.length) {
obj.attribs.push(params);
}
return obj;
}
2011-11-02 21:09:19 +00:00
template_param
= name:template_param_name "=" c:template_param_text {
return [name, c];
} / c:template_param_text {
return [null, c];
}
2011-11-02 21:09:19 +00:00
tplarg
= "{{{" name:link_target params:("|" p:template_param { return p })* "}}}" {
var obj = {
type: 'SELFCLOSINGTAG',
name: 'templatearg',
attribs: [['argname', name]]
};
if (params && params.length) {
obj.attribs.push(params);
}
return obj;
}
2011-11-02 21:09:19 +00:00
template_param_name
= h:( !"}}" x:([^=|]) { return x } )* { return h.join(''); }
template_param_text
= template_param_text_chunk*
/* = h:( !"}}" x:([^|]) { return x } )* { return h.join(''); }*/
template_param_text_chunk
= comment
/ xmlish_tag
/ extlink
/ template
/ link
/ quote
2011-11-02 21:09:19 +00:00
/ !"}}" x:([^|]) { return x }
link
= "[[" target:link_target text:("|" link_text)* "]]" {
var obj = {
type: 'TAG',
name: 'a',
attribs: [['data-type', 'internal']]
};
2011-11-18 14:48:32 +00:00
obj.attribs.push(['href', target]);
if (text && text.length) {
2011-11-18 14:48:32 +00:00
var textTokens = text[0][1]; // XXX
} else {
var textTokens = [{type: 'TEXT', value: target}];
}
2011-11-18 14:48:32 +00:00
return [obj].concat(textTokens, [{type: 'ENDTAG', name: 'a'}]);
}
2011-11-02 21:09:19 +00:00
link_target
= h:( !"]]" x:([^|]) { return x } )* { return h.join(''); }
link_text
= h:( & { return setFlag('linkdesc'); }
x:inlineline { return x }
)* {
clearFlag('linkdesc')
2011-11-18 14:48:32 +00:00
return h;
}
/ & { clearFlag('linkdesc') } { return null; }
link_end = "]]"
2011-11-02 21:09:19 +00:00
/* Generic quote production for italic and bold, further processed in a token
* stream transformation in doQuotes. Relies on NEWLINE tokens being emitted
* for each line of text to balance quotes per line.
*
* We are not using a simple pair rule here as we need to support mis-nested
* bolds/italics and MediaWiki's special heuristics for apostrophes, which are
* all not context free. */
quote = "''" x:"'"* {
return {
type : 'QUOTE',
value: "''" + x.join('')
}
}
2011-11-02 21:09:19 +00:00
/* Will need to check anything xmlish agains known/allowed HTML tags and
* registered extensions, otherwise fail the match. Should ref be treated as a
* regular extension? */
xmlish_tag = nowiki / ref / references / generic_tag
pre
= "<pre"
attribs:generic_attribute*
">"
ts:(t1:[^<]+ { return {type:'TEXT',value:t1.join('')} }
/ nowiki
/ !"</pre>" t2:. {return {type:'TEXT',value:t2}})+
("</pre>" / eof) {
// return nowiki tags as well?
//console.log('inpre');
return [ {type: 'TAG', name: 'pre', attribs: attribs} ]
.concat(ts, [{type: 'ENDTAG', name: 'pre'}]);
}
/ "</pre>" { return {type: 'TEXT', value: "</pre>"}; }
nowiki
= "<nowiki>" nc:nowiki_content "</nowiki>" { return nc }
/ "<nowiki>" {
// console.log('nowiki fallback');
return [{type: 'TEXT', value: '<nowiki>'}];
}
/ "</nowiki>" { return [{type: 'TEXT', value: '</nowiki>'}]; }
nowiki_content
= ts:( t:[^<]+ { return t.join('') }
/ "<pre" p0:space* p1:[^>]* ">" p2:nowiki_content "</pre>" {
//console.log('nested pre in nowiki');
return ["<pre"].concat(p0, p1, [">"], [p2[0].value], ["</pre>"]).join('');
}
/ (!("</nowiki>" / "</pre>") c:. {return c}) )* {
// return nowiki tags as well?
return [{type: 'TEXT', value: ts.join('')}];
}
// See http://dev.w3.org/html5/spec/Overview.html#syntax-tag-name and
// following paragraphs
generic_tag
= "<" end:"/"? name:[0-9a-zA-Z]+
attribs:generic_attribute*
selfclose:"/"?
">" {
var res = {name: name.join(''), attribs: attribs};
if ( end != '' ) {
res.type = 'ENDTAG';
} else if ( selfclose != '' ) {
res.type = 'SELFCLOSINGTAG';
} else {
res.type = 'TAG';
}
return res;
}
generic_attribute
= s:space*
name:generic_attribute_name
value:(space*
generic_attribute_value)?
{
if ( value !== '' ) {
return [name, value[1]];
} else {
return [name,''];
}
}
// http://dev.w3.org/html5/spec/Overview.html#attributes-0
generic_attribute_name
= n:[^ \t\0/"'>=]+ {
return n.join('');
}
generic_attribute_value
= "=" space* v:att_value {return v}
att_value
= t:[^ \t'"<>=']+ { return [null, t.join('')]; }
/ "'" t:[^'>]+ "'" { return unquote("'", t.join('')); }
/ '"' t:[^">]+ '"' { return unquote('"', t.join('')); }
2011-11-02 21:09:19 +00:00
ref = ref_full / ref_empty
/* Can we do backreferences to genericize this? */
ref_full
= start:ref_start ">" content:ref_content* close:ref_end {
return [
{ type: 'TAG',
name: 'ext',
attribs: [['data-extname', 'ref']]
.concat(start.params, [['data-startws', start.ws]])},
content,
{type: 'ENDTAG', name: 'ref'}
];
2011-11-02 21:09:19 +00:00
}
ref_empty
= start:ref_start close:(space* "/>") {
return [{ type: 'SELFCLOSINGTAG',
name: 'ext',
attribs: [['data-extname', 'ref']]
.concat(start.params
,[['data-startws', start.ws]])
}];
2011-11-02 21:09:19 +00:00
}
ref_start
= "<ref" params:ext_param* ws:space* {
return {
params: params,
ws: ws
};
2011-11-02 21:09:19 +00:00
}
ref_end
= all:("</ref" space* ">") {
return all.join('');
}
ref_content
= !ref_end a:inline { // XXX: ineffective syntactic stop
2011-11-02 21:09:19 +00:00
return a;
}
/* fixme probably have to programatically add these */
references = references_full / references_empty
references_full
= start:references_start ">" content:references_content* close:references_end {
return [
{ type: 'TAG',
name: 'ext',
attribs: [['data-extname', 'references']]
.concat(start.params
,[['data-startws', start.ws]])
},
content,
{ type: 'ENDTAG', name: 'ext' }
];
}
2011-11-02 21:09:19 +00:00
references_empty
= start:references_start close:(space* "/>") {
return
[{ type: 'SELFCLOSINGTAG',
name: 'ext',
attribs: [['data-extname', 'references']]
.concat(start.params
,[['data-startws', start.ws]])
}];
}
2011-11-02 21:09:19 +00:00
references_start
= "<references" params:ext_param* ws:space* {
return {
params: params,
ws: ws
};
}
2011-11-02 21:09:19 +00:00
references_end
= all:("</references" space* ">") {
return all.join('');
}
2011-11-02 21:09:19 +00:00
references_content
= !references_end a:inline {
return a;
}
2011-11-02 21:09:19 +00:00
ext_param
= space* name:ext_param_name "=" val:ext_param_val {
val[0] = name;
return val;
}
2011-11-02 21:09:19 +00:00
ext_param_name
= name:[a-zA-Z0-9-]+ {
return name.join('');
}
ext_param_val
= t:[0-9A-Za-z]+ { return [null, t.join('')]; }
/ "'" t:[^'>]+ "'" { return [null, unquote("'", t.join(''))]; }
/ '"' t:[^">]+ '"' { return [null, unquote('"', t.join(''))]; }
2011-11-02 21:09:19 +00:00
lists = e:(dtdd / li) es:(sol (dtdd / li))*
{
return annotateList( [ { type: 'TAG', name: 'list'} ]
.concat(flatten([e].concat(es))
,[{ type: 'ENDTAG', name: 'list' }]));
2011-11-02 21:09:19 +00:00
}
li = bullets:list_char+
c:inlineline
&eolf
2011-11-02 21:09:19 +00:00
{
return [ { type: 'TAG',
name: 'listItem',
bullets: bullets }
, c ];
2011-11-02 21:09:19 +00:00
}
dtdd = bullets:list_char+
c:(inline_element / (n:[^:\n] { return {type: 'TEXT', value: n}; }))+
2011-11-02 21:09:19 +00:00
":"
d:(inline_element / (n:[^\n] { return {type: 'TEXT', value: n}; }))+
&eolf
2011-11-02 21:09:19 +00:00
{
// reject rule if bullets do not end in semicolon
if (bullets[bullets.length - 1] != ';') {
return null;
} else {
var dtbullets = bullets.slice(0, bullets.length - 1);
dtbullets.push(':');
return [ { type: 'TAG', name: 'listItem', bullets: bullets } ]
.concat( c
,[{ type: 'TAG', name: 'listItem', bullets: dtbullets } ]
, d );
2011-11-02 21:09:19 +00:00
}
}
list_char = [*#:;]
/* Tables */
table
= tas:table_start c:table_caption? b:table_body? table_end {
var res = {type: 'TAG', name: 'table'}
2011-11-02 21:09:19 +00:00
var body = b !== '' ? b : [];
dp("body: " + pp(body));
2011-11-02 21:09:19 +00:00
if (tas.length > 0) {
// FIXME: actually parse and build structure
res.attribs = [['data-unparsed', tas.join('')]];
2011-11-02 21:09:19 +00:00
}
if (c != '') {
var caption = [{type: 'TAG', name: 'caption'}]
.concat(c, [{type: 'ENDTAG', name: 'caption'}]);
} else {
var caption = [];
}
//dp(pp(res));
return [res].concat(caption, body,
[{type: 'ENDTAG', name: 'table'}]);
2011-11-02 21:09:19 +00:00
}
table_start
= sol
"{|"
& { setFlag('table'); return true; }
ta:table_attribs*
space* {
//dp("table_start " + pp(ta) + ", pos:" + pos);
2011-11-02 21:09:19 +00:00
return ta;
}
/ sol "{|" { clearFlag('table'); return null; }
2011-11-02 21:09:19 +00:00
table_attribs
= text / ! inline_breaks !newline .
2011-11-02 21:09:19 +00:00
table_caption
= newline
"|+" c:inline* {
return c;
}
2011-11-02 21:09:19 +00:00
table_body
= & { dp("table_body enter"); return true; }
firstrow:table_firstrow otherrows:table_row* {
/* dp('table first and otherrows: '
* + pp([firstrow].concat(otherrows))); */
2011-11-02 21:09:19 +00:00
return [firstrow].concat(otherrows);
}
/ otherrows:table_row* {
//dp('table otherrows: ' + pp(otherrows));
2011-11-02 21:09:19 +00:00
return otherrows;
}
table_firstrow
= td:table_data+ {
return [{ type: 'TAG', name: 'tr' }]
.concat(td, [{type: 'ENDTAG', name: 'tr'}]);
2011-11-02 21:09:19 +00:00
}
table_row
= & { dp("table row enter"); return true; }
newline
"|-" thtd_attribs? space* td:(table_data / table_header)* {
return [{type: 'TAG', name: 'tr'}]
.concat(td, [{type: 'ENDTAG', name: 'tr'}]);
}
2011-11-02 21:09:19 +00:00
table_data
= & { dp("table_data enter, pos=" + pos); return true; }
("||" / newline "|")
! [}+-]
a:thtd_attribs?
td:(!inline_breaks anyblock)* {
dp("table data result: " + pp(td) + ", attribts: " + pp(a));
return [{ type: 'TAG', name: 'td', attribs: [['data-unparsed', a]]}]
.concat(td, [{type: 'ENDTAG', name: 'td'}]);
}
2011-11-02 21:09:19 +00:00
table_header
= ("!!" / newline "!")
a:thtd_attribs?
c:inline {
return [{type: 'TAG', name: 'th', attribs: [['data-unparsed', a]]}]
.concat(c, [{type: 'ENDTAG', name: 'th'}]);
}
thtd_attribs
// In particular, do not match [|\n]
= a:(text / ! inline_breaks [="':;/,.-] )+ "|" ! [|}+-] {
return a;
}
2011-11-02 21:09:19 +00:00
table_end = newline? "|}" { clearFlag('table'); }
2011-11-02 21:09:19 +00:00
/* Wikidom TODO:
* split off text into content nodes
* convert inlines into annotations
* change contents into children
*
* { text: text,
* annotations: [(normal annotations)],
* maybeannotations: [
* { type: 'something',
* side: MA_START,
* tag: { start: x, length: y }
* }
* ]
* }
* offsets in annotations: presume maybeannotations are actually text
* -> need to transform annotations if match found
* -> format annotations, comments can run to the end (re-opened after
* block-level tags); only closed on table cells, object,?
* -> other annotations (images, templates etc) are limited by block-level
* elements, tightly bound
*
* Block-level elements
* --------------------
* - Need some early clean-up to provide structure and offsets
* - Establish scope limits for some inlines
* - Line-based balanced by construction
* - HTML tags need balancing/ matching / implicit close
* - content in illegal places (e.g. between table and td tags) needs foster
* parenting
* - grammar will match outermost pair if unmatched pairs are recognized as
* tokens (or as text)
* - post-processing needed, but has to be limited by scope
2011-11-02 21:09:19 +00:00
*/
/* Tabs do not mix well with the hybrid production syntax */
/* vim: expandtab:ts=4:sw=4:cindent */