mirror of
https://gerrit.wikimedia.org/r/mediawiki/extensions/VisualEditor
synced 2024-11-16 02:51:50 +00:00
d918fa18ac
* Tokens are now immutable. The progress of transformations is tracked on chunks instead of tokens. Tokenizer output is cached and can be directly returned without a need for cloning. Transforms are required to clone or newly create tokens they are modifying. * Expansions per chunk are now shared between equivalent frames via a cache stored on the chunk itself. Equivalence of frames is not yet ideal though, as right now a hash tree of *unexpanded* arguments is used. This should be switched to a hash of the fully expanded local parameters instead. * There is now a vastly improved maybeSyncReturn wrapper for async transforms that either forwards processing to the iterative transformTokens if the current transform is still ongoing, or manages a recursive transformation if needed. * Parameters for parser functions are now wrapped in abstract Params and ParserValue objects, which support some handy on-demand *value* expansions. Keys are always expanded. Parser functions are converted to use these interfaces, and now properly expand their values in the correct frame. Making this expansion lazier is certainly possible, but would complicate transformTokens and other token-handling machinery. Need to investigate if it would really be worth it. Dead branch elimination is certainly a bigger win overall. * Complex recursive asynchronous expansions should now be closer to correct for both the iterative (transformTokens) and recursive (maybeSyncReturn after transformTokens has returned) code paths. * Performance degraded slightly. There are no micro-optimizations done yet and the shared expansion cache still has a low hit rate. The progress tracking on chunks is not yet perfect, so there are likely a lot of unneeded re-expansions that can be easily eliminated. There is also more debug tracing right now. Obama currently expands in 54 seconds on my laptop. Change-Id: I4a603f3d3c70ca657ebda9fbb8570269f943d6b6
339 lines
10 KiB
JavaScript
339 lines
10 KiB
JavaScript
/**
|
|
* This module assembles parser pipelines from parser stages with
|
|
* asynchronous communnication between stages based on events. Apart from the
|
|
* default pipeline which converts WikiText to HTML DOM, it also provides
|
|
* sub-pipelines for the processing of template transclusions.
|
|
*
|
|
* See http://www.mediawiki.org/wiki/Parsoid and
|
|
* http://www.mediawiki.org/wiki/Parsoid/Token_stream_transformations
|
|
* for illustrations of the pipeline architecture.
|
|
*
|
|
* @author Gabriel Wicke <gwicke@wikimedia.org>
|
|
* @author Neil Kandalgaonkar <neilk@wikimedia.org>
|
|
*/
|
|
|
|
// make this global for now
|
|
// XXX: figure out a way to get away without a global for PEG actions!
|
|
$ = require('jquery');
|
|
var events = require( 'events' );
|
|
|
|
var fs = require('fs'),
|
|
path = require('path'),
|
|
PegTokenizer = require('./mediawiki.tokenizer.peg.js').PegTokenizer,
|
|
TokenTransformManager = require('./mediawiki.TokenTransformManager.js'),
|
|
SyncTokenTransformManager = TokenTransformManager.SyncTokenTransformManager,
|
|
AsyncTokenTransformManager = TokenTransformManager.AsyncTokenTransformManager,
|
|
|
|
NoIncludeOnly = require('./ext.core.NoIncludeOnly.js'),
|
|
IncludeOnly = NoIncludeOnly.IncludeOnly,
|
|
NoInclude = NoIncludeOnly.NoInclude,
|
|
OnlyInclude = NoIncludeOnly.OnlyInclude,
|
|
QuoteTransformer = require('./ext.core.QuoteTransformer.js').QuoteTransformer,
|
|
PostExpandParagraphHandler = require('./ext.core.PostExpandParagraphHandler.js')
|
|
.PostExpandParagraphHandler,
|
|
Sanitizer = require('./ext.core.Sanitizer.js').Sanitizer,
|
|
TemplateHandler = require('./ext.core.TemplateHandler.js').TemplateHandler,
|
|
AttributeExpander = require('./ext.core.AttributeExpander.js').AttributeExpander,
|
|
ListHandler = require('./ext.core.ListHandler.js').ListHandler,
|
|
LinkHandler = require('./ext.core.LinkHandler.js'),
|
|
WikiLinkHandler = LinkHandler.WikiLinkHandler,
|
|
ExternalLinkHandler = LinkHandler.ExternalLinkHandler,
|
|
Cite = require('./ext.Cite.js').Cite,
|
|
BehaviorSwitchHandler = require('./ext.core.BehaviorSwitchHandler.js').BehaviorSwitchHandler,
|
|
TreeBuilder = require('./mediawiki.HTML5TreeBuilder.node.js')
|
|
.FauxHTML5.TreeBuilder,
|
|
DOMPostProcessor = require('./mediawiki.DOMPostProcessor.js').DOMPostProcessor;
|
|
|
|
|
|
function ParserPipelineFactory ( env ) {
|
|
this.pipelineCache = {};
|
|
this.env = env;
|
|
}
|
|
|
|
/**
|
|
* Recipe for parser pipelines and -subpipelines, depending on input types.
|
|
*
|
|
* Token stream transformations to register by type and per phase. The
|
|
* possible ranks for individual transformation registrations are [0,1)
|
|
* (excluding 1.0) for sync01, [1,2) for async12 and [2,3) for sync23.
|
|
*
|
|
* Should perhaps be moved to mediawiki.parser.environment.js, so that all
|
|
* configuration can be found in a single place.
|
|
*/
|
|
ParserPipelineFactory.prototype.recipes = {
|
|
// The full wikitext pipeline
|
|
'text/x-mediawiki/full': [
|
|
// Input pipeline including the tokenizer
|
|
'text/x-mediawiki',
|
|
// Final synchronous token transforms and DOM building / processing
|
|
'tokens/x-mediawiki/expanded'
|
|
],
|
|
|
|
// A pipeline from wikitext to expanded tokens. The input pipeline for
|
|
// wikitext.
|
|
'text/x-mediawiki': [
|
|
[ PegTokenizer, [] ],
|
|
'tokens/x-mediawiki'
|
|
],
|
|
|
|
// Synchronous per-input and async token stream transformations. Produces
|
|
// a fully expanded token stream ready for consumption by the
|
|
// tokens/expanded pipeline.
|
|
'tokens/x-mediawiki': [
|
|
// Synchronous in-order per input
|
|
[
|
|
SyncTokenTransformManager,
|
|
[ 1, 'tokens/x-mediawiki' ],
|
|
[
|
|
OnlyInclude,
|
|
IncludeOnly,
|
|
NoInclude
|
|
// Insert TokenCollectors for extensions here (don't expand
|
|
// templates in extension contents); wrap collected tokens in
|
|
// special extension token.
|
|
/* Extension1, */
|
|
/* Extension2, */
|
|
]
|
|
],
|
|
/*
|
|
* Asynchronous out-of-order per input. Each async transform can only
|
|
* operate on a single input token, but can emit multiple output
|
|
* tokens. If multiple tokens need to be collected per-input, then a
|
|
* separate collection transform in sync01 can be used to wrap the
|
|
* collected tokens into a single one later processed in an async12
|
|
* transform.
|
|
*/
|
|
[
|
|
AsyncTokenTransformManager,
|
|
[ 2, 'tokens/x-mediawiki' ],
|
|
[
|
|
TemplateHandler,
|
|
/* ExtensionHandler1, */ // using SFH_OBJECT_ARGS in PHP
|
|
|
|
// Expand attributes after templates to avoid expanding unused branches
|
|
// No expansion of quotes, paragraphs etc in attributes, as in
|
|
// PHP parser- up to text/x-mediawiki/expanded only.
|
|
AttributeExpander,
|
|
|
|
// now all attributes expanded to tokens or string
|
|
|
|
WikiLinkHandler, // more convenient after attribute expansion
|
|
ExternalLinkHandler
|
|
/* ExtensionHandler2, */ // using expanded args
|
|
// Finally expand attributes to plain text
|
|
]
|
|
]
|
|
],
|
|
|
|
// Final stages of main pipeline, operating on fully expanded tokens of
|
|
// potentially mixed origin.
|
|
'tokens/x-mediawiki/expanded': [
|
|
// Synchronous in-order on fully expanded token stream (including
|
|
// expanded templates etc). In order to support mixed input (from
|
|
// wikitext and plain HTML, say) all applicable transforms need to be
|
|
// included here. Input-specific token types avoid any runtime
|
|
// overhead for unused transforms.
|
|
[
|
|
SyncTokenTransformManager,
|
|
[ 3, 'tokens/x-mediawiki/expanded' ],
|
|
[
|
|
// text/wiki-specific tokens
|
|
QuoteTransformer,
|
|
ListHandler,
|
|
|
|
// before transforms that depend on behavior switches
|
|
// examples: toc generation, edit sections
|
|
BehaviorSwitchHandler,
|
|
|
|
// Synchronous extensions
|
|
Cite, // both before and after paragraph handler
|
|
|
|
// Paragraph wrapping
|
|
PostExpandParagraphHandler,
|
|
Sanitizer
|
|
// SkipperUnpacker
|
|
]
|
|
],
|
|
|
|
// Build a tree out of the fully processed token stream
|
|
[ TreeBuilder, [] ],
|
|
|
|
/**
|
|
* Final processing on the HTML DOM.
|
|
*/
|
|
|
|
/* Generic DOM transformer.
|
|
* This currently performs minor tree-dependent clean up like wrapping
|
|
* plain text in paragraphs. For HTML output, it would also be configured
|
|
* to perform more aggressive nesting cleanup.
|
|
*/
|
|
[ DOMPostProcessor, [] ]
|
|
]
|
|
};
|
|
|
|
/**
|
|
* Generic pipeline creation from the above recipes
|
|
*/
|
|
ParserPipelineFactory.prototype.makePipeline = function( type, isInclude, cacheType ) {
|
|
var recipe = this.recipes[type];
|
|
if ( ! recipe ) {
|
|
console.trace();
|
|
throw( 'Error while trying to construct pipeline for ' + type );
|
|
}
|
|
var stages = [];
|
|
for ( var i = 0, l = recipe.length; i < l; i++ ) {
|
|
// create the stage
|
|
var stageData = recipe[i],
|
|
stage;
|
|
|
|
if ( stageData.constructor === String ) {
|
|
// Points to another subpipeline, get it recursively
|
|
stage = this.makePipeline( stageData, isInclude );
|
|
} else {
|
|
stage = Object.create( stageData[0].prototype );
|
|
// call the constructor
|
|
stageData[0].apply( stage, [ this.env, isInclude, this ].concat( stageData[1] ) );
|
|
if ( stageData.length >= 3 ) {
|
|
// Create (and implicitly register) transforms
|
|
var transforms = stageData[2];
|
|
for ( var t = 0; t < transforms.length; t++ ) {
|
|
new transforms[t](stage , isInclude);
|
|
}
|
|
}
|
|
}
|
|
|
|
// connect with previous stage
|
|
if ( i ) {
|
|
stage.addListenersOn( stages[i-1] );
|
|
}
|
|
stages.push( stage );
|
|
}
|
|
//console.warn( 'stages' + stages + JSON.stringify( stages ) );
|
|
return new ParserPipeline(
|
|
stages[0],
|
|
stages[stages.length - 1],
|
|
cacheType ? this.returnPipeline.bind( this, cacheType )
|
|
: null
|
|
);
|
|
};
|
|
|
|
/**
|
|
* Get a subpipeline (not the top-level one) of a given type.
|
|
*
|
|
* Subpipelines are cached as they are frequently created.
|
|
*/
|
|
ParserPipelineFactory.prototype.getPipeline = function ( type, isInclude ) {
|
|
// default to include
|
|
if ( isInclude === undefined ) {
|
|
isInclude = true;
|
|
}
|
|
var pipe,
|
|
cacheType = type;
|
|
if ( ! isInclude ) {
|
|
cacheType += '::noInclude';
|
|
}
|
|
if ( ! this.pipelineCache[cacheType] ) {
|
|
this.pipelineCache[cacheType] = [];
|
|
}
|
|
if ( this.pipelineCache[cacheType].length ) {
|
|
//console.warn( JSON.stringify( this.pipelineCache[cacheType] ));
|
|
return this.pipelineCache[cacheType].pop();
|
|
} else {
|
|
return this.makePipeline( type, isInclude, cacheType );
|
|
}
|
|
};
|
|
|
|
/**
|
|
* Callback called by a pipeline at the end of its processing. Returns the
|
|
* pipeline to the cache.
|
|
*/
|
|
ParserPipelineFactory.prototype.returnPipeline = function ( type, pipe ) {
|
|
pipe.removeAllListeners( 'end' );
|
|
pipe.removeAllListeners( 'chunk' );
|
|
var cache = this.pipelineCache[type];
|
|
if ( cache.length < 8 ) {
|
|
cache.push( pipe );
|
|
}
|
|
};
|
|
|
|
|
|
/******************** ParserPipeline ****************************/
|
|
|
|
/**
|
|
* Wrap some stages into a pipeline. The last member of the pipeline is
|
|
* supposed to emit events, while the first is supposed to support a process()
|
|
* method that sets the pipeline in motion.
|
|
*/
|
|
function ParserPipeline ( first, last, returnToCacheCB ) {
|
|
this.first = first;
|
|
this.last = last;
|
|
|
|
if ( returnToCacheCB ) {
|
|
var self = this;
|
|
this.returnToCacheCB = function () {
|
|
returnToCacheCB( self );
|
|
};
|
|
|
|
// add a callback to return the pipeline back to the cache
|
|
this.last.addListener( 'end', this.returnToCacheCB );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Feed input tokens to the first pipeline stage
|
|
*/
|
|
ParserPipeline.prototype.process = function(input, key) {
|
|
return this.first.process(input, key);
|
|
};
|
|
|
|
/**
|
|
* Set the frame on the last pipeline stage (normally the
|
|
* AsyncTokenTransformManager).
|
|
*/
|
|
ParserPipeline.prototype.setFrame = function(frame, title, args) {
|
|
return this.last.setFrame(frame, title, args);
|
|
};
|
|
|
|
/**
|
|
* Register the first pipeline stage with the last stage from a separate pipeline
|
|
*/
|
|
ParserPipeline.prototype.addListenersOn = function(stage) {
|
|
return this.first.addListenersOn(stage);
|
|
};
|
|
|
|
// Forward the EventEmitter API to this.last
|
|
ParserPipeline.prototype.on = function (ev, cb) {
|
|
return this.last.on(ev, cb);
|
|
};
|
|
ParserPipeline.prototype.once = function (ev, cb) {
|
|
return this.last.once(ev, cb);
|
|
};
|
|
ParserPipeline.prototype.addListener = function(ev, cb) {
|
|
return this.last.addListener(ev, cb);
|
|
};
|
|
ParserPipeline.prototype.removeListener = function(ev, cb) {
|
|
return this.last.removeListener(ev, cb);
|
|
};
|
|
ParserPipeline.prototype.setMaxListeners = function(n) {
|
|
return this.last.setMaxListeners(n);
|
|
};
|
|
ParserPipeline.prototype.listeners = function(ev) {
|
|
return this.last.listeners(ev);
|
|
};
|
|
ParserPipeline.prototype.removeAllListeners = function ( event ) {
|
|
if ( event === 'end' ) {
|
|
this.last.removeAllListeners('end');
|
|
// now re-add the cache callback
|
|
if ( this.returnToCacheCB ) {
|
|
this.last.addListener( 'end', this.returnToCacheCB );
|
|
}
|
|
} else {
|
|
return this.last.removeAllListeners( event );
|
|
}
|
|
};
|
|
|
|
if (typeof module == "object") {
|
|
module.exports.ParserPipeline = ParserPipeline;
|
|
module.exports.ParserPipelineFactory = ParserPipelineFactory;
|
|
}
|