mirror of
https://gerrit.wikimedia.org/r/mediawiki/extensions/VisualEditor
synced 2024-11-15 10:35:48 +00:00
455 lines
15 KiB
JavaScript
455 lines
15 KiB
JavaScript
/**
|
|
* Creates an es.TransactionProcessor object.
|
|
*
|
|
* @class
|
|
* @constructor
|
|
*/
|
|
es.TransactionProcessor = function( model, transaction ) {
|
|
this.model = model;
|
|
this.transaction = transaction;
|
|
this.cursor = 0;
|
|
this.set = [];
|
|
this.clear = [];
|
|
};
|
|
|
|
/* Static Members */
|
|
|
|
es.TransactionProcessor.operationMap = {
|
|
// Retain
|
|
'retain': {
|
|
'commit': function( op ) {
|
|
this.retain( op );
|
|
},
|
|
'rollback': function( op ) {
|
|
this.retain( op );
|
|
}
|
|
},
|
|
// Insert
|
|
'insert': {
|
|
'commit': function( op ) {
|
|
this.insert( op );
|
|
},
|
|
'rollback': function( op ) {
|
|
this.remove( op );
|
|
}
|
|
},
|
|
// Remove
|
|
'remove': {
|
|
'commit': function( op ) {
|
|
this.remove( op );
|
|
},
|
|
'rollback': function( op ) {
|
|
this.insert( op );
|
|
}
|
|
},
|
|
// Change element attributes
|
|
'attribute': {
|
|
'commit': function( op ) {
|
|
this.attribute( op, false );
|
|
},
|
|
'rollback': function( op ) {
|
|
this.attribute( op, true );
|
|
}
|
|
},
|
|
// Change content annotations
|
|
'annotate': {
|
|
'commit': function( op ) {
|
|
this.mark( op, false );
|
|
},
|
|
'rollback': function( op ) {
|
|
this.mark( op, true );
|
|
}
|
|
}
|
|
};
|
|
|
|
/* Static Methods */
|
|
|
|
es.TransactionProcessor.commit = function( doc, transaction ) {
|
|
var tp = new es.TransactionProcessor( doc, transaction );
|
|
tp.process( 'commit' );
|
|
};
|
|
|
|
es.TransactionProcessor.rollback = function( doc, transaction ) {
|
|
var tp = new es.TransactionProcessor( doc, transaction );
|
|
tp.process( 'rollback' );
|
|
};
|
|
|
|
/* Methods */
|
|
|
|
es.TransactionProcessor.prototype.process = function( method ) {
|
|
var operations = this.transaction.getOperations();
|
|
for ( var i = 0, length = operations.length; i < length; i++ ) {
|
|
var operation = operations[i];
|
|
if ( operation.type in es.TransactionProcessor.operationMap ) {
|
|
es.TransactionProcessor.operationMap[operation.type][method].call( this, operation );
|
|
} else {
|
|
throw 'Invalid operation error. Operation type is not supported: ' + operation.type;
|
|
}
|
|
}
|
|
};
|
|
|
|
// TODO: document this. Various arguments are optional or nonoptional in different cases, that's confusing
|
|
// so it needs to be documented well.
|
|
es.TransactionProcessor.prototype.rebuildNodes = function( newData, oldNodes, parent, index ) {
|
|
var newNodes = es.DocumentModel.createNodesFromData( newData ),
|
|
remove = 0;
|
|
if ( oldNodes ) {
|
|
// Determine parent and index if not given
|
|
if ( oldNodes[0] === oldNodes[0].getRoot() ) {
|
|
// We know the values for parent and index in this case
|
|
// and don't have to compute them. Override any parent
|
|
// or index parameter passed.
|
|
parent = oldNodes[0];
|
|
index = 0;
|
|
remove = parent.getChildren().length;
|
|
} else {
|
|
parent = parent || oldNodes[0].getParent();
|
|
index = index || parent.indexOf( oldNodes[0] );
|
|
remove = oldNodes.length;
|
|
}
|
|
// Try to preserve the first node
|
|
if (
|
|
// There must be an old and new node to preserve
|
|
newNodes.length &&
|
|
oldNodes.length &&
|
|
// Node types need to match
|
|
newNodes[0].type === oldNodes[0].type &&
|
|
// Only for leaf nodes
|
|
!newNodes[0].hasChildren()
|
|
) {
|
|
var newNode = newNodes.shift(),
|
|
oldNode = oldNodes.shift();
|
|
// Let's just leave this first node in place and adjust it's length
|
|
var newAttributes = newNode.getElement().attributes,
|
|
oldAttributes = oldNode.getElement().attributes;
|
|
if ( oldAttributes || newAttributes ) {
|
|
oldNode.getElement().attributes = newAttributes;
|
|
}
|
|
oldNode.adjustContentLength( newNode.getContentLength() - oldNode.getContentLength() );
|
|
index++;
|
|
remove--;
|
|
}
|
|
}
|
|
// Try to perform this in a single operation if possible, this reduces the number of UI updates
|
|
// TODO: Introduce a global for max argument length - 1024 is also assumed in es.insertIntoArray
|
|
if ( newNodes.length < 1024 ) {
|
|
parent.splice.apply( parent, [index, remove].concat( newNodes ) );
|
|
} else if ( newNodes.length ) {
|
|
parent.splice.apply( parent, [index, remove] );
|
|
// Safe to call with arbitrary length of newNodes
|
|
es.insertIntoArray( parent, index, newNodes );
|
|
}
|
|
};
|
|
|
|
/**
|
|
* Get the parent node that would be affected by inserting given data into it's child.
|
|
*
|
|
* This is used when inserting data that closes and reopens one or more parent nodes into a child
|
|
* node, which requires rebuilding at a higher level.
|
|
*
|
|
* @method
|
|
* @param {es.DocumentNode} node Child node to start from
|
|
* @param {Array} data Data to inspect for closings
|
|
* @returns {es.DocumentNode} Lowest level parent node being affected
|
|
*/
|
|
es.TransactionProcessor.prototype.getScope = function( node, data ) {
|
|
var i,
|
|
length,
|
|
level = 0,
|
|
max = 0;
|
|
for ( i = 0, length = data.length; i < length; i++ ) {
|
|
if ( typeof data[i].type === 'string' ) {
|
|
level += data[i].type.charAt( 0 ) === '/' ? 1 : -1;
|
|
max = Math.max( max, level );
|
|
}
|
|
}
|
|
if ( max > 0 ) {
|
|
for ( i = 0; i < max - 1; i++ ) {
|
|
node = node.getParent();
|
|
}
|
|
}
|
|
return node;
|
|
};
|
|
|
|
es.TransactionProcessor.prototype.applyAnnotations = function( to, update ) {
|
|
var i,
|
|
j,
|
|
k,
|
|
length,
|
|
annotation,
|
|
changes = 0,
|
|
index;
|
|
// Handle annotations
|
|
if ( this.set.length ) {
|
|
for ( i = 0, length = this.set.length; i < length; i++ ) {
|
|
annotation = this.set[i];
|
|
// Auto-build annotation hash
|
|
if ( annotation.hash === undefined ) {
|
|
annotation.hash = es.DocumentModel.getHash( annotation );
|
|
}
|
|
for ( j = this.cursor; j < to; j++ ) {
|
|
// Auto-convert to array
|
|
if ( es.isArray( this.model.data[j] ) ) {
|
|
this.model.data[j].push( annotation );
|
|
} else {
|
|
this.model.data[j] = [this.model.data[j], annotation];
|
|
}
|
|
}
|
|
}
|
|
changes++;
|
|
}
|
|
if ( this.clear.length ) {
|
|
for ( i = 0, length = this.clear.length; i < length; i++ ) {
|
|
annotation = this.clear[i];
|
|
if ( annotation instanceof RegExp ) {
|
|
for ( j = this.cursor; j < to; j++ ) {
|
|
var matches = es.DocumentModel.getMatchingAnnotations(
|
|
this.model.data[j], annotation
|
|
);
|
|
for ( k = 0; k < matches.length; k++ ) {
|
|
index = this.model.data[j].indexOf( matches[k] );
|
|
if ( index !== -1 ) {
|
|
this.model.data[j].splice( index, 1 );
|
|
}
|
|
}
|
|
// Auto-convert to string
|
|
if ( this.model.data[j].length === 1 ) {
|
|
this.model.data[j] = this.model.data[j][0];
|
|
}
|
|
}
|
|
} else {
|
|
// Auto-build annotation hash
|
|
if ( annotation.hash === undefined ) {
|
|
annotation.hash = es.DocumentModel.getHash( annotation );
|
|
}
|
|
for ( j = this.cursor; j < to; j++ ) {
|
|
index = es.DocumentModel.getIndexOfAnnotation(
|
|
this.model.data[j], annotation
|
|
);
|
|
if ( index !== -1 ) {
|
|
this.model.data[j].splice( index, 1 );
|
|
}
|
|
// Auto-convert to string
|
|
if ( this.model.data[j].length === 1 ) {
|
|
this.model.data[j] = this.model.data[j][0];
|
|
}
|
|
}
|
|
}
|
|
}
|
|
changes++;
|
|
}
|
|
if ( update && changes ) {
|
|
var fromNode = this.model.getNodeFromOffset( this.cursor ),
|
|
toNode = this.model.getNodeFromOffset( to );
|
|
this.model.traverseLeafNodes( function( node ) {
|
|
node.emit( 'update' );
|
|
if ( node === toNode ) {
|
|
return false;
|
|
}
|
|
}, fromNode );
|
|
}
|
|
};
|
|
|
|
es.TransactionProcessor.prototype.retain = function( op ) {
|
|
this.applyAnnotations( this.cursor + op.length, true );
|
|
this.cursor += op.length;
|
|
};
|
|
|
|
es.TransactionProcessor.prototype.insert = function( op ) {
|
|
var node = this.model.getNodeFromOffset( this.cursor ),
|
|
offset;
|
|
if ( node.getParent() === this.model ) {
|
|
offset = this.model.getOffsetFromNode( node );
|
|
} else {
|
|
node = this.getScope( node, op.data );
|
|
offset = this.model.getOffsetFromNode( node );
|
|
}
|
|
if ( es.DocumentModel.containsElementData( op.data ) ) {
|
|
// Perform insert on linear data model
|
|
es.insertIntoArray( this.model.data, this.cursor, op.data );
|
|
this.applyAnnotations( this.cursor + op.data.length );
|
|
// Synchronize model tree
|
|
if ( offset === -1 ) {
|
|
throw 'Invalid offset error. Node is not in model tree';
|
|
}
|
|
this.rebuildNodes(
|
|
this.model.data.slice( offset, offset + node.getElementLength() + op.data.length ),
|
|
[node]
|
|
);
|
|
} else {
|
|
// Perform insert on linear data model
|
|
// TODO this is duplicated from above
|
|
es.insertIntoArray( this.model.data, this.cursor, op.data );
|
|
this.applyAnnotations( this.cursor + op.data.length );
|
|
// Update model tree
|
|
node.adjustContentLength( op.data.length, true );
|
|
node.emit( 'update', this.cursor - offset );
|
|
}
|
|
this.cursor += op.data.length;
|
|
};
|
|
|
|
es.TransactionProcessor.prototype.remove = function( op ) {
|
|
if ( es.DocumentModel.containsElementData( op.data ) ) {
|
|
// Figure out which nodes are covered by the removal
|
|
var ranges = this.model.selectNodes( new es.Range( this.cursor, this.cursor + op.data.length ) );
|
|
|
|
// Build the list of nodes to rebuild and the data to keep
|
|
var oldNodes = [], newData = [], parent = null, index = null, firstKeptNode, lastKeptNode;
|
|
for ( var i = 0; i < ranges.length; i++ ) {
|
|
oldNodes.push( ranges[i].node );
|
|
if ( ranges[i].range !== undefined ) {
|
|
// We have to keep part of this node
|
|
if ( firstKeptNode === undefined ) {
|
|
// This is the first node we're keeping
|
|
firstKeptNode = ranges[i].node;
|
|
}
|
|
// Compute the start and end offset of this node
|
|
// We could do that with getOffsetFromNode() but
|
|
// we already have all the numbers we need so why would we
|
|
var startOffset = ranges[i].globalRange.start - ranges[i].range.start,
|
|
endOffset = startOffset + ranges[i].node.getContentLength(),
|
|
// Get this node's data
|
|
nodeData = this.model.data.slice( startOffset, endOffset );
|
|
// Remove data covered by the range from nodeData
|
|
nodeData.splice( ranges[i].range.start, ranges[i].range.end - ranges[i].range.start );
|
|
// What remains in nodeData is the data we need to keep
|
|
// Append it to newData
|
|
newData = newData.concat( nodeData );
|
|
|
|
lastKeptNode = ranges[i].node;
|
|
}
|
|
}
|
|
|
|
// Surround newData with the right openings and closings if needed
|
|
if ( firstKeptNode !== undefined ) {
|
|
// There are a number of conceptually different cases here,
|
|
// but the algorithm for dealing with them is the same.
|
|
// 1. Removal within one node: firstKeptNode === lastKeptNode
|
|
// 2. Merge of siblings: firstKeptNode.getParent() === lastKeptNode.getParent()
|
|
// 3. Merge of arbitrary depth: firstKeptNode and lastKeptNode have a common ancestor
|
|
// Because #1 and #2 are special cases of #3 (merges with depth=0 and depth=1, respectively),
|
|
// the code below that deals with the general case (#3) and automatically covers
|
|
// #1 and #2 that way as well.
|
|
|
|
// Simultaneously traverse upwards from firstKeptNode and lastKeptNode
|
|
// to find the common ancestor. On our way up, keep the element of each
|
|
// node we visit and verify that the transaction is a valid merge (i.e. it satisfies
|
|
// the merge criteria in prepareRemoval()'s canMerge()).
|
|
// FIXME: The code is essentially the same as canMerge(), merge these algorithms
|
|
var openings = [], closings = [],
|
|
paths = es.DocumentNode.getCommonAncestorPaths( firstKeptNode, lastKeptNode ),
|
|
i, prevN1, prevN2;
|
|
|
|
if ( !paths ) {
|
|
throw 'Removal is not a valid merge: nodes do not have a common ancestor or are not at the same depth';
|
|
}
|
|
for ( i = 0; i < paths.node1Path.length; i++ ) {
|
|
// Verify the element types are equal
|
|
if ( paths.node1Path[i].getElementType() !== paths.node2Path[i].getElementType() ) {
|
|
throw 'Removal is not a valid merge: corresponding parents have different types ( ' +
|
|
paths.node1Path[i].getElementType() + ' vs ' + paths.node2Path[i].getElementType() + ' )';
|
|
}
|
|
// Record the opening of n1 and the closing of n2
|
|
openings.push( paths.node1Path[i].getElement() );
|
|
closings.push( { 'type': '/' + paths.node2Path[i].getElementType() } );
|
|
}
|
|
|
|
// Surround newData with the openings and closings
|
|
newData = openings.reverse().concat( newData, closings );
|
|
|
|
// Rebuild oldNodes if needed
|
|
// This only happens for merges with depth > 1
|
|
prevN1 = paths.node1Path.length ? paths.node1Path[paths.node1Path.length - 1] : null;
|
|
prevN2 = paths.node2Path.length ? paths.node2Path[paths.node2Path.length - 1] : null;
|
|
if ( prevN1 && prevN1 !== oldNodes[0] ) {
|
|
oldNodes = [ prevN1 ];
|
|
parent = paths.commonAncestor;
|
|
index = parent.indexOf( prevN1 ); // Pass to rebuildNodes() so it's not recomputed
|
|
if ( index === -1 ) {
|
|
throw "Tree corruption detected: node isn't in its parent's children array";
|
|
}
|
|
var foundPrevN2 = false;
|
|
for ( var j = index + 1; !foundPrevN2 && j < parent.getChildren().length; j++ ) {
|
|
oldNodes.push( parent.getChildren()[j] );
|
|
foundPrevN2 = parent.getChildren()[j] === prevN2;
|
|
}
|
|
if ( !foundPrevN2 ) {
|
|
throw "Tree corruption detected: node isn't in its parent's children array";
|
|
}
|
|
}
|
|
}
|
|
|
|
// Update the linear model
|
|
this.model.data.splice( this.cursor, op.data.length );
|
|
// Perform the rebuild. This updates the model tree
|
|
this.rebuildNodes( newData, oldNodes, parent, index );
|
|
} else {
|
|
// We're removing content only. Take a shortcut
|
|
// Get the node we are removing content from
|
|
var node = this.model.getNodeFromOffset( this.cursor );
|
|
// Update model tree
|
|
node.adjustContentLength( -op.data.length, true );
|
|
// Update the linear model
|
|
this.model.data.splice( this.cursor, op.data.length );
|
|
// Emit an update so things sync up
|
|
var offset = this.model.getOffsetFromNode( node );
|
|
node.emit( 'update', this.cursor - offset );
|
|
}
|
|
};
|
|
|
|
es.TransactionProcessor.prototype.attribute = function( op, invert ) {
|
|
var element = this.model.data[this.cursor];
|
|
if ( element.type === undefined ) {
|
|
throw 'Invalid element error. Can not set attributes on non-element data.';
|
|
}
|
|
if ( ( op.method === 'set' && !invert ) || ( op.method === 'clear' && invert ) ) {
|
|
// Automatically initialize attributes object
|
|
if ( !element.attributes ) {
|
|
element.attributes = {};
|
|
}
|
|
element.attributes[op.key] = op.value;
|
|
} else if ( ( op.method === 'clear' && !invert ) || ( op.method === 'set' && invert ) ) {
|
|
if ( element.attributes ) {
|
|
delete element.attributes[op.key];
|
|
}
|
|
// Automatically clean up attributes object
|
|
var empty = true;
|
|
for ( var key in element.attributes ) {
|
|
empty = false;
|
|
break;
|
|
}
|
|
if ( empty ) {
|
|
delete element.attributes;
|
|
}
|
|
} else {
|
|
throw 'Invalid method error. Can not operate attributes this way: ' + method;
|
|
}
|
|
var node = this.model.getNodeFromOffset( this.cursor + 1 );
|
|
node.emit( 'update' );
|
|
};
|
|
|
|
es.TransactionProcessor.prototype.mark = function( op, invert ) {
|
|
var target;
|
|
if ( ( op.method === 'set' && !invert ) || ( op.method === 'clear' && invert ) ) {
|
|
target = this.set;
|
|
} else if ( ( op.method === 'clear' && !invert ) || ( op.method === 'set' && invert ) ) {
|
|
target = this.clear;
|
|
} else {
|
|
throw 'Invalid method error. Can not operate attributes this way: ' + method;
|
|
}
|
|
if ( op.bias === 'start' ) {
|
|
target.push( op.annotation );
|
|
} else if ( op.bias === 'stop' ) {
|
|
var index;
|
|
if ( op.annotation instanceof RegExp ) {
|
|
index = target.indexOf( op.annotation );
|
|
} else {
|
|
index = es.DocumentModel.getIndexOfAnnotation( target, op.annotation );
|
|
}
|
|
if ( index === -1 ) {
|
|
throw 'Annotation stack error. Annotation is missing.';
|
|
}
|
|
target.splice( index, 1 );
|
|
}
|
|
};
|