A number of tests have hardcoded expections that pass only in WMF CI where Quibble has LocalSettings.php with $wgScript and $wgArticlePath set a certain way. We could fix these by adding setMwGlobals() in their tests, as we often do, but these are so often forgotten that I'd rather we just add them to TestSetup.php so that it is simply impossible to write a test that that passes locally for you (if you have the same config) but not for someone else. There is a larger project in there somewhere about expanding this slowly such that we basically only pluck DB-settings and extension enablement from LocalSettings and otherwise run the tests with the default settings in PHPUnit. Pretty much by definition, any (other) setting you have in LocalSettings is irrelevant because it either: 1. has no effect on the test (majority, harmless either way), 2. has a custom default via TestSetup.php (which has precedence over LocalSettings.php), 3. is relevant to the code being tested and the test case correctly calls setMwGlobals() to ensure a consistent value during test. 4. is relevant to the tested code but has no override, thus only passes if you happen to have the "right" value set for it (undesirable). Case 4 is already categorically impossible for the most common config settings that influence random code because we give them a value in TestSetup.php. This patch expands that to include $wgScript and $wgArticlePath. Perhaps in the future we can think about a way to do this automatically by either re-applying MainConfigSchema (sans db settings) or by only selectively applying LocalSettings.php in the first place. This patch follows-up I072ddf89562fe, which added a test case in WikitextContentHandlerIntegrationTest.php that assumed "/index.php" as the value of $wgScript. This passes in WMF CI since Quibble uses that value, but the tests failed in most local development installs since those tend to use "/w" instead. Rather than one-off fixing that one test with overrideConfigValues(), switch to a more general fixture, since the precise values don't matter for this test. Bug: T349087 Bug: T277470 Change-Id: If4304b7ca4a838bd892d4516a0b5c6dfbc30986e
2969 lines
96 KiB
PHP
2969 lines
96 KiB
PHP
<?php
|
|
/**
|
|
* Generic backend for the MediaWiki parser test suite, used by both the
|
|
* standalone parserTests.php and the PHPUnit "parsertests" suite.
|
|
*
|
|
* Copyright © 2004, 2010 Brooke Vibber <bvibber@wikimedia.org>
|
|
* https://www.mediawiki.org/
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with this program; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
* http://www.gnu.org/copyleft/gpl.html
|
|
*
|
|
* @todo Make this more independent of the configuration (and if possible the database)
|
|
* @file
|
|
* @ingroup Testing
|
|
*/
|
|
|
|
use MediaWiki\Context\IContextSource;
|
|
use MediaWiki\Context\RequestContext;
|
|
use MediaWiki\HookContainer\HookRunner;
|
|
use MediaWiki\Html\Html;
|
|
use MediaWiki\Interwiki\ClassicInterwikiLookup;
|
|
use MediaWiki\Linker\LinkTarget;
|
|
use MediaWiki\MainConfigNames;
|
|
use MediaWiki\MediaWikiServices;
|
|
use MediaWiki\Page\PageReferenceValue;
|
|
use MediaWiki\Parser\Parser;
|
|
use MediaWiki\Parser\ParserOutput;
|
|
use MediaWiki\Parser\ParserOutputFlags;
|
|
use MediaWiki\Permissions\UltimateAuthority;
|
|
use MediaWiki\Revision\MutableRevisionRecord;
|
|
use MediaWiki\Revision\RevisionRecord;
|
|
use MediaWiki\Revision\SlotRecord;
|
|
use MediaWiki\Title\Title;
|
|
use MediaWiki\Title\TitleValue;
|
|
use MediaWiki\User\User;
|
|
use MediaWiki\User\UserIdentity;
|
|
use MediaWiki\User\UserIdentityValue;
|
|
use MediaWiki\Utils\MWTimestamp;
|
|
use MediaWiki\WikiMap\WikiMap;
|
|
use Psr\Log\NullLogger;
|
|
use Wikimedia\Assert\Assert;
|
|
use Wikimedia\Parsoid\Config\PageConfig;
|
|
use Wikimedia\Parsoid\Config\SiteConfig;
|
|
use Wikimedia\Parsoid\Core\LinkTarget as ParsoidLinkTarget;
|
|
use Wikimedia\Parsoid\Core\SelserData;
|
|
use Wikimedia\Parsoid\DOM\Document;
|
|
use Wikimedia\Parsoid\Ext\ExtensionModule;
|
|
use Wikimedia\Parsoid\ParserTests\Article as ParserTestArticle;
|
|
use Wikimedia\Parsoid\ParserTests\ParserHook as ParsoidParserHook;
|
|
use Wikimedia\Parsoid\ParserTests\RawHTML as ParsoidRawHTML;
|
|
use Wikimedia\Parsoid\ParserTests\StyleTag as ParsoidStyleTag;
|
|
use Wikimedia\Parsoid\ParserTests\Test as ParserTest;
|
|
use Wikimedia\Parsoid\ParserTests\TestFileReader;
|
|
use Wikimedia\Parsoid\ParserTests\TestMode as ParserTestMode;
|
|
use Wikimedia\Parsoid\Parsoid;
|
|
use Wikimedia\Parsoid\Utils\ContentUtils;
|
|
use Wikimedia\Parsoid\Utils\DOMCompat;
|
|
use Wikimedia\Parsoid\Utils\DOMDataUtils;
|
|
use Wikimedia\Parsoid\Utils\DOMUtils;
|
|
use Wikimedia\Rdbms\DBError;
|
|
use Wikimedia\Rdbms\IDatabase;
|
|
use Wikimedia\ScopedCallback;
|
|
use Wikimedia\TestingAccessWrapper;
|
|
|
|
/**
|
|
* @ingroup Testing
|
|
*/
|
|
class ParserTestRunner {
|
|
|
|
/**
|
|
* @var array<string,bool> The status of each setup function
|
|
*/
|
|
private $setupDone = [
|
|
'staticSetup' => false,
|
|
'perTestSetup' => false,
|
|
'setupDatabase' => false,
|
|
'setupUploads' => false,
|
|
];
|
|
|
|
/**
|
|
* @var array (CLI/Config) Options for the test runner
|
|
* See the constructor for documentation
|
|
*/
|
|
private $options;
|
|
|
|
/**
|
|
* @var string[] Set of requested test modes
|
|
*/
|
|
private $requestedTestModes;
|
|
|
|
/**
|
|
* Our connection to the database
|
|
* @var IDatabase
|
|
*/
|
|
private $db;
|
|
|
|
/**
|
|
* @var TestRecorder
|
|
*/
|
|
private $recorder;
|
|
|
|
/**
|
|
* The upload directory, or null to not set up an upload directory
|
|
*
|
|
* @var string|null
|
|
*/
|
|
private $uploadDir = null;
|
|
|
|
/**
|
|
* The name of the file backend to use, or false to use MockFileBackend.
|
|
* @var string|false
|
|
*/
|
|
private $fileBackendName;
|
|
|
|
/**
|
|
* A complete regex for filtering tests.
|
|
* @var string
|
|
*/
|
|
private $regex;
|
|
|
|
/**
|
|
* A list of normalization functions to apply to the expected and actual
|
|
* output.
|
|
* @var string[]
|
|
*/
|
|
private $normalizationFunctions = [];
|
|
|
|
/**
|
|
* Run disabled parser tests
|
|
* @var bool
|
|
*/
|
|
private $runDisabled;
|
|
|
|
/**
|
|
* Disable parse on article insertion
|
|
* @var bool
|
|
*/
|
|
private $disableSaveParse;
|
|
|
|
/**
|
|
* Reuse upload directory
|
|
* @var bool
|
|
*/
|
|
private $keepUploads;
|
|
|
|
/**
|
|
* Did some Parsoid test pass where it was expected to fail?
|
|
* This can happen if the test failure is recorded in the -knownFailures.json file
|
|
* but the test result changed, or functionality changed that causes tests to pass.
|
|
* @var bool
|
|
*/
|
|
public $unexpectedTestPasses = false;
|
|
|
|
/**
|
|
* Table name prefix.
|
|
*/
|
|
public const DB_PREFIX = 'parsertest_';
|
|
|
|
/** @var string */
|
|
private const FILTER_MSG = "Test doesn't match filter";
|
|
|
|
/**
|
|
* Compute the set of valid test runner modes
|
|
*
|
|
* @return string[]
|
|
*/
|
|
public function getRequestedTestModes(): array {
|
|
return $this->requestedTestModes;
|
|
}
|
|
|
|
/**
|
|
* @param TestRecorder $recorder
|
|
* @param array $options
|
|
* - parsoid (bool) if true, run Parsoid tests
|
|
* - testFile (string)
|
|
* If set, the (Parsoid) test file to run tests from.
|
|
* Currently, only used for CLI PHPUnit test runs
|
|
* to avoid running every single test file out there.
|
|
* Legacy parser test runs ignore this option.
|
|
* - wt2html (bool) If true, run Parsoid wt2html tests
|
|
* - wt2wt (bool) If true, run Parsoid wt2wt tests
|
|
* - html2wt (bool) If true, run Parsoid html2wt tests
|
|
* - html2html (bool) If true, run Parsoid html2html tests
|
|
* - selser (bool/"noauto")
|
|
* If true, run Parsoid auto-generated selser tests
|
|
* If "noauto", run Parsoid manual edit selser tests
|
|
* - numchanges (int) number of selser edit tests to generate
|
|
* - changetree (array|null)
|
|
* If not null, run a Parsoid selser edit test with this changetree
|
|
* - updateKnownFailures (bool)
|
|
* If true, *knownFailures.json files are updated
|
|
* - traceFlags (array)
|
|
* (Parsoid-only) Flags for tracing different Parsoid stages
|
|
* - dumpFlags (array)
|
|
* (Parsoid-only) Flags for dumping various pieces of state
|
|
* - norm (array)
|
|
* An array of normalization functions to run on test output
|
|
* to use in legacy parser test runs
|
|
* - regex (string) Regex for filtering tests
|
|
* - run-disabled (bool) If true, run disabled tests
|
|
* - keep-uploads (bool) If true, reuse upload directory
|
|
* - file-backend (string|bool)
|
|
* If false, use MockFileBackend
|
|
* Else name of the file backend to use
|
|
* - disable-save-parse (bool) if true, disable parse on article insertion
|
|
* - update-tests (bool) Update parserTests.txt with results from wt2html fails.
|
|
*
|
|
* NOTE: At this time, Parsoid-specific test options are only handled
|
|
* in PHPUnit mode. A future patch will likely tweak some of this and
|
|
* support these flags no matter how this test runner is instantiated.
|
|
*/
|
|
public function __construct( TestRecorder $recorder, $options = [] ) {
|
|
$this->recorder = $recorder;
|
|
$this->options = $options + [
|
|
'keep-uploads' => false,
|
|
'file-backend' => false,
|
|
'run-disabled' => false,
|
|
'disable-save-parse' => false,
|
|
'upload-dir' => null,
|
|
'regex' => false,
|
|
'norm' => [],
|
|
// Parsoid-specific options
|
|
'parsoid' => false,
|
|
'knownFailures' => true,
|
|
'updateKnownFailures' => false,
|
|
'changetree' => null,
|
|
'update-tests' => false,
|
|
// Options can also match those in ParserTestModes::TEST_MODES
|
|
// but we don't need to initialize those here; they will be
|
|
// accessed via $this->requestedTestModes instead.
|
|
'traceFlags' => [],
|
|
'dumpFlags' => [],
|
|
];
|
|
|
|
// Requested test modes are used for Parsoid tests and ignored for
|
|
// legacy parser tests.
|
|
$this->requestedTestModes = ParserTestMode::requestedTestModes(
|
|
$this->options
|
|
);
|
|
|
|
$normFuncs = $this->options['norm'];
|
|
'@phan-var string[] $normFuncs';
|
|
foreach ( $normFuncs as $func ) {
|
|
if ( in_array( $func, [ 'removeTbody', 'trimWhitespace' ] ) ) {
|
|
$this->normalizationFunctions[] = $func;
|
|
} else {
|
|
$this->recorder->warning(
|
|
"Warning: unknown normalization option \"$func\"\n" );
|
|
}
|
|
}
|
|
|
|
if ( $this->options['regex'] !== false ) {
|
|
$this->regex = $this->options['regex'];
|
|
} else {
|
|
# Matches anything
|
|
$this->regex = '//';
|
|
}
|
|
|
|
$this->keepUploads = (bool)$this->options['keep-uploads'];
|
|
$this->fileBackendName = $this->options['file-backend'];
|
|
$this->runDisabled = (bool)$this->options['run-disabled'];
|
|
$this->disableSaveParse = (bool)$this->options['disable-save-parse'];
|
|
$this->uploadDir = $this->options['upload-dir'];
|
|
}
|
|
|
|
/**
|
|
* @return array
|
|
*/
|
|
public function getOptions(): array {
|
|
return $this->options;
|
|
}
|
|
|
|
/**
|
|
* Get list of filenames to extension and core parser tests
|
|
*
|
|
* @param string[] $dirs
|
|
* @return string[]
|
|
*/
|
|
public static function getParserTestFiles( array $dirs = [] ): array {
|
|
if ( $dirs ) {
|
|
$ptDirs = [];
|
|
foreach ( $dirs as $i => $dir ) {
|
|
if ( !is_dir( $dir ) ) {
|
|
echo "$dir is not a directory. Skipping it.\n";
|
|
continue;
|
|
}
|
|
$ptDirs["_CLI{$i}_"] = $dir;
|
|
}
|
|
} else {
|
|
// Auto-discover core test files
|
|
$ptDirs = [ 'core' => __DIR__ ];
|
|
|
|
// Auto-discover extension parser tests
|
|
$registry = ExtensionRegistry::getInstance();
|
|
foreach ( $registry->getAllThings() as $info ) {
|
|
$dir = dirname( $info['path'] ) . '/tests/parser';
|
|
if ( !is_dir( $dir ) ) {
|
|
continue;
|
|
}
|
|
$ptDirs[ $info['name'] ] = $dir;
|
|
}
|
|
}
|
|
|
|
$files = [];
|
|
foreach ( $ptDirs as $extName => $dir ) {
|
|
$counter = 1;
|
|
$dirIterator = new RecursiveIteratorIterator(
|
|
new RecursiveDirectoryIterator( $dir )
|
|
);
|
|
foreach ( $dirIterator as $fileInfo ) {
|
|
/** @var SplFileInfo $fileInfo */
|
|
if ( str_ends_with( $fileInfo->getFilename(), '.txt' ) ) {
|
|
$name = $extName . '_' . $counter;
|
|
while ( isset( $files[$name] ) ) {
|
|
$counter++;
|
|
$name = $extName . '_' . $counter;
|
|
}
|
|
$files[$name] = $fileInfo->getPathname();
|
|
}
|
|
}
|
|
}
|
|
|
|
return array_unique( $files );
|
|
}
|
|
|
|
public function getRecorder(): TestRecorder {
|
|
return $this->recorder;
|
|
}
|
|
|
|
/**
|
|
* Do any setup which can be done once for all tests, independent of test
|
|
* options, except for database setup.
|
|
*
|
|
* Public setup functions in this class return a ScopedCallback object. When
|
|
* this object is destroyed by going out of scope, teardown of the
|
|
* corresponding test setup is performed.
|
|
*
|
|
* Teardown objects may be chained by passing a ScopedCallback from a
|
|
* previous setup stage as the $nextTeardown parameter. This enforces the
|
|
* convention that teardown actions are taken in reverse order to the
|
|
* corresponding setup actions. When $nextTeardown is specified, a
|
|
* ScopedCallback will be returned which first tears down the current
|
|
* setup stage, and then tears down the previous setup stage which was
|
|
* specified by $nextTeardown.
|
|
*
|
|
* @param ScopedCallback|null $nextTeardown
|
|
* @return ScopedCallback
|
|
*/
|
|
public function staticSetup( $nextTeardown = null ) {
|
|
// A note on coding style:
|
|
|
|
// The general idea here is to keep setup code together with
|
|
// corresponding teardown code, in a fine-grained manner. We have two
|
|
// arrays: $setup and $teardown. The code snippets in the $setup array
|
|
// are executed at the end of the method, before it returns, and the
|
|
// code snippets in the $teardown array are executed in reverse order
|
|
// when the Wikimedia\ScopedCallback object is consumed.
|
|
|
|
// Because it is a common operation to save, set and restore global
|
|
// variables, we have an additional convention: when the array key of
|
|
// $setup is a string, the string is taken to be the name of the global
|
|
// variable, and the element value is taken to be the desired new value.
|
|
|
|
// It's acceptable to just do the setup immediately, instead of adding
|
|
// a closure to $setup, except when the setup action depends on global
|
|
// variable initialisation being done first. In this case, you have to
|
|
// append a closure to $setup after the global variable is appended.
|
|
|
|
// When you add to setup functions in this class, please keep associated
|
|
// setup and teardown actions together in the source code, and please
|
|
// add comments explaining why the setup action is necessary.
|
|
|
|
$setup = [];
|
|
$teardown = [];
|
|
|
|
$teardown[] = $this->markSetupDone( 'staticSetup' );
|
|
|
|
// Some settings which influence HTML output
|
|
$setup['wgSitename'] = 'MediaWiki';
|
|
$setup['wgMetaNamespace'] = "TestWiki";
|
|
$setup['wgServer'] = 'http://example.org';
|
|
$setup['wgServerName'] = 'example.org';
|
|
$setup['wgCapitalLinks'] = true;
|
|
$setup['wgNoFollowLinks'] = true;
|
|
$setup['wgNoFollowDomainExceptions'] = [ 'no-nofollow.org' ];
|
|
$setup['wgExternalLinkTarget'] = false;
|
|
$setup['wgLocaltimezone'] = 'UTC';
|
|
$reset = static function () {
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'UrlUtils' );
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
|
|
// Language and variant settings
|
|
$setup['wgLanguageCode'] = 'en';
|
|
$setup['wgDisableLangConversion'] = false;
|
|
$setup['wgDisableTitleConversion'] = false;
|
|
$setup['wgUsePigLatinVariant'] = false;
|
|
$this->resetLanguageServices( $setup, $teardown, true );
|
|
|
|
// "extra language links"
|
|
// see https://gerrit.wikimedia.org/r/111390
|
|
$setup['wgExtraInterlanguageLinkPrefixes'] = [ 'mul' ];
|
|
|
|
// All FileRepo changes should be done here by injecting services,
|
|
// there should be no need to change global variables.
|
|
MediaWikiServices::getInstance()->disableService( 'RepoGroup' );
|
|
MediaWikiServices::getInstance()->redefineService( 'RepoGroup',
|
|
function () {
|
|
return $this->createRepoGroup();
|
|
}
|
|
);
|
|
$teardown[] = static function () {
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'RepoGroup' );
|
|
};
|
|
|
|
// Set up null lock managers
|
|
$setup['wgLockManagers'] = [ [
|
|
'name' => 'fsLockManager',
|
|
'class' => NullLockManager::class,
|
|
], [
|
|
'name' => 'nullLockManager',
|
|
'class' => NullLockManager::class,
|
|
] ];
|
|
$reset = static function () {
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'LockManagerGroupFactory' );
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
|
|
// This allows article insertion into the prefixed DB
|
|
$setup['wgDefaultExternalStore'] = false;
|
|
|
|
// This might slightly reduce memory usage
|
|
$setup['wgAdaptiveMessageCache'] = true;
|
|
|
|
// This is essential and overrides disabling of database messages in TestSetup
|
|
$setup['wgUseDatabaseMessages'] = true;
|
|
$reset = static function () {
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'MessageCache' );
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
|
|
// It's not necessary to actually convert any files
|
|
$setup['wgSVGConverter'] = 'null';
|
|
$setup['wgSVGConverters'] = [ 'null' => 'echo "1">$output' ];
|
|
|
|
// Fake constant timestamp
|
|
MediaWikiServices::getInstance()->getHookContainer()->register(
|
|
'ParserGetVariableValueTs',
|
|
function ( $parser, &$ts ) {
|
|
$ts = $this->getFakeTimestamp();
|
|
return true;
|
|
}
|
|
);
|
|
|
|
$teardown[] = static function () {
|
|
MediaWikiServices::getInstance()->getHookContainer()->clear( 'ParserGetVariableValueTs' );
|
|
};
|
|
|
|
$this->appendNamespaceSetup( $setup, $teardown );
|
|
|
|
// Set up interwikis and append teardown function
|
|
$this->appendInterwikiSetup( $setup, $teardown );
|
|
|
|
// Set up a mock MediaHandlerFactory
|
|
MediaWikiServices::getInstance()->disableService( 'MediaHandlerFactory' );
|
|
MediaWikiServices::getInstance()->redefineService(
|
|
'MediaHandlerFactory',
|
|
static function ( MediaWikiServices $services ) {
|
|
$handlers = $services->getMainConfig()->get( MainConfigNames::ParserTestMediaHandlers );
|
|
return new MediaHandlerFactory(
|
|
new NullLogger(),
|
|
$handlers
|
|
);
|
|
}
|
|
);
|
|
$teardown[] = static function () {
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'MediaHandlerFactory' );
|
|
};
|
|
|
|
$teardown[] = $this->executeSetupSnippets( $setup );
|
|
|
|
// Schedule teardown snippets in reverse order
|
|
return $this->createTeardownObject( $teardown, $nextTeardown );
|
|
}
|
|
|
|
private function appendNamespaceSetup( &$setup, &$teardown ) {
|
|
// Add a namespace shadowing a interwiki link, to test
|
|
// proper precedence when resolving links. (T53680)
|
|
$setup['wgExtraNamespaces'] = [
|
|
100 => 'MemoryAlpha',
|
|
101 => 'MemoryAlpha_talk'
|
|
];
|
|
// Changing wgExtraNamespaces invalidates caches in NamespaceInfo and any live Language
|
|
// object, both on setup and teardown
|
|
$reset = static function () {
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'MainConfig' );
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'NamespaceInfo' );
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'LanguageFactory' );
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'ContentLanguage' );
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'LanguageConverterFactory' );
|
|
// Don't destroy the link cache, because various things hold a
|
|
// copy of it. Just clear it.
|
|
MediaWikiServices::getInstance()->getLinkCache()->clear();
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
}
|
|
|
|
/**
|
|
* Create a RepoGroup object appropriate for the current configuration
|
|
* @return RepoGroup
|
|
*/
|
|
protected function createRepoGroup() {
|
|
if ( $this->uploadDir ) {
|
|
if ( $this->fileBackendName ) {
|
|
throw new RuntimeException( 'You cannot specify both use-filebackend and upload-dir' );
|
|
}
|
|
$backend = new FSFileBackend( [
|
|
'name' => 'local-backend',
|
|
'wikiId' => WikiMap::getCurrentWikiId(),
|
|
'basePath' => $this->uploadDir,
|
|
'tmpDirectory' => wfTempDir()
|
|
] );
|
|
} elseif ( $this->fileBackendName ) {
|
|
global $wgFileBackends;
|
|
$name = $this->fileBackendName;
|
|
$useConfig = false;
|
|
foreach ( $wgFileBackends as $conf ) {
|
|
if ( $conf['name'] === $name ) {
|
|
$useConfig = $conf;
|
|
}
|
|
}
|
|
if ( $useConfig === false ) {
|
|
throw new RuntimeException( "Unable to find file backend \"$name\"" );
|
|
}
|
|
$useConfig['name'] = 'local-backend'; // swap name
|
|
unset( $useConfig['lockManager'] );
|
|
$class = $useConfig['class'];
|
|
// @phan-suppress-next-line PhanInvalidFQSENInClasslike
|
|
$backend = new $class( $useConfig );
|
|
} else {
|
|
# Replace with a mock. We do not care about generating real
|
|
# files on the filesystem, just need to expose the file
|
|
# informations.
|
|
$backend = new MockFileBackend( [
|
|
'name' => 'local-backend',
|
|
'wikiId' => WikiMap::getCurrentWikiId()
|
|
] );
|
|
}
|
|
|
|
$services = MediaWikiServices::getInstance();
|
|
return new RepoGroup(
|
|
[
|
|
'class' => MockLocalRepo::class,
|
|
'name' => 'local',
|
|
'url' => 'http://example.com/images',
|
|
'hashLevels' => 2,
|
|
'transformVia404' => false,
|
|
'backend' => $backend
|
|
],
|
|
[],
|
|
$services->getMainWANObjectCache(),
|
|
$services->getMimeAnalyzer()
|
|
);
|
|
}
|
|
|
|
/**
|
|
* Execute an array in which elements with integer keys are taken to be
|
|
* callable objects, and other elements are taken to be global variable
|
|
* set operations, with the key giving the variable name and the value
|
|
* giving the new global variable value. A closure is returned which, when
|
|
* executed, sets the global variables back to the values they had before
|
|
* this function was called.
|
|
*
|
|
* @see staticSetup
|
|
*
|
|
* @param array $setup
|
|
* @return closure
|
|
*/
|
|
protected function executeSetupSnippets( $setup ) {
|
|
$saved = [];
|
|
foreach ( $setup as $name => $value ) {
|
|
if ( is_int( $name ) ) {
|
|
$value();
|
|
} else {
|
|
$saved[$name] = $GLOBALS[$name] ?? null;
|
|
$GLOBALS[$name] = $value;
|
|
}
|
|
}
|
|
return function () use ( $saved ) {
|
|
$this->executeSetupSnippets( $saved );
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Take a setup array in the same format as the one given to
|
|
* executeSetupSnippets(), and return a ScopedCallback which, when consumed,
|
|
* executes the snippets in the setup array in reverse order. This is used
|
|
* to create "teardown objects" for the public API.
|
|
*
|
|
* @see staticSetup
|
|
*
|
|
* @param array $teardown The snippet array
|
|
* @param ScopedCallback|null $nextTeardown A ScopedCallback to consume
|
|
* @return ScopedCallback
|
|
*/
|
|
protected function createTeardownObject(
|
|
array $teardown, ?ScopedCallback $nextTeardown = null
|
|
) {
|
|
return new ScopedCallback( function () use ( $teardown, $nextTeardown ) {
|
|
// Schedule teardown snippets in reverse order
|
|
$teardown = array_reverse( $teardown );
|
|
|
|
$this->executeSetupSnippets( $teardown );
|
|
if ( $nextTeardown ) {
|
|
ScopedCallback::consume( $nextTeardown );
|
|
}
|
|
} );
|
|
}
|
|
|
|
/**
|
|
* Set a setupDone flag to indicate that setup has been done, and return
|
|
* the teardown closure. If the flag was already set, throw an exception.
|
|
*
|
|
* @param string $funcName The setup function name
|
|
* @return closure
|
|
*/
|
|
protected function markSetupDone( $funcName ) {
|
|
if ( $this->setupDone[$funcName] ) {
|
|
throw new RuntimeException( "$funcName is already done" );
|
|
}
|
|
$this->setupDone[$funcName] = true;
|
|
return function () use ( $funcName ) {
|
|
$this->setupDone[$funcName] = false;
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Ensure one of the given setup stages has been done, throw an exception otherwise.
|
|
* @param string $funcName
|
|
*/
|
|
protected function checkSetupDone( string $funcName ) {
|
|
if ( !$this->setupDone[$funcName] ) {
|
|
throw new BadMethodCallException( "$funcName must be called before calling " . wfGetCaller() );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Determine whether a particular setup function has been run
|
|
*
|
|
* @param string $funcName
|
|
* @return bool
|
|
*/
|
|
public function isSetupDone( $funcName ) {
|
|
return $this->setupDone[$funcName] ?? false;
|
|
}
|
|
|
|
/**
|
|
* Insert hardcoded interwiki in the lookup table.
|
|
*
|
|
* This function insert a set of well known interwikis that are used in
|
|
* the parser tests. We use the $wgInterwikiCache mechanism to completely
|
|
* replace any other lookup. (Note that the InterwikiLoadPrefix hook
|
|
* isn't used because it doesn't alter the result of
|
|
* Interwiki::getAllPrefixes() and so is incompatible with some users,
|
|
* including Parsoid.)
|
|
* @param array &$setup
|
|
* @param array &$teardown
|
|
*/
|
|
private function appendInterwikiSetup( &$setup, &$teardown ) {
|
|
static $testInterwikis = [
|
|
[
|
|
'iw_prefix' => 'local',
|
|
// This is a "local interwiki" (see wgLocalInterwikis elsewhere in this file)
|
|
'iw_url' => 'http://example.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
// Local interwiki that matches a namespace name (T228616)
|
|
[
|
|
'iw_prefix' => 'project',
|
|
// This is a "local interwiki" (see wgLocalInterwikis elsewhere in this file)
|
|
'iw_url' => 'http://example.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
[
|
|
'iw_prefix' => 'wikipedia',
|
|
'iw_url' => 'http://en.wikipedia.org/wiki/$1',
|
|
'iw_local' => 0,
|
|
],
|
|
[
|
|
'iw_prefix' => 'meatball',
|
|
// this has been updated in the live wikis, but the parser tests
|
|
// expect the old value
|
|
'iw_url' => 'http://www.usemod.com/cgi-bin/mb.pl?$1',
|
|
'iw_local' => 0,
|
|
],
|
|
[
|
|
'iw_prefix' => 'memoryalpha',
|
|
'iw_url' => 'http://www.memory-alpha.org/en/index.php/$1',
|
|
'iw_local' => 0,
|
|
],
|
|
[
|
|
'iw_prefix' => 'zh',
|
|
'iw_url' => 'http://zh.wikipedia.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
[
|
|
'iw_prefix' => 'es',
|
|
'iw_url' => 'http://es.wikipedia.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
[
|
|
'iw_prefix' => 'fr',
|
|
'iw_url' => 'http://fr.wikipedia.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
[
|
|
'iw_prefix' => 'ru',
|
|
'iw_url' => 'http://ru.wikipedia.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
[
|
|
'iw_prefix' => 'mi',
|
|
// This is a "local interwiki" (see wgLocalInterwikis elsewhere in this file)
|
|
'iw_url' => 'http://example.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
[
|
|
'iw_prefix' => 'mul',
|
|
'iw_url' => 'http://wikisource.org/wiki/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
// Additions from Parsoid
|
|
[
|
|
'iw_prefix' => 'en',
|
|
'iw_url' => '//en.wikipedia.org/wiki/$1',
|
|
'iw_local' => 1
|
|
],
|
|
[
|
|
'iw_prefix' => 'stats',
|
|
'iw_url' => 'https://stats.wikimedia.org/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
[
|
|
'iw_prefix' => 'gerrit',
|
|
'iw_url' => 'https://gerrit.wikimedia.org/$1',
|
|
'iw_local' => 1,
|
|
],
|
|
// Added to support T145590#8608455
|
|
[
|
|
'iw_prefix' => 'pmid',
|
|
'iw_url' => '//www.ncbi.nlm.nih.gov/pubmed/$1?dopt=Abstract',
|
|
'iw_local' => 1,
|
|
],
|
|
// Deliberately missing a $1 in the URL to exercise a common
|
|
// misconfiguration.
|
|
[
|
|
'iw_prefix' => 'wikinvest',
|
|
'iw_url' => 'https://meta.wikimedia.org/wiki/Interwiki_map/discontinued#Wikinvest',
|
|
'iw_local' => 1,
|
|
],
|
|
];
|
|
// When running from parserTests.php, database setup happens *after*
|
|
// interwiki setup, and that changes the wiki id. In order to avoid
|
|
// breaking the interwiki cache, use 'global scope' for the interwiki
|
|
// lookup.
|
|
$GLOBAL_SCOPE = 2; // See docs for $wgInterwikiScopes
|
|
$setup['wgInterwikiScopes'] = $GLOBAL_SCOPE;
|
|
$setup['wgInterwikiCache'] =
|
|
ClassicInterwikiLookup::buildCdbHash( $testInterwikis, $GLOBAL_SCOPE );
|
|
$reset = static function () {
|
|
// Reset the service in case any other tests already cached some prefixes.
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'InterwikiLookup' );
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
|
|
// This affects title normalization in links. It invalidates
|
|
// MediaWikiTitleCodec objects.
|
|
// These interwikis should have 'iw_url' that matches wgServer.
|
|
$setup['wgLocalInterwikis'] = [ 'local', 'project', 'mi' ];
|
|
$reset = function () {
|
|
$this->resetTitleServices();
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
}
|
|
|
|
/**
|
|
* Reset the Title-related services that need resetting
|
|
* for each test
|
|
*
|
|
* @todo We need to reset all services on every test
|
|
*/
|
|
private function resetTitleServices() {
|
|
$services = MediaWikiServices::getInstance();
|
|
$services->resetServiceForTesting( 'TitleFormatter' );
|
|
$services->resetServiceForTesting( 'TitleParser' );
|
|
$services->resetServiceForTesting( '_MediaWikiTitleCodec' );
|
|
$services->resetServiceForTesting( 'LinkRenderer' );
|
|
$services->resetServiceForTesting( 'LinkRendererFactory' );
|
|
$services->resetServiceForTesting( 'NamespaceInfo' );
|
|
$services->resetServiceForTesting( 'SpecialPageFactory' );
|
|
}
|
|
|
|
/**
|
|
* Remove last character if it is a newline
|
|
* @param string $s
|
|
* @return string
|
|
*/
|
|
public static function chomp( $s ) {
|
|
if ( str_ends_with( $s, "\n" ) ) {
|
|
return substr( $s, 0, -1 );
|
|
} else {
|
|
return $s;
|
|
}
|
|
}
|
|
|
|
private function skipAllTestsInFile(
|
|
string $filename, TestFileReader $testFileInfo, ParserTestMode $mode, string $skipMessage
|
|
): void {
|
|
foreach ( $testFileInfo->testCases as $test ) {
|
|
$this->recorder->startTest( $test, $mode );
|
|
$this->recorder->skipped( $test, $mode, $skipMessage );
|
|
}
|
|
$this->recorder->endSuite( $filename );
|
|
}
|
|
|
|
/**
|
|
* Run a series of tests listed in the given text files.
|
|
* Each test consists of a brief description, wikitext input,
|
|
* and the expected HTML output.
|
|
*
|
|
* Prints status updates on stdout and counts up the total
|
|
* number and percentage of passed tests.
|
|
*
|
|
* Handles all setup and teardown.
|
|
*
|
|
* @param string[] $filenames
|
|
* @return bool True if passed all tests, false if any tests failed.
|
|
*/
|
|
public function runTestsFromFiles( $filenames ) {
|
|
$ok = false;
|
|
|
|
$teardownGuard = null;
|
|
$teardownGuard = $this->setupDatabase( $teardownGuard );
|
|
$teardownGuard = $this->staticSetup( $teardownGuard );
|
|
$teardownGuard = $this->setupUploads( $teardownGuard );
|
|
|
|
$this->recorder->start();
|
|
try {
|
|
$ok = true;
|
|
$inParsoidMode = $this->options['parsoid'];
|
|
if ( $inParsoidMode ) {
|
|
$legacyMode = null;
|
|
$skipMode = new ParserTestMode( 'parsoid' );
|
|
} else {
|
|
$legacyMode = $skipMode = new ParserTestMode( 'legacy' );
|
|
}
|
|
|
|
foreach ( $filenames as $filename ) {
|
|
$this->recorder->startSuite( $filename );
|
|
|
|
$testFileInfo = TestFileReader::read( $filename,
|
|
static function ( $msg ) {
|
|
wfDeprecatedMsg( $msg, '1.35', false, false );
|
|
}
|
|
);
|
|
|
|
$skipMessage = $this->getFileSkipMessage( !$inParsoidMode, $testFileInfo->fileOptions, $filename );
|
|
if ( $skipMessage !== null ) {
|
|
$this->skipAllTestsInFile( $filename, $testFileInfo, $skipMode, $skipMessage );
|
|
continue;
|
|
}
|
|
|
|
$parsoidTestModeStrs = [];
|
|
if ( $inParsoidMode ) { // Intersect requested modes with test modes enabled in the file
|
|
$parsoidTestModeStrs = $this->computeValidTestModes(
|
|
$this->getRequestedTestModes(), $testFileInfo->fileOptions );
|
|
|
|
if ( !$parsoidTestModeStrs ) {
|
|
$skipMessage = 'No compatible Parsoid modes found for the file';
|
|
$this->skipAllTestsInFile( $filename, $testFileInfo, $skipMode, $skipMessage );
|
|
continue;
|
|
}
|
|
}
|
|
|
|
$this->checkSetupDone( 'staticSetup' );
|
|
$teardown = $this->addArticles( $testFileInfo->articles );
|
|
|
|
// Run tests
|
|
foreach ( $testFileInfo->testCases as $test ) {
|
|
$skipMessage = $this->getTestSkipMessage( $test, !$inParsoidMode );
|
|
if ( $skipMessage !== null ) {
|
|
if ( $skipMessage !== self::FILTER_MSG ) {
|
|
$this->recorder->startTest( $test, $skipMode );
|
|
$this->recorder->skipped( $test, $skipMode, $skipMessage );
|
|
}
|
|
continue;
|
|
}
|
|
|
|
if ( $inParsoidMode ) {
|
|
// calls runTestInternal for each mode
|
|
$passed = $this->runTestInParsoidModes( $test, $parsoidTestModeStrs );
|
|
} else {
|
|
'@phan-var ParserTestMode $legacyMode'; // assert that this is not null
|
|
$passed = $this->runTestInternal( $test, $legacyMode )->isSuccess();
|
|
}
|
|
|
|
$ok = $ok && $passed;
|
|
}
|
|
|
|
// Update tests / known-failures
|
|
if ( $inParsoidMode && $this->options['updateKnownFailures'] ) {
|
|
$this->updateKnownFailures( $filename, $testFileInfo );
|
|
}
|
|
|
|
if ( $this->options['update-tests'] ) {
|
|
$this->updateTests( $filename, $testFileInfo, !$inParsoidMode );
|
|
}
|
|
|
|
// Clean up
|
|
ScopedCallback::consume( $teardown );
|
|
|
|
$this->recorder->endSuite( $filename );
|
|
}
|
|
|
|
$this->recorder->report();
|
|
} catch ( DBError $e ) {
|
|
$this->recorder->warning( $e->getMessage() );
|
|
}
|
|
$this->recorder->end();
|
|
|
|
ScopedCallback::consume( $teardownGuard );
|
|
|
|
return $ok;
|
|
}
|
|
|
|
/**
|
|
* Determine whether the current parser has the hooks registered in it
|
|
* that are required by a file read by TestFileReader.
|
|
* @param array[] $requirements
|
|
* @return bool
|
|
*/
|
|
public function meetsRequirements( $requirements ) {
|
|
foreach ( $requirements as $requirement ) {
|
|
$ok = true;
|
|
switch ( $requirement['type'] ) {
|
|
case 'hook':
|
|
$ok = $this->requireHook( $requirement['name'] );
|
|
break;
|
|
case 'functionHook':
|
|
$ok = $this->requireFunctionHook( $requirement['name'] );
|
|
break;
|
|
}
|
|
if ( !$ok ) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* @param bool $isLegacy
|
|
* @param array $fileOptions
|
|
* @param string $filename
|
|
* @return string|null
|
|
*/
|
|
public function getFileSkipMessage( bool $isLegacy, array $fileOptions, string $filename ): ?string {
|
|
$runnerOpts = $this->getOptions();
|
|
// Verify minimum version #
|
|
$testFormat = intval( $fileOptions['version'] ?? '1' );
|
|
if ( $testFormat < 2 ) {
|
|
throw new RuntimeException(
|
|
"$filename needs an update. Support for the parserTest v1 file format was removed in MediaWiki 1.36"
|
|
);
|
|
}
|
|
|
|
// If any requirements are not met, mark all tests from the file as skipped
|
|
if ( !(
|
|
$isLegacy ||
|
|
isset( $fileOptions['parsoid-compatible'] )
|
|
) ) {
|
|
// Running files in Parsoid integrated mode is opt-in for now.
|
|
return 'not compatible with Parsoid integrated mode';
|
|
} elseif ( !$this->meetsRequirements( $fileOptions['requirements'] ?? [] ) ) {
|
|
return 'required extension not enabled';
|
|
} elseif ( ( $runnerOpts['testFile'] ?? $filename ) !== $filename ) {
|
|
return 'Not the requested test file';
|
|
} else {
|
|
return null;
|
|
}
|
|
}
|
|
|
|
public function getTestSkipMessage( ParserTest $test, bool $isLegacy ): ?string {
|
|
$opts = $test->options;
|
|
|
|
if ( !$test->matchesFilter( [ 'regex' => $this->regex ] ) ) {
|
|
return self::FILTER_MSG;
|
|
}
|
|
|
|
// Skip deprecated preprocessor tests
|
|
if ( isset( $opts['preprocessor'] ) && $opts['preprocessor'] !== 'Preprocessor_Hash' ) {
|
|
wfDeprecated( 'preprocessor=Preprocessor_DOM', '1.36' );
|
|
return "Unsupported preprocessor type";
|
|
}
|
|
if ( $test->wikitext === null ) {
|
|
// Note that /in theory/ we could have pure html2html tests
|
|
// with no wikitext section, but /in practice/ all tests
|
|
// include a wikitext section.
|
|
$test->error( "Test lacks wikitext section", $test->testName );
|
|
}
|
|
// Skip disabled / filtered tests
|
|
if ( isset( $opts['disabled'] ) && !$this->runDisabled ) {
|
|
return "Test disabled";
|
|
}
|
|
|
|
// Skip parsoid-only tests if running in a legacy test mode
|
|
if (
|
|
$test->legacyHtml === null &&
|
|
self::getLegacyMetadataSection( $test ) === null
|
|
) {
|
|
// A Parsoid-only test should have one of the following sections
|
|
if (
|
|
isset( $test->sections['html/parsoid'] ) ||
|
|
isset( $test->sections['html/parsoid+integrated'] ) ||
|
|
isset( $test->sections['html/parsoid+standalone'] ) ||
|
|
isset( $test->sections['wikitext/edited'] ) ||
|
|
self::getParsoidMetadataSection( $test ) !== null
|
|
) {
|
|
if ( $isLegacy ) {
|
|
// Not an error, just skip this test if we're in
|
|
// legacy mode.
|
|
return "Parsoid-only test";
|
|
}
|
|
} else {
|
|
// This test lacks both a legacy html or metadata
|
|
// section and also any parsoid-specific html or
|
|
// metadata section or wikitext/edited section.
|
|
$test->error( "Test lacks html or metadata section", $test->testName );
|
|
}
|
|
}
|
|
return null;
|
|
}
|
|
|
|
public static function getLegacyMetadataSection( ParserTest $test ) {
|
|
return // specific results for legacy parser
|
|
$test->sections['metadata/php'] ??
|
|
// specific results for legacy parser and parsoid integrated mode
|
|
$test->sections['metadata/integrated'] ??
|
|
// generic for all parsers (even standalone)
|
|
$test->sections['metadata'] ??
|
|
// missing (== use legacy combined output format)
|
|
null;
|
|
}
|
|
|
|
public static function getParsoidMetadataSection( ParserTest $test ) {
|
|
return // specific results for parsoid integrated mode
|
|
$test->sections['metadata/parsoid+integrated'] ??
|
|
// specific results for parsoid
|
|
$test->sections['metadata/parsoid'] ??
|
|
// specific results for legacy parser and parsoid integrated mode
|
|
$test->sections['metadata/integrated'] ??
|
|
// generic for all parsers (even standalone)
|
|
$test->sections['metadata'] ??
|
|
// missing (== use legacy combined output format)
|
|
null;
|
|
}
|
|
|
|
/**
|
|
* Compute valid test modes based on requested modes and file-enabled modes
|
|
* @param string[] $testModes
|
|
* @param array $fileOptions
|
|
* @return string[]
|
|
*/
|
|
public function computeValidTestModes( array $testModes, array $fileOptions ): array {
|
|
$modeRestriction = $fileOptions['parsoid-compatible'] ?: false;
|
|
if ( $modeRestriction !== false ) {
|
|
if ( is_string( $modeRestriction ) ) {
|
|
// shorthand
|
|
$modeRestriction = [ $modeRestriction ];
|
|
}
|
|
$testModes = array_values( array_intersect( $testModes, $modeRestriction ) );
|
|
}
|
|
|
|
return $testModes;
|
|
}
|
|
|
|
/**
|
|
* Run the tests from a single file. staticSetup() and setupDatabase()
|
|
* must have been called already.
|
|
*
|
|
* @param ParserTest $t
|
|
* @param string[] $testModeStrs What Parsoid modes to run these these in?
|
|
* @return bool True if passed all modes, false if any mode failed.
|
|
*/
|
|
private function runTestInParsoidModes( ParserTest $t, array $testModeStrs ): bool {
|
|
if ( $this->options['updateKnownFailures'] ) {
|
|
// Reset known failures to ensure we reset newly skipped tests
|
|
$t->knownFailures = [];
|
|
}
|
|
|
|
$ok = true;
|
|
$runner = $this;
|
|
$t->testAllModes( $t->computeTestModes( $testModeStrs ), $this->options,
|
|
function ( ParserTest $test, string $modeStr, array $options ) use ( $runner, $t, &$ok ) {
|
|
// $test could be a clone of $t
|
|
// Ensure that updates to knownFailures in $test are reflected in $t
|
|
$test->knownFailures = &$t->knownFailures;
|
|
$mode = new ParserTestMode( $modeStr, $test->changetree );
|
|
if ( $modeStr === 'selser' && $test->changetree === null ) {
|
|
// This is an auto-edit test with either a CLI changetree
|
|
// or a change tree that should be generated
|
|
$mode = new ParserTestMode( 'selser-auto', json_decode( $runner->options['changetree'] ) );
|
|
}
|
|
$result = $this->runTestInternal( $test, $mode );
|
|
$ok = $ok && $result->isSuccess();
|
|
}
|
|
);
|
|
|
|
return $ok;
|
|
}
|
|
|
|
/**
|
|
* Update known failures JSON file for the parser tests file
|
|
* @param string $filename The parser test file
|
|
* @param TestFileReader $testFileInfo
|
|
*/
|
|
public function updateKnownFailures( string $filename, TestFileReader $testFileInfo ) {
|
|
$testKnownFailures = [];
|
|
foreach ( $testFileInfo->testCases as $t ) {
|
|
if ( $t->knownFailures && $t->testName ) {
|
|
// @phan-suppress-next-line PhanTypeMismatchDimAssignment False positive
|
|
$testKnownFailures[$t->testName] = $t->knownFailures;
|
|
// FIXME: This reduces noise when updateKnownFailures is used
|
|
// with a subset of test modes. But, this also mixes up the selser
|
|
// test results with non-selser ones.
|
|
// ksort( $testKnownFailures[$t->testName] );
|
|
}
|
|
}
|
|
// Sort, otherwise, titles get added above based on the first
|
|
// failing mode, which can make diffs harder to verify when
|
|
// failing modes change.
|
|
ksort( $testKnownFailures );
|
|
// Cast to object to ensure that empty list is encoded as `{}` not `[]`
|
|
$contents = FormatJson::encode( (object)$testKnownFailures, "\t", FormatJson::ALL_OK ) . "\n";
|
|
|
|
if ( file_exists( $testFileInfo->knownFailuresPath ) ) {
|
|
$old = file_get_contents( $testFileInfo->knownFailuresPath );
|
|
} else {
|
|
// If file doesn't exist, use the JSON representation of an
|
|
// empty array, so it compares equal in the case that we
|
|
// end up with an empty array of known failures below.
|
|
$old = "{}";
|
|
}
|
|
|
|
if ( $old !== $contents ) {
|
|
if ( $testFileInfo->knownFailuresPath ) {
|
|
$this->recorder->warning( "Updating known failures file: {$testFileInfo->knownFailuresPath}" );
|
|
file_put_contents( $testFileInfo->knownFailuresPath, $contents );
|
|
} else {
|
|
// To be safe, we don't try to write a file that doesn't
|
|
// (yet) exist. Create an empty file if you need to, and
|
|
// then we'll happily update it for you.
|
|
throw new RuntimeException(
|
|
"Known failures file for $filename does not exist, " .
|
|
"and so won't be updated."
|
|
);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param string $filename The parser test file
|
|
* @param TestFileReader $testFileInfo
|
|
* @param bool $isLegacy
|
|
*/
|
|
public function updateTests(
|
|
string $filename, TestFileReader $testFileInfo, bool $isLegacy
|
|
) {
|
|
$fileContent = file_get_contents( $filename );
|
|
foreach ( $testFileInfo->testCases as $t ) {
|
|
$testName = $t->testName;
|
|
$fail = $t->knownFailures[$isLegacy ? 'legacy' : 'wt2html'] ?? null;
|
|
$html = $isLegacy ? $t->legacyHtml : $t->parsoidHtml;
|
|
if ( $isLegacy && $fail === null ) {
|
|
$fail = $t->knownFailures['metadata'] ?? null;
|
|
$html = self::getLegacyMetadataSection( $t );
|
|
}
|
|
if ( $testName !== null && $fail !== null && $html !== null ) {
|
|
$exp = '/(!!\s*test\s*' .
|
|
preg_quote( $testName, '/' ) .
|
|
'(?:(?!!!\s*end)[\s\S])*' .
|
|
')(' . preg_quote( $html, '/' ) .
|
|
')/m';
|
|
$fileContent = preg_replace_callback(
|
|
$exp,
|
|
static function ( array $matches ) use ( $fail ) {
|
|
return $matches[1] . $fail;
|
|
},
|
|
$fileContent
|
|
);
|
|
}
|
|
}
|
|
file_put_contents( $filename, $fileContent );
|
|
}
|
|
|
|
/**
|
|
* @param string $wikitext
|
|
* @return array
|
|
*/
|
|
private function getRevRecordProperties( string $wikitext ): array {
|
|
return [
|
|
'pageid' => 187, // Some random fake page id
|
|
'revid' => 1337, // see Parser::getRevisionId()
|
|
'timestamp' => $this->getFakeTimestamp(),
|
|
'wikitext' => $wikitext
|
|
];
|
|
}
|
|
|
|
/**
|
|
* Create a mutable rev record for test use.
|
|
*
|
|
* @param LinkTarget $target
|
|
* @param UserIdentity $user
|
|
* @param array $revProps
|
|
* @return RevisionRecord
|
|
*/
|
|
private function createRevRecord( LinkTarget $target, UserIdentity $user, array $revProps ): RevisionRecord {
|
|
$content = new WikitextContent( $revProps['wikitext'] );
|
|
$title = MediaWikiServices::getInstance()->getTitleFactory()->newFromRow( (object)[
|
|
'page_id' => $revProps['pageid'],
|
|
'page_len' => $content->getSize(),
|
|
'page_latest' => $revProps['revid'],
|
|
'page_namespace' => $target->getNamespace(),
|
|
'page_title' => $target->getDBkey(),
|
|
'page_is_redirect' => 0
|
|
] );
|
|
|
|
$revRecord = new MutableRevisionRecord( $title );
|
|
$revRecord->setContent( SlotRecord::MAIN, $content )
|
|
->setUser( $user )
|
|
->setTimestamp( strval( $revProps['timestamp'] ) )
|
|
->setPageId( $title->getArticleID() )
|
|
->setId( $title->getLatestRevID() );
|
|
|
|
return $revRecord;
|
|
}
|
|
|
|
/**
|
|
* Shared code to initialize ParserOptions based on the $test object,
|
|
* used by both the legacy Parser and the Parsoid parser.
|
|
* @param ParserTest $test
|
|
* @param callable(IContextSource,LinkTarget,array):ParserOptions $parserOptionsCallback A callback to create the
|
|
* initial ParserOptions object. This allows for some minor
|
|
* differences in how the legacy Parser and Parsoid create this.
|
|
* @return array<LinkTarget|ParserOptions|int> An array of LinkTarget, ParserOptions, and integer revId.
|
|
* @phan-return array{0:LinkTarget,1:ParserOptions,2:int}
|
|
*/
|
|
private function setupParserOptions( ParserTest $test, callable $parserOptionsCallback ) {
|
|
$opts = $test->options;
|
|
$context = RequestContext::getMain();
|
|
$wikitext = $test->wikitext;
|
|
'@phan-var string $wikitext'; // assert that this is not null
|
|
$revProps = $this->getRevRecordProperties( $wikitext );
|
|
$user = $context->getUser();
|
|
$title = isset( $opts['title'] )
|
|
? MediaWikiServices::getInstance()->getTitleParser()->parseTitle( $opts['title'] )
|
|
: new TitleValue( NS_MAIN, 'Parser test' );
|
|
|
|
$revRecord = null;
|
|
if ( isset( $opts['lastsavedrevision'] ) ) {
|
|
$revRecord = $this->createRevRecord( $title, $user, $revProps );
|
|
$revProps['rev'] = $revRecord;
|
|
// Increment timestamp so that parser tests can distinguish between
|
|
// ParserOptions source and RevisionRecord
|
|
$revProps['timestamp'] += 234;
|
|
}
|
|
|
|
$options = $parserOptionsCallback( $context, $title, $revProps );
|
|
$options->setTimestamp(
|
|
MWTimestamp::convert( TS_MW, $revProps['timestamp'] )
|
|
);
|
|
$options->setUserLang( $context->getLanguage() );
|
|
|
|
if ( isset( $opts['lastsavedrevision'] ) ) {
|
|
$oldCallback = $options->getCurrentRevisionRecordCallback();
|
|
$options->setCurrentRevisionRecordCallback(
|
|
static function ( LinkTarget $link, $parser = null ) use ( $title, $revRecord, $oldCallback ) {
|
|
if ( $link->isSameLinkAs( $title ) ) {
|
|
return $revRecord;
|
|
} else {
|
|
return $oldCallback( $link, $parser );
|
|
}
|
|
}
|
|
);
|
|
}
|
|
|
|
if ( isset( $opts['maxincludesize'] ) ) {
|
|
$options->setMaxIncludeSize( $opts['maxincludesize'] );
|
|
}
|
|
if ( isset( $opts['maxtemplatedepth'] ) ) {
|
|
$options->setMaxTemplateDepth( $opts['maxtemplatedepth'] );
|
|
}
|
|
|
|
return [ $title, $options, $revProps['revid'] ];
|
|
}
|
|
|
|
/**
|
|
* Get a Parser object
|
|
*
|
|
* @return Parser
|
|
*/
|
|
public function getParser() {
|
|
$parserFactory = MediaWikiServices::getInstance()->getParserFactory();
|
|
$parser = $parserFactory->create(); // A fresh parser object.
|
|
ParserTestParserHook::setup( $parser );
|
|
return $parser;
|
|
}
|
|
|
|
/**
|
|
* Run a given wikitext input through either the legacy wiki parser
|
|
* or Parsoid, depending on the given test mode, and compare the
|
|
* output against the expected results.
|
|
*
|
|
* @param ParserTest $test The test parameters
|
|
* @param ParserTestMode $mode The test mode
|
|
* @return ParserTestResult The test results.
|
|
*/
|
|
private function runTestInternal( ParserTest $test, ParserTestMode $mode ): ParserTestResult {
|
|
$this->recorder->startTest( $test, $mode );
|
|
if ( $mode->isLegacy() ) {
|
|
$result = $this->runLegacyTest( $test, $mode );
|
|
} else {
|
|
// Parsoid might skip a test for unsupported features
|
|
$result = $this->runParsoidTest( $test, $mode );
|
|
if ( $result === false ) {
|
|
$this->recorder->skipped( $test, $mode, 'SKIP' );
|
|
return new ParserTestResult( $test, $mode, 'SKIP', 'SKIP' );
|
|
}
|
|
}
|
|
$this->recorder->record( $result );
|
|
return $result;
|
|
}
|
|
|
|
/**
|
|
* Run a given wikitext input through either the legacy wiki parser
|
|
* or Parsoid, depending on the given test mode, and compare the
|
|
* output against the expected results.
|
|
*
|
|
* @param ParserTest $test The test parameters
|
|
* @param ParserTestMode $mode The test mode
|
|
* @return ParserTestResult The test results.
|
|
*/
|
|
public function runTest( ParserTest $test, ParserTestMode $mode ): ParserTestResult {
|
|
$skipMessage = $this->getTestSkipMessage( $test, $mode->isLegacy() );
|
|
if ( $skipMessage !== null ) {
|
|
if ( $skipMessage !== self::FILTER_MSG ) {
|
|
$this->recorder->startTest( $test, $mode );
|
|
$this->recorder->skipped( $test, $mode, $skipMessage );
|
|
}
|
|
return new ParserTestResult( $test, $mode, 'SKIP', 'SKIP' );
|
|
}
|
|
return $this->runTestInternal( $test, $mode );
|
|
}
|
|
|
|
/**
|
|
* Run a given wikitext input through a freshly-constructed instance
|
|
* of the legacy wiki parser, and compare the output against the expected
|
|
* results.
|
|
*
|
|
* Prints status and explanatory messages to stdout.
|
|
*
|
|
* staticSetup() and setupWikiData() must be called before this function
|
|
* is entered.
|
|
*
|
|
* @param ParserTest $test The test parameters
|
|
* @param ParserTestMode $mode The test mode
|
|
* @return ParserTestResult
|
|
*/
|
|
public function runLegacyTest( ParserTest $test, ParserTestMode $mode ): ParserTestResult {
|
|
$desc = ( $test->comment ?? '' ) . $test->testName;
|
|
wfDebug( __METHOD__ . ": running $desc" );
|
|
$teardownGuard = $this->perTestSetup( $test );
|
|
[ $title, $options, $revId ] = $this->setupParserOptions(
|
|
$test,
|
|
static function ( $context, $title, $revProps ) {
|
|
return ParserOptions::newFromContext( $context );
|
|
}
|
|
);
|
|
|
|
$opts = $test->options;
|
|
$local = isset( $opts['local'] );
|
|
$parser = $this->getParser();
|
|
|
|
if ( isset( $opts['styletag'] ) ) {
|
|
// For testing the behavior of <style> (including those deduplicated
|
|
// into <link> tags), add tag hooks to allow them to be generated.
|
|
$parser->setHook( 'style', static function ( $content, $attributes, $parser ) {
|
|
$marker = Parser::MARKER_PREFIX . '-style-' . md5( $content ) . Parser::MARKER_SUFFIX;
|
|
// @phan-suppress-next-line SecurityCheck-XSS
|
|
$parser->getStripState()->addNoWiki( $marker, $content );
|
|
return Html::inlineStyle( $marker, 'all', $attributes );
|
|
} );
|
|
$parser->setHook( 'link', static function ( $content, $attributes, $parser ) {
|
|
return Html::element( 'link', $attributes );
|
|
} );
|
|
}
|
|
|
|
$wikitext = $test->wikitext;
|
|
$output = null;
|
|
$pageReference = new PageReferenceValue( $title->getNamespace(), $title->getDBkey(), PageReferenceValue::LOCAL );
|
|
'@phan-var string $wikitext'; // assert that this is not null
|
|
if ( isset( $opts['pst'] ) ) {
|
|
$out = $parser->preSaveTransform( $wikitext, $pageReference, $options->getUserIdentity(), $options );
|
|
$output = $parser->getOutput();
|
|
} elseif ( isset( $opts['msg'] ) ) {
|
|
$out = $parser->transformMsg( $wikitext, $options, $pageReference );
|
|
} elseif ( isset( $opts['section'] ) ) {
|
|
$section = $opts['section'];
|
|
$out = $parser->getSection( $wikitext, $section );
|
|
} elseif ( isset( $opts['replace'] ) ) {
|
|
$section = $opts['replace'][0];
|
|
$replace = $opts['replace'][1];
|
|
$out = $parser->replaceSection( $wikitext, $section, $replace );
|
|
} elseif ( isset( $opts['comment'] ) ) {
|
|
$out = MediaWikiServices::getInstance()->getCommentFormatter()->format( $wikitext, $title, $local );
|
|
} elseif ( isset( $opts['preload'] ) ) {
|
|
$out = $parser->getPreloadText( $wikitext, $pageReference, $options );
|
|
} else {
|
|
$output = $parser->parse( $wikitext, $pageReference, $options, true, true, $revId );
|
|
if ( isset( $opts['nohtml'] ) ) {
|
|
$out = '';
|
|
} else {
|
|
$out = $output->getText( [
|
|
'allowTOC' => !isset( $opts['notoc'] ),
|
|
'unwrap' => !isset( $opts['wrap'] ),
|
|
'skin' => $this->getSkin( $opts['skin'] ?? 'fallback' ),
|
|
] );
|
|
$out = rtrim( $out );
|
|
}
|
|
}
|
|
|
|
$metadataExpected = self::getLegacyMetadataSection( $test );
|
|
$metadataActual = null;
|
|
if ( $output ) {
|
|
$this->addParserOutputInfo(
|
|
$out, $output, $opts, $title,
|
|
$metadataExpected, $metadataActual
|
|
);
|
|
}
|
|
|
|
ScopedCallback::consume( $teardownGuard );
|
|
|
|
$rawOut = $out;
|
|
$expected = $test->legacyHtml ?? '';
|
|
if ( count( $this->normalizationFunctions ) ) {
|
|
$expected = ParserTestResultNormalizer::normalize(
|
|
$expected, $this->normalizationFunctions );
|
|
$out = ParserTestResultNormalizer::normalize( $out, $this->normalizationFunctions );
|
|
}
|
|
|
|
$testResult = new ParserTestResult( $test, $mode, $expected, $out );
|
|
|
|
if ( $this->options['update-tests'] && !$testResult->isSuccess() ) {
|
|
$test->knownFailures["$mode"] = $rawOut;
|
|
}
|
|
|
|
if ( $testResult->isSuccess() && $metadataExpected !== null ) {
|
|
$testResult = new ParserTestResult( $test, $mode, $metadataExpected, $metadataActual ?? '' );
|
|
|
|
if ( $this->options['update-tests'] && !$testResult->isSuccess() ) {
|
|
$test->knownFailures['metadata'] = $metadataActual;
|
|
}
|
|
}
|
|
|
|
return $testResult;
|
|
}
|
|
|
|
private function getSkin( $name ) {
|
|
static $skinCache = [];
|
|
|
|
$skinCache[$name] ??= MediaWikiServices::getInstance()->getSkinFactory()->makeSkin( $name );
|
|
return $skinCache[$name];
|
|
}
|
|
|
|
/**
|
|
* Add information from the parser output to the result string
|
|
*
|
|
* @param string &$out The "actual" parser output
|
|
* @param ParserOutput $output The "actual" parser metadata
|
|
* @param array $opts Test options
|
|
* @param ParsoidLinkTarget $title
|
|
* @param ?string $metadataExpected The contents of the !!metadata section,
|
|
* or null if it is missing
|
|
* @param ?string &$metadataActual The "actual" metadata output
|
|
*/
|
|
private function addParserOutputInfo(
|
|
&$out, ParserOutput $output, array $opts, ParsoidLinkTarget $title,
|
|
?string $metadataExpected, ?string &$metadataActual
|
|
) {
|
|
$before = [];
|
|
$after = [];
|
|
// The "before" entries may contain HTML.
|
|
if ( isset( $opts['showtitle'] ) ) {
|
|
if ( $output->getTitleText() ) {
|
|
$titleText = $output->getTitleText();
|
|
} else {
|
|
// TitleFormatter doesn't (yet) take ParsoidLinkTarget
|
|
// (which is identical to core's LinkTarget, but phan doesn't
|
|
// know that), so go through TitleValue for now.
|
|
$titleText = MediaWikiServices::getInstance()->getTitleFormatter()->getPrefixedText(
|
|
TitleValue::newFromLinkTarget( $title )
|
|
);
|
|
}
|
|
$before[] = $titleText;
|
|
}
|
|
|
|
if ( isset( $opts['showindicators'] ) ) {
|
|
foreach ( $output->getIndicators() as $id => $content ) {
|
|
$before[] = "$id=$content";
|
|
}
|
|
}
|
|
|
|
if ( isset( $opts['ill'] ) ) {
|
|
$after[] = implode( ' ', $output->getLanguageLinks() );
|
|
}
|
|
|
|
if ( isset( $opts['cat'] ) ) {
|
|
$defaultSortKey = $output->getPageProperty( 'defaultsort' ) ?? '';
|
|
foreach ( $output->getCategoryNames() as $name ) {
|
|
$sortkey = $output->getCategorySortKey( $name ) ?: $defaultSortKey;
|
|
$after[] = "cat=$name sort=$sortkey";
|
|
}
|
|
}
|
|
|
|
if ( isset( $opts['extension'] ) ) {
|
|
$extList = $opts['extension'];
|
|
if ( !is_array( $extList ) ) {
|
|
$extList = [ $extList ];
|
|
}
|
|
foreach ( $extList as $ext ) {
|
|
$after[] = "extension[$ext]=" .
|
|
// XXX should use JsonCodec
|
|
json_encode(
|
|
$output->getExtensionData( $ext ),
|
|
JSON_UNESCAPED_SLASHES | JSON_UNESCAPED_UNICODE | JSON_PRETTY_PRINT
|
|
);
|
|
}
|
|
}
|
|
|
|
if ( isset( $opts['property'] ) ) {
|
|
$propList = $opts['property'];
|
|
if ( !is_array( $propList ) ) {
|
|
$propList = [ $propList ];
|
|
}
|
|
foreach ( $propList as $prop ) {
|
|
$after[] = "property[$prop]=" .
|
|
( $output->getPageProperty( $prop ) ?? '' );
|
|
}
|
|
}
|
|
if ( isset( $opts['showflags'] ) ) {
|
|
$actualFlags = [];
|
|
foreach ( ParserOutputFlags::cases() as $name ) {
|
|
if ( $output->getOutputFlag( $name ) ) {
|
|
$actualFlags[] = $name;
|
|
}
|
|
}
|
|
sort( $actualFlags );
|
|
$after[] = "flags=" . implode( ', ', $actualFlags );
|
|
# In 1.21 we deprecated the use of arbitrary keys for
|
|
# ParserOutput::setFlag() by extensions; if we find anyone
|
|
# still doing that complain about it.
|
|
$oldFlags = array_diff_key(
|
|
TestingAccessWrapper::newFromObject( $output )->mFlags,
|
|
array_fill_keys( ParserOutputFlags::cases(), true )
|
|
);
|
|
if ( $oldFlags ) {
|
|
wfDeprecated( 'Arbitrary flags in ParserOutput', '1.39' );
|
|
}
|
|
}
|
|
if ( isset( $opts['showtocdata'] ) ) {
|
|
$tocData = $output->getTOCData();
|
|
if ( $tocData !== null ) {
|
|
$after[] = $tocData->prettyPrint();
|
|
}
|
|
}
|
|
if ( isset( $opts['showmedia'] ) ) {
|
|
$after[] = 'images=' . implode( ', ', array_keys( $output->getImages() ) );
|
|
}
|
|
if ( $metadataExpected === null ) {
|
|
// legacy format, add $before and $after to $out
|
|
if ( $before ) {
|
|
$before = implode( "\n", $before );
|
|
$out = "$before\n$out";
|
|
}
|
|
if ( $after ) {
|
|
if ( $out && !str_ends_with( $out, "\n" ) ) {
|
|
$out .= "\n";
|
|
}
|
|
$out .= implode( "\n", $after );
|
|
}
|
|
} else {
|
|
$metadataActual = implode( "\n", array_merge( $before, $after ) );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* This processes test results and updates the known failures info for the test
|
|
*
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @param string|null|callable $rawExpected
|
|
* If null, indicates that the normalizer should look for the expected
|
|
* output in the $test object.
|
|
* @param string $rawActual
|
|
* @param callable $normalizer normalizer of expected & actual output strings
|
|
* @return ParserTestResult|false
|
|
*/
|
|
private function processResults(
|
|
ParserTest $test, ParserTestMode $mode, $rawExpected, string $rawActual, callable $normalizer
|
|
) {
|
|
if ( $mode->isCachingMode() ) {
|
|
return false;
|
|
}
|
|
|
|
if ( !$this->options['knownFailures'] ) {
|
|
$expectedFailure = null;
|
|
} else {
|
|
$expectedFailure = $test->knownFailures["$mode"] ?? null;
|
|
}
|
|
|
|
$expectedToFail = $expectedFailure !== null;
|
|
$knownFailureChanged = $expectedToFail && $expectedFailure !== $rawActual;
|
|
|
|
if ( is_callable( $rawExpected ) ) {
|
|
$rawExpected = $rawExpected();
|
|
}
|
|
[ $actual, $expected ] = $normalizer( $rawActual, $rawExpected, false /* standalone */ );
|
|
$passed = $actual === $expected;
|
|
|
|
$unexpectedPass = $expectedToFail && $passed;
|
|
$unexpectedFail = !$expectedToFail && !$passed;
|
|
|
|
if ( $unexpectedPass ) {
|
|
$this->recorder->warning( "{$test->testName}: $mode: EXPECTED TO FAIL, BUT PASSED!" );
|
|
}
|
|
|
|
if ( $this->options['updateKnownFailures'] && (
|
|
$knownFailureChanged || $unexpectedFail || $unexpectedPass
|
|
) ) {
|
|
if ( $unexpectedPass ) {
|
|
unset( $test->knownFailures["$mode"] );
|
|
} else {
|
|
if ( $knownFailureChanged ) {
|
|
$this->recorder->warning( "{$test->testName}: $mode: KNOWN FAILURE CHANGED!" );
|
|
}
|
|
$test->knownFailures["$mode"] = $rawActual;
|
|
}
|
|
}
|
|
|
|
if ( $this->options['update-tests'] && !$passed ) {
|
|
$test->knownFailures["$mode"] = $rawActual;
|
|
}
|
|
|
|
if ( $unexpectedPass ) {
|
|
if ( !$this->options['updateKnownFailures'] ) {
|
|
$this->unexpectedTestPasses = true;
|
|
}
|
|
} elseif ( $expectedToFail && !$knownFailureChanged ) {
|
|
// Don't flag failures noisily when nothing really changed
|
|
$expected = $expectedFailure;
|
|
$actual = $rawActual;
|
|
}
|
|
|
|
return new ParserTestResult( $test, $mode, $expected, $actual );
|
|
}
|
|
|
|
private function fetchCachedDoc( $parsoid, $pageConfig, $test ) {
|
|
// If cachedBODYstr is not already set, create it here.
|
|
if ( $test->cachedBODYstr === null ) {
|
|
$this->wt2html( $parsoid, $pageConfig, $test, new ParserTestMode( 'cache' ) );
|
|
}
|
|
$doc = DOMUtils::parseHTML( $test->cachedBODYstr, true );
|
|
// XXX applyChanges calls functions in WTUtils which assume we
|
|
// have a DataBag associated with the document.
|
|
DOMDataUtils::prepareDoc( $doc );
|
|
return $doc;
|
|
}
|
|
|
|
private function fetchCachedWt( $parsoid, $pageConfig, $test ) {
|
|
// If cachedWTstr is not already set, set it here.
|
|
if ( ( $test->cachedWTstr ?? null ) === null ) {
|
|
$this->html2wt( $parsoid, $pageConfig, $test, new ParserTestMode( 'cache' ) );
|
|
}
|
|
return $test->cachedWTstr;
|
|
}
|
|
|
|
/**
|
|
* Run wt2html on the test
|
|
*
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
private function wt2html( Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode ) {
|
|
$html = $test->sections['html/parsoid+integrated'] ?? $test->parsoidHtml;
|
|
Assert::invariant(
|
|
$test->wikitext !== null,
|
|
"All tests include a wikitext section"
|
|
);
|
|
|
|
if ( $html === null && !$mode->isCachingMode() && !isset( $test->options['nohtml'] ) ) {
|
|
// Nothing to test, but if mode is 'cache' we're executing this
|
|
// in order to set cachedBODYstr (say, for a wt2wt test)
|
|
// If the 'nohtml' option is set, we're executing this in order
|
|
// to test metadata.
|
|
return false;
|
|
}
|
|
|
|
$metadata = new ParserOutput( 'parsoid' );
|
|
$origOut = $parsoid->wikitext2html( $pageConfig, [
|
|
'body_only' => true,
|
|
'wrapSections' => $test->options['parsoid']['wrapSections'] ?? false,
|
|
'traceFlags' => $this->options['traceFlags'],
|
|
'dumpFlags' => $this->options['dumpFlags']
|
|
], $headers, $metadata );
|
|
|
|
if ( isset( $test->options['nohtml'] ) ) {
|
|
// Suppress HTML (presumably because we want to test the metadata)
|
|
$origOut = '';
|
|
}
|
|
|
|
$metadataExpected = self::getParsoidMetadataSection( $test );
|
|
$metadataActual = null;
|
|
$titleParser = MediaWikiServices::getInstance()->getTitleParser();
|
|
$this->addParserOutputInfo(
|
|
$origOut, $metadata, $test->options,
|
|
$pageConfig->getLinkTarget(),
|
|
$metadataExpected, $metadataActual
|
|
);
|
|
|
|
$test->cachedBODYstr = $origOut;
|
|
|
|
$testResult = $this->processResults(
|
|
$test, $mode,
|
|
// Passing 'null' for expected output here asks normalizeHTML
|
|
// to look it up for us, based on parsoid-only & standalone status
|
|
isset( $test->options['nohtml'] ) ? '' : null,
|
|
$origOut, [ $test, "normalizeHTML" ]
|
|
);
|
|
if ( $metadataExpected !== null && !$mode->isCachingMode() ) {
|
|
$nullNormalizer = static function ( $actual, $expected, $ignore ) {
|
|
return [ $actual, $expected ];
|
|
};
|
|
$metadataResult = $this->processResults(
|
|
$test, new ParserTestMode( 'metadata' ), $metadataExpected, $metadataActual ?? '', $nullNormalizer
|
|
);
|
|
if ( $testResult === false || $testResult->isSuccess() ) {
|
|
// Ensure both results have to be successful
|
|
$testResult = $metadataResult;
|
|
}
|
|
}
|
|
return $testResult;
|
|
}
|
|
|
|
/**
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
private function wt2wt( Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode ) {
|
|
$html = $test->sections['html/parsoid+integrated'] ?? $test->parsoidHtml;
|
|
if ( $html === null ) {
|
|
// We currently don't run this in standalone mode.
|
|
// The expectation is to add html/parsoid sections
|
|
// if we want to run these tests.
|
|
return false;
|
|
}
|
|
|
|
Assert::invariant(
|
|
$test->wikitext !== null,
|
|
"All tests include a wikitext section"
|
|
);
|
|
|
|
// Handle a 'changes' option if present.
|
|
$doc = $this->fetchCachedDoc( $parsoid, $pageConfig, $test );
|
|
$testManualChanges = $test->options['parsoid']['changes'] ?? null;
|
|
if ( $testManualChanges ) {
|
|
$test->applyManualChanges( $doc );
|
|
}
|
|
|
|
$origWT = $parsoid->dom2wikitext( $pageConfig, $doc );
|
|
if ( isset( $test->options['parsoid']['changes'] ) ) {
|
|
$expectedWT = $test->sections['wikitext/edited'];
|
|
} else {
|
|
$expectedWT = $test->wikitext;
|
|
}
|
|
|
|
return $this->processResults(
|
|
$test, $mode, $expectedWT, $origWT, [ $test, "normalizeWT" ]
|
|
);
|
|
}
|
|
|
|
/**
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
private function html2wt( Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode ) {
|
|
$html = $test->sections['html/parsoid+integrated'] ?? $test->parsoidHtml;
|
|
if ( $html === null ) {
|
|
// Although ::getTestSkipMessage checks that parsoid-only tests
|
|
// have html sections, this test would have (eg) only an
|
|
// html/parsoid+standalone section, and we're executing in
|
|
// integrated mode.
|
|
return false; // Skip. Nothing to test.
|
|
}
|
|
if ( $test->wikitext === null && !$mode->isCachingMode() ) {
|
|
// If mode is 'cache' we're executing this in order to
|
|
// set cachedWTstr.
|
|
throw new RuntimeException( 'Error in the test setup' );
|
|
}
|
|
|
|
$test->cachedWTstr = $origWT = $parsoid->html2wikitext( $pageConfig, $html );
|
|
|
|
return $this->processResults(
|
|
$test, $mode, $test->wikitext ?? '', $origWT, [ $test, "normalizeWT" ]
|
|
);
|
|
}
|
|
|
|
/**
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
private function html2html( Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode ) {
|
|
$html = $test->sections['html/parsoid+integrated'] ?? $test->parsoidHtml;
|
|
if ( $html === null ) {
|
|
// Although ::getTestSkipMessage checks that parsoid-only tests
|
|
// have html sections, this test would have (eg) only an
|
|
// html/parsoid+standalone section, and we're executing in
|
|
// integrated mode.
|
|
return false; // Skip. Nothing to test.
|
|
}
|
|
|
|
$wt = $this->fetchCachedWt( $parsoid, $pageConfig, $test );
|
|
|
|
// Construct a fresh PageConfig object with $wt
|
|
$oldWt = $test->wikitext;
|
|
$test->wikitext = $wt;
|
|
$pageConfig = $this->setupParserOptionsAndBuildPageConfig( $test );
|
|
$test->wikitext = $oldWt;
|
|
|
|
$newHtml = $parsoid->wikitext2html( $pageConfig, [
|
|
'body_only' => true,
|
|
'wrapSections' => $test->options['parsoid']['wrapSections'] ?? false,
|
|
] );
|
|
|
|
return $this->processResults(
|
|
$test, $mode, $test->cachedNormalizedHTML, $newHtml, [ $test, "normalizeHTML" ]
|
|
);
|
|
}
|
|
|
|
/**
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
private function selser( Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode ) {
|
|
$html = $test->sections['html/parsoid+integrated'] ?? $test->parsoidHtml;
|
|
if ( $html === null ) {
|
|
// We currently don't run this in standalone mode.
|
|
// The expectation is to add html/parsoid sections
|
|
// if we want to run these tests.
|
|
return false;
|
|
}
|
|
|
|
Assert::invariant(
|
|
$test->wikitext !== null,
|
|
"All tests include a wikitext section"
|
|
);
|
|
|
|
if ( $test->changetree === [ 'manual' ] && !isset( $test->options['parsoid']['changes'] ) ) {
|
|
throw new RuntimeException( 'Error in the test setup!' );
|
|
}
|
|
|
|
// Apply edits to the HTML.
|
|
// Always serialize to string and reparse before passing to selser/wt2wt.
|
|
$doc = $this->fetchCachedDoc( $parsoid, $pageConfig, $test );
|
|
if ( $test->changetree === [ 'manual' ] ) {
|
|
$test->applyManualChanges( $doc );
|
|
$expectedWT = $test->sections['wikitext/edited'];
|
|
} else {
|
|
// $test->changetree === [ 5 ]
|
|
$changetree = $test->changetree;
|
|
'@phan-var array $changetree'; // assert that this is not null
|
|
$test->applyChanges( [], $doc, $changetree );
|
|
$expectedWT = $test->wikitext;
|
|
}
|
|
$editedHTML = ContentUtils::toXML( DOMCompat::getBody( $doc ) );
|
|
|
|
// Run selser on edited doc
|
|
$selserData = new SelserData( $test->wikitext, $test->cachedBODYstr );
|
|
$origWT = $parsoid->html2wikitext( $pageConfig, $editedHTML, [], $selserData );
|
|
|
|
if ( $test->changetree === [ 5 ] ) {
|
|
$origWT = preg_replace( '/<!--' . ParserTest::STATIC_RANDOM_STRING . '-->/', '', $origWT );
|
|
}
|
|
|
|
return $this->processResults(
|
|
$test, $mode, $expectedWT, $origWT, [ $test, "normalizeWT" ]
|
|
);
|
|
}
|
|
|
|
/**
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @param Document $doc
|
|
* @return array
|
|
*/
|
|
private function runSelserEditTest(
|
|
Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode, Document $doc
|
|
): array {
|
|
$test->applyChanges( [], $doc, $test->changetree );
|
|
$editedHTML = ContentUtils::toXML( DOMCompat::getBody( $doc ) );
|
|
|
|
// Run selser on edited doc
|
|
$selserData = new SelserData( $test->wikitext, $test->cachedBODYstr );
|
|
$origWT = $parsoid->html2wikitext( $pageConfig, $editedHTML, [], $selserData );
|
|
|
|
$ptResult = $this->processResults(
|
|
$test, $mode,
|
|
static function () use ( $parsoid, $pageConfig, $editedHTML ): string {
|
|
return $parsoid->html2wikitext( $pageConfig, $editedHTML );
|
|
},
|
|
$origWT, [ $test, "normalizeWT" ]
|
|
);
|
|
|
|
return [ $ptResult->actual, $ptResult->expected ];
|
|
}
|
|
|
|
/**
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
private function selserAutoEdit( Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode ) {
|
|
$html = $test->sections['html/parsoid+integrated'] ?? $test->parsoidHtml;
|
|
if ( $html === null ) {
|
|
// We currently don't run this in standalone mode.
|
|
// The expectation is to add html/parsoid sections
|
|
// if we want to run these tests.
|
|
return false;
|
|
}
|
|
Assert::invariant(
|
|
$test->wikitext !== null,
|
|
"All tests include a wikitext section"
|
|
);
|
|
|
|
// Apply edits to the HTML.
|
|
// Always serialize to string and reparse before passing to selser/wt2wt.
|
|
$doc = $this->fetchCachedDoc( $parsoid, $pageConfig, $test );
|
|
if ( !$test->changetree ) {
|
|
$test->changetree = $test->generateChanges( $doc );
|
|
if ( !$test->changetree ) {
|
|
// No more changes to make
|
|
return false;
|
|
}
|
|
$mode = new ParserTestMode( 'selser', $test->changetree );
|
|
}
|
|
[ $out, $expected ] = $this->runSelserEditTest( $parsoid, $pageConfig, $test, $mode, $doc );
|
|
return new ParserTestResult( $test, $mode, $expected, $out );
|
|
}
|
|
|
|
/**
|
|
* @param Parsoid $parsoid
|
|
* @param PageConfig $pageConfig
|
|
* @param ParserTest $test
|
|
* @param ParserTestMode $mode
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
private function selserAutoEditComposite( Parsoid $parsoid, PageConfig $pageConfig, ParserTest $test, ParserTestMode $mode ) {
|
|
$html = $test->sections['html/parsoid+integrated'] ?? $test->parsoidHtml;
|
|
if ( $html === null ) {
|
|
// We currently don't run this in standalone mode.
|
|
// The expectation is to add html/parsoid sections
|
|
// if we want to run these tests.
|
|
return false;
|
|
}
|
|
Assert::invariant(
|
|
$test->wikitext !== null,
|
|
"All tests include a wikitext section"
|
|
);
|
|
|
|
if ( $test->changetree ) {
|
|
// Apply edits to the HTML.
|
|
// Always serialize to string and reparse before passing to selser/wt2wt.
|
|
$doc = $this->fetchCachedDoc( $parsoid, $pageConfig, $test );
|
|
Assert::invariant(
|
|
$mode->changetree === $test->changetree,
|
|
"changetree should be consistent with mode"
|
|
);
|
|
[ $out, $expected ] = $this->runSelserEditTest( $parsoid, $pageConfig, $test, $mode, $doc );
|
|
return new ParserTestResult( $test, $mode, $expected, $out );
|
|
} else {
|
|
// this mode is a composite of multiple selser tests
|
|
$mode = new ParserTestMode( "selserAutoEdits" );
|
|
$runnerOpts = $this->getOptions();
|
|
$numChanges = $runnerOpts['numchanges'] ?: 20; // default in Parsoid
|
|
$bufOut = "";
|
|
$bufExpected = "";
|
|
for ( $i = 0; $i < $numChanges; $i++ ) {
|
|
// Apply edits to the HTML.
|
|
// Always serialize to string and reparse before passing to selser/wt2wt.
|
|
$doc = $this->fetchCachedDoc( $parsoid, $pageConfig, $test );
|
|
$test->seed = $i . '';
|
|
$test->changetree = $test->generateChanges( $doc );
|
|
if ( $test->changetree ) {
|
|
// new mode with the generated changetree
|
|
$nmode = new ParserTestMode( 'selser', $test->changetree );
|
|
[ $out, $expected ] = $this->runSelserEditTest( $parsoid, $pageConfig, $test, $nmode, $doc );
|
|
$testTitle = "TEST: {$test->testName} ($nmode)\n";
|
|
$bufOut .= $testTitle;
|
|
$bufExpected .= $testTitle;
|
|
$bufOut .= "RESULT: $out\n";
|
|
$bufExpected .= "RESULT: $expected\n";
|
|
}
|
|
// $test->changetree can be [] which is a NOP for testing
|
|
// but not a NOP for duplicate change tree tests.
|
|
if ( $test->isDuplicateChangeTree( $test->changetree ) ) {
|
|
// Once we get a duplicate change tree, we can no longer
|
|
// generate and run new tests. So, be done now!
|
|
break;
|
|
} else {
|
|
$test->selserChangeTrees[$i] = $test->changetree;
|
|
}
|
|
}
|
|
return new ParserTestResult( $test, $mode, $bufExpected, $bufOut );
|
|
}
|
|
}
|
|
|
|
private function setupParserOptionsAndBuildPageConfig( ParserTest $test ): PageConfig {
|
|
$services = MediaWikiServices::getInstance();
|
|
$pageConfigFactory = $services->get( 'ParsoidPageConfigFactory' );
|
|
$pageConfig = null;
|
|
$runner = $this;
|
|
$this->setupParserOptions(
|
|
$test,
|
|
static function ( $context, $title, $revProps ) use ( $runner, $pageConfigFactory, &$pageConfig ) {
|
|
$user = $context->getUser();
|
|
$revRecord = $revProps['rev'] ?? null;
|
|
if ( !$revRecord ) {
|
|
// Unlike the legacy parser which doesn't need an actual revrecord to parse
|
|
// wikitext, Parsoid creates a PageConfig which needs an actual revrecord.
|
|
// So create a fake mutable on here.
|
|
$revRecord = $runner->createRevRecord( $title, $user, $revProps );
|
|
}
|
|
$page = MediaWikiServices::getInstance()->getTitleFactory()->newFromLinkTarget( $title );
|
|
$pageConfig = $pageConfigFactory->create(
|
|
$page, $user, $revRecord, null, null
|
|
);
|
|
return $pageConfig->getParserOptions();
|
|
} );
|
|
'@phan-var PageConfig $pageConfig'; // assert that this is not null
|
|
return $pageConfig;
|
|
}
|
|
|
|
/**
|
|
* Helper function to register a Parsoid extension module in such a
|
|
* way that it can be cleaned up after the test is complete.
|
|
* @param SiteConfig $siteConfig
|
|
* @param class-string<ExtensionModule> $moduleName
|
|
* @return callable
|
|
*/
|
|
private static function registerExtensionModule( SiteConfig $siteConfig, string $moduleName ): callable {
|
|
$id = $siteConfig->registerExtensionModule( $moduleName );
|
|
return static function () use ( $siteConfig, $id ) {
|
|
$siteConfig->unregisterExtensionModule( $id );
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Run a given wikitext input through a freshly-constructed Parsoid parser,
|
|
* running in 'integrated' mode, and compare the output against the
|
|
* expected results.
|
|
*
|
|
* Prints status and explanatory messages to stdout.
|
|
*
|
|
* staticSetup() and setupWikiData() must be called before this function
|
|
* is entered.
|
|
*
|
|
* @param ParserTest $test The test parameters:
|
|
* @param ParserTestMode $mode Parsoid test mode to run (including specific
|
|
* changetree to run selser test in, if applicable)
|
|
*
|
|
* @return ParserTestResult|false false if skipped
|
|
*/
|
|
public function runParsoidTest( ParserTest $test, ParserTestMode $mode ) {
|
|
wfDebug( __METHOD__ . ": running {$test->testName} [$mode]" );
|
|
$opts = $test->options;
|
|
|
|
// Skip tests targetting features Parsoid doesn't (yet) support
|
|
// @todo T270312
|
|
if ( isset( $opts['styletag'] ) || isset( $opts['pst'] ) ||
|
|
isset( $opts['msg'] ) || isset( $opts['section'] ) ||
|
|
isset( $opts['replace'] ) || isset( $opts['comment'] ) ||
|
|
isset( $opts['preload'] )
|
|
) {
|
|
return false;
|
|
}
|
|
|
|
$teardownGuard = $this->perTestSetup( $test );
|
|
$pageConfig = $this->setupParserOptionsAndBuildPageConfig( $test );
|
|
|
|
$teardown = [];
|
|
|
|
// Register any special extensions required by this test case
|
|
$services = MediaWikiServices::getInstance();
|
|
$siteConfig = $services->get( 'ParsoidSiteConfig' );
|
|
$siteConfig->getLogger()->log( 'debug', "\n------ LOGS for {$test->testName} [$mode] ------" );
|
|
$teardown[] = self::registerExtensionModule( $siteConfig, ParsoidParserHook::class );
|
|
if ( ( $test->options['wgrawhtml'] ?? null ) === '1' ) {
|
|
$teardown[] = self::registerExtensionModule( $siteConfig, ParsoidRawHTML::class );
|
|
}
|
|
if ( isset( $test->options['styletag'] ) ) {
|
|
$teardown[] = self::registerExtensionModule( $siteConfig, ParsoidStyleTag::class );
|
|
}
|
|
// unregister these after this test
|
|
$teardownGuard = $this->createTeardownObject( $teardown, $teardownGuard );
|
|
|
|
// Create the Parsoid object. (This is cheap, since the SiteConfig
|
|
// and DataAccess are cached by the ServiceContainer.)
|
|
$dataAccess = $services->get( 'ParsoidDataAccess' );
|
|
$parsoid = new Parsoid( $siteConfig, $dataAccess );
|
|
switch ( $mode->mode ) {
|
|
case 'wt2html':
|
|
case 'wt2html+integrated':
|
|
$res = $this->wt2html( $parsoid, $pageConfig, $test, $mode );
|
|
break;
|
|
|
|
case 'wt2wt':
|
|
$res = $this->wt2wt( $parsoid, $pageConfig, $test, $mode );
|
|
break;
|
|
|
|
case 'html2wt':
|
|
$res = $this->html2wt( $parsoid, $pageConfig, $test, $mode );
|
|
break;
|
|
|
|
case 'html2html':
|
|
$res = $this->html2html( $parsoid, $pageConfig, $test, $mode );
|
|
break;
|
|
|
|
case 'selser':
|
|
$test->changetree = $mode->changetree;
|
|
$res = $this->selser( $parsoid, $pageConfig, $test, $mode );
|
|
$test->changetree = null; // Reset after each selser test
|
|
break;
|
|
|
|
case 'selser-auto-composite':
|
|
$test->changetree = $mode->changetree;
|
|
$res = $this->selserAutoEditComposite( $parsoid, $pageConfig, $test, $mode );
|
|
$test->changetree = null; // Reset after each selser test
|
|
break;
|
|
|
|
case 'selser-auto':
|
|
$test->changetree = $mode->changetree;
|
|
$res = $this->selserAutoEdit( $parsoid, $pageConfig, $test, $mode );
|
|
if ( $res === false && !$test->changetree ) {
|
|
// TEMPORARY HACK
|
|
// If we don't run any selser tests, ensure changetree is not null
|
|
// because we are going to call Test::isDuplicateChangeTree() on it
|
|
// This ensures that we continue to crash if change tree is null
|
|
// for any other reason!
|
|
$test->changetree = [];
|
|
}
|
|
// Don't reset changetree here -- it is used to detect duplicate trees
|
|
// and stop selser test generation in Test.php::testAllModes
|
|
break;
|
|
|
|
default:
|
|
// Unsupported Mode
|
|
$res = false;
|
|
break;
|
|
}
|
|
|
|
ScopedCallback::consume( $teardownGuard );
|
|
return $res;
|
|
}
|
|
|
|
/**
|
|
* Use a regex to find out the value of an option
|
|
* @param string $key Name of option val to retrieve
|
|
* @param array $opts Options array to look in
|
|
* @param mixed $default Default value returned if not found
|
|
* @return mixed
|
|
*/
|
|
private static function getOptionValue( $key, $opts, $default ) {
|
|
$key = strtolower( $key );
|
|
return $opts[$key] ?? $default;
|
|
}
|
|
|
|
/**
|
|
* Do any required setup which is dependent on test options.
|
|
*
|
|
* @see staticSetup() for more information about setup/teardown
|
|
*
|
|
* @param ParserTest $test Test info supplied by TestFileReader
|
|
* @param callable|null $nextTeardown
|
|
* @return ScopedCallback
|
|
*/
|
|
public function perTestSetup( ParserTest $test, $nextTeardown = null ) {
|
|
$teardown = [];
|
|
|
|
$this->checkSetupDone( 'setupDatabase' );
|
|
$teardown[] = $this->markSetupDone( 'perTestSetup' );
|
|
|
|
$opts = $test->options;
|
|
$mwServices = MediaWikiServices::getInstance();
|
|
|
|
// Find out values for some special options.
|
|
$langCode =
|
|
self::getOptionValue( 'language', $opts, 'en' );
|
|
$variant = // old deprecated option name
|
|
self::getOptionValue( 'variant', $opts, false );
|
|
$variant =
|
|
self::getOptionValue( 'htmlVariantLanguage', $opts, $variant );
|
|
$maxtoclevel =
|
|
self::getOptionValue( 'wgMaxTocLevel', $opts, 999 );
|
|
$linkHolderBatchSize =
|
|
self::getOptionValue( 'wgLinkHolderBatchSize', $opts, 1000 );
|
|
|
|
$setup = [
|
|
'wgEnableUploads' => self::getOptionValue( 'wgEnableUploads', $opts, true ),
|
|
'wgLanguageCode' => $langCode,
|
|
'wgRawHtml' => self::getOptionValue( 'wgRawHtml', $opts, false ),
|
|
'wgNamespacesWithSubpages' => array_fill_keys(
|
|
$mwServices->getNamespaceInfo()->getValidNamespaces(),
|
|
isset( $opts['subpage'] )
|
|
),
|
|
'wgMaxTocLevel' => $maxtoclevel,
|
|
'wgAllowExternalImages' => self::getOptionValue( 'wgAllowExternalImages', $opts, true ),
|
|
'wgThumbLimits' => [ 0, 0, 0, 0, 0, self::getOptionValue( 'thumbsize', $opts, 180 ) ],
|
|
'wgDefaultLanguageVariant' => $variant,
|
|
'wgLinkHolderBatchSize' => $linkHolderBatchSize,
|
|
// Set as a JSON object like:
|
|
// wgEnableMagicLinks={"ISBN":false, "PMID":false, "RFC":false}
|
|
'wgEnableMagicLinks' => self::getOptionValue( 'wgEnableMagicLinks', $opts, [] )
|
|
+ [ 'ISBN' => true, 'PMID' => true, 'RFC' => true ],
|
|
// Test with legacy encoding by default until HTML5 is very stable and default
|
|
'wgFragmentMode' => [ 'legacy' ],
|
|
];
|
|
|
|
if ( isset( $opts['externallinktarget'] ) ) {
|
|
$setup['wgExternalLinkTarget'] = self::getOptionValue( 'externallinktarget', $opts, '' );
|
|
}
|
|
|
|
$nonIncludable = self::getOptionValue( 'wgNonincludableNamespaces', $opts, false );
|
|
if ( $nonIncludable !== false ) {
|
|
$setup['wgNonincludableNamespaces'] = [ $nonIncludable ];
|
|
}
|
|
|
|
if ( $test->config ) {
|
|
foreach ( $test->config as $var => $value ) {
|
|
$setup[$var] = $value;
|
|
}
|
|
}
|
|
|
|
/** @since 1.20 */
|
|
( new HookRunner( $mwServices->getHookContainer() ) )->onParserTestGlobals( $setup );
|
|
|
|
// Take care of the content language and variant
|
|
$resetVariant = $variant !== false || isset( $test->config['wgUsePigLatinVariant'] );
|
|
if ( $langCode !== 'en' || $resetVariant ) {
|
|
$this->resetLanguageServices( $setup, $teardown, $resetVariant );
|
|
}
|
|
|
|
// Other services that can be configured per test
|
|
$reset = function () use ( $mwServices ) {
|
|
// Configurations changed above
|
|
$this->resetTitleServices();
|
|
$mwServices->resetServiceForTesting( 'MagicWordFactory' );
|
|
$mwServices->resetServiceForTesting( 'ParserFactory' );
|
|
// Depends on $wgParserEnableLegacyMediaDOM
|
|
$mwServices->resetServiceForTesting( 'Tidy' );
|
|
// The SiteConfig depends on various services that reset above,
|
|
// so reset it as well.
|
|
// T310283: be more selective about resetting SiteConfig if
|
|
// performance is a concern.
|
|
$mwServices->resetServiceForTesting( 'ParsoidSiteConfig' );
|
|
// Depends on the legacy ParserFactory
|
|
$mwServices->resetServiceForTesting( 'ParsoidDataAccess' );
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
|
|
// Clear language conversion tables
|
|
$teardown[] = static function () use ( $mwServices, $langCode ) {
|
|
$lang = $mwServices->getLanguageFactory()->getLanguage( $langCode );
|
|
$wrapper = TestingAccessWrapper::newFromObject(
|
|
$mwServices->getLanguageConverterFactory()
|
|
->getLanguageConverter( $lang )
|
|
);
|
|
@$wrapper->reloadTables();
|
|
};
|
|
|
|
// Reset OOUI auto-increment IDs (T345515)
|
|
OOUI\Tag::resetElementId();
|
|
|
|
$teardown[] = $this->executeSetupSnippets( $setup );
|
|
|
|
return $this->createTeardownObject( $teardown, $nextTeardown );
|
|
}
|
|
|
|
private function resetLanguageServices( array &$setup, array &$teardown, $resetVariant = false ) {
|
|
$mwServices = MediaWikiServices::getInstance();
|
|
$reset = static function () use ( $mwServices, $resetVariant ) {
|
|
$mwServices->resetServiceForTesting( 'ContentLanguage' );
|
|
|
|
if ( $resetVariant ) {
|
|
// If !!config touches $wgUsePigLatinVariant, these need to be reset
|
|
$mwServices->resetServiceForTesting( 'LanguageConverterFactory' );
|
|
$mwServices->resetServiceForTesting( 'LanguageFactory' );
|
|
$mwServices->resetServiceForTesting( 'LanguageNameUtils' );
|
|
$mwServices->resetServiceForTesting( 'LocalisationCache' );
|
|
}
|
|
|
|
// Content language invalidates a lot of services.
|
|
$mwServices->resetServiceForTesting( 'UserOptionsManager' );
|
|
$mwServices->resetServiceForTesting( 'UserOptionsLookup' );
|
|
$mwServices->resetServiceForTesting( 'MessageCache' );
|
|
|
|
// Purge the cached Language object
|
|
$context = RequestContext::getMain();
|
|
$context->setUser( $context->getUser() );
|
|
};
|
|
$setup[] = $reset;
|
|
$teardown[] = $reset;
|
|
}
|
|
|
|
/**
|
|
* Set up temporary DB tables.
|
|
*
|
|
* For best performance, call this once only for all tests. However, it can
|
|
* be called at the start of each test if more isolation is desired.
|
|
*
|
|
*
|
|
* Do not call this function from a MediaWikiIntegrationTestCase subclass,
|
|
* since MediaWikiIntegrationTestCase does its own DB setup.
|
|
*
|
|
* @see staticSetup() for more information about setup/teardown
|
|
*
|
|
* @param ScopedCallback|null $nextTeardown The next teardown object
|
|
* @return ScopedCallback The teardown object
|
|
*/
|
|
public function setupDatabase( $nextTeardown = null ) {
|
|
global $wgDBprefix;
|
|
|
|
$lb = MediaWikiServices::getInstance()->getDBLoadBalancer();
|
|
$this->db = $lb->getConnectionInternal( DB_PRIMARY );
|
|
|
|
$suspiciousPrefixes = [ self::DB_PREFIX, MediaWikiIntegrationTestCase::DB_PREFIX ];
|
|
if ( in_array( $wgDBprefix, $suspiciousPrefixes ) ) {
|
|
throw new RuntimeException( "\$wgDBprefix=$wgDBprefix suggests DB setup is already done" );
|
|
}
|
|
|
|
$teardown = [];
|
|
$teardown[] = $this->markSetupDone( 'setupDatabase' );
|
|
|
|
// Set up a test DB just for parser tests
|
|
MediaWikiIntegrationTestCase::setupAllTestDBs(
|
|
$this->db,
|
|
self::DB_PREFIX,
|
|
true // postgres requires that we use temporary tables
|
|
);
|
|
MediaWikiIntegrationTestCase::resetNonServiceCaches();
|
|
$teardown[] = static function () {
|
|
MediaWikiIntegrationTestCase::teardownTestDB();
|
|
};
|
|
|
|
MediaWikiIntegrationTestCase::installMockMwServices();
|
|
$teardown[] = static function () {
|
|
MediaWikiIntegrationTestCase::restoreMwServices();
|
|
};
|
|
|
|
// Wipe some DB query result caches on setup and teardown
|
|
$reset = static function () {
|
|
$services = MediaWikiServices::getInstance();
|
|
$services->getLinkCache()->clear();
|
|
|
|
// Clear the message cache
|
|
$services->getMessageCache()->clear();
|
|
};
|
|
$reset();
|
|
$teardown[] = $reset;
|
|
return $this->createTeardownObject( $teardown, $nextTeardown );
|
|
}
|
|
|
|
/**
|
|
* Add data about uploads to the new test DB, and set up the upload
|
|
* directory. This should be called after setupDatabase().
|
|
*
|
|
* @param ScopedCallback|null $nextTeardown The next teardown object
|
|
* @return ScopedCallback The teardown object
|
|
*/
|
|
public function setupUploads( $nextTeardown = null ) {
|
|
$teardown = [];
|
|
|
|
$this->checkSetupDone( 'setupDatabase' );
|
|
$teardown[] = $this->markSetupDone( 'setupUploads' );
|
|
|
|
// Create the files in the upload directory (or pretend to create them
|
|
// in a MockFileBackend). Append teardown callback.
|
|
$teardown[] = $this->setupUploadBackend();
|
|
|
|
// Create a authority
|
|
$performer = new UltimateAuthority( new UserIdentityValue( 0, User::MAINTENANCE_SCRIPT_USER ) );
|
|
|
|
// Register the uploads in the database
|
|
$localRepo = MediaWikiServices::getInstance()->getRepoGroup()->getLocalRepo();
|
|
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'Foobar.jpg' ) );
|
|
# note that the size/width/height/bits/etc of the file
|
|
# are actually set by inspecting the file itself; the arguments
|
|
# to recordUpload3 have no effect. That said, we try to make things
|
|
# match up so it is less confusing to readers of the code & tests.
|
|
$image->recordUpload3(
|
|
'',
|
|
'Upload of some lame file', 'Some lame file',
|
|
$performer,
|
|
[
|
|
'size' => 7881,
|
|
'width' => 1941,
|
|
'height' => 220,
|
|
'bits' => 8,
|
|
'media_type' => MEDIATYPE_BITMAP,
|
|
'mime' => 'image/jpeg',
|
|
'metadata' => [],
|
|
'sha1' => Wikimedia\base_convert( '1', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20010115123500' )
|
|
);
|
|
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'Thumb.png' ) );
|
|
# again, note that size/width/height below are ignored; see above.
|
|
$image->recordUpload3(
|
|
'',
|
|
'Upload of some lame thumbnail',
|
|
'Some lame thumbnail',
|
|
$performer,
|
|
[
|
|
'size' => 22589,
|
|
'width' => 135,
|
|
'height' => 135,
|
|
'bits' => 8,
|
|
'media_type' => MEDIATYPE_BITMAP,
|
|
'mime' => 'image/png',
|
|
'metadata' => [],
|
|
'sha1' => Wikimedia\base_convert( '2', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20130225203040' )
|
|
);
|
|
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'Foobar.svg' ) );
|
|
$image->recordUpload3(
|
|
'',
|
|
'Upload of some lame SVG',
|
|
'Some lame SVG',
|
|
$performer,
|
|
[
|
|
'size' => 12345,
|
|
'width' => 240,
|
|
'height' => 180,
|
|
'bits' => 0,
|
|
'media_type' => MEDIATYPE_DRAWING,
|
|
'mime' => 'image/svg+xml',
|
|
'metadata' => [
|
|
'version' => SvgHandler::SVG_METADATA_VERSION,
|
|
'width' => 240,
|
|
'height' => 180,
|
|
'originalWidth' => '100%',
|
|
'originalHeight' => '100%',
|
|
'translations' => [
|
|
'en' => SVGReader::LANG_FULL_MATCH,
|
|
'ru' => SVGReader::LANG_FULL_MATCH,
|
|
],
|
|
],
|
|
'sha1' => Wikimedia\base_convert( '', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20010115123500' )
|
|
);
|
|
|
|
# This image will be prohibited via the list in [[MediaWiki:Bad image list]]
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'Bad.jpg' ) );
|
|
$image->recordUpload3(
|
|
'',
|
|
'zomgnotcensored',
|
|
'Borderline image',
|
|
$performer,
|
|
[
|
|
'size' => 12345,
|
|
'width' => 320,
|
|
'height' => 240,
|
|
'bits' => 24,
|
|
'media_type' => MEDIATYPE_BITMAP,
|
|
'mime' => 'image/jpeg',
|
|
'metadata' => [],
|
|
'sha1' => Wikimedia\base_convert( '3', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20010115123500' )
|
|
);
|
|
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'Hi-ho.jpg' ) );
|
|
$image->recordUpload3(
|
|
'',
|
|
'Hi',
|
|
'ho',
|
|
$performer,
|
|
[
|
|
'size' => 7881,
|
|
'width' => 1941,
|
|
'height' => 220,
|
|
'bits' => 8,
|
|
'media_type' => MEDIATYPE_BITMAP,
|
|
'mime' => 'image/jpeg',
|
|
'metadata' => [],
|
|
'sha1' => Wikimedia\base_convert( '1', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20010115123500' )
|
|
);
|
|
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'Video.ogv' ) );
|
|
$image->recordUpload3(
|
|
'',
|
|
'A pretty movie',
|
|
'Will it play',
|
|
$performer,
|
|
[
|
|
'size' => 12345,
|
|
'width' => 320,
|
|
'height' => 240,
|
|
'bits' => 0,
|
|
'media_type' => MEDIATYPE_VIDEO,
|
|
'mime' => 'application/ogg',
|
|
'metadata' => [],
|
|
'sha1' => Wikimedia\base_convert( '', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20010115123500' )
|
|
);
|
|
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'Audio.oga' ) );
|
|
$image->recordUpload3(
|
|
'',
|
|
'An awesome hitsong',
|
|
'Will it play',
|
|
$performer,
|
|
[
|
|
'size' => 12345,
|
|
'width' => 0,
|
|
'height' => 0,
|
|
'bits' => 0,
|
|
'media_type' => MEDIATYPE_AUDIO,
|
|
'mime' => 'application/ogg',
|
|
'metadata' => [],
|
|
'sha1' => Wikimedia\base_convert( '', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20010115123500' )
|
|
);
|
|
|
|
# A DjVu file
|
|
$image = $localRepo->newFile( new TitleValue( NS_FILE, 'LoremIpsum.djvu' ) );
|
|
$djvuMetadata = [
|
|
'data' => [
|
|
'pages' => [
|
|
[ 'height' => 3508, 'width' => 2480, 'dpi' => 300, 'gamma' => 2.2 ],
|
|
[ 'height' => 3508, 'width' => 2480, 'dpi' => 300, 'gamma' => 2.2 ],
|
|
[ 'height' => 3508, 'width' => 2480, 'dpi' => 300, 'gamma' => 2.2 ],
|
|
[ 'height' => 3508, 'width' => 2480, 'dpi' => 300, 'gamma' => 2.2 ],
|
|
[ 'height' => 3508, 'width' => 2480, 'dpi' => 300, 'gamma' => 2.2 ],
|
|
],
|
|
],
|
|
];
|
|
$image->recordUpload3(
|
|
'',
|
|
'Upload a DjVu',
|
|
'A DjVu',
|
|
$performer,
|
|
[
|
|
'size' => 3249,
|
|
'width' => 2480,
|
|
'height' => 3508,
|
|
'bits' => 0,
|
|
'media_type' => MEDIATYPE_OFFICE,
|
|
'mime' => 'image/vnd.djvu',
|
|
'metadata' => $djvuMetadata,
|
|
'sha1' => Wikimedia\base_convert( '', 16, 36, 31 ),
|
|
'fileExists' => true
|
|
],
|
|
$this->db->timestamp( '20010115123600' )
|
|
);
|
|
|
|
return $this->createTeardownObject( $teardown, $nextTeardown );
|
|
}
|
|
|
|
/**
|
|
* Upload test files to the backend created by createRepoGroup().
|
|
*
|
|
* @return callable The teardown callback
|
|
*/
|
|
private function setupUploadBackend() {
|
|
global $IP;
|
|
|
|
$repo = MediaWikiServices::getInstance()->getRepoGroup()->getLocalRepo();
|
|
$base = $repo->getZonePath( 'public' );
|
|
$backend = $repo->getBackend();
|
|
$backend->prepare( [ 'dir' => "$base/3/3a" ] );
|
|
$backend->quickStore( [
|
|
'src' => "$IP/tests/phpunit/data/parser/headbg.jpg",
|
|
'dst' => "$base/3/3a/Foobar.jpg"
|
|
] );
|
|
$backend->prepare( [ 'dir' => "$base/e/ea" ] );
|
|
$backend->quickStore( [
|
|
'src' => "$IP/tests/phpunit/data/parser/wiki.png",
|
|
'dst' => "$base/e/ea/Thumb.png"
|
|
] );
|
|
$backend->prepare( [ 'dir' => "$base/0/09" ] );
|
|
$backend->quickStore( [
|
|
'src' => "$IP/tests/phpunit/data/parser/headbg.jpg",
|
|
'dst' => "$base/0/09/Bad.jpg"
|
|
] );
|
|
$backend->prepare( [ 'dir' => "$base/5/5f" ] );
|
|
$backend->quickStore( [
|
|
'src' => "$IP/tests/phpunit/data/parser/LoremIpsum.djvu",
|
|
'dst' => "$base/5/5f/LoremIpsum.djvu"
|
|
] );
|
|
|
|
// No helpful SVG file to copy, so make one ourselves
|
|
$data = '<?xml version="1.0" encoding="utf-8"?>' .
|
|
'<svg xmlns="http://www.w3.org/2000/svg"' .
|
|
' version="1.1" width="240" height="180"/>';
|
|
|
|
$backend->prepare( [ 'dir' => "$base/f/ff" ] );
|
|
$backend->quickCreate( [
|
|
'content' => $data, 'dst' => "$base/f/ff/Foobar.svg"
|
|
] );
|
|
|
|
return function () use ( $backend ) {
|
|
if ( $backend instanceof MockFileBackend ) {
|
|
// In memory backend, so dont bother cleaning them up.
|
|
return;
|
|
}
|
|
$this->teardownUploadBackend();
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Remove the dummy uploads directory
|
|
*/
|
|
private function teardownUploadBackend() {
|
|
if ( $this->keepUploads ) {
|
|
return;
|
|
}
|
|
|
|
$public = MediaWikiServices::getInstance()->getRepoGroup()->getLocalRepo()
|
|
->getZonePath( 'public' );
|
|
|
|
$this->deleteFiles(
|
|
[
|
|
"$public/3/3a/Foobar.jpg",
|
|
"$public/e/ea/Thumb.png",
|
|
"$public/0/09/Bad.jpg",
|
|
"$public/5/5f/LoremIpsum.djvu",
|
|
"$public/f/ff/Foobar.svg",
|
|
"$public/0/00/Video.ogv",
|
|
"$public/4/41/Audio.oga",
|
|
]
|
|
);
|
|
}
|
|
|
|
/**
|
|
* Delete the specified files and their parent directories
|
|
* @param string[] $files File backend URIs mwstore://...
|
|
*/
|
|
private function deleteFiles( $files ) {
|
|
// Delete the files
|
|
$backend = MediaWikiServices::getInstance()->getRepoGroup()->getLocalRepo()->getBackend();
|
|
foreach ( $files as $file ) {
|
|
$backend->quickDelete( [ 'src' => $file ], [ 'force' => 1 ] );
|
|
}
|
|
|
|
// Delete the parent directories
|
|
foreach ( $files as $file ) {
|
|
$tmp = FileBackend::parentStoragePath( $file );
|
|
while ( $tmp ) {
|
|
if ( !$backend->clean( [ 'dir' => $tmp ] )->isOK() ) {
|
|
break;
|
|
}
|
|
$tmp = FileBackend::parentStoragePath( $tmp );
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Add articles to the test DB.
|
|
*
|
|
* @see staticSetup() for more information about setup/teardown
|
|
*
|
|
* @param ParserTestArticle[] $articles Article info array from TestFileReader
|
|
* @param ?ScopedCallback $nextTeardown The next teardown object
|
|
* @return ScopedCallback The teardown object
|
|
*/
|
|
public function addArticles(
|
|
array $articles, ?ScopedCallback $nextTeardown = null
|
|
): ScopedCallback {
|
|
$this->checkSetupDone( 'setupDatabase' );
|
|
$this->checkSetupDone( 'staticSetup' );
|
|
|
|
foreach ( $articles as $info ) {
|
|
$this->addArticle( $info->title, $info->text, $info->filename, $info->lineNumStart );
|
|
}
|
|
|
|
$teardown = [];
|
|
$services = MediaWikiServices::getInstance();
|
|
|
|
// Wipe WANObjectCache process cache, which is invalidated by article insertion
|
|
// due to T144706
|
|
$services->getMainWANObjectCache()->clearProcessCache();
|
|
// Avoid reuse of lazy-loaded Title::mArticleId after page creation
|
|
Title::clearCaches();
|
|
|
|
$reset = static function () use ( $services ) {
|
|
// Reset the service so that any "MediaWiki:bad image list" articles
|
|
// added get fetched
|
|
$services->resetServiceForTesting( 'BadFileLookup' );
|
|
// Depends on BadFileLookup, also reset here in case it no longer
|
|
// depends on the legacy ParserFactory which reset per test.
|
|
$services->resetServiceForTesting( 'ParsoidDataAccess' );
|
|
};
|
|
$reset();
|
|
|
|
$teardown[] = $reset;
|
|
$teardown[] = function () use ( $articles ) {
|
|
$this->cleanupArticles( $articles );
|
|
};
|
|
|
|
return $this->createTeardownObject( $teardown, $nextTeardown );
|
|
}
|
|
|
|
/**
|
|
* Remove articles from the test DB. This prevents independent parser
|
|
* test files from having conflicts when they choose the same names
|
|
* for article or template test fixtures.
|
|
*
|
|
* @param ParserTestArticle[] $articles Article info array from TestFileReader
|
|
*/
|
|
public function cleanupArticles( $articles ) {
|
|
$this->checkSetupDone( 'setupDatabase' );
|
|
$this->checkSetupDone( 'staticSetup' );
|
|
$deleter = new UltimateAuthority( new UserIdentityValue( 0, User::MAINTENANCE_SCRIPT_USER ) );
|
|
$wikiPageFactory = MediaWikiServices::getInstance()->getWikiPageFactory();
|
|
$titleParser = MediaWikiServices::getInstance()->getTitleParser();
|
|
$delPageFactory = MediaWikiServices::getInstance()->getDeletePageFactory();
|
|
foreach ( $articles as $info ) {
|
|
$name = self::chomp( $info->title );
|
|
$title = $titleParser->parseTitle( $name );
|
|
$page = $wikiPageFactory->newFromLinkTarget( $title );
|
|
$delPageFactory->newDeletePage( $page, $deleter )->deleteUnsafe( 'cleaning up' );
|
|
}
|
|
|
|
// Clear the static cache that Title class maintains.
|
|
// This ensures that Parsoid test runs that follow legacy parser test runs
|
|
// don't reuse titles. This matters because it looks like legacy test run
|
|
// and Parsoid test run differ in the number of articles they create in the db.
|
|
// We need to investigate that separately, but given that they differ, titles
|
|
// will get different article and revision ids across test runs.
|
|
// While we could add this to resetTitleServices(), there is really
|
|
// no reason to clear this for every test. Sufficient to clear this
|
|
// once per test file.
|
|
// Also the LinkCache is cleared which holds some information about titles
|
|
Title::clearCaches();
|
|
}
|
|
|
|
/**
|
|
* Insert a temporary test article
|
|
*
|
|
* @todo Refactor to share more code w/ ::editPage
|
|
*
|
|
* @param string $name The title, including any prefix
|
|
* @param string $text The article text
|
|
* @param string $file The input file name
|
|
* @param int|string $line The input line number, for reporting errors
|
|
*/
|
|
private function addArticle( $name, $text, $file, $line ) {
|
|
$text = self::chomp( $text );
|
|
$name = self::chomp( $name );
|
|
|
|
$services = MediaWikiServices::getInstance();
|
|
|
|
$title = $services->getTitleParser()->parseTitle( $name );
|
|
wfDebug( __METHOD__ . ": adding $name" );
|
|
|
|
if ( $title === null ) {
|
|
throw new RuntimeException( "invalid title '$name' at $file:$line\n" );
|
|
}
|
|
|
|
$performer = new UltimateAuthority( new UserIdentityValue( 0, User::MAINTENANCE_SCRIPT_USER ) );
|
|
|
|
$page = $services->getWikiPageFactory()->newFromLinkTarget( $title );
|
|
$page->loadPageData( IDBAccessObject::READ_LATEST );
|
|
$newContent = $page->getContentHandler()->unserializeContent( $text );
|
|
|
|
if ( $page->exists() ) {
|
|
$content = $page->getContent( RevisionRecord::RAW );
|
|
// Only reject the title, if the content/content model is different.
|
|
// This makes it easier to create Template:(( or Template:)) in different extensions
|
|
if ( $newContent->equals( $content ) ) {
|
|
return;
|
|
}
|
|
throw new RuntimeException(
|
|
"duplicate article '$name' with different content at $file:$line\n"
|
|
);
|
|
}
|
|
|
|
// Optionally use mock parser, to make debugging of actual parser tests simpler.
|
|
// But initialise the MessageCache clone first, don't let MessageCache
|
|
// get a reference to the mock object.
|
|
if ( $this->disableSaveParse ) {
|
|
$services->getMessageCache()->getParser();
|
|
$services->disableService( 'Parser' );
|
|
$services->disableService( 'ParserFactory' );
|
|
$services->redefineService(
|
|
'Parser',
|
|
static function () {
|
|
return new ParserTestMockParser;
|
|
}
|
|
);
|
|
$restore = static function () {
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'Parser' );
|
|
MediaWikiServices::getInstance()->resetServiceForTesting( 'ParserFactory' );
|
|
};
|
|
} else {
|
|
$restore = false;
|
|
}
|
|
$status = null;
|
|
try {
|
|
$status = $page->doUserEditContent(
|
|
$newContent,
|
|
$performer,
|
|
'',
|
|
EDIT_NEW | EDIT_SUPPRESS_RC | EDIT_INTERNAL
|
|
);
|
|
} finally {
|
|
if ( $restore ) {
|
|
$restore();
|
|
}
|
|
}
|
|
|
|
if ( !$status->isOK() ) {
|
|
throw new RuntimeException( $status->__toString() );
|
|
}
|
|
|
|
// an edit always attempt to purge backlink links such as history
|
|
// pages. That is unnecessary.
|
|
$jobQueueGroup = $services->getJobQueueGroup();
|
|
$jobQueueGroup->get( 'htmlCacheUpdate' )->delete();
|
|
// WikiPages::doEditUpdates randomly adds RC purges
|
|
$jobQueueGroup->get( 'recentChangesUpdate' )->delete();
|
|
|
|
// The RepoGroup cache is invalidated by the creation of file redirects
|
|
if ( $title->inNamespace( NS_FILE ) ) {
|
|
$services->getRepoGroup()->clearCache( $title );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Check if a hook is installed
|
|
*
|
|
* @param string $name
|
|
* @return bool True if tag hook is present
|
|
*/
|
|
public function requireHook( $name ) {
|
|
$parser = MediaWikiServices::getInstance()->getParser();
|
|
|
|
if ( preg_match( '/^[Ee]xtension:(.*)$/', $name, $matches ) ) {
|
|
$extName = $matches[1];
|
|
if ( ExtensionRegistry::getInstance()->isLoaded( $extName ) ) {
|
|
return true;
|
|
} else {
|
|
$this->recorder->warning( " Skipping this test suite because it requires the '$extName' " .
|
|
"extension, which isn't loaded." );
|
|
return false;
|
|
}
|
|
}
|
|
if ( in_array( $name, $parser->getTags(), true ) ) {
|
|
return true;
|
|
} else {
|
|
$this->recorder->warning( " Skipping this test suite because it requires the '$name' hook, " .
|
|
"which isn't provided by any loaded extension." );
|
|
return false;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Check if a function hook is installed
|
|
*
|
|
* @param string $name
|
|
* @return bool True if function hook is present
|
|
*/
|
|
public function requireFunctionHook( $name ) {
|
|
$parser = MediaWikiServices::getInstance()->getParser();
|
|
|
|
if ( in_array( $name, $parser->getFunctionHooks(), true ) ) {
|
|
return true;
|
|
} else {
|
|
$this->recorder->warning( " This test suite requires the '$name' function " .
|
|
"hook extension, skipping." );
|
|
return false;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Fake constant timestamp to make sure time-related parser
|
|
* functions give a persistent value.
|
|
*
|
|
* - Parser::expandMagicVariable (via ParserGetVariableValueTs hook)
|
|
* - Parser::preSaveTransform (via ParserOptions)
|
|
* @return int Fake constant timestamp.
|
|
*/
|
|
private function getFakeTimestamp() {
|
|
// parsed as '1970-01-01T00:02:03Z'
|
|
return 123;
|
|
}
|
|
}
|