We already use CONN_TRX_AUTOCOMMIT for mysql/postgres and a separate DB file (e.g. "server" config) for sqlite (via the installer). For any legacy case where the main DB is still used with sqlite, the removed method was unlikely to be very effective. Ignore DBO_TRX/DBO_DEFAULT in JobQueueDB "server" config for good measure, similar to SqlBagOStuff. This removes more direct uses of IDatabase::setFlags(), a method which does not play well with DBConnRef. Bug: T311090 Change-Id: Ia8457dea2ed30539e23345f89cb6b382be442975
955 lines
27 KiB
PHP
955 lines
27 KiB
PHP
<?php
|
|
/**
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with this program; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
* http://www.gnu.org/copyleft/gpl.html
|
|
*
|
|
* @file
|
|
*/
|
|
use MediaWiki\MediaWikiServices;
|
|
use Wikimedia\Rdbms\DBConnectionError;
|
|
use Wikimedia\Rdbms\DBError;
|
|
use Wikimedia\Rdbms\IDatabase;
|
|
use Wikimedia\Rdbms\IMaintainableDatabase;
|
|
use Wikimedia\Rdbms\IReadableDatabase;
|
|
use Wikimedia\Rdbms\RawSQLValue;
|
|
use Wikimedia\Rdbms\SelectQueryBuilder;
|
|
use Wikimedia\Rdbms\ServerInfo;
|
|
|
|
/**
|
|
* Database-backed job queue storage.
|
|
*
|
|
* @since 1.21
|
|
* @ingroup JobQueue
|
|
*/
|
|
class JobQueueDB extends JobQueue {
|
|
/* seconds to cache info without re-validating */
|
|
private const CACHE_TTL_SHORT = 30;
|
|
/* seconds a job can live once claimed */
|
|
private const MAX_AGE_PRUNE = 7 * 24 * 3600;
|
|
/**
|
|
* Used for job_random, the highest safe 32-bit signed integer.
|
|
* Equivalent to `( 2 ** 31 ) - 1` on 64-bit.
|
|
*/
|
|
private const MAX_JOB_RANDOM = 2_147_483_647;
|
|
/* maximum number of rows to skip */
|
|
private const MAX_OFFSET = 255;
|
|
|
|
/** @var IMaintainableDatabase|DBError|null */
|
|
protected $conn;
|
|
|
|
/** @var array|null Server configuration array */
|
|
protected $server;
|
|
/** @var string|null Name of an external DB cluster or null for the local DB cluster */
|
|
protected $cluster;
|
|
|
|
/**
|
|
* Additional parameters include:
|
|
* - server : Server configuration array for Database::factory. Overrides "cluster".
|
|
* - cluster : The name of an external cluster registered via LBFactory.
|
|
* If not specified, the primary DB cluster for the wiki will be used.
|
|
* This can be overridden with a custom cluster so that DB handles will
|
|
* be retrieved via LBFactory::getExternalLB() and getConnection().
|
|
* @param array $params
|
|
*/
|
|
protected function __construct( array $params ) {
|
|
parent::__construct( $params );
|
|
|
|
if ( isset( $params['server'] ) ) {
|
|
$this->server = $params['server'];
|
|
// Always use autocommit mode, even if DBO_TRX is configured
|
|
$this->server['flags'] ??= 0;
|
|
$this->server['flags'] &= ~( IDatabase::DBO_TRX | IDatabase::DBO_DEFAULT );
|
|
} elseif ( isset( $params['cluster'] ) && is_string( $params['cluster'] ) ) {
|
|
$this->cluster = $params['cluster'];
|
|
}
|
|
}
|
|
|
|
protected function supportedOrders() {
|
|
return [ 'random', 'timestamp', 'fifo' ];
|
|
}
|
|
|
|
protected function optimalOrder() {
|
|
return 'random';
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doIsEmpty()
|
|
* @return bool
|
|
*/
|
|
protected function doIsEmpty() {
|
|
$dbr = $this->getReplicaDB();
|
|
try {
|
|
// unclaimed job
|
|
$found = (bool)$dbr->newSelectQueryBuilder()
|
|
->select( '1' )
|
|
->from( 'job' )
|
|
->where( [ 'job_cmd' => $this->type, 'job_token' => '' ] )
|
|
->caller( __METHOD__ )->fetchField();
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
|
|
return !$found;
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doGetSize()
|
|
* @return int
|
|
*/
|
|
protected function doGetSize() {
|
|
$key = $this->getCacheKey( 'size' );
|
|
|
|
$size = $this->wanCache->get( $key );
|
|
if ( is_int( $size ) ) {
|
|
return $size;
|
|
}
|
|
|
|
$dbr = $this->getReplicaDB();
|
|
try {
|
|
$size = $dbr->newSelectQueryBuilder()
|
|
->from( 'job' )
|
|
->where( [ 'job_cmd' => $this->type, 'job_token' => '' ] )
|
|
->caller( __METHOD__ )->fetchRowCount();
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
$this->wanCache->set( $key, $size, self::CACHE_TTL_SHORT );
|
|
|
|
return $size;
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doGetAcquiredCount()
|
|
* @return int
|
|
*/
|
|
protected function doGetAcquiredCount() {
|
|
if ( $this->claimTTL <= 0 ) {
|
|
return 0; // no acknowledgements
|
|
}
|
|
|
|
$key = $this->getCacheKey( 'acquiredcount' );
|
|
|
|
$count = $this->wanCache->get( $key );
|
|
if ( is_int( $count ) ) {
|
|
return $count;
|
|
}
|
|
|
|
$dbr = $this->getReplicaDB();
|
|
try {
|
|
$count = $dbr->newSelectQueryBuilder()
|
|
->from( 'job' )
|
|
->where( [
|
|
'job_cmd' => $this->type,
|
|
$dbr->expr( 'job_token', '!=', '' ),
|
|
] )
|
|
->caller( __METHOD__ )->fetchRowCount();
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
$this->wanCache->set( $key, $count, self::CACHE_TTL_SHORT );
|
|
|
|
return $count;
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doGetAbandonedCount()
|
|
* @return int
|
|
* @throws JobQueueConnectionError
|
|
* @throws JobQueueError
|
|
*/
|
|
protected function doGetAbandonedCount() {
|
|
if ( $this->claimTTL <= 0 ) {
|
|
return 0; // no acknowledgements
|
|
}
|
|
|
|
$key = $this->getCacheKey( 'abandonedcount' );
|
|
|
|
$count = $this->wanCache->get( $key );
|
|
if ( is_int( $count ) ) {
|
|
return $count;
|
|
}
|
|
|
|
$dbr = $this->getReplicaDB();
|
|
try {
|
|
$count = $dbr->newSelectQueryBuilder()
|
|
->from( 'job' )
|
|
->where(
|
|
[
|
|
'job_cmd' => $this->type,
|
|
$dbr->expr( 'job_token', '!=', '' ),
|
|
$dbr->expr( 'job_attempts', '>=', $this->maxTries ),
|
|
]
|
|
)
|
|
->caller( __METHOD__ )->fetchRowCount();
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
|
|
$this->wanCache->set( $key, $count, self::CACHE_TTL_SHORT );
|
|
|
|
return $count;
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doBatchPush()
|
|
* @param IJobSpecification[] $jobs
|
|
* @param int $flags
|
|
* @throws DBError|Exception
|
|
* @return void
|
|
*/
|
|
protected function doBatchPush( array $jobs, $flags ) {
|
|
$dbw = $this->getPrimaryDB();
|
|
// In general, there will be two cases here:
|
|
// a) sqlite; DB connection is probably a regular round-aware handle.
|
|
// If the connection is busy with a transaction, then defer the job writes
|
|
// until right before the main round commit step. Any errors that bubble
|
|
// up will rollback the main commit round.
|
|
// b) mysql/postgres; DB connection is generally a separate CONN_TRX_AUTOCOMMIT handle.
|
|
// No transaction is active nor will be started by writes, so enqueue the jobs
|
|
// now so that any errors will show up immediately as the interface expects. Any
|
|
// errors that bubble up will rollback the main commit round.
|
|
$fname = __METHOD__;
|
|
$dbw->onTransactionPreCommitOrIdle(
|
|
function ( IDatabase $dbw ) use ( $jobs, $flags, $fname ) {
|
|
$this->doBatchPushInternal( $dbw, $jobs, $flags, $fname );
|
|
},
|
|
$fname
|
|
);
|
|
}
|
|
|
|
/**
|
|
* This function should *not* be called outside of JobQueueDB
|
|
*
|
|
* @suppress SecurityCheck-SQLInjection Bug in phan-taint-check handling bulk inserts
|
|
* @param IDatabase $dbw
|
|
* @param IJobSpecification[] $jobs
|
|
* @param int $flags
|
|
* @param string $method
|
|
* @throws DBError
|
|
* @return void
|
|
*/
|
|
public function doBatchPushInternal( IDatabase $dbw, array $jobs, $flags, $method ) {
|
|
if ( $jobs === [] ) {
|
|
return;
|
|
}
|
|
|
|
$rowSet = []; // (sha1 => job) map for jobs that are de-duplicated
|
|
$rowList = []; // list of jobs for jobs that are not de-duplicated
|
|
foreach ( $jobs as $job ) {
|
|
$row = $this->insertFields( $job, $dbw );
|
|
if ( $job->ignoreDuplicates() ) {
|
|
$rowSet[$row['job_sha1']] = $row;
|
|
} else {
|
|
$rowList[] = $row;
|
|
}
|
|
}
|
|
|
|
if ( $flags & self::QOS_ATOMIC ) {
|
|
$dbw->startAtomic( $method ); // wrap all the job additions in one transaction
|
|
}
|
|
try {
|
|
// Strip out any duplicate jobs that are already in the queue...
|
|
if ( count( $rowSet ) ) {
|
|
$res = $dbw->newSelectQueryBuilder()
|
|
->select( 'job_sha1' )
|
|
->from( 'job' )
|
|
->where(
|
|
[
|
|
// No job_type condition since it's part of the job_sha1 hash
|
|
'job_sha1' => array_map( 'strval', array_keys( $rowSet ) ),
|
|
'job_token' => '' // unclaimed
|
|
]
|
|
)
|
|
->caller( $method )->fetchResultSet();
|
|
foreach ( $res as $row ) {
|
|
wfDebug( "Job with hash '{$row->job_sha1}' is a duplicate." );
|
|
unset( $rowSet[$row->job_sha1] ); // already enqueued
|
|
}
|
|
}
|
|
// Build the full list of job rows to insert
|
|
$rows = array_merge( $rowList, array_values( $rowSet ) );
|
|
// Insert the job rows in chunks to avoid replica DB lag...
|
|
foreach ( array_chunk( $rows, 50 ) as $rowBatch ) {
|
|
$dbw->newInsertQueryBuilder()
|
|
->insertInto( 'job' )
|
|
->rows( $rowBatch )
|
|
->caller( $method )->execute();
|
|
}
|
|
$this->incrStats( 'inserts', $this->type, count( $rows ) );
|
|
$this->incrStats( 'dupe_inserts', $this->type,
|
|
count( $rowSet ) + count( $rowList ) - count( $rows )
|
|
);
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
if ( $flags & self::QOS_ATOMIC ) {
|
|
$dbw->endAtomic( $method );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doPop()
|
|
* @return RunnableJob|false
|
|
*/
|
|
protected function doPop() {
|
|
$job = false; // job popped off
|
|
try {
|
|
$uuid = wfRandomString( 32 ); // pop attempt
|
|
do { // retry when our row is invalid or deleted as a duplicate
|
|
// Try to reserve a row in the DB...
|
|
if ( in_array( $this->order, [ 'fifo', 'timestamp' ] ) ) {
|
|
$row = $this->claimOldest( $uuid );
|
|
} else { // random first
|
|
$rand = mt_rand( 0, self::MAX_JOB_RANDOM ); // encourage concurrent UPDATEs
|
|
$gte = (bool)mt_rand( 0, 1 ); // find rows with rand before/after $rand
|
|
$row = $this->claimRandom( $uuid, $rand, $gte );
|
|
}
|
|
// Check if we found a row to reserve...
|
|
if ( !$row ) {
|
|
break; // nothing to do
|
|
}
|
|
$this->incrStats( 'pops', $this->type );
|
|
|
|
// Get the job object from the row...
|
|
$job = $this->jobFromRow( $row );
|
|
break; // done
|
|
} while ( true );
|
|
|
|
if ( !$job || mt_rand( 0, 9 ) == 0 ) {
|
|
// Handled jobs that need to be recycled/deleted;
|
|
// any recycled jobs will be picked up next attempt
|
|
$this->recycleAndDeleteStaleJobs();
|
|
}
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
|
|
return $job;
|
|
}
|
|
|
|
/**
|
|
* Reserve a row with a single UPDATE without holding row locks over RTTs...
|
|
*
|
|
* @param string $uuid 32 char hex string
|
|
* @param int $rand Random unsigned integer (31 bits)
|
|
* @param bool $gte Search for job_random >= $random (otherwise job_random <= $random)
|
|
* @return stdClass|false Row|false
|
|
*/
|
|
protected function claimRandom( $uuid, $rand, $gte ) {
|
|
$dbw = $this->getPrimaryDB();
|
|
// Check cache to see if the queue has <= OFFSET items
|
|
$tinyQueue = $this->wanCache->get( $this->getCacheKey( 'small' ) );
|
|
|
|
$invertedDirection = false; // whether one job_random direction was already scanned
|
|
// This uses a replication safe method for acquiring jobs. One could use UPDATE+LIMIT
|
|
// instead, but that either uses ORDER BY (in which case it deadlocks in MySQL) or is
|
|
// not replication safe. Due to https://bugs.mysql.com/bug.php?id=6980, subqueries cannot
|
|
// be used here with MySQL.
|
|
do {
|
|
if ( $tinyQueue ) { // queue has <= MAX_OFFSET rows
|
|
// For small queues, using OFFSET will overshoot and return no rows more often.
|
|
// Instead, this uses job_random to pick a row (possibly checking both directions).
|
|
$row = $dbw->newSelectQueryBuilder()
|
|
->select( self::selectFields() )
|
|
->from( 'job' )
|
|
->where(
|
|
[
|
|
'job_cmd' => $this->type,
|
|
'job_token' => '', // unclaimed
|
|
$dbw->expr( 'job_random', $gte ? '>=' : '<=', $rand )
|
|
]
|
|
)
|
|
->orderBy(
|
|
'job_random',
|
|
$gte ? SelectQueryBuilder::SORT_ASC : SelectQueryBuilder::SORT_DESC
|
|
)
|
|
->caller( __METHOD__ )->fetchRow();
|
|
if ( !$row && !$invertedDirection ) {
|
|
$gte = !$gte;
|
|
$invertedDirection = true;
|
|
continue; // try the other direction
|
|
}
|
|
} else { // table *may* have >= MAX_OFFSET rows
|
|
// T44614: "ORDER BY job_random" with a job_random inequality causes high CPU
|
|
// in MySQL if there are many rows for some reason. This uses a small OFFSET
|
|
// instead of job_random for reducing excess claim retries.
|
|
$row = $dbw->newSelectQueryBuilder()
|
|
->select( self::selectFields() )
|
|
->from( 'job' )
|
|
->where(
|
|
[
|
|
'job_cmd' => $this->type,
|
|
'job_token' => '', // unclaimed
|
|
]
|
|
)
|
|
->offset( mt_rand( 0, self::MAX_OFFSET ) )
|
|
->caller( __METHOD__ )->fetchRow();
|
|
if ( !$row ) {
|
|
$tinyQueue = true; // we know the queue must have <= MAX_OFFSET rows
|
|
$this->wanCache->set( $this->getCacheKey( 'small' ), 1, 30 );
|
|
continue; // use job_random
|
|
}
|
|
}
|
|
|
|
if ( !$row ) {
|
|
break;
|
|
}
|
|
|
|
$dbw->newUpdateQueryBuilder()
|
|
->update( 'job' ) // update by PK
|
|
->set( [
|
|
'job_token' => $uuid,
|
|
'job_token_timestamp' => $dbw->timestamp(),
|
|
'job_attempts' => new RawSQLValue( 'job_attempts+1' ),
|
|
] )
|
|
->where( [
|
|
'job_cmd' => $this->type,
|
|
'job_id' => $row->job_id,
|
|
'job_token' => ''
|
|
] )
|
|
->caller( __METHOD__ )->execute();
|
|
|
|
// This might get raced out by another runner when claiming the previously
|
|
// selected row. The use of job_random should minimize this problem, however.
|
|
if ( !$dbw->affectedRows() ) {
|
|
$row = false; // raced out
|
|
}
|
|
} while ( !$row );
|
|
|
|
return $row;
|
|
}
|
|
|
|
/**
|
|
* Reserve a row with a single UPDATE without holding row locks over RTTs...
|
|
*
|
|
* @param string $uuid 32 char hex string
|
|
* @return stdClass|false Row|false
|
|
*/
|
|
protected function claimOldest( $uuid ) {
|
|
$dbw = $this->getPrimaryDB();
|
|
|
|
$row = false; // the row acquired
|
|
do {
|
|
if ( $dbw->getType() === 'mysql' ) {
|
|
// Per https://bugs.mysql.com/bug.php?id=6980, we can't use subqueries on the
|
|
// same table being changed in an UPDATE query in MySQL (gives Error: 1093).
|
|
// Postgres has no such limitation. However, MySQL offers an
|
|
// alternative here by supporting ORDER BY + LIMIT for UPDATE queries.
|
|
$dbw->query( "UPDATE {$dbw->tableName( 'job' )} " .
|
|
"SET " .
|
|
"job_token = {$dbw->addQuotes( $uuid ) }, " .
|
|
"job_token_timestamp = {$dbw->addQuotes( $dbw->timestamp() )}, " .
|
|
"job_attempts = job_attempts+1 " .
|
|
"WHERE ( " .
|
|
"job_cmd = {$dbw->addQuotes( $this->type )} " .
|
|
"AND job_token = {$dbw->addQuotes( '' )} " .
|
|
") ORDER BY job_id ASC LIMIT 1",
|
|
__METHOD__
|
|
);
|
|
} else {
|
|
// Use a subquery to find the job, within an UPDATE to claim it.
|
|
// This uses as much of the DB wrapper functions as possible.
|
|
$qb = $dbw->newSelectQueryBuilder()
|
|
->select( 'job_id' )
|
|
->from( 'job' )
|
|
->where( [ 'job_cmd' => $this->type, 'job_token' => '' ] )
|
|
->orderBy( 'job_id', SelectQueryBuilder::SORT_ASC )
|
|
->limit( 1 );
|
|
|
|
$dbw->newUpdateQueryBuilder()
|
|
->update( 'job' )
|
|
->set( [
|
|
'job_token' => $uuid,
|
|
'job_token_timestamp' => $dbw->timestamp(),
|
|
'job_attempts' => new RawSQLValue( 'job_attempts+1' ),
|
|
] )
|
|
->where( [ 'job_id' => new RawSQLValue( '(' . $qb->getSQL() . ')' ) ] )
|
|
->caller( __METHOD__ )->execute();
|
|
}
|
|
|
|
if ( !$dbw->affectedRows() ) {
|
|
break;
|
|
}
|
|
|
|
// Fetch any row that we just reserved...
|
|
$row = $dbw->newSelectQueryBuilder()
|
|
->select( self::selectFields() )
|
|
->from( 'job' )
|
|
->where( [ 'job_cmd' => $this->type, 'job_token' => $uuid ] )
|
|
->caller( __METHOD__ )->fetchRow();
|
|
if ( !$row ) { // raced out by duplicate job removal
|
|
wfDebug( "Row deleted as duplicate by another process." );
|
|
}
|
|
} while ( !$row );
|
|
|
|
return $row;
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doAck()
|
|
* @param RunnableJob $job
|
|
* @throws JobQueueConnectionError
|
|
* @throws JobQueueError
|
|
*/
|
|
protected function doAck( RunnableJob $job ) {
|
|
$id = $job->getMetadata( 'id' );
|
|
if ( $id === null ) {
|
|
throw new UnexpectedValueException( "Job of type '{$job->getType()}' has no ID." );
|
|
}
|
|
|
|
$dbw = $this->getPrimaryDB();
|
|
try {
|
|
// Delete a row with a single DELETE without holding row locks over RTTs...
|
|
$dbw->newDeleteQueryBuilder()
|
|
->deleteFrom( 'job' )
|
|
->where( [ 'job_cmd' => $this->type, 'job_id' => $id ] )
|
|
->caller( __METHOD__ )->execute();
|
|
|
|
$this->incrStats( 'acks', $this->type );
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doDeduplicateRootJob()
|
|
* @param IJobSpecification $job
|
|
* @throws JobQueueConnectionError
|
|
* @return bool
|
|
*/
|
|
protected function doDeduplicateRootJob( IJobSpecification $job ) {
|
|
// Callers should call JobQueueGroup::push() before this method so that if the
|
|
// insert fails, the de-duplication registration will be aborted. Since the insert
|
|
// is deferred till "transaction idle", do the same here, so that the ordering is
|
|
// maintained. Having only the de-duplication registration succeed would cause
|
|
// jobs to become no-ops without any actual jobs that made them redundant.
|
|
$dbw = $this->getPrimaryDB();
|
|
$dbw->onTransactionCommitOrIdle(
|
|
function () use ( $job ) {
|
|
parent::doDeduplicateRootJob( $job );
|
|
},
|
|
__METHOD__
|
|
);
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doDelete()
|
|
* @return bool
|
|
*/
|
|
protected function doDelete() {
|
|
$dbw = $this->getPrimaryDB();
|
|
try {
|
|
$dbw->newDeleteQueryBuilder()
|
|
->deleteFrom( 'job' )
|
|
->where( [ 'job_cmd' => $this->type ] )
|
|
->caller( __METHOD__ )->execute();
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::doWaitForBackups()
|
|
* @return void
|
|
*/
|
|
protected function doWaitForBackups() {
|
|
if ( $this->server ) {
|
|
return; // not using LBFactory instance
|
|
}
|
|
|
|
$lbFactory = MediaWikiServices::getInstance()->getDBLoadBalancerFactory();
|
|
$lbFactory->waitForReplication();
|
|
}
|
|
|
|
/**
|
|
* @return void
|
|
*/
|
|
protected function doFlushCaches() {
|
|
foreach ( [ 'size', 'acquiredcount' ] as $type ) {
|
|
$this->wanCache->delete( $this->getCacheKey( $type ) );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::getAllQueuedJobs()
|
|
* @return Iterator<RunnableJob>
|
|
*/
|
|
public function getAllQueuedJobs() {
|
|
return $this->getJobIterator( [ 'job_cmd' => $this->getType(), 'job_token' => '' ] );
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::getAllAcquiredJobs()
|
|
* @return Iterator<RunnableJob>
|
|
*/
|
|
public function getAllAcquiredJobs() {
|
|
$dbr = $this->getReplicaDB();
|
|
return $this->getJobIterator( [ 'job_cmd' => $this->getType(), $dbr->expr( 'job_token', '>', '' ) ] );
|
|
}
|
|
|
|
/**
|
|
* @see JobQueue::getAllAbandonedJobs()
|
|
* @return Iterator<RunnableJob>
|
|
*/
|
|
public function getAllAbandonedJobs() {
|
|
$dbr = $this->getReplicaDB();
|
|
return $this->getJobIterator( [
|
|
'job_cmd' => $this->getType(),
|
|
$dbr->expr( 'job_token', '>', '' ),
|
|
$dbr->expr( 'job_attempts', '>=', intval( $this->maxTries ) ),
|
|
] );
|
|
}
|
|
|
|
/**
|
|
* @param array $conds Query conditions
|
|
* @return Iterator<RunnableJob>
|
|
*/
|
|
protected function getJobIterator( array $conds ) {
|
|
$dbr = $this->getReplicaDB();
|
|
$qb = $dbr->newSelectQueryBuilder()
|
|
->select( self::selectFields() )
|
|
->from( 'job' )
|
|
->where( $conds );
|
|
try {
|
|
return new MappedIterator(
|
|
$qb->caller( __METHOD__ )->fetchResultSet(),
|
|
function ( $row ) {
|
|
return $this->jobFromRow( $row );
|
|
}
|
|
);
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
}
|
|
|
|
public function getCoalesceLocationInternal() {
|
|
if ( $this->server ) {
|
|
return null; // not using the LBFactory instance
|
|
}
|
|
|
|
return is_string( $this->cluster )
|
|
? "DBCluster:{$this->cluster}:{$this->domain}"
|
|
: "LBFactory:{$this->domain}";
|
|
}
|
|
|
|
protected function doGetSiblingQueuesWithJobs( array $types ) {
|
|
$dbr = $this->getReplicaDB();
|
|
// @note: this does not check whether the jobs are claimed or not.
|
|
// This is useful so JobQueueGroup::pop() also sees queues that only
|
|
// have stale jobs. This lets recycleAndDeleteStaleJobs() re-enqueue
|
|
// failed jobs so that they can be popped again for that edge case.
|
|
$res = $dbr->newSelectQueryBuilder()
|
|
->select( 'job_cmd' )
|
|
->distinct()
|
|
->from( 'job' )
|
|
->where( [ 'job_cmd' => $types ] )
|
|
->caller( __METHOD__ )->fetchResultSet();
|
|
|
|
$types = [];
|
|
foreach ( $res as $row ) {
|
|
$types[] = $row->job_cmd;
|
|
}
|
|
|
|
return $types;
|
|
}
|
|
|
|
protected function doGetSiblingQueueSizes( array $types ) {
|
|
$dbr = $this->getReplicaDB();
|
|
|
|
$res = $dbr->newSelectQueryBuilder()
|
|
->select( [ 'job_cmd', 'count' => 'COUNT(*)' ] )
|
|
->from( 'job' )
|
|
->where( [ 'job_cmd' => $types ] )
|
|
->groupBy( 'job_cmd' )
|
|
->caller( __METHOD__ )->fetchResultSet();
|
|
|
|
$sizes = [];
|
|
foreach ( $res as $row ) {
|
|
$sizes[$row->job_cmd] = (int)$row->count;
|
|
}
|
|
|
|
return $sizes;
|
|
}
|
|
|
|
/**
|
|
* Recycle or destroy any jobs that have been claimed for too long
|
|
*
|
|
* @return int Number of jobs recycled/deleted
|
|
*/
|
|
public function recycleAndDeleteStaleJobs() {
|
|
$now = time();
|
|
$count = 0; // affected rows
|
|
$dbw = $this->getPrimaryDB();
|
|
|
|
try {
|
|
if ( !$dbw->lock( "jobqueue-recycle-{$this->type}", __METHOD__, 1 ) ) {
|
|
return $count; // already in progress
|
|
}
|
|
|
|
// Remove claims on jobs acquired for too long if enabled...
|
|
if ( $this->claimTTL > 0 ) {
|
|
$claimCutoff = $dbw->timestamp( $now - $this->claimTTL );
|
|
// Get the IDs of jobs that have be claimed but not finished after too long.
|
|
// These jobs can be recycled into the queue by expiring the claim. Selecting
|
|
// the IDs first means that the UPDATE can be done by primary key (less deadlocks).
|
|
$res = $dbw->newSelectQueryBuilder()
|
|
->select( 'job_id' )
|
|
->from( 'job' )
|
|
->where(
|
|
[
|
|
'job_cmd' => $this->type,
|
|
$dbw->expr( 'job_token', '!=', '' ), // was acquired
|
|
$dbw->expr( 'job_token_timestamp', '<', $claimCutoff ), // stale
|
|
$dbw->expr( 'job_attempts', '<', $this->maxTries ), // retries left
|
|
]
|
|
)
|
|
->caller( __METHOD__ )->fetchResultSet();
|
|
$ids = array_map(
|
|
static function ( $o ) {
|
|
return $o->job_id;
|
|
}, iterator_to_array( $res )
|
|
);
|
|
if ( count( $ids ) ) {
|
|
// Reset job_token for these jobs so that other runners will pick them up.
|
|
// Set the timestamp to the current time, as it is useful to now that the job
|
|
// was already tried before (the timestamp becomes the "released" time).
|
|
$dbw->newUpdateQueryBuilder()
|
|
->update( 'job' )
|
|
->set( [
|
|
'job_token' => '',
|
|
'job_token_timestamp' => $dbw->timestamp( $now ) // time of release
|
|
] )
|
|
->where( [
|
|
'job_id' => $ids,
|
|
$dbw->expr( 'job_token', '!=', '' ),
|
|
] )
|
|
->caller( __METHOD__ )->execute();
|
|
|
|
$affected = $dbw->affectedRows();
|
|
$count += $affected;
|
|
$this->incrStats( 'recycles', $this->type, $affected );
|
|
}
|
|
}
|
|
|
|
// Just destroy any stale jobs...
|
|
$pruneCutoff = $dbw->timestamp( $now - self::MAX_AGE_PRUNE );
|
|
$qb = $dbw->newSelectQueryBuilder()
|
|
->select( 'job_id' )
|
|
->from( 'job' )
|
|
->where(
|
|
[
|
|
'job_cmd' => $this->type,
|
|
$dbw->expr( 'job_token', '!=', '' ), // was acquired
|
|
$dbw->expr( 'job_token_timestamp', '<', $pruneCutoff ) // stale
|
|
]
|
|
);
|
|
if ( $this->claimTTL > 0 ) { // only prune jobs attempted too many times...
|
|
$qb->andWhere( $dbw->expr( 'job_attempts', '>=', $this->maxTries ) );
|
|
}
|
|
// Get the IDs of jobs that are considered stale and should be removed. Selecting
|
|
// the IDs first means that the UPDATE can be done by primary key (less deadlocks).
|
|
$res = $qb->caller( __METHOD__ )->fetchResultSet();
|
|
$ids = array_map(
|
|
static function ( $o ) {
|
|
return $o->job_id;
|
|
}, iterator_to_array( $res )
|
|
);
|
|
if ( count( $ids ) ) {
|
|
$dbw->newDeleteQueryBuilder()
|
|
->deleteFrom( 'job' )
|
|
->where( [ 'job_id' => $ids ] )
|
|
->caller( __METHOD__ )->execute();
|
|
$affected = $dbw->affectedRows();
|
|
$count += $affected;
|
|
$this->incrStats( 'abandons', $this->type, $affected );
|
|
}
|
|
|
|
$dbw->unlock( "jobqueue-recycle-{$this->type}", __METHOD__ );
|
|
} catch ( DBError $e ) {
|
|
throw $this->getDBException( $e );
|
|
}
|
|
|
|
return $count;
|
|
}
|
|
|
|
/**
|
|
* @param IJobSpecification $job
|
|
* @param IReadableDatabase $db
|
|
* @return array
|
|
*/
|
|
protected function insertFields( IJobSpecification $job, IReadableDatabase $db ) {
|
|
return [
|
|
// Fields that describe the nature of the job
|
|
'job_cmd' => $job->getType(),
|
|
'job_namespace' => $job->getParams()['namespace'] ?? NS_SPECIAL,
|
|
'job_title' => $job->getParams()['title'] ?? '',
|
|
'job_params' => self::makeBlob( $job->getParams() ),
|
|
// Additional job metadata
|
|
'job_timestamp' => $db->timestamp(),
|
|
'job_sha1' => Wikimedia\base_convert(
|
|
sha1( serialize( $job->getDeduplicationInfo() ) ),
|
|
16, 36, 31
|
|
),
|
|
'job_random' => mt_rand( 0, self::MAX_JOB_RANDOM )
|
|
];
|
|
}
|
|
|
|
/**
|
|
* @throws JobQueueConnectionError
|
|
* @return IDatabase
|
|
*/
|
|
protected function getReplicaDB() {
|
|
try {
|
|
return $this->getDB( DB_REPLICA );
|
|
} catch ( DBConnectionError $e ) {
|
|
throw new JobQueueConnectionError( "DBConnectionError:" . $e->getMessage() );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @throws JobQueueConnectionError
|
|
* @return IMaintainableDatabase
|
|
* @since 1.37
|
|
*/
|
|
protected function getPrimaryDB() {
|
|
try {
|
|
return $this->getDB( DB_PRIMARY );
|
|
} catch ( DBConnectionError $e ) {
|
|
throw new JobQueueConnectionError( "DBConnectionError:" . $e->getMessage() );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param int $index (DB_REPLICA/DB_PRIMARY)
|
|
* @return IMaintainableDatabase
|
|
*/
|
|
protected function getDB( $index ) {
|
|
if ( $this->server ) {
|
|
if ( $this->conn instanceof IDatabase ) {
|
|
return $this->conn;
|
|
} elseif ( $this->conn instanceof DBError ) {
|
|
throw $this->conn;
|
|
}
|
|
|
|
try {
|
|
$this->conn = MediaWikiServices::getInstance()->getDatabaseFactory()->create(
|
|
$this->server['type'],
|
|
$this->server
|
|
);
|
|
} catch ( DBError $e ) {
|
|
$this->conn = $e;
|
|
throw $e;
|
|
}
|
|
|
|
return $this->conn;
|
|
} else {
|
|
$lbFactory = MediaWikiServices::getInstance()->getDBLoadBalancerFactory();
|
|
$lb = is_string( $this->cluster )
|
|
? $lbFactory->getExternalLB( $this->cluster )
|
|
: $lbFactory->getMainLB( $this->domain );
|
|
|
|
if ( $lb->getServerType( ServerInfo::WRITER_INDEX ) !== 'sqlite' ) {
|
|
// Keep a separate connection to avoid contention and deadlocks;
|
|
// However, SQLite has the opposite behavior due to DB-level locking.
|
|
$flags = $lb::CONN_TRX_AUTOCOMMIT;
|
|
} else {
|
|
// Jobs insertion will be deferred until the PRESEND stage to reduce contention.
|
|
$flags = 0;
|
|
}
|
|
|
|
return $lb->getMaintenanceConnectionRef( $index, [], $this->domain, $flags );
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param string $property
|
|
* @return string
|
|
*/
|
|
private function getCacheKey( $property ) {
|
|
$cluster = is_string( $this->cluster ) ? $this->cluster : 'main';
|
|
|
|
return $this->wanCache->makeGlobalKey(
|
|
'jobqueue',
|
|
$this->domain,
|
|
$cluster,
|
|
$this->type,
|
|
$property
|
|
);
|
|
}
|
|
|
|
/**
|
|
* @param array|false $params
|
|
* @return string
|
|
*/
|
|
protected static function makeBlob( $params ) {
|
|
if ( $params !== false ) {
|
|
return serialize( $params );
|
|
} else {
|
|
return '';
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param stdClass $row
|
|
* @return RunnableJob
|
|
*/
|
|
protected function jobFromRow( $row ) {
|
|
$params = ( (string)$row->job_params !== '' ) ? unserialize( $row->job_params ) : [];
|
|
if ( !is_array( $params ) ) { // this shouldn't happen
|
|
throw new UnexpectedValueException(
|
|
"Could not unserialize job with ID '{$row->job_id}'." );
|
|
}
|
|
|
|
$params += [ 'namespace' => $row->job_namespace, 'title' => $row->job_title ];
|
|
$job = $this->factoryJob( $row->job_cmd, $params );
|
|
$job->setMetadata( 'id', $row->job_id );
|
|
$job->setMetadata( 'timestamp', $row->job_timestamp );
|
|
|
|
return $job;
|
|
}
|
|
|
|
/**
|
|
* @param DBError $e
|
|
* @return JobQueueError
|
|
*/
|
|
protected function getDBException( DBError $e ) {
|
|
return new JobQueueError( get_class( $e ) . ": " . $e->getMessage() );
|
|
}
|
|
|
|
/**
|
|
* Return the list of job fields that should be selected.
|
|
* @since 1.23
|
|
* @return array
|
|
*/
|
|
public static function selectFields() {
|
|
return [
|
|
'job_id',
|
|
'job_cmd',
|
|
'job_namespace',
|
|
'job_title',
|
|
'job_timestamp',
|
|
'job_params',
|
|
'job_random',
|
|
'job_attempts',
|
|
'job_token',
|
|
'job_token_timestamp',
|
|
'job_sha1',
|
|
];
|
|
}
|
|
}
|