includes/jobqueue/JobQueueDB.php
<?php
/**
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License along
* with this program; if not, write to the Free Software Foundation, Inc.,
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
* http://www.gnu.org/copyleft/gpl.html
*
* @file
*/
use MediaWiki\MediaWikiServices;
use Wikimedia\Rdbms\DBConnectionError;
use Wikimedia\Rdbms\DBError;
use Wikimedia\Rdbms\IDatabase;
use Wikimedia\Rdbms\IMaintainableDatabase;
use Wikimedia\Rdbms\SelectQueryBuilder;
use Wikimedia\ScopedCallback;
/**
* Database-backed job queue storage.
*
* @since 1.21
* @ingroup JobQueue
*/
class JobQueueDB extends JobQueue {
/* seconds to cache info without re-validating */
private const CACHE_TTL_SHORT = 30;
/* seconds a job can live once claimed */
private const MAX_AGE_PRUNE = 7 * 24 * 3600;
/**
* Used for job_random, the highest safe 32-bit signed integer.
* Equivalent to `( 2 ** 31 ) - 1` on 64-bit.
*/
private const MAX_JOB_RANDOM = 2_147_483_647;
/* maximum number of rows to skip */
private const MAX_OFFSET = 255;
/** @var IMaintainableDatabase|DBError|null */
protected $conn;
/** @var array|null Server configuration array */
protected $server;
/** @var string|null Name of an external DB cluster or null for the local DB cluster */
protected $cluster;
/**
* Additional parameters include:
* - server : Server configuration array for Database::factory. Overrides "cluster".
* - cluster : The name of an external cluster registered via LBFactory.
* If not specified, the primary DB cluster for the wiki will be used.
* This can be overridden with a custom cluster so that DB handles will
* be retrieved via LBFactory::getExternalLB() and getConnection().
* @param array $params
*/
protected function __construct( array $params ) {
parent::__construct( $params );
if ( isset( $params['server'] ) ) {
$this->server = $params['server'];
} elseif ( isset( $params['cluster'] ) && is_string( $params['cluster'] ) ) {
$this->cluster = $params['cluster'];
}
}
protected function supportedOrders() {
return [ 'random', 'timestamp', 'fifo' ];
}
protected function optimalOrder() {
return 'random';
}
/**
* @see JobQueue::doIsEmpty()
* @return bool
*/
protected function doIsEmpty() {
$dbr = $this->getReplicaDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbr );
try {
// unclaimed job
$found = (bool)$dbr->newSelectQueryBuilder()
->select( '1' )
->from( 'job' )
->where( [ 'job_cmd' => $this->type, 'job_token' => '' ] )
->caller( __METHOD__ )->fetchField();
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
return !$found;
}
/**
* @see JobQueue::doGetSize()
* @return int
*/
protected function doGetSize() {
$key = $this->getCacheKey( 'size' );
$size = $this->wanCache->get( $key );
if ( is_int( $size ) ) {
return $size;
}
$dbr = $this->getReplicaDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbr );
try {
$size = $dbr->newSelectQueryBuilder()
->from( 'job' )
->where( [ 'job_cmd' => $this->type, 'job_token' => '' ] )
->caller( __METHOD__ )->fetchRowCount();
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
$this->wanCache->set( $key, $size, self::CACHE_TTL_SHORT );
return $size;
}
/**
* @see JobQueue::doGetAcquiredCount()
* @return int
*/
protected function doGetAcquiredCount() {
if ( $this->claimTTL <= 0 ) {
return 0; // no acknowledgements
}
$key = $this->getCacheKey( 'acquiredcount' );
$count = $this->wanCache->get( $key );
if ( is_int( $count ) ) {
return $count;
}
$dbr = $this->getReplicaDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbr );
try {
$count = $dbr->newSelectQueryBuilder()
->from( 'job' )
->where( [
'job_cmd' => $this->type,
$dbr->expr( 'job_token', '!=', '' ),
] )
->caller( __METHOD__ )->fetchRowCount();
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
$this->wanCache->set( $key, $count, self::CACHE_TTL_SHORT );
return $count;
}
/**
* @see JobQueue::doGetAbandonedCount()
* @return int
* @throws JobQueueConnectionError
* @throws JobQueueError
*/
protected function doGetAbandonedCount() {
if ( $this->claimTTL <= 0 ) {
return 0; // no acknowledgements
}
$key = $this->getCacheKey( 'abandonedcount' );
$count = $this->wanCache->get( $key );
if ( is_int( $count ) ) {
return $count;
}
$dbr = $this->getReplicaDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbr );
try {
$count = $dbr->newSelectQueryBuilder()
->from( 'job' )
->where(
[
'job_cmd' => $this->type,
$dbr->expr( 'job_token', '!=', '' ),
$dbr->expr( 'job_attempts', '>=', $this->maxTries ),
]
)
->caller( __METHOD__ )->fetchRowCount();
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
$this->wanCache->set( $key, $count, self::CACHE_TTL_SHORT );
return $count;
}
/**
* @see JobQueue::doBatchPush()
* @param IJobSpecification[] $jobs
* @param int $flags
* @throws DBError|Exception
* @return void
*/
protected function doBatchPush( array $jobs, $flags ) {
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
// In general, there will be two cases here:
// a) sqlite; DB connection is probably a regular round-aware handle.
// If the connection is busy with a transaction, then defer the job writes
// until right before the main round commit step. Any errors that bubble
// up will rollback the main commit round.
// b) mysql/postgres; DB connection is generally a separate CONN_TRX_AUTOCOMMIT handle.
// No transaction is active nor will be started by writes, so enqueue the jobs
// now so that any errors will show up immediately as the interface expects. Any
// errors that bubble up will rollback the main commit round.
$fname = __METHOD__;
$dbw->onTransactionPreCommitOrIdle(
function ( IDatabase $dbw ) use ( $jobs, $flags, $fname ) {
$this->doBatchPushInternal( $dbw, $jobs, $flags, $fname );
},
$fname
);
}
/**
* This function should *not* be called outside of JobQueueDB
*
* @suppress SecurityCheck-SQLInjection Bug in phan-taint-check handling bulk inserts
* @param IDatabase $dbw
* @param IJobSpecification[] $jobs
* @param int $flags
* @param string $method
* @throws DBError
* @return void
*/
public function doBatchPushInternal( IDatabase $dbw, array $jobs, $flags, $method ) {
if ( $jobs === [] ) {
return;
}
$rowSet = []; // (sha1 => job) map for jobs that are de-duplicated
$rowList = []; // list of jobs for jobs that are not de-duplicated
foreach ( $jobs as $job ) {
$row = $this->insertFields( $job, $dbw );
if ( $job->ignoreDuplicates() ) {
$rowSet[$row['job_sha1']] = $row;
} else {
$rowList[] = $row;
}
}
if ( $flags & self::QOS_ATOMIC ) {
$dbw->startAtomic( $method ); // wrap all the job additions in one transaction
}
try {
// Strip out any duplicate jobs that are already in the queue...
if ( count( $rowSet ) ) {
$res = $dbw->newSelectQueryBuilder()
->select( 'job_sha1' )
->from( 'job' )
->where(
[
// No job_type condition since it's part of the job_sha1 hash
'job_sha1' => array_map( 'strval', array_keys( $rowSet ) ),
'job_token' => '' // unclaimed
]
)
->caller( $method )->fetchResultSet();
foreach ( $res as $row ) {
wfDebug( "Job with hash '{$row->job_sha1}' is a duplicate." );
unset( $rowSet[$row->job_sha1] ); // already enqueued
}
}
// Build the full list of job rows to insert
$rows = array_merge( $rowList, array_values( $rowSet ) );
// Insert the job rows in chunks to avoid replica DB lag...
foreach ( array_chunk( $rows, 50 ) as $rowBatch ) {
$dbw->newInsertQueryBuilder()
->insertInto( 'job' )
->rows( $rowBatch )
->caller( $method )->execute();
}
$this->incrStats( 'inserts', $this->type, count( $rows ) );
$this->incrStats( 'dupe_inserts', $this->type,
count( $rowSet ) + count( $rowList ) - count( $rows )
);
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
if ( $flags & self::QOS_ATOMIC ) {
$dbw->endAtomic( $method );
}
}
/**
* @see JobQueue::doPop()
* @return RunnableJob|false
*/
protected function doPop() {
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
$job = false; // job popped off
try {
$uuid = wfRandomString( 32 ); // pop attempt
do { // retry when our row is invalid or deleted as a duplicate
// Try to reserve a row in the DB...
if ( in_array( $this->order, [ 'fifo', 'timestamp' ] ) ) {
$row = $this->claimOldest( $uuid );
} else { // random first
$rand = mt_rand( 0, self::MAX_JOB_RANDOM ); // encourage concurrent UPDATEs
$gte = (bool)mt_rand( 0, 1 ); // find rows with rand before/after $rand
$row = $this->claimRandom( $uuid, $rand, $gte );
}
// Check if we found a row to reserve...
if ( !$row ) {
break; // nothing to do
}
$this->incrStats( 'pops', $this->type );
// Get the job object from the row...
$job = $this->jobFromRow( $row );
break; // done
} while ( true );
if ( !$job || mt_rand( 0, 9 ) == 0 ) {
// Handled jobs that need to be recycled/deleted;
// any recycled jobs will be picked up next attempt
$this->recycleAndDeleteStaleJobs();
}
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
return $job;
}
/**
* Reserve a row with a single UPDATE without holding row locks over RTTs...
*
* @param string $uuid 32 char hex string
* @param int $rand Random unsigned integer (31 bits)
* @param bool $gte Search for job_random >= $random (otherwise job_random <= $random)
* @return stdClass|false Row|false
*/
protected function claimRandom( $uuid, $rand, $gte ) {
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
// Check cache to see if the queue has <= OFFSET items
$tinyQueue = $this->wanCache->get( $this->getCacheKey( 'small' ) );
$invertedDirection = false; // whether one job_random direction was already scanned
// This uses a replication safe method for acquiring jobs. One could use UPDATE+LIMIT
// instead, but that either uses ORDER BY (in which case it deadlocks in MySQL) or is
// not replication safe. Due to https://bugs.mysql.com/bug.php?id=6980, subqueries cannot
// be used here with MySQL.
do {
if ( $tinyQueue ) { // queue has <= MAX_OFFSET rows
// For small queues, using OFFSET will overshoot and return no rows more often.
// Instead, this uses job_random to pick a row (possibly checking both directions).
$row = $dbw->newSelectQueryBuilder()
->select( self::selectFields() )
->from( 'job' )
->where(
[
'job_cmd' => $this->type,
'job_token' => '', // unclaimed
$dbw->expr( 'job_random', $gte ? '>=' : '<=', $rand )
]
)
->orderBy(
'job_random',
$gte ? SelectQueryBuilder::SORT_ASC : SelectQueryBuilder::SORT_DESC
)
->caller( __METHOD__ )->fetchRow();
if ( !$row && !$invertedDirection ) {
$gte = !$gte;
$invertedDirection = true;
continue; // try the other direction
}
} else { // table *may* have >= MAX_OFFSET rows
// T44614: "ORDER BY job_random" with a job_random inequality causes high CPU
// in MySQL if there are many rows for some reason. This uses a small OFFSET
// instead of job_random for reducing excess claim retries.
$row = $dbw->newSelectQueryBuilder()
->select( self::selectFields() )
->from( 'job' )
->where(
[
'job_cmd' => $this->type,
'job_token' => '', // unclaimed
]
)
->offset( mt_rand( 0, self::MAX_OFFSET ) )
->caller( __METHOD__ )->fetchRow();
if ( !$row ) {
$tinyQueue = true; // we know the queue must have <= MAX_OFFSET rows
$this->wanCache->set( $this->getCacheKey( 'small' ), 1, 30 );
continue; // use job_random
}
}
if ( !$row ) {
break;
}
$dbw->newUpdateQueryBuilder()
->update( 'job' ) // update by PK
->set( [
'job_token' => $uuid,
'job_token_timestamp' => $dbw->timestamp(),
'job_attempts = job_attempts+1'
] )
->where( [
'job_cmd' => $this->type,
'job_id' => $row->job_id,
'job_token' => ''
] )
->caller( __METHOD__ )->execute();
// This might get raced out by another runner when claiming the previously
// selected row. The use of job_random should minimize this problem, however.
if ( !$dbw->affectedRows() ) {
$row = false; // raced out
}
} while ( !$row );
return $row;
}
/**
* Reserve a row with a single UPDATE without holding row locks over RTTs...
*
* @param string $uuid 32 char hex string
* @return stdClass|false Row|false
*/
protected function claimOldest( $uuid ) {
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
$row = false; // the row acquired
do {
if ( $dbw->getType() === 'mysql' ) {
// Per https://bugs.mysql.com/bug.php?id=6980, we can't use subqueries on the
// same table being changed in an UPDATE query in MySQL (gives Error: 1093).
// Postgres has no such limitation. However, MySQL offers an
// alternative here by supporting ORDER BY + LIMIT for UPDATE queries.
$dbw->query( "UPDATE {$dbw->tableName( 'job' )} " .
"SET " .
"job_token = {$dbw->addQuotes( $uuid ) }, " .
"job_token_timestamp = {$dbw->addQuotes( $dbw->timestamp() )}, " .
"job_attempts = job_attempts+1 " .
"WHERE ( " .
"job_cmd = {$dbw->addQuotes( $this->type )} " .
"AND job_token = {$dbw->addQuotes( '' )} " .
") ORDER BY job_id ASC LIMIT 1",
__METHOD__
);
} else {
// Use a subquery to find the job, within an UPDATE to claim it.
// This uses as much of the DB wrapper functions as possible.
$qb = $dbw->newSelectQueryBuilder()
->select( 'job_id' )
->from( 'job' )
->where( [ 'job_cmd' => $this->type, 'job_token' => '' ] )
->orderBy( 'job_id', SelectQueryBuilder::SORT_ASC )
->limit( 1 );
$dbw->newUpdateQueryBuilder()
->update( 'job' )
->set( [
'job_token' => $uuid,
'job_token_timestamp' => $dbw->timestamp(),
'job_attempts = job_attempts+1'
] )
->where( [ 'job_id = (' . $qb->getSQL() . ')' ] )
->caller( __METHOD__ )->execute();
}
if ( !$dbw->affectedRows() ) {
break;
}
// Fetch any row that we just reserved...
$row = $dbw->newSelectQueryBuilder()
->select( self::selectFields() )
->from( 'job' )
->where( [ 'job_cmd' => $this->type, 'job_token' => $uuid ] )
->caller( __METHOD__ )->fetchRow();
if ( !$row ) { // raced out by duplicate job removal
wfDebug( "Row deleted as duplicate by another process." );
}
} while ( !$row );
return $row;
}
/**
* @see JobQueue::doAck()
* @param RunnableJob $job
* @throws JobQueueConnectionError
* @throws JobQueueError
*/
protected function doAck( RunnableJob $job ) {
$id = $job->getMetadata( 'id' );
if ( $id === null ) {
throw new UnexpectedValueException( "Job of type '{$job->getType()}' has no ID." );
}
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
try {
// Delete a row with a single DELETE without holding row locks over RTTs...
$dbw->newDeleteQueryBuilder()
->deleteFrom( 'job' )
->where( [ 'job_cmd' => $this->type, 'job_id' => $id ] )
->caller( __METHOD__ )->execute();
$this->incrStats( 'acks', $this->type );
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
}
/**
* @see JobQueue::doDeduplicateRootJob()
* @param IJobSpecification $job
* @throws JobQueueConnectionError
* @return bool
*/
protected function doDeduplicateRootJob( IJobSpecification $job ) {
// Callers should call JobQueueGroup::push() before this method so that if the
// insert fails, the de-duplication registration will be aborted. Since the insert
// is deferred till "transaction idle", do the same here, so that the ordering is
// maintained. Having only the de-duplication registration succeed would cause
// jobs to become no-ops without any actual jobs that made them redundant.
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
$dbw->onTransactionCommitOrIdle(
function () use ( $job ) {
parent::doDeduplicateRootJob( $job );
},
__METHOD__
);
return true;
}
/**
* @see JobQueue::doDelete()
* @return bool
*/
protected function doDelete() {
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
try {
$dbw->newDeleteQueryBuilder()
->deleteFrom( 'job' )
->where( [ 'job_cmd' => $this->type ] )
->caller( __METHOD__ )->execute();
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
return true;
}
/**
* @see JobQueue::doWaitForBackups()
* @return void
*/
protected function doWaitForBackups() {
if ( $this->server ) {
return; // not using LBFactory instance
}
$lbFactory = MediaWikiServices::getInstance()->getDBLoadBalancerFactory();
$lbFactory->waitForReplication();
}
/**
* @return void
*/
protected function doFlushCaches() {
foreach ( [ 'size', 'acquiredcount' ] as $type ) {
$this->wanCache->delete( $this->getCacheKey( $type ) );
}
}
/**
* @see JobQueue::getAllQueuedJobs()
* @return Iterator<RunnableJob>
*/
public function getAllQueuedJobs() {
return $this->getJobIterator( [ 'job_cmd' => $this->getType(), 'job_token' => '' ] );
}
/**
* @see JobQueue::getAllAcquiredJobs()
* @return Iterator<RunnableJob>
*/
public function getAllAcquiredJobs() {
return $this->getJobIterator( [ 'job_cmd' => $this->getType(), "job_token > ''" ] );
}
/**
* @see JobQueue::getAllAbandonedJobs()
* @return Iterator<RunnableJob>
*/
public function getAllAbandonedJobs() {
return $this->getJobIterator( [
'job_cmd' => $this->getType(),
"job_token > ''",
"job_attempts >= " . intval( $this->maxTries )
] );
}
/**
* @param array $conds Query conditions
* @return Iterator<RunnableJob>
*/
protected function getJobIterator( array $conds ) {
$dbr = $this->getReplicaDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbr );
$qb = $dbr->newSelectQueryBuilder()
->select( self::selectFields() )
->from( 'job' )
->where( $conds );
try {
return new MappedIterator(
$qb->caller( __METHOD__ )->fetchResultSet(),
function ( $row ) {
return $this->jobFromRow( $row );
}
);
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
}
public function getCoalesceLocationInternal() {
if ( $this->server ) {
return null; // not using the LBFactory instance
}
return is_string( $this->cluster )
? "DBCluster:{$this->cluster}:{$this->domain}"
: "LBFactory:{$this->domain}";
}
protected function doGetSiblingQueuesWithJobs( array $types ) {
$dbr = $this->getReplicaDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbr );
// @note: this does not check whether the jobs are claimed or not.
// This is useful so JobQueueGroup::pop() also sees queues that only
// have stale jobs. This lets recycleAndDeleteStaleJobs() re-enqueue
// failed jobs so that they can be popped again for that edge case.
$res = $dbr->newSelectQueryBuilder()
->select( 'job_cmd' )
->distinct()
->from( 'job' )
->where( [ 'job_cmd' => $types ] )
->caller( __METHOD__ )->fetchResultSet();
$types = [];
foreach ( $res as $row ) {
$types[] = $row->job_cmd;
}
return $types;
}
protected function doGetSiblingQueueSizes( array $types ) {
$dbr = $this->getReplicaDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbr );
$res = $dbr->newSelectQueryBuilder()
->select( [ 'job_cmd', 'count' => 'COUNT(*)' ] )
->from( 'job' )
->where( [ 'job_cmd' => $types ] )
->groupBy( 'job_cmd' )
->caller( __METHOD__ )->fetchResultSet();
$sizes = [];
foreach ( $res as $row ) {
$sizes[$row->job_cmd] = (int)$row->count;
}
return $sizes;
}
/**
* Recycle or destroy any jobs that have been claimed for too long
*
* @return int Number of jobs recycled/deleted
*/
public function recycleAndDeleteStaleJobs() {
$now = time();
$count = 0; // affected rows
$dbw = $this->getPrimaryDB();
/** @noinspection PhpUnusedLocalVariableInspection */
$scope = $this->getScopedNoTrxFlag( $dbw );
try {
if ( !$dbw->lock( "jobqueue-recycle-{$this->type}", __METHOD__, 1 ) ) {
return $count; // already in progress
}
// Remove claims on jobs acquired for too long if enabled...
if ( $this->claimTTL > 0 ) {
$claimCutoff = $dbw->timestamp( $now - $this->claimTTL );
// Get the IDs of jobs that have be claimed but not finished after too long.
// These jobs can be recycled into the queue by expiring the claim. Selecting
// the IDs first means that the UPDATE can be done by primary key (less deadlocks).
$res = $dbw->newSelectQueryBuilder()
->select( 'job_id' )
->from( 'job' )
->where(
[
'job_cmd' => $this->type,
$dbw->expr( 'job_token', '!=', '' ), // was acquired
$dbw->expr( 'job_token_timestamp', '<', $claimCutoff ), // stale
$dbw->expr( 'job_attempts', '<', $this->maxTries ), // retries left
]
)
->caller( __METHOD__ )->fetchResultSet();
$ids = array_map(
static function ( $o ) {
return $o->job_id;
}, iterator_to_array( $res )
);
if ( count( $ids ) ) {
// Reset job_token for these jobs so that other runners will pick them up.
// Set the timestamp to the current time, as it is useful to now that the job
// was already tried before (the timestamp becomes the "released" time).
$dbw->newUpdateQueryBuilder()
->update( 'job' )
->set( [
'job_token' => '',
'job_token_timestamp' => $dbw->timestamp( $now ) // time of release
] )
->where( [
'job_id' => $ids,
$dbw->expr( 'job_token', '!=', '' ),
] )
->caller( __METHOD__ )->execute();
$affected = $dbw->affectedRows();
$count += $affected;
$this->incrStats( 'recycles', $this->type, $affected );
}
}
// Just destroy any stale jobs...
$pruneCutoff = $dbw->timestamp( $now - self::MAX_AGE_PRUNE );
$qb = $dbw->newSelectQueryBuilder()
->select( 'job_id' )
->from( 'job' )
->where(
[
'job_cmd' => $this->type,
$dbw->expr( 'job_token', '!=', '' ), // was acquired
$dbw->expr( 'job_token_timestamp', '<', $pruneCutoff ) // stale
]
);
if ( $this->claimTTL > 0 ) { // only prune jobs attempted too many times...
$qb->andWhere( $dbw->expr( 'job_attempts', '>=', $this->maxTries ) );
}
// Get the IDs of jobs that are considered stale and should be removed. Selecting
// the IDs first means that the UPDATE can be done by primary key (less deadlocks).
$res = $qb->caller( __METHOD__ )->fetchResultSet();
$ids = array_map(
static function ( $o ) {
return $o->job_id;
}, iterator_to_array( $res )
);
if ( count( $ids ) ) {
$dbw->newDeleteQueryBuilder()
->deleteFrom( 'job' )
->where( [ 'job_id' => $ids ] )
->caller( __METHOD__ )->execute();
$affected = $dbw->affectedRows();
$count += $affected;
$this->incrStats( 'abandons', $this->type, $affected );
}
$dbw->unlock( "jobqueue-recycle-{$this->type}", __METHOD__ );
} catch ( DBError $e ) {
throw $this->getDBException( $e );
}
return $count;
}
/**
* @param IJobSpecification $job
* @param IDatabase $db
* @return array
*/
protected function insertFields( IJobSpecification $job, IDatabase $db ) {
return [
// Fields that describe the nature of the job
'job_cmd' => $job->getType(),
'job_namespace' => $job->getParams()['namespace'] ?? NS_SPECIAL,
'job_title' => $job->getParams()['title'] ?? '',
'job_params' => self::makeBlob( $job->getParams() ),
// Additional job metadata
'job_timestamp' => $db->timestamp(),
'job_sha1' => Wikimedia\base_convert(
sha1( serialize( $job->getDeduplicationInfo() ) ),
16, 36, 31
),
'job_random' => mt_rand( 0, self::MAX_JOB_RANDOM )
];
}
/**
* @throws JobQueueConnectionError
* @return IDatabase
*/
protected function getReplicaDB() {
try {
return $this->getDB( DB_REPLICA );
} catch ( DBConnectionError $e ) {
throw new JobQueueConnectionError( "DBConnectionError:" . $e->getMessage() );
}
}
/**
* @throws JobQueueConnectionError
* @return IMaintainableDatabase
* @since 1.37
*/
protected function getPrimaryDB() {
try {
return $this->getDB( DB_PRIMARY );
} catch ( DBConnectionError $e ) {
throw new JobQueueConnectionError( "DBConnectionError:" . $e->getMessage() );
}
}
/**
* @param int $index (DB_REPLICA/DB_PRIMARY)
* @return IMaintainableDatabase
*/
protected function getDB( $index ) {
if ( $this->server ) {
if ( $this->conn instanceof IDatabase ) {
return $this->conn;
} elseif ( $this->conn instanceof DBError ) {
throw $this->conn;
}
try {
$this->conn = MediaWikiServices::getInstance()->getDatabaseFactory()->create(
$this->server['type'],
$this->server
);
} catch ( DBError $e ) {
$this->conn = $e;
throw $e;
}
return $this->conn;
} else {
$lbFactory = MediaWikiServices::getInstance()->getDBLoadBalancerFactory();
$lb = is_string( $this->cluster )
? $lbFactory->getExternalLB( $this->cluster )
: $lbFactory->getMainLB( $this->domain );
if ( $lb->getServerType( $lb->getWriterIndex() ) !== 'sqlite' ) {
// Keep a separate connection to avoid contention and deadlocks;
// However, SQLite has the opposite behavior due to DB-level locking.
$flags = $lb::CONN_TRX_AUTOCOMMIT;
} else {
// Jobs insertion will be deferred until the PRESEND stage to reduce contention.
$flags = 0;
}
return $lb->getMaintenanceConnectionRef( $index, [], $this->domain, $flags );
}
}
/**
* @param IDatabase $db
* @return ScopedCallback
*/
private function getScopedNoTrxFlag( IDatabase $db ) {
$autoTrx = $db->getFlag( DBO_TRX ); // get current setting
$db->clearFlag( DBO_TRX ); // make each query its own transaction
return new ScopedCallback( static function () use ( $db, $autoTrx ) {
if ( $autoTrx ) {
$db->setFlag( DBO_TRX ); // restore old setting
}
} );
}
/**
* @param string $property
* @return string
*/
private function getCacheKey( $property ) {
$cluster = is_string( $this->cluster ) ? $this->cluster : 'main';
return $this->wanCache->makeGlobalKey(
'jobqueue',
$this->domain,
$cluster,
$this->type,
$property
);
}
/**
* @param array|false $params
* @return string
*/
protected static function makeBlob( $params ) {
if ( $params !== false ) {
return serialize( $params );
} else {
return '';
}
}
/**
* @param stdClass $row
* @return RunnableJob
*/
protected function jobFromRow( $row ) {
$params = ( (string)$row->job_params !== '' ) ? unserialize( $row->job_params ) : [];
if ( !is_array( $params ) ) { // this shouldn't happen
throw new UnexpectedValueException(
"Could not unserialize job with ID '{$row->job_id}'." );
}
$params += [ 'namespace' => $row->job_namespace, 'title' => $row->job_title ];
$job = $this->factoryJob( $row->job_cmd, $params );
$job->setMetadata( 'id', $row->job_id );
$job->setMetadata( 'timestamp', $row->job_timestamp );
return $job;
}
/**
* @param DBError $e
* @return JobQueueError
*/
protected function getDBException( DBError $e ) {
return new JobQueueError( get_class( $e ) . ": " . $e->getMessage() );
}
/**
* Return the list of job fields that should be selected.
* @since 1.23
* @return array
*/
public static function selectFields() {
return [
'job_id',
'job_cmd',
'job_namespace',
'job_title',
'job_timestamp',
'job_params',
'job_random',
'job_attempts',
'job_token',
'job_token_timestamp',
'job_sha1',
];
}
}