includes/jobqueue/JobQueueRedis.php
<?php
/**
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License along
* with this program; if not, write to the Free Software Foundation, Inc.,
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
* http://www.gnu.org/copyleft/gpl.html
*
* @file
*/
use MediaWiki\Logger\LoggerFactory;
use MediaWiki\WikiMap\WikiMap;
use Psr\Log\LoggerInterface;
/**
* Redis-backed job queue storage.
*
* This is a faster and less resource-intensive job queue than JobQueueDB.
* All data for a queue using this class is placed into one redis server.
*
* When used on a wiki farm, you can optionally use the `redisJobRunnerService` background
* service from the `mediawiki/services/jobrunner.git` repository, to run jobs from a central
* system rather than per-wiki via one of the default job runners (e.g. maintenance/runJobs.php).
*
* There are eight main redis keys (per queue) used to track jobs:
* - l-unclaimed : A list of job IDs used for ready unclaimed jobs
* - z-claimed : A sorted set of (job ID, UNIX timestamp as score) used for job retries
* - z-abandoned : A sorted set of (job ID, UNIX timestamp as score) used for broken jobs
* - z-delayed : A sorted set of (job ID, UNIX timestamp as score) used for delayed jobs
* - h-idBySha1 : A hash of (SHA1 => job ID) for unclaimed jobs used for de-duplication
* - h-sha1ById : A hash of (job ID => SHA1) for unclaimed jobs used for de-duplication
* - h-attempts : A hash of (job ID => attempt count) used for job claiming/retries
* - h-data : A hash of (job ID => serialized blobs) for job storage
*
* A job ID can be in only one of z-delayed, l-unclaimed, z-claimed, and z-abandoned.
* If an ID appears in any of those lists, it should have a h-data entry for its ID.
* If a job has a SHA1 de-duplication value and its ID is in l-unclaimed or z-delayed, then
* there should be no other such jobs with that SHA1. Every h-idBySha1 entry has an h-sha1ById
* entry and every h-sha1ById must refer to an ID that is l-unclaimed. If a job has its
* ID in z-claimed or z-abandoned, then it must also have an h-attempts entry for its ID.
*
* The following keys are used to track queue states:
* - s-queuesWithJobs : A set of all queues with non-abandoned jobs
*
* The background service takes care of undelaying, recycling, and pruning jobs as well as
* removing s-queuesWithJobs entries as queues empty.
*
* Additionally, "rootjob:* keys track "root jobs" used for additional de-duplication.
* Aside from root job keys, all keys have no expiry, and are only removed when jobs are run.
* All the keys are prefixed with the relevant wiki ID information.
*
* This class requires Redis 2.6 or later as it uses Lua scripting for fast atomic operations.
* Additionally, it should be noted that redis has different persistence modes, such
* as rdb snapshots, journaling, or no persistence. Appropriate configuration should be
* made on the servers based on what queues are using it and what tolerance they have.
*
* @since 1.22
* @ingroup JobQueue
* @ingroup Redis
*/
class JobQueueRedis extends JobQueue {
/** @var RedisConnectionPool */
protected $redisPool;
/** @var LoggerInterface */
protected $logger;
/** @var string Server address */
protected $server;
/** @var string Compression method to use */
protected $compression;
private const MAX_PUSH_SIZE = 25; // avoid tying up the server
/**
* @param array $params Possible keys:
* - redisConfig : An array of parameters to RedisConnectionPool::__construct().
* Note that the serializer option is ignored as "none" is always used.
* - redisServer : A hostname/port combination or the absolute path of a UNIX socket.
* If a hostname is specified but no port, the standard port number
* 6379 will be used. Required.
* - compression : The type of compression to use; one of (none,gzip).
* - daemonized : Set to true if the redisJobRunnerService runs in the background.
* This will disable job recycling/undelaying from the MediaWiki side
* to avoid redundancy and out-of-sync configuration.
*/
public function __construct( array $params ) {
parent::__construct( $params );
$params['redisConfig']['serializer'] = 'none'; // make it easy to use Lua
$this->server = $params['redisServer'];
$this->compression = $params['compression'] ?? 'none';
$this->redisPool = RedisConnectionPool::singleton( $params['redisConfig'] );
if ( empty( $params['daemonized'] ) ) {
throw new InvalidArgumentException(
"Non-daemonized mode is no longer supported. Please install the " .
"mediawiki/services/jobrunner service and update \$wgJobTypeConf as needed." );
}
$this->logger = LoggerFactory::getInstance( 'redis' );
}
protected function supportedOrders() {
return [ 'timestamp', 'fifo' ];
}
protected function optimalOrder() {
return 'fifo';
}
protected function supportsDelayedJobs() {
return true;
}
/**
* @see JobQueue::doIsEmpty()
* @return bool
* @throws JobQueueError
*/
protected function doIsEmpty() {
return $this->doGetSize() == 0;
}
/**
* @see JobQueue::doGetSize()
* @return int
* @throws JobQueueError
*/
protected function doGetSize() {
$conn = $this->getConnection();
try {
return $conn->lLen( $this->getQueueKey( 'l-unclaimed' ) );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @see JobQueue::doGetAcquiredCount()
* @return int
* @throws JobQueueError
*/
protected function doGetAcquiredCount() {
$conn = $this->getConnection();
try {
$conn->multi( Redis::PIPELINE );
$conn->zCard( $this->getQueueKey( 'z-claimed' ) );
$conn->zCard( $this->getQueueKey( 'z-abandoned' ) );
return array_sum( $conn->exec() );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @see JobQueue::doGetDelayedCount()
* @return int
* @throws JobQueueError
*/
protected function doGetDelayedCount() {
$conn = $this->getConnection();
try {
return $conn->zCard( $this->getQueueKey( 'z-delayed' ) );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @see JobQueue::doGetAbandonedCount()
* @return int
* @throws JobQueueError
*/
protected function doGetAbandonedCount() {
$conn = $this->getConnection();
try {
return $conn->zCard( $this->getQueueKey( 'z-abandoned' ) );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @see JobQueue::doBatchPush()
* @param IJobSpecification[] $jobs
* @param int $flags
* @return void
* @throws JobQueueError
*/
protected function doBatchPush( array $jobs, $flags ) {
// Convert the jobs into field maps (de-duplicated against each other)
$items = []; // (job ID => job fields map)
foreach ( $jobs as $job ) {
$item = $this->getNewJobFields( $job );
if ( strlen( $item['sha1'] ) ) { // hash identifier => de-duplicate
$items[$item['sha1']] = $item;
} else {
$items[$item['uuid']] = $item;
}
}
if ( $items === [] ) {
return; // nothing to do
}
$conn = $this->getConnection();
try {
// Actually push the non-duplicate jobs into the queue...
if ( $flags & self::QOS_ATOMIC ) {
$batches = [ $items ]; // all or nothing
} else {
$batches = array_chunk( $items, self::MAX_PUSH_SIZE );
}
$failed = 0;
$pushed = 0;
foreach ( $batches as $itemBatch ) {
$added = $this->pushBlobs( $conn, $itemBatch );
if ( is_int( $added ) ) {
$pushed += $added;
} else {
$failed += count( $itemBatch );
}
}
$this->incrStats( 'inserts', $this->type, count( $items ) );
$this->incrStats( 'inserts_actual', $this->type, $pushed );
$this->incrStats( 'dupe_inserts', $this->type,
count( $items ) - $failed - $pushed );
if ( $failed > 0 ) {
$err = "Could not insert {$failed} {$this->type} job(s).";
wfDebugLog( 'JobQueue', $err );
throw new RedisException( $err );
}
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @param RedisConnRef $conn
* @param array[] $items List of results from JobQueueRedis::getNewJobFields()
* @return int Number of jobs inserted (duplicates are ignored)
* @throws RedisException
*/
protected function pushBlobs( RedisConnRef $conn, array $items ) {
$args = [ $this->encodeQueueName() ];
// Next args come in 4s ([id, sha1, rtime, blob [, id, sha1, rtime, blob ... ] ] )
foreach ( $items as $item ) {
$args[] = (string)$item['uuid'];
$args[] = (string)$item['sha1'];
$args[] = (string)$item['rtimestamp'];
$args[] = (string)$this->serialize( $item );
}
static $script =
/** @lang Lua */
<<<LUA
local kUnclaimed, kSha1ById, kIdBySha1, kDelayed, kData, kQwJobs = unpack(KEYS)
-- First argument is the queue ID
local queueId = ARGV[1]
-- Next arguments all come in 4s (one per job)
local variadicArgCount = #ARGV - 1
if variadicArgCount % 4 ~= 0 then
return redis.error_reply('Unmatched arguments')
end
-- Insert each job into this queue as needed
local pushed = 0
for i = 2,#ARGV,4 do
local id,sha1,rtimestamp,blob = ARGV[i],ARGV[i+1],ARGV[i+2],ARGV[i+3]
if sha1 == '' or redis.call('hExists',kIdBySha1,sha1) == 0 then
if 1*rtimestamp > 0 then
-- Insert into delayed queue (release time as score)
redis.call('zAdd',kDelayed,rtimestamp,id)
else
-- Insert into unclaimed queue
redis.call('lPush',kUnclaimed,id)
end
if sha1 ~= '' then
redis.call('hSet',kSha1ById,id,sha1)
redis.call('hSet',kIdBySha1,sha1,id)
end
redis.call('hSet',kData,id,blob)
pushed = pushed + 1
end
end
-- Mark this queue as having jobs
redis.call('sAdd',kQwJobs,queueId)
return pushed
LUA;
return $conn->luaEval( $script,
array_merge(
[
$this->getQueueKey( 'l-unclaimed' ), # KEYS[1]
$this->getQueueKey( 'h-sha1ById' ), # KEYS[2]
$this->getQueueKey( 'h-idBySha1' ), # KEYS[3]
$this->getQueueKey( 'z-delayed' ), # KEYS[4]
$this->getQueueKey( 'h-data' ), # KEYS[5]
$this->getGlobalKey( 's-queuesWithJobs' ), # KEYS[6]
],
$args
),
6 # number of first argument(s) that are keys
);
}
/**
* @see JobQueue::doPop()
* @return RunnableJob|false
* @throws JobQueueError
*/
protected function doPop() {
$job = false;
$conn = $this->getConnection();
try {
do {
$blob = $this->popAndAcquireBlob( $conn );
if ( !is_string( $blob ) ) {
break; // no jobs; nothing to do
}
$this->incrStats( 'pops', $this->type );
$item = $this->unserialize( $blob );
if ( $item === false ) {
wfDebugLog( 'JobQueue', "Could not unserialize {$this->type} job." );
continue;
}
// If $item is invalid, the runner loop recycling will cleanup as needed
$job = $this->getJobFromFields( $item ); // may be false
} while ( !$job ); // job may be false if invalid
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return $job;
}
/**
* @param RedisConnRef $conn
* @return array Serialized string or false
* @throws RedisException
*/
protected function popAndAcquireBlob( RedisConnRef $conn ) {
static $script =
/** @lang Lua */
<<<LUA
local kUnclaimed, kSha1ById, kIdBySha1, kClaimed, kAttempts, kData = unpack(KEYS)
local rTime = unpack(ARGV)
-- Pop an item off the queue
local id = redis.call('rPop',kUnclaimed)
if not id then
return false
end
-- Allow new duplicates of this job
local sha1 = redis.call('hGet',kSha1ById,id)
if sha1 then redis.call('hDel',kIdBySha1,sha1) end
redis.call('hDel',kSha1ById,id)
-- Mark the jobs as claimed and return it
redis.call('zAdd',kClaimed,rTime,id)
redis.call('hIncrBy',kAttempts,id,1)
return redis.call('hGet',kData,id)
LUA;
return $conn->luaEval( $script,
[
$this->getQueueKey( 'l-unclaimed' ), # KEYS[1]
$this->getQueueKey( 'h-sha1ById' ), # KEYS[2]
$this->getQueueKey( 'h-idBySha1' ), # KEYS[3]
$this->getQueueKey( 'z-claimed' ), # KEYS[4]
$this->getQueueKey( 'h-attempts' ), # KEYS[5]
$this->getQueueKey( 'h-data' ), # KEYS[6]
time(), # ARGV[1] (injected to be replication-safe)
],
6 # number of first argument(s) that are keys
);
}
/**
* @see JobQueue::doAck()
* @param RunnableJob $job
* @return RunnableJob|bool
* @throws UnexpectedValueException
* @throws JobQueueError
*/
protected function doAck( RunnableJob $job ) {
$uuid = $job->getMetadata( 'uuid' );
if ( $uuid === null ) {
throw new UnexpectedValueException( "Job of type '{$job->getType()}' has no UUID." );
}
$conn = $this->getConnection();
try {
static $script =
/** @lang Lua */
<<<LUA
local kClaimed, kAttempts, kData = unpack(KEYS)
local id = unpack(ARGV)
-- Unmark the job as claimed
local removed = redis.call('zRem',kClaimed,id)
-- Check if the job was recycled
if removed == 0 then
return 0
end
-- Delete the retry data
redis.call('hDel',kAttempts,id)
-- Delete the job data itself
return redis.call('hDel',kData,id)
LUA;
$res = $conn->luaEval( $script,
[
$this->getQueueKey( 'z-claimed' ), # KEYS[1]
$this->getQueueKey( 'h-attempts' ), # KEYS[2]
$this->getQueueKey( 'h-data' ), # KEYS[3]
$uuid # ARGV[1]
],
3 # number of first argument(s) that are keys
);
if ( !$res ) {
wfDebugLog( 'JobQueue', "Could not acknowledge {$this->type} job $uuid." );
return false;
}
$this->incrStats( 'acks', $this->type );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return true;
}
/**
* @see JobQueue::doDeduplicateRootJob()
* @param IJobSpecification $job
* @return bool
* @throws JobQueueError
*/
protected function doDeduplicateRootJob( IJobSpecification $job ) {
if ( !$job->hasRootJobParams() ) {
throw new LogicException( "Cannot register root job; missing parameters." );
}
$params = $job->getRootJobParams();
$key = $this->getRootJobCacheKey( $params['rootJobSignature'], $job->getType() );
$conn = $this->getConnection();
try {
$timestamp = $conn->get( $key ); // last known timestamp of such a root job
if ( $timestamp && $timestamp >= $params['rootJobTimestamp'] ) {
return true; // a newer version of this root job was enqueued
}
// Update the timestamp of the last root job started at the location...
return $conn->set( $key, $params['rootJobTimestamp'], self::ROOTJOB_TTL ); // 2 weeks
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @see JobQueue::doIsRootJobOldDuplicate()
* @param IJobSpecification $job
* @return bool
* @throws JobQueueError
*/
protected function doIsRootJobOldDuplicate( IJobSpecification $job ) {
if ( !$job->hasRootJobParams() ) {
return false; // job has no de-duplication info
}
$params = $job->getRootJobParams();
$conn = $this->getConnection();
try {
// Get the last time this root job was enqueued
$timestamp = $conn->get( $this->getRootJobCacheKey( $params['rootJobSignature'], $job->getType() ) );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
// Check if a new root job was started at the location after this one's...
return ( $timestamp && $timestamp > $params['rootJobTimestamp'] );
}
/**
* @see JobQueue::doDelete()
* @return bool
* @throws JobQueueError
*/
protected function doDelete() {
static $props = [ 'l-unclaimed', 'z-claimed', 'z-abandoned',
'z-delayed', 'h-idBySha1', 'h-sha1ById', 'h-attempts', 'h-data' ];
$conn = $this->getConnection();
try {
$keys = [];
foreach ( $props as $prop ) {
$keys[] = $this->getQueueKey( $prop );
}
$ok = ( $conn->del( $keys ) !== false );
$conn->sRem( $this->getGlobalKey( 's-queuesWithJobs' ), $this->encodeQueueName() );
return $ok;
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @see JobQueue::getAllQueuedJobs()
* @return Iterator<RunnableJob>
* @throws JobQueueError
*/
public function getAllQueuedJobs() {
$conn = $this->getConnection();
try {
$uids = $conn->lRange( $this->getQueueKey( 'l-unclaimed' ), 0, -1 );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return $this->getJobIterator( $conn, $uids );
}
/**
* @see JobQueue::getAllDelayedJobs()
* @return Iterator<RunnableJob>
* @throws JobQueueError
*/
public function getAllDelayedJobs() {
$conn = $this->getConnection();
try {
$uids = $conn->zRange( $this->getQueueKey( 'z-delayed' ), 0, -1 );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return $this->getJobIterator( $conn, $uids );
}
/**
* @see JobQueue::getAllAcquiredJobs()
* @return Iterator<RunnableJob>
* @throws JobQueueError
*/
public function getAllAcquiredJobs() {
$conn = $this->getConnection();
try {
$uids = $conn->zRange( $this->getQueueKey( 'z-claimed' ), 0, -1 );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return $this->getJobIterator( $conn, $uids );
}
/**
* @see JobQueue::getAllAbandonedJobs()
* @return Iterator<RunnableJob>
* @throws JobQueueError
*/
public function getAllAbandonedJobs() {
$conn = $this->getConnection();
try {
$uids = $conn->zRange( $this->getQueueKey( 'z-abandoned' ), 0, -1 );
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return $this->getJobIterator( $conn, $uids );
}
/**
* @param RedisConnRef $conn
* @param array $uids List of job UUIDs
* @return MappedIterator<RunnableJob>
*/
protected function getJobIterator( RedisConnRef $conn, array $uids ) {
return new MappedIterator(
$uids,
function ( $uid ) use ( $conn ) {
return $this->getJobFromUidInternal( $uid, $conn );
},
[ 'accept' => static function ( $job ) {
return is_object( $job );
} ]
);
}
public function getCoalesceLocationInternal() {
return "RedisServer:" . $this->server;
}
protected function doGetSiblingQueuesWithJobs( array $types ) {
return array_keys( array_filter( $this->doGetSiblingQueueSizes( $types ) ) );
}
protected function doGetSiblingQueueSizes( array $types ) {
$sizes = []; // (type => size)
$types = array_values( $types ); // reindex
$conn = $this->getConnection();
try {
$conn->multi( Redis::PIPELINE );
foreach ( $types as $type ) {
$conn->lLen( $this->getQueueKey( 'l-unclaimed', $type ) );
}
$res = $conn->exec();
if ( is_array( $res ) ) {
foreach ( $res as $i => $size ) {
$sizes[$types[$i]] = $size;
}
}
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return $sizes;
}
/**
* This function should not be called outside JobQueueRedis
*
* @param string $uid
* @param RedisConnRef|Redis $conn
* @return RunnableJob|false Returns false if the job does not exist
* @throws JobQueueError
* @throws UnexpectedValueException
*/
public function getJobFromUidInternal( $uid, $conn ) {
try {
$data = $conn->hGet( $this->getQueueKey( 'h-data' ), $uid );
if ( $data === false ) {
return false; // not found
}
$item = $this->unserialize( $data );
if ( !is_array( $item ) ) { // this shouldn't happen
throw new UnexpectedValueException( "Could not unserialize job with ID '$uid'." );
}
$params = $item['params'];
$params += [ 'namespace' => $item['namespace'], 'title' => $item['title'] ];
$job = $this->factoryJob( $item['type'], $params );
$job->setMetadata( 'uuid', $item['uuid'] );
$job->setMetadata( 'timestamp', $item['timestamp'] );
// Add in attempt count for debugging at showJobs.php
$job->setMetadata( 'attempts',
$conn->hGet( $this->getQueueKey( 'h-attempts' ), $uid ) );
return $job;
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
}
/**
* @return array List of (wiki,type) tuples for queues with non-abandoned jobs
* @throws JobQueueConnectionError
* @throws JobQueueError
*/
public function getServerQueuesWithJobs() {
$queues = [];
$conn = $this->getConnection();
try {
$set = $conn->sMembers( $this->getGlobalKey( 's-queuesWithJobs' ) );
foreach ( $set as $queue ) {
$queues[] = $this->decodeQueueName( $queue );
}
} catch ( RedisException $e ) {
throw $this->handleErrorAndMakeException( $conn, $e );
}
return $queues;
}
/**
* @param IJobSpecification $job
* @return array
*/
protected function getNewJobFields( IJobSpecification $job ) {
return [
// Fields that describe the nature of the job
'type' => $job->getType(),
'namespace' => $job->getParams()['namespace'] ?? NS_SPECIAL,
'title' => $job->getParams()['title'] ?? '',
'params' => $job->getParams(),
// Some jobs cannot run until a "release timestamp"
'rtimestamp' => $job->getReleaseTimestamp() ?: 0,
// Additional job metadata
'uuid' => $this->idGenerator->newRawUUIDv4(),
'sha1' => $job->ignoreDuplicates()
? Wikimedia\base_convert( sha1( serialize( $job->getDeduplicationInfo() ) ), 16, 36, 31 )
: '',
'timestamp' => time() // UNIX timestamp
];
}
/**
* @param array $fields
* @return RunnableJob|false
*/
protected function getJobFromFields( array $fields ) {
$params = $fields['params'];
$params += [ 'namespace' => $fields['namespace'], 'title' => $fields['title'] ];
$job = $this->factoryJob( $fields['type'], $params );
$job->setMetadata( 'uuid', $fields['uuid'] );
$job->setMetadata( 'timestamp', $fields['timestamp'] );
return $job;
}
/**
* @param array $fields
* @return string Serialized and possibly compressed version of $fields
*/
protected function serialize( array $fields ) {
$blob = serialize( $fields );
if ( $this->compression === 'gzip'
&& strlen( $blob ) >= 1024
&& function_exists( 'gzdeflate' )
) {
$object = (object)[ 'blob' => gzdeflate( $blob ), 'enc' => 'gzip' ];
$blobz = serialize( $object );
return ( strlen( $blobz ) < strlen( $blob ) ) ? $blobz : $blob;
} else {
return $blob;
}
}
/**
* @param string $blob
* @return array|false Unserialized version of $blob or false
*/
protected function unserialize( $blob ) {
$fields = unserialize( $blob );
if ( is_object( $fields ) ) {
if ( $fields->enc === 'gzip' && function_exists( 'gzinflate' ) ) {
$fields = unserialize( gzinflate( $fields->blob ) );
} else {
$fields = false;
}
}
return is_array( $fields ) ? $fields : false;
}
/**
* Get a connection to the server that handles all sub-queues for this queue
*
* @return RedisConnRef|Redis
* @throws JobQueueConnectionError
*/
protected function getConnection() {
$conn = $this->redisPool->getConnection( $this->server, $this->logger );
if ( !$conn ) {
throw new JobQueueConnectionError(
"Unable to connect to redis server {$this->server}." );
}
return $conn;
}
/**
* @param RedisConnRef $conn
* @param RedisException $e
* @return JobQueueError
*/
protected function handleErrorAndMakeException( RedisConnRef $conn, $e ) {
$this->redisPool->handleError( $conn, $e );
return new JobQueueError( "Redis server error: {$e->getMessage()}\n" );
}
/**
* @return string JSON
*/
private function encodeQueueName() {
return json_encode( [ $this->type, $this->domain ] );
}
/**
* @param string $name JSON
* @return array (type, wiki)
*/
private function decodeQueueName( $name ) {
return json_decode( $name );
}
/**
* @param string $name
* @return string
*/
private function getGlobalKey( $name ) {
$parts = [ 'global', 'jobqueue', $name ];
foreach ( $parts as $part ) {
if ( !preg_match( '/[a-zA-Z0-9_-]+/', $part ) ) {
throw new InvalidArgumentException( "Key part characters are out of range." );
}
}
return implode( ':', $parts );
}
/**
* @param string $prop
* @param string|null $type Override this for sibling queues
* @return string
*/
private function getQueueKey( $prop, $type = null ) {
$type = is_string( $type ) ? $type : $this->type;
// Use wiki ID for b/c
$keyspace = WikiMap::getWikiIdFromDbDomain( $this->domain );
$parts = [ $keyspace, 'jobqueue', $type, $prop ];
// Parts are typically ASCII, but encode to escape ":"
return implode( ':', array_map( 'rawurlencode', $parts ) );
}
}