1
0
Fork 0
mirror of https://github.com/mperham/sidekiq.git synced 2022-11-09 13:52:34 -05:00
mperham--sidekiq/lib/sidekiq/scheduled.rb

102 lines
3.9 KiB
Ruby
Raw Normal View History

require 'sidekiq'
require 'sidekiq/util'
require 'sidekiq/actor'
2014-10-06 11:53:06 -04:00
require 'sidekiq/api'
module Sidekiq
module Scheduled
2014-06-07 23:57:33 -04:00
INITIAL_WAIT = 10
##
2014-10-06 11:53:25 -04:00
# The Poller checks Redis every N seconds for jobs in the retry or scheduled
# set have passed their timestamp and should be enqueued. If so, it
2014-10-06 11:53:25 -04:00
# just pops the job back onto its original queue so the
# workers can pick it up like any other job.
class Poller
include Util
include Actor
SETS = %w(retry schedule)
def poll(first_time=false)
watchdog('scheduling poller thread died!') do
initial_wait if first_time
begin
2014-10-06 11:53:25 -04:00
# A job's "score" in Redis is the time at which it should be processed.
# Just check Redis for the set of jobs with a timestamp before now.
now = Time.now.to_f.to_s
Sidekiq.redis do |conn|
SETS.each do |sorted_set|
# Get the next item in the queue if it's score (time to execute) is <= now.
# We need to go through the list one at a time to reduce the risk of something
# going wrong between the time jobs are popped from the scheduled queue and when
# they are pushed onto a work queue and losing the jobs.
2014-10-06 11:53:25 -04:00
while job = conn.zrangebyscore(sorted_set, '-inf', now, :limit => [0, 1]).first do
# Pop item off the queue and add it to the work queue. If the job can't be popped from
# the queue, it's because another process already popped it so we can move on to the
# next one.
2014-10-06 11:53:25 -04:00
if conn.zrem(sorted_set, job)
Sidekiq::Client.push(Sidekiq.load_json(job))
logger.debug { "enqueued #{sorted_set}: #{job}" }
end
end
end
end
rescue => ex
# Most likely a problem with redis networking.
# Punt and try again at the next interval
logger.error ex.message
logger.error ex.backtrace.first
end
# Randomizing scales the interval by half since
# on average calling `rand` returns 0.5.
# We make up for this by doubling the interval
after(poll_interval * 2 * rand) { poll }
end
end
private
# We do our best to tune poll_interval to the size of the active Sidekiq
# cluster. If you have 30 processes and poll every 15 seconds, that means one
# Sidekiq is checking Redis every 0.5 seconds - way too often for most people
# and really bad if the retry or scheduled sets are large.
#
# Instead try to avoid polling more than once every 15 seconds. If you have
# 30 Sidekiq processes, we'll set poll_interval to 30 * 15 * 2 or 900 seconds.
# To keep things statistically random, we'll sleep a random amount between
# 0 and 900 seconds for each poll or 450 seconds on average. Otherwise restarting
# all your Sidekiq processes at the same time will lead to them all polling at
# the same time: the thundering herd problem.
#
# We only do this if poll_interval is unset (the default).
2012-06-19 16:39:20 -04:00
def poll_interval
Sidekiq.options[:poll_interval] ||= begin
2014-10-06 11:58:37 -04:00
ps = Sidekiq::ProcessSet.new
pcount = ps.size
pcount = 1 if pcount == 0
pcount * 15
end
2012-06-19 16:39:20 -04:00
end
def initial_wait
begin
2014-06-07 23:57:33 -04:00
# Have all processes sleep between 10-15 seconds. 10 seconds
# to give time for the heartbeat to register and 5 random seconds
# to ensure they don't all hit Redis at the same time.
sleep(INITIAL_WAIT)
2014-06-07 23:57:33 -04:00
sleep(5 * rand)
rescue Celluloid::Task::TerminatedError
# Hit Ctrl-C when Sidekiq is finished booting and we have a chance
# to get here.
end
end
end
end
end