2012-02-08 17:04:02 -08:00
|
|
|
require 'sidekiq/util'
|
2013-05-10 20:43:53 -07:00
|
|
|
require 'sidekiq/actor'
|
2012-02-08 17:04:02 -08:00
|
|
|
|
2012-03-17 13:41:53 -07:00
|
|
|
require 'sidekiq/middleware/server/retry_jobs'
|
2012-02-25 13:43:53 -08:00
|
|
|
require 'sidekiq/middleware/server/logging'
|
2012-02-19 13:02:32 -08:00
|
|
|
|
2012-01-25 13:32:51 -08:00
|
|
|
module Sidekiq
|
2012-06-12 21:55:06 -07:00
|
|
|
##
|
|
|
|
# The Processor receives a message from the Manager and actually
|
|
|
|
# processes it. It instantiates the worker, runs the middleware
|
|
|
|
# chain and then calls Sidekiq::Worker#perform.
|
2012-01-25 13:32:51 -08:00
|
|
|
class Processor
|
2014-03-23 23:02:36 -04:00
|
|
|
# To prevent a memory leak, ensure that stats expire. However, they should take up a minimal amount of storage
|
|
|
|
# so keep them around for a long time
|
|
|
|
STATS_TIMEOUT = 24 * 60 * 60 * 365 * 5
|
2013-05-31 09:02:27 -07:00
|
|
|
|
2012-01-26 12:45:04 -08:00
|
|
|
include Util
|
2013-05-10 20:43:53 -07:00
|
|
|
include Actor
|
2012-01-25 13:32:51 -08:00
|
|
|
|
2012-02-19 13:02:32 -08:00
|
|
|
def self.default_middleware
|
|
|
|
Middleware::Chain.new do |m|
|
2012-02-25 13:43:53 -08:00
|
|
|
m.add Middleware::Server::Logging
|
2012-03-17 23:04:31 -07:00
|
|
|
m.add Middleware::Server::RetryJobs
|
2014-04-24 21:37:20 -07:00
|
|
|
if defined?(::ActiveRecord::Base)
|
|
|
|
require 'sidekiq/middleware/server/active_record'
|
|
|
|
m.add Sidekiq::Middleware::Server::ActiveRecord
|
|
|
|
end
|
2012-02-19 13:02:32 -08:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2013-06-10 22:20:15 -07:00
|
|
|
attr_accessor :proxy_id
|
2013-03-26 22:56:49 -07:00
|
|
|
|
2012-02-08 17:04:02 -08:00
|
|
|
def initialize(boss)
|
|
|
|
@boss = boss
|
2012-02-04 16:53:09 -08:00
|
|
|
end
|
|
|
|
|
2013-01-05 21:17:08 -08:00
|
|
|
def process(work)
|
|
|
|
msgstr = work.message
|
|
|
|
queue = work.queue_name
|
2012-06-29 20:37:45 -07:00
|
|
|
|
2014-02-01 20:49:42 -08:00
|
|
|
@boss.async.real_thread(proxy_id, Thread.current)
|
|
|
|
|
2015-09-09 12:32:41 -07:00
|
|
|
ack = false
|
2014-02-01 20:49:42 -08:00
|
|
|
begin
|
|
|
|
msg = Sidekiq.load_json(msgstr)
|
2015-09-21 14:58:18 -07:00
|
|
|
klass = msg['class'.freeze].constantize
|
2014-02-01 20:49:42 -08:00
|
|
|
worker = klass.new
|
2015-09-21 14:58:18 -07:00
|
|
|
worker.jid = msg['jid'.freeze]
|
2014-02-01 20:49:42 -08:00
|
|
|
|
|
|
|
stats(worker, msg, queue) do
|
|
|
|
Sidekiq.server_middleware.invoke(worker, msg, queue) do
|
2015-09-09 12:32:41 -07:00
|
|
|
# Only ack if we either attempted to start this job or
|
|
|
|
# successfully completed it. This prevents us from
|
|
|
|
# losing jobs if a middleware raises an exception before yielding
|
|
|
|
ack = true
|
2015-09-21 14:58:18 -07:00
|
|
|
execute_job(worker, cloned(msg['args'.freeze]))
|
2012-08-16 18:12:25 -07:00
|
|
|
end
|
2012-02-10 23:16:12 -08:00
|
|
|
end
|
2015-09-09 12:32:41 -07:00
|
|
|
ack = true
|
2014-02-01 20:49:42 -08:00
|
|
|
rescue Sidekiq::Shutdown
|
|
|
|
# Had to force kill this job because it didn't finish
|
|
|
|
# within the timeout. Don't acknowledge the work since
|
|
|
|
# we didn't properly finish it.
|
|
|
|
ack = false
|
|
|
|
rescue Exception => ex
|
|
|
|
handle_exception(ex, msg || { :message => msgstr })
|
|
|
|
raise
|
|
|
|
ensure
|
|
|
|
work.acknowledge if ack
|
2012-02-04 16:53:09 -08:00
|
|
|
end
|
2013-05-11 16:12:01 -07:00
|
|
|
|
2012-11-03 19:56:06 -07:00
|
|
|
@boss.async.processor_done(current_actor)
|
2012-01-25 13:32:51 -08:00
|
|
|
end
|
2012-01-29 14:35:16 -08:00
|
|
|
|
2013-06-07 22:15:13 -07:00
|
|
|
def inspect
|
|
|
|
"<Processor##{object_id.to_s(16)}>"
|
|
|
|
end
|
|
|
|
|
2014-09-09 19:57:39 -06:00
|
|
|
def execute_job(worker, cloned_args)
|
|
|
|
worker.perform(*cloned_args)
|
|
|
|
end
|
|
|
|
|
2013-03-26 22:55:07 -07:00
|
|
|
private
|
|
|
|
|
2014-03-02 16:36:00 -08:00
|
|
|
def thread_identity
|
2014-03-19 20:12:12 -07:00
|
|
|
@str ||= Thread.current.object_id.to_s(36)
|
2012-02-10 23:16:12 -08:00
|
|
|
end
|
|
|
|
|
|
|
|
def stats(worker, msg, queue)
|
2014-02-24 20:06:48 -08:00
|
|
|
# Do not conflate errors from the job with errors caused by updating
|
|
|
|
# stats so calling code can react appropriately
|
2014-02-24 16:08:57 -05:00
|
|
|
retry_and_suppress_exceptions do
|
2014-02-24 20:06:48 -08:00
|
|
|
hash = Sidekiq.dump_json({:queue => queue, :payload => msg, :run_at => Time.now.to_i })
|
2014-03-07 22:41:10 -08:00
|
|
|
Sidekiq.redis do |conn|
|
2014-02-24 14:58:35 -05:00
|
|
|
conn.multi do
|
2014-03-19 20:12:12 -07:00
|
|
|
conn.hmset("#{identity}:workers", thread_identity, hash)
|
2014-06-11 11:17:40 -07:00
|
|
|
conn.expire("#{identity}:workers", 60*60*4)
|
2014-02-24 14:58:35 -05:00
|
|
|
end
|
2012-02-11 13:14:03 -08:00
|
|
|
end
|
2012-02-10 23:16:12 -08:00
|
|
|
end
|
|
|
|
|
2015-09-21 14:58:18 -07:00
|
|
|
nowdate = Time.now.utc.strftime("%Y-%m-%d".freeze)
|
2012-02-10 23:16:12 -08:00
|
|
|
begin
|
|
|
|
yield
|
2012-05-12 13:23:23 -07:00
|
|
|
rescue Exception
|
2014-02-24 16:08:57 -05:00
|
|
|
retry_and_suppress_exceptions do
|
2015-09-21 14:58:18 -07:00
|
|
|
failed = "stat:failed:#{nowdate}"
|
2014-03-07 22:41:10 -08:00
|
|
|
Sidekiq.redis do |conn|
|
2014-12-18 09:15:52 -08:00
|
|
|
conn.multi do
|
2015-09-21 14:58:18 -07:00
|
|
|
conn.incrby("stat:failed".freeze, 1)
|
2014-02-24 14:58:35 -05:00
|
|
|
conn.incrby(failed, 1)
|
2014-12-18 09:15:52 -08:00
|
|
|
conn.expire(failed, STATS_TIMEOUT)
|
2014-02-24 14:58:35 -05:00
|
|
|
end
|
2012-02-11 13:14:03 -08:00
|
|
|
end
|
2012-02-10 23:16:12 -08:00
|
|
|
end
|
|
|
|
raise
|
|
|
|
ensure
|
2014-02-24 16:08:57 -05:00
|
|
|
retry_and_suppress_exceptions do
|
2015-09-21 14:58:18 -07:00
|
|
|
processed = "stat:processed:#{nowdate}"
|
2014-03-07 22:41:10 -08:00
|
|
|
Sidekiq.redis do |conn|
|
2014-12-18 09:15:52 -08:00
|
|
|
conn.multi do
|
2014-03-19 20:12:12 -07:00
|
|
|
conn.hdel("#{identity}:workers", thread_identity)
|
2015-09-21 14:58:18 -07:00
|
|
|
conn.incrby("stat:processed".freeze, 1)
|
2014-02-24 14:58:35 -05:00
|
|
|
conn.incrby(processed, 1)
|
2014-12-18 09:15:52 -08:00
|
|
|
conn.expire(processed, STATS_TIMEOUT)
|
2014-02-24 14:58:35 -05:00
|
|
|
end
|
2012-02-11 13:14:03 -08:00
|
|
|
end
|
2012-02-10 23:16:12 -08:00
|
|
|
end
|
|
|
|
end
|
2012-08-04 12:11:46 -07:00
|
|
|
end
|
2012-02-10 23:16:12 -08:00
|
|
|
|
2014-01-27 20:29:19 -08:00
|
|
|
# Deep clone the arguments passed to the worker so that if
|
2012-08-04 12:11:46 -07:00
|
|
|
# the message fails, what is pushed back onto Redis hasn't
|
|
|
|
# been mutated by the worker.
|
|
|
|
def cloned(ary)
|
2014-01-27 20:29:19 -08:00
|
|
|
Marshal.load(Marshal.dump(ary))
|
2012-02-10 23:16:12 -08:00
|
|
|
end
|
2014-02-24 14:58:35 -05:00
|
|
|
|
2014-02-24 16:10:07 -05:00
|
|
|
# If an exception occurs in the block passed to this method, that block will be retried up to max_retries times.
|
2014-02-24 14:58:35 -05:00
|
|
|
# All exceptions will be swallowed and logged.
|
2015-05-19 19:08:54 -07:00
|
|
|
def retry_and_suppress_exceptions(max_retries = 5)
|
2014-02-24 14:58:35 -05:00
|
|
|
retry_count = 0
|
|
|
|
begin
|
|
|
|
yield
|
2014-02-24 16:10:58 -05:00
|
|
|
rescue => e
|
2014-02-24 14:58:35 -05:00
|
|
|
retry_count += 1
|
|
|
|
if retry_count <= max_retries
|
2014-02-24 16:08:57 -05:00
|
|
|
Sidekiq.logger.debug {"Suppressing and retrying error: #{e.inspect}"}
|
2015-05-20 10:31:16 -07:00
|
|
|
pause_for_recovery(retry_count)
|
2014-02-24 14:58:35 -05:00
|
|
|
retry
|
|
|
|
else
|
2014-02-24 20:47:44 -08:00
|
|
|
handle_exception(e, { :message => "Exhausted #{max_retries} retries"})
|
2014-02-24 14:58:35 -05:00
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
2015-05-19 15:05:09 -07:00
|
|
|
|
2015-05-20 10:31:16 -07:00
|
|
|
def pause_for_recovery(retry_count)
|
2015-05-19 15:05:09 -07:00
|
|
|
sleep(retry_count)
|
|
|
|
end
|
2012-01-25 13:32:51 -08:00
|
|
|
end
|
|
|
|
end
|