mirror of
https://github.com/rails/rails.git
synced 2022-11-09 12:12:34 -05:00
116 lines
4 KiB
Ruby
116 lines
4 KiB
Ruby
# frozen_string_literal: true
|
|
|
|
require "securerandom"
|
|
require "concurrent/scheduled_task"
|
|
require "concurrent/executor/thread_pool_executor"
|
|
require "concurrent/utility/processor_counter"
|
|
|
|
module ActiveJob
|
|
module QueueAdapters
|
|
# == Active Job Async adapter
|
|
#
|
|
# The Async adapter runs jobs with an in-process thread pool.
|
|
#
|
|
# This is the default queue adapter. It's well-suited for dev/test since
|
|
# it doesn't need an external infrastructure, but it's a poor fit for
|
|
# production since it drops pending jobs on restart.
|
|
#
|
|
# To use this adapter, set queue adapter to +:async+:
|
|
#
|
|
# config.active_job.queue_adapter = :async
|
|
#
|
|
# To configure the adapter's thread pool, instantiate the adapter and
|
|
# pass your own config:
|
|
#
|
|
# config.active_job.queue_adapter = ActiveJob::QueueAdapters::AsyncAdapter.new \
|
|
# min_threads: 1,
|
|
# max_threads: 2 * Concurrent.processor_count,
|
|
# idletime: 600.seconds
|
|
#
|
|
# The adapter uses a {Concurrent Ruby}[https://github.com/ruby-concurrency/concurrent-ruby] thread pool to schedule and execute
|
|
# jobs. Since jobs share a single thread pool, long-running jobs will block
|
|
# short-lived jobs. Fine for dev/test; bad for production.
|
|
class AsyncAdapter
|
|
# See {Concurrent::ThreadPoolExecutor}[https://ruby-concurrency.github.io/concurrent-ruby/master/Concurrent/ThreadPoolExecutor.html] for executor options.
|
|
def initialize(**executor_options)
|
|
@scheduler = Scheduler.new(**executor_options)
|
|
end
|
|
|
|
def enqueue(job) #:nodoc:
|
|
@scheduler.enqueue JobWrapper.new(job), queue_name: job.queue_name
|
|
end
|
|
|
|
def enqueue_at(job, timestamp) #:nodoc:
|
|
@scheduler.enqueue_at JobWrapper.new(job), timestamp, queue_name: job.queue_name
|
|
end
|
|
|
|
# Gracefully stop processing jobs. Finishes in-progress work and handles
|
|
# any new jobs following the executor's fallback policy (`caller_runs`).
|
|
# Waits for termination by default. Pass `wait: false` to continue.
|
|
def shutdown(wait: true) #:nodoc:
|
|
@scheduler.shutdown wait: wait
|
|
end
|
|
|
|
# Used for our test suite.
|
|
def immediate=(immediate) #:nodoc:
|
|
@scheduler.immediate = immediate
|
|
end
|
|
|
|
# Note that we don't actually need to serialize the jobs since we're
|
|
# performing them in-process, but we do so anyway for parity with other
|
|
# adapters and deployment environments. Otherwise, serialization bugs
|
|
# may creep in undetected.
|
|
class JobWrapper #:nodoc:
|
|
def initialize(job)
|
|
job.provider_job_id = SecureRandom.uuid
|
|
@job_data = job.serialize
|
|
end
|
|
|
|
def perform
|
|
Base.execute @job_data
|
|
end
|
|
end
|
|
|
|
class Scheduler #:nodoc:
|
|
DEFAULT_EXECUTOR_OPTIONS = {
|
|
min_threads: 0,
|
|
max_threads: Concurrent.processor_count,
|
|
auto_terminate: true,
|
|
idletime: 60, # 1 minute
|
|
max_queue: 0, # unlimited
|
|
fallback_policy: :caller_runs # shouldn't matter -- 0 max queue
|
|
}.freeze
|
|
|
|
attr_accessor :immediate
|
|
|
|
def initialize(**options)
|
|
self.immediate = false
|
|
@immediate_executor = Concurrent::ImmediateExecutor.new
|
|
@async_executor = Concurrent::ThreadPoolExecutor.new(DEFAULT_EXECUTOR_OPTIONS.merge(options))
|
|
end
|
|
|
|
def enqueue(job, queue_name:)
|
|
executor.post(job, &:perform)
|
|
end
|
|
|
|
def enqueue_at(job, timestamp, queue_name:)
|
|
delay = timestamp - Time.current.to_f
|
|
if delay > 0
|
|
Concurrent::ScheduledTask.execute(delay, args: [job], executor: executor, &:perform)
|
|
else
|
|
enqueue(job, queue_name: queue_name)
|
|
end
|
|
end
|
|
|
|
def shutdown(wait: true)
|
|
@async_executor.shutdown
|
|
@async_executor.wait_for_termination if wait
|
|
end
|
|
|
|
def executor
|
|
immediate ? @immediate_executor : @async_executor
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|