2011-09-27 13:53:45 -04:00
|
|
|
require 'stringio'
|
2011-09-27 12:23:03 -04:00
|
|
|
|
2011-09-22 22:24:43 -04:00
|
|
|
require 'puma/thread_pool'
|
2011-09-27 12:23:03 -04:00
|
|
|
require 'puma/const'
|
2011-09-27 16:52:50 -04:00
|
|
|
require 'puma/events'
|
2011-12-01 17:33:34 -05:00
|
|
|
require 'puma/null_io'
|
2012-03-29 18:12:59 -04:00
|
|
|
require 'puma/compat'
|
2012-07-23 13:26:52 -04:00
|
|
|
require 'puma/reactor'
|
|
|
|
require 'puma/client'
|
2012-08-02 18:03:52 -04:00
|
|
|
require 'puma/binder'
|
|
|
|
require 'puma/delegation'
|
|
|
|
require 'puma/accept_nonblock'
|
2013-02-05 01:31:40 -05:00
|
|
|
require 'puma/util'
|
2011-09-27 12:23:03 -04:00
|
|
|
|
2012-01-08 14:37:59 -05:00
|
|
|
require 'puma/puma_http11'
|
2011-09-27 12:23:03 -04:00
|
|
|
|
2013-12-14 05:54:23 -05:00
|
|
|
unless Puma.const_defined? "IOBuffer"
|
2012-08-11 18:20:38 -04:00
|
|
|
require 'puma/io_buffer'
|
|
|
|
end
|
|
|
|
|
2011-09-27 12:23:03 -04:00
|
|
|
require 'socket'
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2011-09-22 22:24:43 -04:00
|
|
|
module Puma
|
2011-12-01 18:23:14 -05:00
|
|
|
|
|
|
|
# The HTTP Server itself. Serves out a single Rack app.
|
2011-09-18 16:02:34 -04:00
|
|
|
class Server
|
|
|
|
|
2011-09-22 22:24:43 -04:00
|
|
|
include Puma::Const
|
2012-08-02 18:03:52 -04:00
|
|
|
extend Puma::Delegation
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2011-09-27 17:33:17 -04:00
|
|
|
attr_reader :thread
|
|
|
|
attr_reader :events
|
2011-09-23 01:14:39 -04:00
|
|
|
attr_accessor :app
|
|
|
|
|
2011-09-27 17:33:17 -04:00
|
|
|
attr_accessor :min_threads
|
|
|
|
attr_accessor :max_threads
|
2011-10-05 00:11:10 -04:00
|
|
|
attr_accessor :persistent_timeout
|
2011-12-05 13:07:01 -05:00
|
|
|
attr_accessor :auto_trim_time
|
2015-05-19 10:14:30 -04:00
|
|
|
attr_accessor :reaping_time
|
2014-01-30 17:55:44 -05:00
|
|
|
attr_accessor :first_data_timeout
|
2011-09-23 01:14:39 -04:00
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# Create a server for the rack app +app+.
|
|
|
|
#
|
|
|
|
# +events+ is an object which will be called when certain error events occur
|
|
|
|
# to be handled. See Puma::Events for the list of current methods to implement.
|
2011-09-18 16:02:34 -04:00
|
|
|
#
|
2011-12-01 18:23:14 -05:00
|
|
|
# Server#run returns a thread that you can join on to wait for the server
|
2015-10-17 14:28:21 -04:00
|
|
|
# to do its work.
|
2011-09-18 16:02:34 -04:00
|
|
|
#
|
2013-07-15 17:29:10 -04:00
|
|
|
def initialize(app, events=Events.stdio, options={})
|
2011-09-27 17:33:17 -04:00
|
|
|
@app = app
|
|
|
|
@events = events
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2013-02-05 01:31:40 -05:00
|
|
|
@check, @notify = Puma::Util.pipe
|
2011-09-27 13:53:45 -04:00
|
|
|
|
2011-12-05 12:01:19 -05:00
|
|
|
@status = :stop
|
2011-09-23 01:14:39 -04:00
|
|
|
|
2011-09-27 17:33:17 -04:00
|
|
|
@min_threads = 0
|
|
|
|
@max_threads = 16
|
2011-12-05 13:07:01 -05:00
|
|
|
@auto_trim_time = 1
|
2015-05-19 10:14:30 -04:00
|
|
|
@reaping_time = 1
|
2011-09-24 03:26:17 -04:00
|
|
|
|
2011-09-27 17:33:17 -04:00
|
|
|
@thread = nil
|
|
|
|
@thread_pool = nil
|
2011-09-27 14:29:20 -04:00
|
|
|
|
2016-07-19 15:22:27 -04:00
|
|
|
@persistent_timeout = options.fetch(:persistent_timeout, PERSISTENT_TIMEOUT)
|
2011-10-05 00:11:10 -04:00
|
|
|
|
2012-08-02 18:03:52 -04:00
|
|
|
@binder = Binder.new(events)
|
2013-02-09 13:12:13 -05:00
|
|
|
@own_binder = true
|
|
|
|
|
2012-08-09 19:54:55 -04:00
|
|
|
@first_data_timeout = FIRST_DATA_TIMEOUT
|
2012-04-11 01:17:49 -04:00
|
|
|
|
2013-07-01 19:48:17 -04:00
|
|
|
@leak_stack_on_error = true
|
|
|
|
|
2013-07-15 17:29:10 -04:00
|
|
|
@options = options
|
2015-01-20 09:46:33 -05:00
|
|
|
@queue_requests = options[:queue_requests].nil? ? true : options[:queue_requests]
|
2013-07-15 17:29:10 -04:00
|
|
|
|
2012-01-13 18:33:05 -05:00
|
|
|
ENV['RACK_ENV'] ||= "development"
|
2013-08-07 19:36:04 -04:00
|
|
|
|
|
|
|
@mode = :http
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
|
2013-07-01 19:48:17 -04:00
|
|
|
attr_accessor :binder, :leak_stack_on_error
|
2012-08-02 18:03:52 -04:00
|
|
|
|
|
|
|
forward :add_tcp_listener, :@binder
|
|
|
|
forward :add_ssl_listener, :@binder
|
|
|
|
forward :add_unix_listener, :@binder
|
2016-02-04 11:15:41 -05:00
|
|
|
forward :connected_port, :@binder
|
2012-08-02 18:03:52 -04:00
|
|
|
|
2013-02-09 13:12:13 -05:00
|
|
|
def inherit_binder(bind)
|
|
|
|
@binder = bind
|
|
|
|
@own_binder = false
|
|
|
|
end
|
|
|
|
|
2013-08-07 19:36:04 -04:00
|
|
|
def tcp_mode!
|
|
|
|
@mode = :tcp
|
|
|
|
end
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# On Linux, use TCP_CORK to better control how the TCP stack
|
|
|
|
# packetizes our stream. This improves both latency and throughput.
|
|
|
|
#
|
2011-12-01 14:40:30 -05:00
|
|
|
if RUBY_PLATFORM =~ /linux/
|
|
|
|
# 6 == Socket::IPPROTO_TCP
|
|
|
|
# 3 == TCP_CORK
|
|
|
|
# 1/0 == turn on/off
|
|
|
|
def cork_socket(socket)
|
2014-01-25 22:32:54 -05:00
|
|
|
begin
|
|
|
|
socket.setsockopt(6, 3, 1) if socket.kind_of? TCPSocket
|
|
|
|
rescue IOError, SystemCallError
|
|
|
|
end
|
2011-12-01 14:40:30 -05:00
|
|
|
end
|
|
|
|
|
|
|
|
def uncork_socket(socket)
|
2013-06-18 02:07:50 -04:00
|
|
|
begin
|
|
|
|
socket.setsockopt(6, 3, 0) if socket.kind_of? TCPSocket
|
2014-01-25 22:32:54 -05:00
|
|
|
rescue IOError, SystemCallError
|
2013-06-18 02:07:50 -04:00
|
|
|
end
|
2011-12-01 14:40:30 -05:00
|
|
|
end
|
|
|
|
else
|
|
|
|
def cork_socket(socket)
|
|
|
|
end
|
|
|
|
|
|
|
|
def uncork_socket(socket)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2011-12-05 14:15:44 -05:00
|
|
|
def backlog
|
|
|
|
@thread_pool and @thread_pool.backlog
|
|
|
|
end
|
|
|
|
|
|
|
|
def running
|
|
|
|
@thread_pool and @thread_pool.spawned
|
|
|
|
end
|
|
|
|
|
2013-08-07 19:36:04 -04:00
|
|
|
# Lopez Mode == raw tcp apps
|
|
|
|
|
|
|
|
def run_lopez_mode(background=true)
|
|
|
|
@thread_pool = ThreadPool.new(@min_threads,
|
|
|
|
@max_threads,
|
|
|
|
Hash) do |client, tl|
|
|
|
|
|
|
|
|
io = client.to_io
|
|
|
|
addr = io.peeraddr.last
|
|
|
|
|
|
|
|
if addr.empty?
|
|
|
|
# Set unix socket addrs to localhost
|
|
|
|
addr = "127.0.0.1:0"
|
|
|
|
else
|
|
|
|
addr = "#{addr}:#{io.peeraddr[1]}"
|
|
|
|
end
|
|
|
|
|
|
|
|
env = { 'thread' => tl, REMOTE_ADDR => addr }
|
|
|
|
|
|
|
|
begin
|
|
|
|
@app.call env, client.to_io
|
|
|
|
rescue Object => e
|
|
|
|
STDERR.puts "! Detected exception at toplevel: #{e.message} (#{e.class})"
|
|
|
|
STDERR.puts e.backtrace
|
|
|
|
end
|
|
|
|
|
|
|
|
client.close unless env['detach']
|
|
|
|
end
|
|
|
|
|
2013-09-13 12:56:39 -04:00
|
|
|
@events.fire :state, :running
|
|
|
|
|
2013-08-07 19:36:04 -04:00
|
|
|
if background
|
|
|
|
@thread = Thread.new { handle_servers_lopez_mode }
|
|
|
|
return @thread
|
|
|
|
else
|
2013-08-08 19:21:22 -04:00
|
|
|
handle_servers_lopez_mode
|
2013-08-07 19:36:04 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def handle_servers_lopez_mode
|
|
|
|
begin
|
|
|
|
check = @check
|
|
|
|
sockets = [check] + @binder.ios
|
|
|
|
pool = @thread_pool
|
|
|
|
|
|
|
|
while @status == :run
|
|
|
|
begin
|
|
|
|
ios = IO.select sockets
|
|
|
|
ios.first.each do |sock|
|
|
|
|
if sock == check
|
|
|
|
break if handle_check
|
|
|
|
else
|
|
|
|
begin
|
|
|
|
if io = sock.accept_nonblock
|
2014-03-28 11:26:11 -04:00
|
|
|
client = Client.new io, nil
|
|
|
|
pool << client
|
2013-08-07 19:36:04 -04:00
|
|
|
end
|
|
|
|
rescue SystemCallError
|
2015-08-04 01:09:17 -04:00
|
|
|
# nothing
|
|
|
|
rescue Errno::ECONNABORTED
|
|
|
|
# client closed the socket even before accept
|
|
|
|
io.close rescue nil
|
2013-08-07 19:36:04 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
rescue Object => e
|
|
|
|
@events.unknown_error self, e, "Listen loop"
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2013-09-13 12:56:39 -04:00
|
|
|
@events.fire :state, @status
|
|
|
|
|
2013-08-07 19:36:04 -04:00
|
|
|
graceful_shutdown if @status == :stop || @status == :restart
|
|
|
|
|
|
|
|
rescue Exception => e
|
|
|
|
STDERR.puts "Exception handling servers: #{e.message} (#{e.class})"
|
|
|
|
STDERR.puts e.backtrace
|
|
|
|
ensure
|
|
|
|
@check.close
|
|
|
|
@notify.close
|
|
|
|
|
|
|
|
if @status != :restart and @own_binder
|
|
|
|
@binder.close
|
|
|
|
end
|
|
|
|
end
|
2013-09-13 12:56:39 -04:00
|
|
|
|
|
|
|
@events.fire :state, :done
|
2013-08-07 19:36:04 -04:00
|
|
|
end
|
2012-07-19 19:25:20 -04:00
|
|
|
# Runs the server.
|
2011-12-01 18:23:14 -05:00
|
|
|
#
|
2012-07-19 19:25:20 -04:00
|
|
|
# If +background+ is true (the default) then a thread is spun
|
|
|
|
# up in the background to handle requests. Otherwise requests
|
|
|
|
# are handled synchronously.
|
|
|
|
#
|
|
|
|
def run(background=true)
|
2011-09-24 03:19:22 -04:00
|
|
|
BasicSocket.do_not_reverse_lookup = true
|
|
|
|
|
2013-09-13 12:56:39 -04:00
|
|
|
@events.fire :state, :booting
|
|
|
|
|
2011-12-05 12:01:19 -05:00
|
|
|
@status = :run
|
2011-09-27 17:33:17 -04:00
|
|
|
|
2013-08-07 19:36:04 -04:00
|
|
|
if @mode == :tcp
|
|
|
|
return run_lopez_mode(background)
|
|
|
|
end
|
|
|
|
|
2015-01-20 13:29:31 -05:00
|
|
|
queue_requests = @queue_requests
|
|
|
|
|
2012-08-11 18:09:09 -04:00
|
|
|
@thread_pool = ThreadPool.new(@min_threads,
|
|
|
|
@max_threads,
|
|
|
|
IOBuffer) do |client, buffer|
|
2016-07-24 16:26:14 -04:00
|
|
|
|
|
|
|
# Advertise this server into the thread
|
|
|
|
Thread.current[ThreadLocalKey] = self
|
|
|
|
|
2012-07-23 20:00:06 -04:00
|
|
|
process_now = false
|
|
|
|
|
|
|
|
begin
|
2015-01-20 13:29:31 -05:00
|
|
|
if queue_requests
|
2015-01-20 07:20:39 -05:00
|
|
|
process_now = client.eagerly_finish
|
|
|
|
else
|
|
|
|
client.finish
|
|
|
|
process_now = true
|
|
|
|
end
|
2015-01-13 23:11:26 -05:00
|
|
|
rescue MiniSSL::SSLError => e
|
|
|
|
ssl_socket = client.io
|
|
|
|
addr = ssl_socket.peeraddr.last
|
|
|
|
cert = ssl_socket.peercert
|
|
|
|
|
|
|
|
client.close
|
|
|
|
|
|
|
|
@events.ssl_error self, addr, cert, e
|
2012-07-23 20:00:06 -04:00
|
|
|
rescue HttpParserError => e
|
2012-09-06 01:09:42 -04:00
|
|
|
client.write_400
|
2012-07-23 20:08:11 -04:00
|
|
|
client.close
|
2012-09-06 01:09:42 -04:00
|
|
|
|
2012-07-23 20:00:06 -04:00
|
|
|
@events.parse_error self, client.env, e
|
2013-06-01 17:20:45 -04:00
|
|
|
rescue ConnectionError
|
2012-07-23 20:00:06 -04:00
|
|
|
client.close
|
2012-07-23 17:29:33 -04:00
|
|
|
else
|
2012-07-23 20:00:06 -04:00
|
|
|
if process_now
|
2012-08-11 18:09:09 -04:00
|
|
|
process_client client, buffer
|
2012-07-23 20:00:06 -04:00
|
|
|
else
|
2012-08-09 19:54:55 -04:00
|
|
|
client.set_timeout @first_data_timeout
|
2012-07-23 20:00:06 -04:00
|
|
|
@reactor.add client
|
|
|
|
end
|
2012-07-23 17:29:33 -04:00
|
|
|
end
|
2011-09-27 17:33:17 -04:00
|
|
|
end
|
|
|
|
|
2014-11-26 02:18:30 -05:00
|
|
|
@thread_pool.clean_thread_locals = @options[:clean_thread_locals]
|
|
|
|
|
2015-01-20 13:29:31 -05:00
|
|
|
if queue_requests
|
2015-01-20 07:20:39 -05:00
|
|
|
@reactor = Reactor.new self, @thread_pool
|
|
|
|
@reactor.run_in_thread
|
|
|
|
end
|
2012-07-23 13:26:52 -04:00
|
|
|
|
2015-05-19 10:14:30 -04:00
|
|
|
if @reaping_time
|
|
|
|
@thread_pool.auto_reap!(@reaping_time)
|
|
|
|
end
|
|
|
|
|
2011-12-05 13:07:01 -05:00
|
|
|
if @auto_trim_time
|
|
|
|
@thread_pool.auto_trim!(@auto_trim_time)
|
|
|
|
end
|
|
|
|
|
2013-09-13 12:56:39 -04:00
|
|
|
@events.fire :state, :running
|
|
|
|
|
2012-07-19 19:25:20 -04:00
|
|
|
if background
|
|
|
|
@thread = Thread.new { handle_servers }
|
|
|
|
return @thread
|
|
|
|
else
|
|
|
|
handle_servers
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def handle_servers
|
|
|
|
begin
|
|
|
|
check = @check
|
2012-08-02 18:03:52 -04:00
|
|
|
sockets = [check] + @binder.ios
|
2012-07-19 19:25:20 -04:00
|
|
|
pool = @thread_pool
|
2015-01-20 13:29:31 -05:00
|
|
|
queue_requests = @queue_requests
|
2012-07-19 19:25:20 -04:00
|
|
|
|
2016-01-06 13:12:09 -05:00
|
|
|
remote_addr_value = nil
|
|
|
|
remote_addr_header = nil
|
|
|
|
|
|
|
|
case @options[:remote_address]
|
|
|
|
when :value
|
|
|
|
remote_addr_value = @options[:remote_address_value]
|
|
|
|
when :header
|
|
|
|
remote_addr_header = @options[:remote_address_header]
|
|
|
|
end
|
|
|
|
|
2012-07-19 19:25:20 -04:00
|
|
|
while @status == :run
|
|
|
|
begin
|
|
|
|
ios = IO.select sockets
|
|
|
|
ios.first.each do |sock|
|
|
|
|
if sock == check
|
|
|
|
break if handle_check
|
|
|
|
else
|
2012-08-01 13:11:27 -04:00
|
|
|
begin
|
|
|
|
if io = sock.accept_nonblock
|
2014-03-28 11:26:11 -04:00
|
|
|
client = Client.new io, @binder.env(sock)
|
2016-01-06 13:12:09 -05:00
|
|
|
if remote_addr_value
|
|
|
|
client.peerip = remote_addr_value
|
|
|
|
elsif remote_addr_header
|
|
|
|
client.remote_addr_header = remote_addr_header
|
|
|
|
end
|
|
|
|
|
2014-03-28 11:26:11 -04:00
|
|
|
pool << client
|
2015-01-20 13:29:31 -05:00
|
|
|
pool.wait_until_not_full unless queue_requests
|
2012-08-01 13:11:27 -04:00
|
|
|
end
|
2012-08-11 18:09:09 -04:00
|
|
|
rescue SystemCallError
|
2015-08-04 01:09:17 -04:00
|
|
|
# nothing
|
|
|
|
rescue Errno::ECONNABORTED
|
|
|
|
# client closed the socket even before accept
|
|
|
|
io.close rescue nil
|
2012-08-01 13:11:27 -04:00
|
|
|
end
|
2011-09-24 03:19:22 -04:00
|
|
|
end
|
|
|
|
end
|
2012-07-19 19:25:20 -04:00
|
|
|
rescue Object => e
|
|
|
|
@events.unknown_error self, e, "Listen loop"
|
2011-09-24 03:19:22 -04:00
|
|
|
end
|
2012-07-19 19:25:20 -04:00
|
|
|
end
|
2011-12-05 12:01:19 -05:00
|
|
|
|
2013-09-13 12:56:39 -04:00
|
|
|
@events.fire :state, @status
|
|
|
|
|
2012-09-09 18:16:39 -04:00
|
|
|
graceful_shutdown if @status == :stop || @status == :restart
|
2015-01-20 13:29:31 -05:00
|
|
|
if queue_requests
|
2016-07-25 20:20:17 -04:00
|
|
|
@reactor.clear!
|
2015-01-20 07:20:39 -05:00
|
|
|
@reactor.shutdown
|
|
|
|
end
|
2013-03-18 19:20:59 -04:00
|
|
|
rescue Exception => e
|
|
|
|
STDERR.puts "Exception handling servers: #{e.message} (#{e.class})"
|
|
|
|
STDERR.puts e.backtrace
|
2012-07-19 19:25:20 -04:00
|
|
|
ensure
|
2013-02-05 01:31:40 -05:00
|
|
|
@check.close
|
|
|
|
@notify.close
|
|
|
|
|
2013-02-09 13:12:13 -05:00
|
|
|
if @status != :restart and @own_binder
|
2012-08-02 18:03:52 -04:00
|
|
|
@binder.close
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
end
|
2013-09-13 12:56:39 -04:00
|
|
|
|
|
|
|
@events.fire :state, :done
|
2011-09-24 03:19:22 -04:00
|
|
|
end
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# :nodoc:
|
2011-09-24 03:19:22 -04:00
|
|
|
def handle_check
|
2014-02-17 12:07:17 -05:00
|
|
|
cmd = @check.read(1)
|
2011-09-24 03:19:22 -04:00
|
|
|
|
|
|
|
case cmd
|
|
|
|
when STOP_COMMAND
|
2011-12-05 12:01:19 -05:00
|
|
|
@status = :stop
|
|
|
|
return true
|
|
|
|
when HALT_COMMAND
|
|
|
|
@status = :halt
|
2011-09-24 03:19:22 -04:00
|
|
|
return true
|
2012-04-04 11:38:22 -04:00
|
|
|
when RESTART_COMMAND
|
|
|
|
@status = :restart
|
|
|
|
return true
|
2011-09-24 03:19:22 -04:00
|
|
|
end
|
|
|
|
|
|
|
|
return false
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# Given a connection on +client+, handle the incoming requests.
|
|
|
|
#
|
|
|
|
# This method support HTTP Keep-Alive so it may, depending on if the client
|
|
|
|
# indicates that it supports keep alive, wait for another request before
|
|
|
|
# returning.
|
|
|
|
#
|
2012-08-11 18:09:09 -04:00
|
|
|
def process_client(client, buffer)
|
2011-09-18 16:02:34 -04:00
|
|
|
begin
|
2016-07-24 23:58:16 -04:00
|
|
|
|
2016-07-25 01:10:13 -04:00
|
|
|
if client.env[HTTP_EXPECT] == CONTINUE
|
|
|
|
client.io << HTTP_11_100
|
2016-07-24 23:58:16 -04:00
|
|
|
end
|
|
|
|
|
2016-02-25 16:35:47 -05:00
|
|
|
clean_thread_locals = @options[:clean_thread_locals]
|
2012-07-23 13:26:52 -04:00
|
|
|
close_socket = true
|
2011-12-08 15:17:45 -05:00
|
|
|
|
2012-07-23 13:26:52 -04:00
|
|
|
while true
|
2012-08-11 18:09:09 -04:00
|
|
|
case handle_request(client, buffer)
|
2012-07-23 13:26:52 -04:00
|
|
|
when false
|
|
|
|
return
|
|
|
|
when :async
|
|
|
|
close_socket = false
|
|
|
|
return
|
|
|
|
when true
|
2015-01-20 09:46:33 -05:00
|
|
|
return unless @queue_requests
|
2012-08-11 18:09:09 -04:00
|
|
|
buffer.reset
|
|
|
|
|
2016-02-25 16:35:47 -05:00
|
|
|
ThreadPool.clean_thread_locals if clean_thread_locals
|
|
|
|
|
2012-09-09 22:51:36 -04:00
|
|
|
unless client.reset(@status == :run)
|
2012-07-23 13:26:52 -04:00
|
|
|
close_socket = false
|
|
|
|
client.set_timeout @persistent_timeout
|
|
|
|
@reactor.add client
|
|
|
|
return
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
2011-12-01 18:23:14 -05:00
|
|
|
|
|
|
|
# The client disconnected while we were reading data
|
2013-06-01 17:24:49 -04:00
|
|
|
rescue ConnectionError
|
2011-12-01 18:23:14 -05:00
|
|
|
# Swallow them. The ensure tries to close +client+ down
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2015-01-13 23:11:26 -05:00
|
|
|
# SSL handshake error
|
|
|
|
rescue MiniSSL::SSLError => e
|
2016-04-07 14:07:26 -04:00
|
|
|
lowlevel_error(e, client.env)
|
|
|
|
|
2015-01-13 23:11:26 -05:00
|
|
|
ssl_socket = client.io
|
|
|
|
addr = ssl_socket.peeraddr.last
|
|
|
|
cert = ssl_socket.peercert
|
|
|
|
|
|
|
|
close_socket = true
|
|
|
|
|
|
|
|
@events.ssl_error self, addr, cert, e
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# The client doesn't know HTTP well
|
2011-09-18 16:02:34 -04:00
|
|
|
rescue HttpParserError => e
|
2016-04-07 14:07:26 -04:00
|
|
|
lowlevel_error(e, client.env)
|
|
|
|
|
2012-09-06 01:09:42 -04:00
|
|
|
client.write_400
|
|
|
|
|
2012-07-23 13:26:52 -04:00
|
|
|
@events.parse_error self, client.env, e
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# Server error
|
2011-09-27 16:52:50 -04:00
|
|
|
rescue StandardError => e
|
2016-04-07 14:07:26 -04:00
|
|
|
lowlevel_error(e, client.env)
|
|
|
|
|
2012-09-06 01:09:42 -04:00
|
|
|
client.write_500
|
|
|
|
|
2012-04-10 18:06:49 -04:00
|
|
|
@events.unknown_error self, e, "Read"
|
2011-09-18 16:02:34 -04:00
|
|
|
|
|
|
|
ensure
|
2012-11-18 13:19:22 -05:00
|
|
|
buffer.reset
|
|
|
|
|
2011-09-18 16:02:34 -04:00
|
|
|
begin
|
2012-01-08 20:27:57 -05:00
|
|
|
client.close if close_socket
|
2011-09-27 16:52:50 -04:00
|
|
|
rescue IOError, SystemCallError
|
2011-09-18 16:02:34 -04:00
|
|
|
# Already closed
|
2011-09-27 16:52:50 -04:00
|
|
|
rescue StandardError => e
|
2012-04-10 18:06:49 -04:00
|
|
|
@events.unknown_error self, e, "Client"
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# Given a Hash +env+ for the request read from +client+, add
|
|
|
|
# and fixup keys to comply with Rack's env guidelines.
|
|
|
|
#
|
2011-09-27 14:29:20 -04:00
|
|
|
def normalize_env(env, client)
|
|
|
|
if host = env[HTTP_HOST]
|
2011-09-24 03:19:22 -04:00
|
|
|
if colon = host.index(":")
|
2011-09-27 14:29:20 -04:00
|
|
|
env[SERVER_NAME] = host[0, colon]
|
2012-03-29 18:26:03 -04:00
|
|
|
env[SERVER_PORT] = host[colon+1, host.bytesize]
|
2011-09-24 03:19:22 -04:00
|
|
|
else
|
2011-09-27 14:29:20 -04:00
|
|
|
env[SERVER_NAME] = host
|
2013-02-19 22:31:38 -05:00
|
|
|
env[SERVER_PORT] = default_server_port(env)
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
2012-04-28 12:52:48 -04:00
|
|
|
else
|
|
|
|
env[SERVER_NAME] = LOCALHOST
|
2013-02-19 22:31:38 -05:00
|
|
|
env[SERVER_PORT] = default_server_port(env)
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
|
2011-09-27 14:29:20 -04:00
|
|
|
unless env[REQUEST_PATH]
|
2011-09-24 03:19:22 -04:00
|
|
|
# it might be a dumbass full host request header
|
2011-09-27 14:29:20 -04:00
|
|
|
uri = URI.parse(env[REQUEST_URI])
|
|
|
|
env[REQUEST_PATH] = uri.path
|
2011-09-24 02:26:13 -04:00
|
|
|
|
2011-09-27 14:29:20 -04:00
|
|
|
raise "No REQUEST PATH" unless env[REQUEST_PATH]
|
2016-04-07 13:47:04 -04:00
|
|
|
|
|
|
|
env[QUERY_STRING] = uri.query
|
2011-09-24 02:26:13 -04:00
|
|
|
end
|
|
|
|
|
2011-10-14 15:15:35 -04:00
|
|
|
env[PATH_INFO] = env[REQUEST_PATH]
|
|
|
|
|
2011-09-24 03:19:22 -04:00
|
|
|
# From http://www.ietf.org/rfc/rfc3875 :
|
|
|
|
# "Script authors should be aware that the REMOTE_ADDR and
|
|
|
|
# REMOTE_HOST meta-variables (see sections 4.1.8 and 4.1.9)
|
|
|
|
# may not identify the ultimate source of the request.
|
|
|
|
# They identify the client for the immediate request to the
|
|
|
|
# server; that client may be a proxy, gateway, or other
|
|
|
|
# intermediary acting on behalf of the actual source client."
|
|
|
|
#
|
2013-07-09 02:08:42 -04:00
|
|
|
|
2015-01-20 13:34:14 -05:00
|
|
|
unless env.key?(REMOTE_ADDR)
|
2015-07-02 11:29:26 -04:00
|
|
|
begin
|
2016-01-06 13:12:09 -05:00
|
|
|
addr = client.peerip
|
2015-07-02 11:29:26 -04:00
|
|
|
rescue Errno::ENOTCONN
|
|
|
|
# Client disconnects can result in an inability to get the
|
|
|
|
# peeraddr from the socket; default to localhost.
|
|
|
|
addr = LOCALHOST_IP
|
|
|
|
end
|
2013-07-09 02:08:42 -04:00
|
|
|
|
2015-01-20 13:34:14 -05:00
|
|
|
# Set unix socket addrs to localhost
|
2015-07-02 11:29:26 -04:00
|
|
|
addr = LOCALHOST_IP if addr.empty?
|
2013-07-09 02:08:42 -04:00
|
|
|
|
2015-01-20 13:34:14 -05:00
|
|
|
env[REMOTE_ADDR] = addr
|
|
|
|
end
|
2011-09-24 02:26:13 -04:00
|
|
|
end
|
|
|
|
|
2013-02-19 22:31:38 -05:00
|
|
|
def default_server_port(env)
|
2015-01-13 23:11:26 -05:00
|
|
|
return PORT_443 if env[HTTPS_KEY] == 'on' || env[HTTPS_KEY] == 'https'
|
2013-02-19 22:23:02 -05:00
|
|
|
env['HTTP_X_FORWARDED_PROTO'] == 'https' ? PORT_443 : PORT_80
|
|
|
|
end
|
|
|
|
|
2016-02-25 22:41:26 -05:00
|
|
|
# Given the request +env+ from +client+ and a partial request body
|
|
|
|
# in +body+, finish reading the body if there is one and invoke
|
|
|
|
# the rack app. Then construct the response and write it back to
|
|
|
|
# +client+
|
|
|
|
#
|
|
|
|
# +cl+ is the previously fetched Content-Length header if there
|
|
|
|
# was one. This is an optimization to keep from having to look
|
|
|
|
# it up again.
|
|
|
|
#
|
|
|
|
def handle_request(req, lines)
|
|
|
|
env = req.env
|
2012-07-23 13:26:52 -04:00
|
|
|
client = req.io
|
|
|
|
|
2016-02-25 22:41:26 -05:00
|
|
|
normalize_env env, req
|
|
|
|
|
|
|
|
env[PUMA_SOCKET] = client
|
|
|
|
|
|
|
|
if env[HTTPS_KEY] && client.peercert
|
|
|
|
env[PUMA_PEERCERT] = client.peercert
|
|
|
|
end
|
|
|
|
|
|
|
|
env[HIJACK_P] = true
|
|
|
|
env[HIJACK] = req
|
|
|
|
|
2012-07-23 13:26:52 -04:00
|
|
|
body = req.body
|
2011-09-24 02:26:13 -04:00
|
|
|
|
2013-06-18 17:27:46 -04:00
|
|
|
head = env[REQUEST_METHOD] == HEAD
|
|
|
|
|
2016-02-25 22:41:26 -05:00
|
|
|
env[RACK_INPUT] = body
|
|
|
|
env[RACK_URL_SCHEME] = env[HTTPS_KEY] ? HTTPS : HTTP
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# A rack extension. If the app writes #call'ables to this
|
|
|
|
# array, we will invoke them when the request is done.
|
|
|
|
#
|
2016-02-25 22:41:26 -05:00
|
|
|
after_reply = env[RACK_AFTER_REPLY] = []
|
2011-10-21 00:44:34 -04:00
|
|
|
|
2011-09-27 14:29:20 -04:00
|
|
|
begin
|
2016-02-25 22:41:26 -05:00
|
|
|
begin
|
|
|
|
status, headers, res_body = @app.call(env)
|
|
|
|
|
|
|
|
return :async if req.hijacked
|
|
|
|
|
|
|
|
status = status.to_i
|
|
|
|
|
|
|
|
if status == -1
|
|
|
|
unless headers.empty? and res_body == []
|
|
|
|
raise "async response must have empty headers and body"
|
|
|
|
end
|
|
|
|
|
|
|
|
return :async
|
|
|
|
end
|
2016-04-07 13:25:10 -04:00
|
|
|
rescue ThreadPool::ForceShutdown => e
|
|
|
|
@events.log "Detected force shutdown of a thread, returning 503"
|
|
|
|
@events.unknown_error self, e, "Rack app"
|
|
|
|
|
|
|
|
status = 503
|
|
|
|
headers = {}
|
|
|
|
res_body = ["Request was internally terminated early\n"]
|
|
|
|
|
2016-07-24 16:16:29 -04:00
|
|
|
rescue Exception => e
|
2016-04-12 13:42:22 -04:00
|
|
|
@events.unknown_error self, e, "Rack app", env
|
2016-02-25 22:41:26 -05:00
|
|
|
|
|
|
|
status, headers, res_body = lowlevel_error(e, env)
|
|
|
|
end
|
|
|
|
|
2011-09-30 11:30:37 -04:00
|
|
|
content_length = nil
|
2013-06-18 17:27:46 -04:00
|
|
|
no_body = head
|
2011-09-30 11:30:37 -04:00
|
|
|
|
2011-10-18 02:12:56 -04:00
|
|
|
if res_body.kind_of? Array and res_body.size == 1
|
2012-03-29 18:26:03 -04:00
|
|
|
content_length = res_body[0].bytesize
|
2011-09-30 11:30:37 -04:00
|
|
|
end
|
|
|
|
|
2011-12-01 14:40:30 -05:00
|
|
|
cork_socket client
|
|
|
|
|
2012-08-11 18:09:09 -04:00
|
|
|
line_ending = LINE_END
|
|
|
|
colon = COLON
|
|
|
|
|
2015-02-19 08:43:17 -05:00
|
|
|
http_11 = if env[HTTP_VERSION] == HTTP_11
|
2011-12-01 17:16:18 -05:00
|
|
|
allow_chunked = true
|
2016-01-06 13:11:37 -05:00
|
|
|
keep_alive = env.fetch(HTTP_CONNECTION, "").downcase != CLOSE
|
2011-12-01 17:16:18 -05:00
|
|
|
include_keepalive_header = false
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# An optimization. The most common response is 200, so we can
|
|
|
|
# reply with the proper 200 status without having to compute
|
|
|
|
# the response header.
|
|
|
|
#
|
2011-12-01 17:16:18 -05:00
|
|
|
if status == 200
|
2012-08-11 18:09:09 -04:00
|
|
|
lines << HTTP_11_200
|
2011-12-01 17:16:18 -05:00
|
|
|
else
|
2012-08-11 18:09:09 -04:00
|
|
|
lines.append "HTTP/1.1 ", status.to_s, " ",
|
2013-07-05 17:40:34 -04:00
|
|
|
fetch_status_code(status), line_ending
|
2011-12-21 12:28:55 -05:00
|
|
|
|
2013-06-18 17:27:46 -04:00
|
|
|
no_body ||= status < 200 || STATUS_WITH_NO_ENTITY_BODY[status]
|
2011-12-01 17:16:18 -05:00
|
|
|
end
|
2015-02-19 08:43:17 -05:00
|
|
|
true
|
2011-12-01 17:16:18 -05:00
|
|
|
else
|
|
|
|
allow_chunked = false
|
2016-01-06 13:11:37 -05:00
|
|
|
keep_alive = env.fetch(HTTP_CONNECTION, "").downcase == KEEP_ALIVE
|
2011-12-01 17:16:18 -05:00
|
|
|
include_keepalive_header = keep_alive
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# Same optimization as above for HTTP/1.1
|
|
|
|
#
|
2011-12-01 17:16:18 -05:00
|
|
|
if status == 200
|
2012-08-11 18:09:09 -04:00
|
|
|
lines << HTTP_10_200
|
2011-12-01 17:16:18 -05:00
|
|
|
else
|
2012-08-11 18:09:09 -04:00
|
|
|
lines.append "HTTP/1.0 ", status.to_s, " ",
|
2013-07-05 17:40:34 -04:00
|
|
|
fetch_status_code(status), line_ending
|
2011-12-21 12:28:55 -05:00
|
|
|
|
2013-06-18 17:27:46 -04:00
|
|
|
no_body ||= status < 200 || STATUS_WITH_NO_ENTITY_BODY[status]
|
2011-12-01 17:16:18 -05:00
|
|
|
end
|
2015-02-19 08:43:17 -05:00
|
|
|
false
|
2011-12-01 17:16:18 -05:00
|
|
|
end
|
2011-09-30 11:30:37 -04:00
|
|
|
|
2013-02-06 01:39:16 -05:00
|
|
|
response_hijack = nil
|
|
|
|
|
2011-09-24 02:26:13 -04:00
|
|
|
headers.each do |k, vs|
|
2016-01-06 13:11:37 -05:00
|
|
|
case k.downcase
|
2011-12-01 17:16:18 -05:00
|
|
|
when CONTENT_LENGTH2
|
2011-10-17 16:56:45 -04:00
|
|
|
content_length = vs
|
|
|
|
next
|
2011-12-01 17:16:18 -05:00
|
|
|
when TRANSFER_ENCODING
|
2011-10-23 02:06:27 -04:00
|
|
|
allow_chunked = false
|
|
|
|
content_length = nil
|
2013-02-06 01:39:16 -05:00
|
|
|
when HIJACK
|
|
|
|
response_hijack = vs
|
|
|
|
next
|
2011-10-17 16:56:45 -04:00
|
|
|
end
|
|
|
|
|
2013-08-08 12:21:55 -04:00
|
|
|
if vs.respond_to?(:to_s)
|
|
|
|
vs.to_s.split(NEWLINE).each do |v|
|
|
|
|
lines.append k, colon, v, line_ending
|
|
|
|
end
|
|
|
|
else
|
|
|
|
lines.append k, colon, line_ending
|
2011-09-24 02:26:13 -04:00
|
|
|
end
|
|
|
|
end
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2015-02-19 08:43:17 -05:00
|
|
|
if include_keepalive_header
|
|
|
|
lines << CONNECTION_KEEP_ALIVE
|
|
|
|
elsif http_11 && !keep_alive
|
|
|
|
lines << CONNECTION_CLOSE
|
|
|
|
end
|
|
|
|
|
2011-12-13 08:45:12 -05:00
|
|
|
if no_body
|
2013-07-06 00:35:26 -04:00
|
|
|
if content_length and status != 204
|
|
|
|
lines.append CONTENT_LENGTH_S, content_length.to_s, line_ending
|
|
|
|
end
|
|
|
|
|
2012-08-11 18:09:09 -04:00
|
|
|
lines << line_ending
|
2012-11-29 14:32:50 -05:00
|
|
|
fast_write client, lines.to_s
|
2011-12-13 08:45:12 -05:00
|
|
|
return keep_alive
|
|
|
|
end
|
2016-01-06 13:11:37 -05:00
|
|
|
|
2015-06-08 12:39:49 -04:00
|
|
|
if content_length
|
|
|
|
lines.append CONTENT_LENGTH_S, content_length.to_s, line_ending
|
|
|
|
chunked = false
|
2015-06-10 14:21:56 -04:00
|
|
|
elsif !response_hijack and allow_chunked
|
2015-06-08 12:39:49 -04:00
|
|
|
lines << TRANSFER_ENCODING_CHUNKED
|
|
|
|
chunked = true
|
2011-10-17 16:56:45 -04:00
|
|
|
end
|
|
|
|
|
2012-08-11 18:09:09 -04:00
|
|
|
lines << line_ending
|
|
|
|
|
2012-11-29 14:32:50 -05:00
|
|
|
fast_write client, lines.to_s
|
2011-09-18 16:02:34 -04:00
|
|
|
|
2013-02-06 01:39:16 -05:00
|
|
|
if response_hijack
|
|
|
|
response_hijack.call client
|
|
|
|
return :async
|
|
|
|
end
|
|
|
|
|
2013-07-06 00:52:15 -04:00
|
|
|
begin
|
|
|
|
res_body.each do |part|
|
|
|
|
if chunked
|
2015-07-17 19:31:33 -04:00
|
|
|
next if part.bytesize.zero?
|
2014-07-12 20:49:44 -04:00
|
|
|
fast_write client, part.bytesize.to_s(16)
|
|
|
|
fast_write client, line_ending
|
2013-07-06 00:52:15 -04:00
|
|
|
fast_write client, part
|
2014-07-12 20:49:44 -04:00
|
|
|
fast_write client, line_ending
|
2013-07-06 00:52:15 -04:00
|
|
|
else
|
|
|
|
fast_write client, part
|
|
|
|
end
|
2011-09-30 11:30:37 -04:00
|
|
|
|
2013-07-06 00:52:15 -04:00
|
|
|
client.flush
|
|
|
|
end
|
2011-09-30 11:30:37 -04:00
|
|
|
|
2013-07-06 00:52:15 -04:00
|
|
|
if chunked
|
2014-07-12 20:49:44 -04:00
|
|
|
fast_write client, CLOSE_CHUNKED
|
2013-07-06 00:52:15 -04:00
|
|
|
client.flush
|
|
|
|
end
|
|
|
|
rescue SystemCallError, IOError
|
|
|
|
raise ConnectionError, "Connection error detected during write"
|
2011-09-30 11:30:37 -04:00
|
|
|
end
|
|
|
|
|
2011-09-18 16:02:34 -04:00
|
|
|
ensure
|
2011-12-01 14:40:30 -05:00
|
|
|
uncork_socket client
|
|
|
|
|
2011-09-24 02:26:13 -04:00
|
|
|
body.close
|
2015-04-21 11:48:13 -04:00
|
|
|
req.tempfile.unlink if req.tempfile
|
2011-09-24 02:26:13 -04:00
|
|
|
res_body.close if res_body.respond_to? :close
|
2011-10-21 00:44:34 -04:00
|
|
|
|
|
|
|
after_reply.each { |o| o.call }
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
2011-09-30 11:30:37 -04:00
|
|
|
|
|
|
|
return keep_alive
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
|
2013-07-05 17:40:34 -04:00
|
|
|
def fetch_status_code(status)
|
|
|
|
HTTP_STATUS_CODES.fetch(status) { 'CUSTOM' }
|
|
|
|
end
|
|
|
|
private :fetch_status_code
|
|
|
|
|
2016-04-07 14:22:15 -04:00
|
|
|
# Given the request +env+ from +client+ and the partial body +body+
|
2011-12-01 18:23:14 -05:00
|
|
|
# plus a potential Content-Length value +cl+, finish reading
|
|
|
|
# the body and return it.
|
|
|
|
#
|
|
|
|
# If the body is larger than MAX_BODY, a Tempfile object is used
|
|
|
|
# for the body, otherwise a StringIO is used.
|
|
|
|
#
|
2011-11-22 13:45:58 -05:00
|
|
|
def read_body(env, client, body, cl)
|
|
|
|
content_length = cl.to_i
|
2011-09-24 03:19:22 -04:00
|
|
|
|
2012-03-29 18:26:03 -04:00
|
|
|
remain = content_length - body.bytesize
|
2011-09-24 03:19:22 -04:00
|
|
|
|
|
|
|
return StringIO.new(body) if remain <= 0
|
|
|
|
|
|
|
|
# Use a Tempfile if there is a lot of data left
|
|
|
|
if remain > MAX_BODY
|
|
|
|
stream = Tempfile.new(Const::PUMA_TMP_BASE)
|
|
|
|
stream.binmode
|
|
|
|
else
|
2012-05-15 18:19:28 -04:00
|
|
|
# The body[0,0] trick is to get an empty string in the same
|
|
|
|
# encoding as body.
|
|
|
|
stream = StringIO.new body[0,0]
|
2011-09-24 03:19:22 -04:00
|
|
|
end
|
|
|
|
|
2012-05-15 18:19:28 -04:00
|
|
|
stream.write body
|
|
|
|
|
2011-09-24 03:19:22 -04:00
|
|
|
# Read an odd sized chunk so we can read even sized ones
|
|
|
|
# after this
|
|
|
|
chunk = client.readpartial(remain % CHUNK_SIZE)
|
|
|
|
|
|
|
|
# No chunk means a closed socket
|
|
|
|
unless chunk
|
|
|
|
stream.close
|
|
|
|
return nil
|
|
|
|
end
|
|
|
|
|
|
|
|
remain -= stream.write(chunk)
|
|
|
|
|
|
|
|
# Raed the rest of the chunks
|
|
|
|
while remain > 0
|
|
|
|
chunk = client.readpartial(CHUNK_SIZE)
|
|
|
|
unless chunk
|
|
|
|
stream.close
|
|
|
|
return nil
|
|
|
|
end
|
|
|
|
|
|
|
|
remain -= stream.write(chunk)
|
|
|
|
end
|
|
|
|
|
|
|
|
stream.rewind
|
|
|
|
|
|
|
|
return stream
|
|
|
|
end
|
|
|
|
|
2011-12-01 18:23:14 -05:00
|
|
|
# A fallback rack response if +@app+ raises as exception.
|
|
|
|
#
|
2016-02-25 15:58:37 -05:00
|
|
|
def lowlevel_error(e, env)
|
2014-02-17 12:07:17 -05:00
|
|
|
if handler = @options[:lowlevel_error_handler]
|
2016-02-25 16:33:32 -05:00
|
|
|
if handler.arity == 1
|
|
|
|
return handler.call(e)
|
|
|
|
else
|
|
|
|
return handler.call(e, env)
|
|
|
|
end
|
2014-02-17 12:07:17 -05:00
|
|
|
end
|
|
|
|
|
2013-07-01 19:48:17 -04:00
|
|
|
if @leak_stack_on_error
|
|
|
|
[500, {}, ["Puma caught this error: #{e.message} (#{e.class})\n#{e.backtrace.join("\n")}"]]
|
|
|
|
else
|
2015-07-15 10:59:34 -04:00
|
|
|
[500, {}, ["An unhandled lowlevel error occurred. The application logs may have details.\n"]]
|
2013-07-01 19:48:17 -04:00
|
|
|
end
|
2011-09-24 03:26:17 -04:00
|
|
|
end
|
|
|
|
|
2011-09-24 03:19:22 -04:00
|
|
|
# Wait for all outstanding requests to finish.
|
2011-12-01 18:23:14 -05:00
|
|
|
#
|
2011-09-24 03:19:22 -04:00
|
|
|
def graceful_shutdown
|
2015-04-11 01:52:38 -04:00
|
|
|
if @options[:shutdown_debug]
|
|
|
|
threads = Thread.list
|
|
|
|
total = threads.size
|
|
|
|
|
2015-04-11 02:01:06 -04:00
|
|
|
pid = Process.pid
|
|
|
|
|
|
|
|
$stdout.syswrite "#{pid}: === Begin thread backtrace dump ===\n"
|
2015-04-11 01:52:38 -04:00
|
|
|
|
|
|
|
threads.each_with_index do |t,i|
|
2015-04-11 02:01:06 -04:00
|
|
|
$stdout.syswrite "#{pid}: Thread #{i+1}/#{total}: #{t.inspect}\n"
|
|
|
|
$stdout.syswrite "#{pid}: #{t.backtrace.join("\n#{pid}: ")}\n\n"
|
2015-04-11 01:52:38 -04:00
|
|
|
end
|
2015-04-11 02:01:06 -04:00
|
|
|
$stdout.syswrite "#{pid}: === End thread backtrace dump ===\n"
|
2015-04-11 01:52:38 -04:00
|
|
|
end
|
|
|
|
|
2013-07-15 17:29:10 -04:00
|
|
|
if @options[:drain_on_shutdown]
|
|
|
|
count = 0
|
|
|
|
|
|
|
|
while true
|
|
|
|
ios = IO.select @binder.ios, nil, nil, 0
|
|
|
|
break unless ios
|
|
|
|
|
|
|
|
ios.first.each do |sock|
|
|
|
|
begin
|
|
|
|
if io = sock.accept_nonblock
|
|
|
|
count += 1
|
2014-03-28 11:26:11 -04:00
|
|
|
client = Client.new io, @binder.env(sock)
|
|
|
|
@thread_pool << client
|
2013-07-15 17:29:10 -04:00
|
|
|
end
|
|
|
|
rescue SystemCallError
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
@events.debug "Drained #{count} additional connections."
|
|
|
|
end
|
|
|
|
|
2016-04-07 13:25:10 -04:00
|
|
|
if @thread_pool
|
|
|
|
if timeout = @options[:force_shutdown_after]
|
|
|
|
@thread_pool.shutdown timeout.to_i
|
|
|
|
else
|
|
|
|
@thread_pool.shutdown
|
|
|
|
end
|
|
|
|
end
|
2011-09-24 03:19:22 -04:00
|
|
|
end
|
2011-09-23 01:14:39 -04:00
|
|
|
|
|
|
|
# Stops the acceptor thread and then causes the worker threads to finish
|
|
|
|
# off the request queue before finally exiting.
|
2011-12-01 18:23:14 -05:00
|
|
|
#
|
2011-09-23 01:14:39 -04:00
|
|
|
def stop(sync=false)
|
2013-05-31 13:26:28 -04:00
|
|
|
begin
|
|
|
|
@notify << STOP_COMMAND
|
|
|
|
rescue IOError
|
|
|
|
# The server, in another thread, is shutting down
|
|
|
|
end
|
2011-09-23 01:14:39 -04:00
|
|
|
|
2011-09-27 17:33:17 -04:00
|
|
|
@thread.join if @thread && sync
|
2011-09-23 01:14:39 -04:00
|
|
|
end
|
2011-12-05 12:01:19 -05:00
|
|
|
|
|
|
|
def halt(sync=false)
|
2013-05-31 13:26:28 -04:00
|
|
|
begin
|
|
|
|
@notify << HALT_COMMAND
|
|
|
|
rescue IOError
|
|
|
|
# The server, in another thread, is shutting down
|
|
|
|
end
|
2011-12-05 12:01:19 -05:00
|
|
|
|
|
|
|
@thread.join if @thread && sync
|
|
|
|
end
|
2012-04-04 11:38:22 -04:00
|
|
|
|
|
|
|
def begin_restart
|
2013-05-31 13:26:28 -04:00
|
|
|
begin
|
|
|
|
@notify << RESTART_COMMAND
|
|
|
|
rescue IOError
|
|
|
|
# The server, in another thread, is shutting down
|
|
|
|
end
|
2012-04-04 11:38:22 -04:00
|
|
|
end
|
2012-11-29 14:32:50 -05:00
|
|
|
|
|
|
|
def fast_write(io, str)
|
2013-03-12 18:51:48 -04:00
|
|
|
n = 0
|
2013-03-18 19:44:38 -04:00
|
|
|
while true
|
2013-03-12 18:51:48 -04:00
|
|
|
begin
|
|
|
|
n = io.syswrite str
|
|
|
|
rescue Errno::EAGAIN, Errno::EWOULDBLOCK
|
2014-02-04 11:46:13 -05:00
|
|
|
if !IO.select(nil, [io], nil, WRITE_TIMEOUT)
|
|
|
|
raise ConnectionError, "Socket timeout writing data"
|
|
|
|
end
|
|
|
|
|
2013-03-12 18:51:48 -04:00
|
|
|
retry
|
2013-09-11 12:52:31 -04:00
|
|
|
rescue Errno::EPIPE, SystemCallError, IOError
|
2014-02-04 11:46:13 -05:00
|
|
|
raise ConnectionError, "Socket timeout writing data"
|
2013-03-12 18:51:48 -04:00
|
|
|
end
|
2012-11-29 15:39:42 -05:00
|
|
|
|
2013-03-12 18:51:48 -04:00
|
|
|
return if n == str.bytesize
|
|
|
|
str = str.byteslice(n..-1)
|
2012-11-29 15:39:42 -05:00
|
|
|
end
|
2012-11-29 14:32:50 -05:00
|
|
|
end
|
|
|
|
private :fast_write
|
2016-07-24 16:26:14 -04:00
|
|
|
|
|
|
|
ThreadLocalKey = :puma_server
|
|
|
|
|
|
|
|
def self.current
|
|
|
|
Thread.current[ThreadLocalKey]
|
|
|
|
end
|
|
|
|
|
|
|
|
def shutting_down?
|
|
|
|
@status == :stop || @status == :restart
|
|
|
|
end
|
2011-09-18 16:02:34 -04:00
|
|
|
end
|
|
|
|
end
|