1
0
Fork 0
mirror of https://github.com/puma/puma.git synced 2022-11-09 13:48:40 -05:00

Bump to 2.0.1

This commit is contained in:
Evan Phoenix 2013-04-30 11:17:49 -07:00
parent 31e3584abf
commit ba83a60807
3 changed files with 9 additions and 3 deletions

View file

@ -1,3 +1,9 @@
=== 2.0.1 / 2013-04-30
* 1 bug fix:
* Fix not starting on JRuby properly
=== 2.0.0 / 2013-04-29
RailsConf 2013 edition!

View file

@ -28,7 +28,7 @@ module Puma
# too taxing on performance.
module Const
PUMA_VERSION = VERSION = "2.0.0".freeze
PUMA_VERSION = VERSION = "2.0.1".freeze
FAST_TRACK_KA_TIMEOUT = 0.2

View file

@ -2,11 +2,11 @@
Gem::Specification.new do |s|
s.name = "puma"
s.version = "2.0.0"
s.version = "2.0.1"
s.required_rubygems_version = Gem::Requirement.new(">= 0") if s.respond_to? :required_rubygems_version=
s.authors = ["Evan Phoenix"]
s.date = "2013-04-29"
s.date = "2013-04-30"
s.description = "Puma is a simple, fast, and highly concurrent HTTP 1.1 server for Ruby web applications. It can be used with any application that supports Rack, and is considered the replacement for Webrick and Mongrel. It was designed to be the go-to server for [Rubinius](http://rubini.us), but also works well with JRuby and MRI. Puma is intended for use in both development and production environments.\n\nUnder the hood, Puma processes requests using a C-optimized Ragel extension (inherited from Mongrel) that provides fast, accurate HTTP 1.1 protocol parsing in a portable way. Puma then serves the request in a thread from an internal thread pool (which you can control). This allows Puma to provide real concurrency for your web application!\n\nWith Rubinius 2.0, Puma will utilize all cores on your CPU with real threads, meaning you won't have to spawn multiple processes to increase throughput. You can expect to see a similar benefit from JRuby.\n\nOn MRI, there is a Global Interpreter Lock (GIL) that ensures only one thread can be run at a time. But if you're doing a lot of blocking IO (such as HTTP calls to external APIs like Twitter), Puma still improves MRI's throughput by allowing blocking IO to be run concurrently (EventMachine-based servers such as Thin turn off this ability, requiring you to use special libraries). Your mileage may vary. In order to get the best throughput, it is highly recommended that you use a Ruby implementation with real threads like [Rubinius](http://rubini.us) or [JRuby](http://jruby.org)."
s.email = ["evan@phx.io"]
s.executables = ["puma", "pumactl"]