mirror of
https://github.com/ruby/ruby.git
synced 2022-11-09 12:17:21 -05:00
44264b4fee
Pick from dfbb5a3811
333 lines
9.2 KiB
Ruby
333 lines
9.2 KiB
Ruby
# frozen_string_literal: true
|
|
require_relative "../rubygems"
|
|
require_relative "request"
|
|
require_relative "request/connection_pools"
|
|
require_relative "s3_uri_signer"
|
|
require_relative "uri_formatter"
|
|
require_relative "uri"
|
|
require_relative "user_interaction"
|
|
|
|
##
|
|
# RemoteFetcher handles the details of fetching gems and gem information from
|
|
# a remote source.
|
|
|
|
class Gem::RemoteFetcher
|
|
include Gem::UserInteraction
|
|
|
|
##
|
|
# A FetchError exception wraps up the various possible IO and HTTP failures
|
|
# that could happen while downloading from the internet.
|
|
|
|
class FetchError < Gem::Exception
|
|
##
|
|
# The URI which was being accessed when the exception happened.
|
|
|
|
attr_accessor :uri, :original_uri
|
|
|
|
def initialize(message, uri)
|
|
uri = Gem::Uri.new(uri)
|
|
|
|
super uri.redact_credentials_from(message)
|
|
|
|
@original_uri = uri.to_s
|
|
@uri = uri.redacted.to_s
|
|
end
|
|
|
|
def to_s # :nodoc:
|
|
"#{super} (#{uri})"
|
|
end
|
|
end
|
|
|
|
##
|
|
# A FetchError that indicates that the reason for not being
|
|
# able to fetch data was that the host could not be contacted
|
|
|
|
class UnknownHostError < FetchError
|
|
end
|
|
deprecate_constant(:UnknownHostError)
|
|
|
|
@fetcher = nil
|
|
|
|
##
|
|
# Cached RemoteFetcher instance.
|
|
|
|
def self.fetcher
|
|
@fetcher ||= self.new Gem.configuration[:http_proxy]
|
|
end
|
|
|
|
attr_accessor :headers
|
|
|
|
##
|
|
# Initialize a remote fetcher using the source URI and possible proxy
|
|
# information.
|
|
#
|
|
# +proxy+
|
|
# * [String]: explicit specification of proxy; overrides any environment
|
|
# variable setting
|
|
# * nil: respect environment variables (HTTP_PROXY, HTTP_PROXY_USER,
|
|
# HTTP_PROXY_PASS)
|
|
# * <tt>:no_proxy</tt>: ignore environment variables and _don't_ use a proxy
|
|
#
|
|
# +headers+: A set of additional HTTP headers to be sent to the server when
|
|
# fetching the gem.
|
|
|
|
def initialize(proxy=nil, dns=nil, headers={})
|
|
require_relative "core_ext/tcpsocket_init" if Gem.configuration.ipv4_fallback_enabled
|
|
require "net/http"
|
|
require "stringio"
|
|
require "uri"
|
|
|
|
Socket.do_not_reverse_lookup = true
|
|
|
|
@proxy = proxy
|
|
@pools = {}
|
|
@pool_lock = Thread::Mutex.new
|
|
@cert_files = Gem::Request.get_cert_files
|
|
|
|
@headers = headers
|
|
end
|
|
|
|
##
|
|
# Given a name and requirement, downloads this gem into cache and returns the
|
|
# filename. Returns nil if the gem cannot be located.
|
|
#--
|
|
# Should probably be integrated with #download below, but that will be a
|
|
# larger, more encompassing effort. -erikh
|
|
|
|
def download_to_cache(dependency)
|
|
found, _ = Gem::SpecFetcher.fetcher.spec_for_dependency dependency
|
|
|
|
return if found.empty?
|
|
|
|
spec, source = found.max_by {|(s,_)| s.version }
|
|
|
|
download spec, source.uri
|
|
end
|
|
|
|
##
|
|
# Moves the gem +spec+ from +source_uri+ to the cache dir unless it is
|
|
# already there. If the source_uri is local the gem cache dir copy is
|
|
# always replaced.
|
|
|
|
def download(spec, source_uri, install_dir = Gem.dir)
|
|
install_cache_dir = File.join install_dir, "cache"
|
|
cache_dir =
|
|
if Dir.pwd == install_dir # see fetch_command
|
|
install_dir
|
|
elsif File.writable?(install_cache_dir) || (File.writable?(install_dir) && (!File.exist?(install_cache_dir)))
|
|
install_cache_dir
|
|
else
|
|
File.join Gem.user_dir, "cache"
|
|
end
|
|
|
|
gem_file_name = File.basename spec.cache_file
|
|
local_gem_path = File.join cache_dir, gem_file_name
|
|
|
|
require "fileutils"
|
|
FileUtils.mkdir_p cache_dir rescue nil unless File.exist? cache_dir
|
|
|
|
source_uri = Gem::Uri.new(source_uri)
|
|
|
|
scheme = source_uri.scheme
|
|
|
|
# URI.parse gets confused by MS Windows paths with forward slashes.
|
|
scheme = nil if scheme =~ /^[a-z]$/i
|
|
|
|
# REFACTOR: split this up and dispatch on scheme (eg download_http)
|
|
# REFACTOR: be sure to clean up fake fetcher when you do this... cleaner
|
|
case scheme
|
|
when "http", "https", "s3" then
|
|
unless File.exist? local_gem_path
|
|
begin
|
|
verbose "Downloading gem #{gem_file_name}"
|
|
|
|
remote_gem_path = source_uri + "gems/#{gem_file_name}"
|
|
|
|
self.cache_update_path remote_gem_path, local_gem_path
|
|
rescue FetchError
|
|
raise if spec.original_platform == spec.platform
|
|
|
|
alternate_name = "#{spec.original_name}.gem"
|
|
|
|
verbose "Failed, downloading gem #{alternate_name}"
|
|
|
|
remote_gem_path = source_uri + "gems/#{alternate_name}"
|
|
|
|
self.cache_update_path remote_gem_path, local_gem_path
|
|
end
|
|
end
|
|
when "file" then
|
|
begin
|
|
path = source_uri.path
|
|
path = File.dirname(path) if File.extname(path) == ".gem"
|
|
|
|
remote_gem_path = Gem::Util.correct_for_windows_path(File.join(path, "gems", gem_file_name))
|
|
|
|
FileUtils.cp(remote_gem_path, local_gem_path)
|
|
rescue Errno::EACCES
|
|
local_gem_path = source_uri.to_s
|
|
end
|
|
|
|
verbose "Using local gem #{local_gem_path}"
|
|
when nil then # TODO test for local overriding cache
|
|
source_path = if Gem.win_platform? && source_uri.scheme &&
|
|
!source_uri.path.include?(":")
|
|
"#{source_uri.scheme}:#{source_uri.path}"
|
|
else
|
|
source_uri.path
|
|
end
|
|
|
|
source_path = Gem::UriFormatter.new(source_path).unescape
|
|
|
|
begin
|
|
FileUtils.cp source_path, local_gem_path unless
|
|
File.identical?(source_path, local_gem_path)
|
|
rescue Errno::EACCES
|
|
local_gem_path = source_uri.to_s
|
|
end
|
|
|
|
verbose "Using local gem #{local_gem_path}"
|
|
else
|
|
raise ArgumentError, "unsupported URI scheme #{source_uri.scheme}"
|
|
end
|
|
|
|
local_gem_path
|
|
end
|
|
|
|
##
|
|
# File Fetcher. Dispatched by +fetch_path+. Use it instead.
|
|
|
|
def fetch_file(uri, *_)
|
|
Gem.read_binary Gem::Util.correct_for_windows_path uri.path
|
|
end
|
|
|
|
##
|
|
# HTTP Fetcher. Dispatched by +fetch_path+. Use it instead.
|
|
|
|
def fetch_http(uri, last_modified = nil, head = false, depth = 0)
|
|
fetch_type = head ? Net::HTTP::Head : Net::HTTP::Get
|
|
response = request uri, fetch_type, last_modified do |req|
|
|
headers.each {|k,v| req.add_field(k,v) }
|
|
end
|
|
|
|
case response
|
|
when Net::HTTPOK, Net::HTTPNotModified then
|
|
response.uri = uri
|
|
head ? response : response.body
|
|
when Net::HTTPMovedPermanently, Net::HTTPFound, Net::HTTPSeeOther,
|
|
Net::HTTPTemporaryRedirect then
|
|
raise FetchError.new("too many redirects", uri) if depth > 10
|
|
|
|
unless location = response["Location"]
|
|
raise FetchError.new("redirecting but no redirect location was given", uri)
|
|
end
|
|
location = Gem::Uri.new location
|
|
|
|
if https?(uri) && !https?(location)
|
|
raise FetchError.new("redirecting to non-https resource: #{location}", uri)
|
|
end
|
|
|
|
fetch_http(location, last_modified, head, depth + 1)
|
|
else
|
|
raise FetchError.new("bad response #{response.message} #{response.code}", uri)
|
|
end
|
|
end
|
|
|
|
alias :fetch_https :fetch_http
|
|
|
|
##
|
|
# Downloads +uri+ and returns it as a String.
|
|
|
|
def fetch_path(uri, mtime = nil, head = false)
|
|
uri = Gem::Uri.new uri
|
|
|
|
unless uri.scheme
|
|
raise ArgumentError, "uri scheme is invalid: #{uri.scheme.inspect}"
|
|
end
|
|
|
|
data = send "fetch_#{uri.scheme}", uri, mtime, head
|
|
|
|
if data && !head && uri.to_s.end_with?(".gz")
|
|
begin
|
|
data = Gem::Util.gunzip data
|
|
rescue Zlib::GzipFile::Error
|
|
raise FetchError.new("server did not return a valid file", uri)
|
|
end
|
|
end
|
|
|
|
data
|
|
rescue Timeout::Error, IOError, SocketError, SystemCallError,
|
|
*(OpenSSL::SSL::SSLError if Gem::HAVE_OPENSSL) => e
|
|
raise FetchError.new("#{e.class}: #{e}", uri)
|
|
end
|
|
|
|
def fetch_s3(uri, mtime = nil, head = false)
|
|
begin
|
|
public_uri = s3_uri_signer(uri).sign
|
|
rescue Gem::S3URISigner::ConfigurationError, Gem::S3URISigner::InstanceProfileError => e
|
|
raise FetchError.new(e.message, "s3://#{uri.host}")
|
|
end
|
|
fetch_https public_uri, mtime, head
|
|
end
|
|
|
|
# we have our own signing code here to avoid a dependency on the aws-sdk gem
|
|
def s3_uri_signer(uri)
|
|
Gem::S3URISigner.new(uri)
|
|
end
|
|
|
|
##
|
|
# Downloads +uri+ to +path+ if necessary. If no path is given, it just
|
|
# passes the data.
|
|
|
|
def cache_update_path(uri, path = nil, update = true)
|
|
mtime = path && File.stat(path).mtime rescue nil
|
|
|
|
data = fetch_path(uri, mtime)
|
|
|
|
if data.nil? # indicates the server returned 304 Not Modified
|
|
return Gem.read_binary(path)
|
|
end
|
|
|
|
if update && path
|
|
Gem.write_binary(path, data)
|
|
end
|
|
|
|
data
|
|
end
|
|
|
|
##
|
|
# Performs a Net::HTTP request of type +request_class+ on +uri+ returning
|
|
# a Net::HTTP response object. request maintains a table of persistent
|
|
# connections to reduce connect overhead.
|
|
|
|
def request(uri, request_class, last_modified = nil)
|
|
proxy = proxy_for @proxy, uri
|
|
pool = pools_for(proxy).pool_for uri
|
|
|
|
request = Gem::Request.new uri, request_class, last_modified, pool
|
|
|
|
request.fetch do |req|
|
|
yield req if block_given?
|
|
end
|
|
end
|
|
|
|
def https?(uri)
|
|
uri.scheme.downcase == "https"
|
|
end
|
|
|
|
def close_all
|
|
@pools.each_value {|pool| pool.close_all }
|
|
end
|
|
|
|
private
|
|
|
|
def proxy_for(proxy, uri)
|
|
Gem::Request.proxy_uri(proxy || Gem::Request.get_proxy_from_env(uri.scheme))
|
|
end
|
|
|
|
def pools_for(proxy)
|
|
@pool_lock.synchronize do
|
|
@pools[proxy] ||= Gem::Request::ConnectionPools.new proxy, @cert_files
|
|
end
|
|
end
|
|
end
|