You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
327 lines
9.8 KiB
327 lines
9.8 KiB
# frozen_string_literal: true |
|
|
|
require 'ipaddr' |
|
require 'socket' |
|
require 'resolv' |
|
|
|
# Use our own timeout class to avoid using HTTP.rb's timeout block |
|
# around the Socket#open method, since we use our own timeout blocks inside |
|
# that method |
|
# |
|
# Also changes how the read timeout behaves so that it is cumulative (closer |
|
# to HTTP::Timeout::Global, but still having distinct timeouts for other |
|
# operation types) |
|
class PerOperationWithDeadline < HTTP::Timeout::PerOperation |
|
READ_DEADLINE = 30 |
|
|
|
def initialize(*args) |
|
super |
|
|
|
@read_deadline = options.fetch(:read_deadline, READ_DEADLINE) |
|
end |
|
|
|
def connect(socket_class, host, port, nodelay = false) |
|
@socket = socket_class.open(host, port) |
|
@socket.setsockopt(Socket::IPPROTO_TCP, Socket::TCP_NODELAY, 1) if nodelay |
|
end |
|
|
|
# Reset deadline when the connection is re-used for different requests |
|
def reset_counter |
|
@deadline = nil |
|
end |
|
|
|
# Read data from the socket |
|
def readpartial(size, buffer = nil) |
|
@deadline ||= Process.clock_gettime(Process::CLOCK_MONOTONIC) + @read_deadline |
|
|
|
timeout = false |
|
loop do |
|
result = @socket.read_nonblock(size, buffer, exception: false) |
|
|
|
return :eof if result.nil? |
|
|
|
remaining_time = @deadline - Process.clock_gettime(Process::CLOCK_MONOTONIC) |
|
raise HTTP::TimeoutError, "Read timed out after #{@read_timeout} seconds" if timeout |
|
raise HTTP::TimeoutError, "Read timed out after a total of #{@read_deadline} seconds" if remaining_time <= 0 |
|
return result if result != :wait_readable |
|
|
|
# marking the socket for timeout. Why is this not being raised immediately? |
|
# it seems there is some race-condition on the network level between calling |
|
# #read_nonblock and #wait_readable, in which #read_nonblock signalizes waiting |
|
# for reads, and when waiting for x seconds, it returns nil suddenly without completing |
|
# the x seconds. In a normal case this would be a timeout on wait/read, but it can |
|
# also mean that the socket has been closed by the server. Therefore we "mark" the |
|
# socket for timeout and try to read more bytes. If it returns :eof, it's all good, no |
|
# timeout. Else, the first timeout was a proper timeout. |
|
# This hack has to be done because io/wait#wait_readable doesn't provide a value for when |
|
# the socket is closed by the server, and HTTP::Parser doesn't provide the limit for the chunks. |
|
timeout = true unless @socket.to_io.wait_readable([remaining_time, @read_timeout].min) |
|
end |
|
end |
|
end |
|
|
|
class Request |
|
REQUEST_TARGET = '(request-target)' |
|
|
|
# We enforce a 5s timeout on DNS resolving, 5s timeout on socket opening |
|
# and 5s timeout on the TLS handshake, meaning the worst case should take |
|
# about 15s in total |
|
TIMEOUT = { connect_timeout: 5, read_timeout: 10, write_timeout: 10, read_deadline: 30 }.freeze |
|
|
|
include RoutingHelper |
|
|
|
def initialize(verb, url, **options) |
|
raise ArgumentError if url.blank? |
|
|
|
@verb = verb |
|
@url = Addressable::URI.parse(url).normalize |
|
@http_client = options.delete(:http_client) |
|
@options = options.merge(socket_class: use_proxy? ? ProxySocket : Socket) |
|
@options = @options.merge(timeout_class: PerOperationWithDeadline, timeout_options: TIMEOUT) |
|
@options = @options.merge(Rails.configuration.x.http_client_proxy) if use_proxy? |
|
@headers = {} |
|
|
|
raise Mastodon::HostValidationError, 'Instance does not support hidden service connections' if block_hidden_service? |
|
|
|
set_common_headers! |
|
set_digest! if options.key?(:body) |
|
end |
|
|
|
def on_behalf_of(account, key_id_format = :uri, sign_with: nil) |
|
raise ArgumentError, 'account must not be nil' if account.nil? |
|
|
|
@account = account |
|
@keypair = sign_with.present? ? OpenSSL::PKey::RSA.new(sign_with) : @account.keypair |
|
@key_id_format = key_id_format |
|
|
|
self |
|
end |
|
|
|
def add_headers(new_headers) |
|
@headers.merge!(new_headers) |
|
self |
|
end |
|
|
|
def perform |
|
begin |
|
response = http_client.public_send(@verb, @url.to_s, @options.merge(headers: headers)) |
|
rescue => e |
|
raise e.class, "#{e.message} on #{@url}", e.backtrace[0] |
|
end |
|
|
|
begin |
|
response = response.extend(ClientLimit) |
|
|
|
# If we are using a persistent connection, we have to |
|
# read every response to be able to move forward at all. |
|
# However, simply calling #to_s or #flush may not be safe, |
|
# as the response body, if malicious, could be too big |
|
# for our memory. So we use the #body_with_limit method |
|
response.body_with_limit if http_client.persistent? |
|
|
|
yield response if block_given? |
|
ensure |
|
http_client.close unless http_client.persistent? |
|
end |
|
end |
|
|
|
def headers |
|
(@account ? @headers.merge('Signature' => signature) : @headers).without(REQUEST_TARGET) |
|
end |
|
|
|
class << self |
|
def valid_url?(url) |
|
begin |
|
parsed_url = Addressable::URI.parse(url) |
|
rescue Addressable::URI::InvalidURIError |
|
return false |
|
end |
|
|
|
%w(http https).include?(parsed_url.scheme) && parsed_url.host.present? |
|
end |
|
|
|
def http_client |
|
HTTP.use(:auto_inflate).follow(max_hops: 3) |
|
end |
|
end |
|
|
|
private |
|
|
|
def set_common_headers! |
|
@headers[REQUEST_TARGET] = "#{@verb} #{@url.path}" |
|
@headers['User-Agent'] = Mastodon::Version.user_agent |
|
@headers['Host'] = @url.host |
|
@headers['Date'] = Time.now.utc.httpdate |
|
@headers['Accept-Encoding'] = 'gzip' if @verb != :head |
|
end |
|
|
|
def set_digest! |
|
@headers['Digest'] = "SHA-256=#{Digest::SHA256.base64digest(@options[:body])}" |
|
end |
|
|
|
def signature |
|
algorithm = 'rsa-sha256' |
|
signature = Base64.strict_encode64(@keypair.sign(OpenSSL::Digest.new('SHA256'), signed_string)) |
|
|
|
"keyId=\"#{key_id}\",algorithm=\"#{algorithm}\",headers=\"#{signed_headers.keys.join(' ').downcase}\",signature=\"#{signature}\"" |
|
end |
|
|
|
def signed_string |
|
signed_headers.map { |key, value| "#{key.downcase}: #{value}" }.join("\n") |
|
end |
|
|
|
def signed_headers |
|
@headers.without('User-Agent', 'Accept-Encoding') |
|
end |
|
|
|
def key_id |
|
case @key_id_format |
|
when :acct |
|
@account.to_webfinger_s |
|
when :uri |
|
[ActivityPub::TagManager.instance.uri_for(@account), '#main-key'].join |
|
end |
|
end |
|
|
|
def http_client |
|
@http_client ||= Request.http_client |
|
end |
|
|
|
def use_proxy? |
|
Rails.configuration.x.http_client_proxy.present? |
|
end |
|
|
|
def block_hidden_service? |
|
!Rails.configuration.x.access_to_hidden_service && /\.(onion|i2p)$/.match?(@url.host) |
|
end |
|
|
|
module ClientLimit |
|
def body_with_limit(limit = 1.megabyte) |
|
raise Mastodon::LengthValidationError if content_length.present? && content_length > limit |
|
|
|
if charset.nil? |
|
encoding = Encoding::BINARY |
|
else |
|
begin |
|
encoding = Encoding.find(charset) |
|
rescue ArgumentError |
|
encoding = Encoding::BINARY |
|
end |
|
end |
|
|
|
contents = String.new(encoding: encoding) |
|
|
|
while (chunk = readpartial) |
|
contents << chunk |
|
chunk.clear |
|
|
|
raise Mastodon::LengthValidationError if contents.bytesize > limit |
|
end |
|
|
|
contents |
|
end |
|
end |
|
|
|
class Socket < TCPSocket |
|
class << self |
|
def open(host, *args) |
|
outer_e = nil |
|
port = args.first |
|
|
|
addresses = [] |
|
begin |
|
addresses = [IPAddr.new(host)] |
|
rescue IPAddr::InvalidAddressError |
|
Resolv::DNS.open do |dns| |
|
dns.timeouts = 5 |
|
addresses = dns.getaddresses(host).take(2) |
|
end |
|
end |
|
|
|
socks = [] |
|
addr_by_socket = {} |
|
|
|
addresses.each do |address| |
|
begin |
|
check_private_address(address) |
|
|
|
sock = ::Socket.new(address.is_a?(Resolv::IPv6) ? ::Socket::AF_INET6 : ::Socket::AF_INET, ::Socket::SOCK_STREAM, 0) |
|
sockaddr = ::Socket.pack_sockaddr_in(port, address.to_s) |
|
|
|
sock.setsockopt(::Socket::IPPROTO_TCP, ::Socket::TCP_NODELAY, 1) |
|
|
|
sock.connect_nonblock(sockaddr) |
|
|
|
# If that hasn't raised an exception, we somehow managed to connect |
|
# immediately, close pending sockets and return immediately |
|
socks.each(&:close) |
|
return sock |
|
rescue IO::WaitWritable |
|
socks << sock |
|
addr_by_socket[sock] = sockaddr |
|
rescue => e |
|
outer_e = e |
|
end |
|
end |
|
|
|
until socks.empty? |
|
_, available_socks, = IO.select(nil, socks, nil, Request::TIMEOUT[:connect]) |
|
|
|
if available_socks.nil? |
|
socks.each(&:close) |
|
raise HTTP::TimeoutError, "Connect timed out after #{Request::TIMEOUT[:connect]} seconds" |
|
end |
|
|
|
available_socks.each do |sock| |
|
socks.delete(sock) |
|
|
|
begin |
|
sock.connect_nonblock(addr_by_socket[sock]) |
|
rescue Errno::EISCONN |
|
# Do nothing |
|
rescue => e |
|
sock.close |
|
outer_e = e |
|
next |
|
end |
|
|
|
socks.each(&:close) |
|
return sock |
|
end |
|
end |
|
|
|
if outer_e |
|
raise outer_e |
|
else |
|
raise SocketError, "No address for #{host}" |
|
end |
|
end |
|
|
|
alias new open |
|
|
|
def check_private_address(address) |
|
addr = IPAddr.new(address.to_s) |
|
return if private_address_exceptions.any? { |range| range.include?(addr) } |
|
raise Mastodon::HostValidationError if PrivateAddressCheck.private_address?(addr) |
|
end |
|
|
|
def private_address_exceptions |
|
@private_address_exceptions = begin |
|
(ENV['ALLOWED_PRIVATE_ADDRESSES'] || '').split(',').map { |addr| IPAddr.new(addr) } |
|
end |
|
end |
|
end |
|
end |
|
|
|
class ProxySocket < Socket |
|
class << self |
|
def check_private_address(_address) |
|
# Accept connections to private addresses as HTTP proxies will usually |
|
# be on local addresses |
|
nil |
|
end |
|
end |
|
end |
|
|
|
private_constant :ClientLimit, :Socket, :ProxySocket |
|
end
|
|
|