In preperation for creating requester and proxy unit tests:
o Tidied up some of the requester and proxy
o Partially de-coupled requester from proxy
o Fixed minor bugs:
o is_valid_uri was not implemented correctly
o http scheme validation had "http" instead of "https"
152 lines
6.1 KiB
Ruby
152 lines
6.1 KiB
Ruby
#
|
|
# Copyright 2011 Wade Alcorn wade@bindshell.net
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
module BeEF
|
|
module Extension
|
|
module Proxy
|
|
class Proxy
|
|
|
|
HB = BeEF::Core::Models::HookedBrowser
|
|
H = BeEF::Core::Models::Http
|
|
@response = nil
|
|
|
|
# Multi-threaded Tunneling Proxy: listens on host:port configured in extensions/proxy/config.yaml
|
|
# and forwards requests to the hooked browser using the Requester component.
|
|
def initialize
|
|
@conf = BeEF::Core::Configuration.instance
|
|
@proxy_server = TCPServer.new(@conf.get('beef.extension.proxy.address'), @conf.get('beef.extension.proxy.port'))
|
|
|
|
loop do
|
|
proxy = @proxy_server.accept
|
|
Thread.new proxy, &method(:handle_request)
|
|
end
|
|
end
|
|
|
|
def handle_request socket
|
|
request_line = socket.readline
|
|
|
|
method = request_line[/^\w+/]
|
|
url = request_line[/^\w+\s+(\S+)/, 1]
|
|
version = request_line[/HTTP\/(1\.\d)\s*$/, 1]
|
|
|
|
# We're overwriting the URI::Parser UNRESERVED regex to prevent BAD URI errors when sending attack vectors (see tolerant_parser)
|
|
tolerant_parser = URI::Parser.new(:UNRESERVED => BeEF::Core::Configuration.instance.get("beef.extension.requester.uri_unreserved_chars"))
|
|
uri = tolerant_parser.parse(url)
|
|
|
|
raw_request = request_line
|
|
content_length = 0
|
|
|
|
loop do
|
|
line = socket.readline
|
|
|
|
if line =~ /^Content-Length:\s+(\d+)\s*$/
|
|
content_length = $1.to_i
|
|
end
|
|
|
|
if line.strip.empty?
|
|
# read data still in the socket, exactly <content_length> bytes
|
|
if content_length >= 0
|
|
raw_request += "\r\n" + socket.read(content_length)
|
|
end
|
|
break
|
|
else
|
|
raw_request += line
|
|
end
|
|
end
|
|
|
|
# Saves the new HTTP request to the db. It will be processed by the PreHookCallback of the requester component.
|
|
# IDs are created and incremented automatically by DataMapper.
|
|
http = H.new(
|
|
:request => raw_request,
|
|
:method => method,
|
|
:domain => uri.host,
|
|
:port => uri.port,
|
|
:path => uri.path,
|
|
:request_date => Time.now,
|
|
:hooked_browser_id => self.get_tunneling_proxy
|
|
)
|
|
http.save
|
|
print_debug("[PROXY] --> Forwarding request ##{http.id}: domain[#{http.domain}:#{http.port}], method[#{http.method}], path[#{http.path}]")
|
|
|
|
# Wait for the HTTP response to be stored in the db.
|
|
# TODO: re-implement this with EventMachine or with the Observer pattern.
|
|
while H.first(:id => http.id).has_ran != "complete"
|
|
sleep 0.5
|
|
end
|
|
@response = H.first(:id => http.id)
|
|
print_debug "[PROXY] <-- Response for request ##{@response.id} to [#{@response.path}] on domain [#{@response.domain}:#{@response.port}] correctly processed"
|
|
|
|
response_body = @response['response_data']
|
|
response_status = @response['response_status_code']
|
|
headers = @response['response_headers']
|
|
|
|
# The following is needed to forward back some of the original HTTP response headers obtained via XHR calls.
|
|
# Original XHR response headers are stored in extension_proxy_http table (response_headers column),
|
|
# but we are forwarding back only some of them (Server, X-.. - like X-Powered-By -, Content-Type, ... ).
|
|
# Some of the original response headers need to be removed, like encoding and cache related: for example
|
|
# about encoding, the original response headers says that the content-length is 1000 as the response is gzipped,
|
|
# but the final content-length forwarded back by the proxy is clearly bigger. Date header follows the same way.
|
|
response_headers = ""
|
|
if (response_status != -1 && response_status != 0)
|
|
headers.each_line do |line|
|
|
# stripping the Encoding, Cache and other headers
|
|
header_key = line.split(': ')[0]
|
|
header_value = line.split(': ')[1]
|
|
if !header_key.nil? &&
|
|
header_key != "Content-Encoding" &&
|
|
header_key != "Keep-Alive" &&
|
|
header_key != "Cache-Control" &&
|
|
header_key != "Vary" &&
|
|
header_key != "Pragma" &&
|
|
header_key != "Connection" &&
|
|
header_key != "Expires" &&
|
|
header_key != "Accept-Ranges" &&
|
|
header_key != "Date"
|
|
if header_value.nil?
|
|
#headers_hash[header_key] = ""
|
|
else
|
|
# update Content-Length with the valid one
|
|
if header_key == "Content-Length"
|
|
header_value = response_body.size
|
|
response_headers += "Content-Length: #{header_value}\r\n"
|
|
else
|
|
response_headers += line
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|
|
|
|
res = "HTTP/#{version} #{response_status}\r\n#{response_headers}\r\n\r\n#{response_body}"
|
|
socket.write(res)
|
|
socket.close
|
|
end
|
|
|
|
def get_tunneling_proxy
|
|
proxy_browser = HB.first(:is_proxy => true)
|
|
if (proxy_browser != nil)
|
|
proxy_browser_id = proxy_browser.id.to_s
|
|
else
|
|
proxy_browser_id = 1
|
|
print_debug "[PROXY] Proxy browser not set. Defaulting to browser id #1"
|
|
end
|
|
proxy_browser_id
|
|
end
|
|
end
|
|
end
|
|
end
|
|
end
|
|
|