naiveproxy/net/proxy/proxy_script_fetcher_impl.cc
2018-01-28 13:32:06 -05:00

394 lines
13 KiB
C++

// Copyright (c) 2012 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "net/proxy/proxy_script_fetcher_impl.h"
#include "base/compiler_specific.h"
#include "base/location.h"
#include "base/logging.h"
#include "base/metrics/histogram_macros.h"
#include "base/single_thread_task_runner.h"
#include "base/strings/string_util.h"
#include "base/threading/thread_task_runner_handle.h"
#include "net/base/data_url.h"
#include "net/base/io_buffer.h"
#include "net/base/load_flags.h"
#include "net/base/net_errors.h"
#include "net/base/net_string_util.h"
#include "net/base/request_priority.h"
#include "net/cert/cert_status_flags.h"
#include "net/http/http_response_headers.h"
#include "net/traffic_annotation/network_traffic_annotation.h"
#include "net/url_request/url_request_context.h"
// TODO(eroman):
// - Support auth-prompts (http://crbug.com/77366)
namespace net {
namespace {
// The maximum size (in bytes) allowed for a PAC script. Responses exceeding
// this will fail with ERR_FILE_TOO_BIG.
const int kDefaultMaxResponseBytes = 1048576; // 1 megabyte
// The maximum duration (in milliseconds) allowed for fetching the PAC script.
// Responses exceeding this will fail with ERR_TIMED_OUT.
const int kDefaultMaxDurationMs = 300000; // 5 minutes
// Returns true if |mime_type| is one of the known PAC mime type.
bool IsPacMimeType(const std::string& mime_type) {
static const char * const kSupportedPacMimeTypes[] = {
"application/x-ns-proxy-autoconfig",
"application/x-javascript-config",
};
for (size_t i = 0; i < arraysize(kSupportedPacMimeTypes); ++i) {
if (base::LowerCaseEqualsASCII(mime_type, kSupportedPacMimeTypes[i]))
return true;
}
return false;
}
// Converts |bytes| (which is encoded by |charset|) to UTF16, saving the resul
// to |*utf16|.
// If |charset| is empty, then we don't know what it was and guess.
void ConvertResponseToUTF16(const std::string& charset,
const std::string& bytes,
base::string16* utf16) {
const char* codepage;
if (charset.empty()) {
// Assume ISO-8859-1 if no charset was specified.
codepage = kCharsetLatin1;
} else {
// Otherwise trust the charset that was provided.
codepage = charset.c_str();
}
// Be generous in the conversion -- if any characters lie outside of |charset|
// (i.e. invalid), then substitute them with U+FFFD rather than failing.
ConvertToUTF16WithSubstitutions(bytes, codepage, utf16);
}
} // namespace
ProxyScriptFetcherImpl::ProxyScriptFetcherImpl(
URLRequestContext* url_request_context)
: url_request_context_(url_request_context),
buf_(new IOBuffer(kBufSize)),
next_id_(0),
cur_request_id_(0),
result_code_(OK),
result_text_(NULL),
max_response_bytes_(kDefaultMaxResponseBytes),
max_duration_(base::TimeDelta::FromMilliseconds(kDefaultMaxDurationMs)),
weak_factory_(this) {
DCHECK(url_request_context);
}
ProxyScriptFetcherImpl::~ProxyScriptFetcherImpl() {
// The URLRequest's destructor will cancel the outstanding request, and
// ensure that the delegate (this) is not called again.
}
base::TimeDelta ProxyScriptFetcherImpl::SetTimeoutConstraint(
base::TimeDelta timeout) {
base::TimeDelta prev = max_duration_;
max_duration_ = timeout;
return prev;
}
size_t ProxyScriptFetcherImpl::SetSizeConstraint(size_t size_bytes) {
size_t prev = max_response_bytes_;
max_response_bytes_ = size_bytes;
return prev;
}
void ProxyScriptFetcherImpl::OnResponseCompleted(URLRequest* request,
int net_error) {
DCHECK_EQ(request, cur_request_.get());
// Use |result_code_| as the request's error if we have already set it to
// something specific.
if (result_code_ == OK && net_error != OK)
result_code_ = net_error;
FetchCompleted();
}
int ProxyScriptFetcherImpl::Fetch(
const GURL& url, base::string16* text, const CompletionCallback& callback) {
// It is invalid to call Fetch() while a request is already in progress.
DCHECK(!cur_request_.get());
DCHECK(!callback.is_null());
DCHECK(text);
if (!url_request_context_)
return ERR_CONTEXT_SHUT_DOWN;
// Handle base-64 encoded data-urls that contain custom PAC scripts.
if (url.SchemeIs("data")) {
std::string mime_type;
std::string charset;
std::string data;
if (!DataURL::Parse(url, &mime_type, &charset, &data))
return ERR_FAILED;
ConvertResponseToUTF16(charset, data, text);
return OK;
}
DCHECK(fetch_start_time_.is_null());
fetch_start_time_ = base::TimeTicks::Now();
net::NetworkTrafficAnnotationTag traffic_annotation =
net::DefineNetworkTrafficAnnotation("proxy_script_fetcher", R"(
semantics {
sender: "Proxy Service"
description:
"Fetches candidate URLs for proxy auto-config (PAC) scripts. This "
"may be carried out as part of the web proxy auto-discovery "
"protocol, or because an explicit PAC script is specified by the "
"proxy settings. The source of these URLs may be user-specified "
"(when part of proxy settings), or may be provided by the network "
"(DNS or DHCP based discovery). Note that a user may not be using "
"a proxy, but determining that (i.e. auto-detect) may cause these "
"fetches."
trigger:
"PAC URLs may be fetched on initial start, every time the network "
"changes, whenever the proxy settings change, or periodically on a "
"timer to check for changes."
data: "None."
destination: OTHER
}
policy {
cookies_allowed: YES
cookies_store: "user"
setting:
"This feature cannot be disabled by settings. This request is only "
"made if the effective proxy settings include either auto-detect, "
"or specify a PAC script."
policy_exception_justification: "Not implemented."
})");
// Use highest priority, so if socket pools are being used for other types of
// requests, PAC requests are aren't blocked on them.
cur_request_ = url_request_context_->CreateRequest(url, MAXIMUM_PRIORITY,
this, traffic_annotation);
cur_request_->set_method("GET");
// Make sure that the PAC script is downloaded using a direct connection,
// to avoid circular dependencies (fetching is a part of proxy resolution).
// Also disable the use of the disk cache. The cache is disabled so that if
// the user switches networks we don't potentially use the cached response
// from old network when we should in fact be re-fetching on the new network.
// If the PAC script is hosted on an HTTPS server we bypass revocation
// checking in order to avoid a circular dependency when attempting to fetch
// the OCSP response or CRL. We could make the revocation check go direct but
// the proxy might be the only way to the outside world. IGNORE_LIMITS is
// used to avoid blocking proxy resolution on other network requests.
cur_request_->SetLoadFlags(LOAD_BYPASS_PROXY | LOAD_DISABLE_CACHE |
LOAD_DISABLE_CERT_REVOCATION_CHECKING |
LOAD_IGNORE_LIMITS);
// Save the caller's info for notification on completion.
callback_ = callback;
result_text_ = text;
bytes_read_so_far_.clear();
// Post a task to timeout this request if it takes too long.
cur_request_id_ = ++next_id_;
base::ThreadTaskRunnerHandle::Get()->PostDelayedTask(
FROM_HERE, base::Bind(&ProxyScriptFetcherImpl::OnTimeout,
weak_factory_.GetWeakPtr(), cur_request_id_),
max_duration_);
// Start the request.
cur_request_->Start();
return ERR_IO_PENDING;
}
void ProxyScriptFetcherImpl::Cancel() {
// ResetCurRequestState will free the URLRequest, which will cause
// cancellation.
ResetCurRequestState();
}
URLRequestContext* ProxyScriptFetcherImpl::GetRequestContext() const {
return url_request_context_;
}
void ProxyScriptFetcherImpl::OnShutdown() {
url_request_context_ = nullptr;
if (cur_request_) {
result_code_ = ERR_CONTEXT_SHUT_DOWN;
FetchCompleted();
}
}
void ProxyScriptFetcherImpl::OnAuthRequired(URLRequest* request,
AuthChallengeInfo* auth_info) {
DCHECK_EQ(request, cur_request_.get());
// TODO(eroman): http://crbug.com/77366
LOG(WARNING) << "Auth required to fetch PAC script, aborting.";
result_code_ = ERR_NOT_IMPLEMENTED;
request->CancelAuth();
}
void ProxyScriptFetcherImpl::OnSSLCertificateError(URLRequest* request,
const SSLInfo& ssl_info,
bool fatal) {
DCHECK_EQ(request, cur_request_.get());
// Revocation check failures are not fatal.
if (IsCertStatusMinorError(ssl_info.cert_status)) {
request->ContinueDespiteLastError();
return;
}
LOG(WARNING) << "SSL certificate error when fetching PAC script, aborting.";
// Certificate errors are in same space as net errors.
result_code_ = MapCertStatusToNetError(ssl_info.cert_status);
request->Cancel();
}
void ProxyScriptFetcherImpl::OnResponseStarted(URLRequest* request,
int net_error) {
DCHECK_EQ(request, cur_request_.get());
DCHECK_NE(ERR_IO_PENDING, net_error);
if (net_error != OK) {
OnResponseCompleted(request, net_error);
return;
}
// Require HTTP responses to have a success status code.
if (request->url().SchemeIsHTTPOrHTTPS()) {
// NOTE about status codes: We are like Firefox 3 in this respect.
// {IE 7, Safari 3, Opera 9.5} do not care about the status code.
if (request->GetResponseCode() != 200) {
VLOG(1) << "Fetched PAC script had (bad) status line: "
<< request->response_headers()->GetStatusLine();
result_code_ = ERR_PAC_STATUS_NOT_OK;
request->Cancel();
return;
}
// NOTE about mime types: We do not enforce mime types on PAC files.
// This is for compatibility with {IE 7, Firefox 3, Opera 9.5}. We will
// however log mismatches to help with debugging.
std::string mime_type;
cur_request_->GetMimeType(&mime_type);
if (!IsPacMimeType(mime_type)) {
VLOG(1) << "Fetched PAC script does not have a proper mime type: "
<< mime_type;
}
}
ReadBody(request);
}
void ProxyScriptFetcherImpl::OnReadCompleted(URLRequest* request,
int num_bytes) {
DCHECK_NE(ERR_IO_PENDING, num_bytes);
DCHECK_EQ(request, cur_request_.get());
if (ConsumeBytesRead(request, num_bytes)) {
// Keep reading.
ReadBody(request);
}
}
void ProxyScriptFetcherImpl::ReadBody(URLRequest* request) {
// Read as many bytes as are available synchronously.
while (true) {
int num_bytes = request->Read(buf_.get(), kBufSize);
if (num_bytes == ERR_IO_PENDING)
return;
if (num_bytes < 0) {
OnResponseCompleted(request, num_bytes);
return;
}
if (!ConsumeBytesRead(request, num_bytes))
return;
}
}
bool ProxyScriptFetcherImpl::ConsumeBytesRead(URLRequest* request,
int num_bytes) {
if (num_bytes <= 0) {
// Error while reading, or EOF.
OnResponseCompleted(request, num_bytes);
return false;
}
// Enforce maximum size bound.
if (num_bytes + bytes_read_so_far_.size() >
static_cast<size_t>(max_response_bytes_)) {
result_code_ = ERR_FILE_TOO_BIG;
request->Cancel();
return false;
}
if (bytes_read_so_far_.empty()) {
DCHECK(fetch_time_to_first_byte_.is_null());
fetch_time_to_first_byte_ = base::TimeTicks::Now();
}
bytes_read_so_far_.append(buf_->data(), num_bytes);
return true;
}
void ProxyScriptFetcherImpl::FetchCompleted() {
if (result_code_ == OK) {
// Calculate duration of time for proxy script fetch to complete.
DCHECK(!fetch_start_time_.is_null());
DCHECK(!fetch_time_to_first_byte_.is_null());
UMA_HISTOGRAM_MEDIUM_TIMES("Net.ProxyScriptFetcher.SuccessDuration",
base::TimeTicks::Now() - fetch_start_time_);
UMA_HISTOGRAM_MEDIUM_TIMES("Net.ProxyScriptFetcher.FirstByteDuration",
fetch_time_to_first_byte_ - fetch_start_time_);
// The caller expects the response to be encoded as UTF16.
std::string charset;
cur_request_->GetCharset(&charset);
ConvertResponseToUTF16(charset, bytes_read_so_far_, result_text_);
} else {
// On error, the caller expects empty string for bytes.
result_text_->clear();
}
int result_code = result_code_;
CompletionCallback callback = callback_;
ResetCurRequestState();
callback.Run(result_code);
}
void ProxyScriptFetcherImpl::ResetCurRequestState() {
cur_request_.reset();
cur_request_id_ = 0;
callback_.Reset();
result_code_ = OK;
result_text_ = NULL;
fetch_start_time_ = base::TimeTicks();
fetch_time_to_first_byte_ = base::TimeTicks();
}
void ProxyScriptFetcherImpl::OnTimeout(int id) {
// Timeout tasks may outlive the URLRequest they reference. Make sure it
// is still applicable.
if (cur_request_id_ != id)
return;
DCHECK(cur_request_.get());
result_code_ = ERR_TIMED_OUT;
FetchCompleted();
}
} // namespace net