mirror of
https://github.com/klzgrad/naiveproxy.git
synced 2024-12-01 01:36:09 +03:00
1642 lines
58 KiB
C++
1642 lines
58 KiB
C++
|
// Copyright (c) 2012 The Chromium Authors. All rights reserved.
|
||
|
// Use of this source code is governed by a BSD-style license that can be
|
||
|
// found in the LICENSE file.
|
||
|
|
||
|
#include "net/proxy/proxy_service.h"
|
||
|
|
||
|
#include <algorithm>
|
||
|
#include <cmath>
|
||
|
#include <utility>
|
||
|
|
||
|
#include "base/bind.h"
|
||
|
#include "base/bind_helpers.h"
|
||
|
#include "base/compiler_specific.h"
|
||
|
#include "base/location.h"
|
||
|
#include "base/logging.h"
|
||
|
#include "base/macros.h"
|
||
|
#include "base/memory/weak_ptr.h"
|
||
|
#include "base/single_thread_task_runner.h"
|
||
|
#include "base/strings/string_util.h"
|
||
|
#include "base/threading/thread_task_runner_handle.h"
|
||
|
#include "base/time/time.h"
|
||
|
#include "base/values.h"
|
||
|
#include "net/base/completion_callback.h"
|
||
|
#include "net/base/net_errors.h"
|
||
|
#include "net/base/proxy_delegate.h"
|
||
|
#include "net/base/url_util.h"
|
||
|
#include "net/log/net_log_capture_mode.h"
|
||
|
#include "net/log/net_log_event_type.h"
|
||
|
#include "net/log/net_log_with_source.h"
|
||
|
#include "net/proxy/dhcp_proxy_script_fetcher.h"
|
||
|
#include "net/proxy/multi_threaded_proxy_resolver.h"
|
||
|
#include "net/proxy/proxy_config_service_fixed.h"
|
||
|
#include "net/proxy/proxy_resolver.h"
|
||
|
#include "net/proxy/proxy_resolver_factory.h"
|
||
|
#include "net/proxy/proxy_script_decider.h"
|
||
|
#include "net/proxy/proxy_script_fetcher.h"
|
||
|
#include "net/url_request/url_request_context.h"
|
||
|
#include "url/gurl.h"
|
||
|
|
||
|
#if defined(OS_WIN)
|
||
|
#include "net/proxy/proxy_config_service_win.h"
|
||
|
#include "net/proxy/proxy_resolver_winhttp.h"
|
||
|
#elif defined(OS_IOS)
|
||
|
#include "net/proxy/proxy_config_service_ios.h"
|
||
|
#include "net/proxy/proxy_resolver_mac.h"
|
||
|
#elif defined(OS_MACOSX)
|
||
|
#include "net/proxy/proxy_config_service_mac.h"
|
||
|
#include "net/proxy/proxy_resolver_mac.h"
|
||
|
#elif defined(OS_LINUX) && !defined(OS_CHROMEOS)
|
||
|
#include "net/proxy/proxy_config_service_linux.h"
|
||
|
#elif defined(OS_ANDROID)
|
||
|
#include "net/proxy/proxy_config_service_android.h"
|
||
|
#endif
|
||
|
|
||
|
using base::TimeDelta;
|
||
|
using base::TimeTicks;
|
||
|
|
||
|
namespace net {
|
||
|
|
||
|
namespace {
|
||
|
|
||
|
const size_t kDefaultNumPacThreads = 4;
|
||
|
|
||
|
// When the IP address changes we don't immediately re-run proxy auto-config.
|
||
|
// Instead, we wait for |kDelayAfterNetworkChangesMs| before
|
||
|
// attempting to re-valuate proxy auto-config.
|
||
|
//
|
||
|
// During this time window, any resolve requests sent to the ProxyService will
|
||
|
// be queued. Once we have waited the required amount of them, the proxy
|
||
|
// auto-config step will be run, and the queued requests resumed.
|
||
|
//
|
||
|
// The reason we play this game is that our signal for detecting network
|
||
|
// changes (NetworkChangeNotifier) may fire *before* the system's networking
|
||
|
// dependencies are fully configured. This is a problem since it means if
|
||
|
// we were to run proxy auto-config right away, it could fail due to spurious
|
||
|
// DNS failures. (see http://crbug.com/50779 for more details.)
|
||
|
//
|
||
|
// By adding the wait window, we give things a better chance to get properly
|
||
|
// set up. Network failures can happen at any time though, so we additionally
|
||
|
// poll the PAC script for changes, which will allow us to recover from these
|
||
|
// sorts of problems.
|
||
|
const int64_t kDelayAfterNetworkChangesMs = 2000;
|
||
|
|
||
|
// This is the default policy for polling the PAC script.
|
||
|
//
|
||
|
// In response to a failure, the poll intervals are:
|
||
|
// 0: 8 seconds (scheduled on timer)
|
||
|
// 1: 32 seconds
|
||
|
// 2: 2 minutes
|
||
|
// 3+: 4 hours
|
||
|
//
|
||
|
// In response to a success, the poll intervals are:
|
||
|
// 0+: 12 hours
|
||
|
//
|
||
|
// Only the 8 second poll is scheduled on a timer, the rest happen in response
|
||
|
// to network activity (and hence will take longer than the written time).
|
||
|
//
|
||
|
// Explanation for these values:
|
||
|
//
|
||
|
// TODO(eroman): These values are somewhat arbitrary, and need to be tuned
|
||
|
// using some histograms data. Trying to be conservative so as not to break
|
||
|
// existing setups when deployed. A simple exponential retry scheme would be
|
||
|
// more elegant, but places more load on server.
|
||
|
//
|
||
|
// The motivation for trying quickly after failures (8 seconds) is to recover
|
||
|
// from spurious network failures, which are common after the IP address has
|
||
|
// just changed (like DNS failing to resolve). The next 32 second boundary is
|
||
|
// to try and catch other VPN weirdness which anecdotally I have seen take
|
||
|
// 10+ seconds for some users.
|
||
|
//
|
||
|
// The motivation for re-trying after a success is to check for possible
|
||
|
// content changes to the script, or to the WPAD auto-discovery results. We are
|
||
|
// not very aggressive with these checks so as to minimize the risk of
|
||
|
// overloading existing PAC setups. Moreover it is unlikely that PAC scripts
|
||
|
// change very frequently in existing setups. More research is needed to
|
||
|
// motivate what safe values are here, and what other user agents do.
|
||
|
//
|
||
|
// Comparison to other browsers:
|
||
|
//
|
||
|
// In Firefox the PAC URL is re-tried on failures according to
|
||
|
// network.proxy.autoconfig_retry_interval_min and
|
||
|
// network.proxy.autoconfig_retry_interval_max. The defaults are 5 seconds and
|
||
|
// 5 minutes respectively. It doubles the interval at each attempt.
|
||
|
//
|
||
|
// TODO(eroman): Figure out what Internet Explorer does.
|
||
|
class DefaultPollPolicy : public ProxyService::PacPollPolicy {
|
||
|
public:
|
||
|
DefaultPollPolicy() {}
|
||
|
|
||
|
Mode GetNextDelay(int initial_error,
|
||
|
TimeDelta current_delay,
|
||
|
TimeDelta* next_delay) const override {
|
||
|
if (initial_error != OK) {
|
||
|
// Re-try policy for failures.
|
||
|
const int kDelay1Seconds = 8;
|
||
|
const int kDelay2Seconds = 32;
|
||
|
const int kDelay3Seconds = 2 * 60; // 2 minutes
|
||
|
const int kDelay4Seconds = 4 * 60 * 60; // 4 Hours
|
||
|
|
||
|
// Initial poll.
|
||
|
if (current_delay < TimeDelta()) {
|
||
|
*next_delay = TimeDelta::FromSeconds(kDelay1Seconds);
|
||
|
return MODE_USE_TIMER;
|
||
|
}
|
||
|
switch (current_delay.InSeconds()) {
|
||
|
case kDelay1Seconds:
|
||
|
*next_delay = TimeDelta::FromSeconds(kDelay2Seconds);
|
||
|
return MODE_START_AFTER_ACTIVITY;
|
||
|
case kDelay2Seconds:
|
||
|
*next_delay = TimeDelta::FromSeconds(kDelay3Seconds);
|
||
|
return MODE_START_AFTER_ACTIVITY;
|
||
|
default:
|
||
|
*next_delay = TimeDelta::FromSeconds(kDelay4Seconds);
|
||
|
return MODE_START_AFTER_ACTIVITY;
|
||
|
}
|
||
|
} else {
|
||
|
// Re-try policy for succeses.
|
||
|
*next_delay = TimeDelta::FromHours(12);
|
||
|
return MODE_START_AFTER_ACTIVITY;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
DISALLOW_COPY_AND_ASSIGN(DefaultPollPolicy);
|
||
|
};
|
||
|
|
||
|
// Config getter that always returns direct settings.
|
||
|
class ProxyConfigServiceDirect : public ProxyConfigService {
|
||
|
public:
|
||
|
// ProxyConfigService implementation:
|
||
|
void AddObserver(Observer* observer) override {}
|
||
|
void RemoveObserver(Observer* observer) override {}
|
||
|
ConfigAvailability GetLatestProxyConfig(ProxyConfig* config) override {
|
||
|
*config = ProxyConfig::CreateDirect();
|
||
|
config->set_source(PROXY_CONFIG_SOURCE_UNKNOWN);
|
||
|
return CONFIG_VALID;
|
||
|
}
|
||
|
};
|
||
|
|
||
|
// Proxy resolver that fails every time.
|
||
|
class ProxyResolverNull : public ProxyResolver {
|
||
|
public:
|
||
|
ProxyResolverNull() {}
|
||
|
|
||
|
// ProxyResolver implementation.
|
||
|
int GetProxyForURL(const GURL& url,
|
||
|
ProxyInfo* results,
|
||
|
const CompletionCallback& callback,
|
||
|
std::unique_ptr<Request>* request,
|
||
|
const NetLogWithSource& net_log) override {
|
||
|
return ERR_NOT_IMPLEMENTED;
|
||
|
}
|
||
|
|
||
|
};
|
||
|
|
||
|
// ProxyResolver that simulates a PAC script which returns
|
||
|
// |pac_string| for every single URL.
|
||
|
class ProxyResolverFromPacString : public ProxyResolver {
|
||
|
public:
|
||
|
explicit ProxyResolverFromPacString(const std::string& pac_string)
|
||
|
: pac_string_(pac_string) {}
|
||
|
|
||
|
int GetProxyForURL(const GURL& url,
|
||
|
ProxyInfo* results,
|
||
|
const CompletionCallback& callback,
|
||
|
std::unique_ptr<Request>* request,
|
||
|
const NetLogWithSource& net_log) override {
|
||
|
results->UsePacString(pac_string_);
|
||
|
return OK;
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
const std::string pac_string_;
|
||
|
};
|
||
|
|
||
|
// Creates ProxyResolvers using a platform-specific implementation.
|
||
|
class ProxyResolverFactoryForSystem : public MultiThreadedProxyResolverFactory {
|
||
|
public:
|
||
|
explicit ProxyResolverFactoryForSystem(size_t max_num_threads)
|
||
|
: MultiThreadedProxyResolverFactory(max_num_threads,
|
||
|
false /*expects_pac_bytes*/) {}
|
||
|
|
||
|
std::unique_ptr<ProxyResolverFactory> CreateProxyResolverFactory() override {
|
||
|
#if defined(OS_WIN)
|
||
|
return std::make_unique<ProxyResolverFactoryWinHttp>();
|
||
|
#elif defined(OS_MACOSX)
|
||
|
return std::make_unique<ProxyResolverFactoryMac>();
|
||
|
#else
|
||
|
NOTREACHED();
|
||
|
return NULL;
|
||
|
#endif
|
||
|
}
|
||
|
|
||
|
static bool IsSupported() {
|
||
|
#if defined(OS_WIN) || defined(OS_MACOSX)
|
||
|
return true;
|
||
|
#else
|
||
|
return false;
|
||
|
#endif
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
DISALLOW_COPY_AND_ASSIGN(ProxyResolverFactoryForSystem);
|
||
|
};
|
||
|
|
||
|
class ProxyResolverFactoryForNullResolver : public ProxyResolverFactory {
|
||
|
public:
|
||
|
ProxyResolverFactoryForNullResolver() : ProxyResolverFactory(false) {}
|
||
|
|
||
|
// ProxyResolverFactory overrides.
|
||
|
int CreateProxyResolver(
|
||
|
const scoped_refptr<ProxyResolverScriptData>& pac_script,
|
||
|
std::unique_ptr<ProxyResolver>* resolver,
|
||
|
const net::CompletionCallback& callback,
|
||
|
std::unique_ptr<Request>* request) override {
|
||
|
resolver->reset(new ProxyResolverNull());
|
||
|
return OK;
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
DISALLOW_COPY_AND_ASSIGN(ProxyResolverFactoryForNullResolver);
|
||
|
};
|
||
|
|
||
|
class ProxyResolverFactoryForPacResult : public ProxyResolverFactory {
|
||
|
public:
|
||
|
explicit ProxyResolverFactoryForPacResult(const std::string& pac_string)
|
||
|
: ProxyResolverFactory(false), pac_string_(pac_string) {}
|
||
|
|
||
|
// ProxyResolverFactory override.
|
||
|
int CreateProxyResolver(
|
||
|
const scoped_refptr<ProxyResolverScriptData>& pac_script,
|
||
|
std::unique_ptr<ProxyResolver>* resolver,
|
||
|
const net::CompletionCallback& callback,
|
||
|
std::unique_ptr<Request>* request) override {
|
||
|
resolver->reset(new ProxyResolverFromPacString(pac_string_));
|
||
|
return OK;
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
const std::string pac_string_;
|
||
|
|
||
|
DISALLOW_COPY_AND_ASSIGN(ProxyResolverFactoryForPacResult);
|
||
|
};
|
||
|
|
||
|
// Returns NetLog parameters describing a proxy configuration change.
|
||
|
std::unique_ptr<base::Value> NetLogProxyConfigChangedCallback(
|
||
|
const ProxyConfig* old_config,
|
||
|
const ProxyConfig* new_config,
|
||
|
NetLogCaptureMode /* capture_mode */) {
|
||
|
std::unique_ptr<base::DictionaryValue> dict(new base::DictionaryValue());
|
||
|
// The "old_config" is optional -- the first notification will not have
|
||
|
// any "previous" configuration.
|
||
|
if (old_config->is_valid())
|
||
|
dict->Set("old_config", old_config->ToValue());
|
||
|
dict->Set("new_config", new_config->ToValue());
|
||
|
return std::move(dict);
|
||
|
}
|
||
|
|
||
|
std::unique_ptr<base::Value> NetLogBadProxyListCallback(
|
||
|
const ProxyRetryInfoMap* retry_info,
|
||
|
NetLogCaptureMode /* capture_mode */) {
|
||
|
auto dict = std::make_unique<base::DictionaryValue>();
|
||
|
auto list = std::make_unique<base::ListValue>();
|
||
|
|
||
|
for (ProxyRetryInfoMap::const_iterator iter = retry_info->begin();
|
||
|
iter != retry_info->end(); ++iter) {
|
||
|
list->AppendString(iter->first);
|
||
|
}
|
||
|
dict->Set("bad_proxy_list", std::move(list));
|
||
|
return std::move(dict);
|
||
|
}
|
||
|
|
||
|
// Returns NetLog parameters on a successfuly proxy resolution.
|
||
|
std::unique_ptr<base::Value> NetLogFinishedResolvingProxyCallback(
|
||
|
const ProxyInfo* result,
|
||
|
NetLogCaptureMode /* capture_mode */) {
|
||
|
std::unique_ptr<base::DictionaryValue> dict(new base::DictionaryValue());
|
||
|
dict->SetString("pac_string", result->ToPacString());
|
||
|
return std::move(dict);
|
||
|
}
|
||
|
|
||
|
#if defined(OS_CHROMEOS)
|
||
|
class UnsetProxyConfigService : public ProxyConfigService {
|
||
|
public:
|
||
|
UnsetProxyConfigService() {}
|
||
|
~UnsetProxyConfigService() override {}
|
||
|
|
||
|
void AddObserver(Observer* observer) override {}
|
||
|
void RemoveObserver(Observer* observer) override {}
|
||
|
ConfigAvailability GetLatestProxyConfig(ProxyConfig* config) override {
|
||
|
return CONFIG_UNSET;
|
||
|
}
|
||
|
};
|
||
|
#endif
|
||
|
|
||
|
// Returns a sanitized copy of |url| which is safe to pass on to a PAC script.
|
||
|
// The method for sanitizing is determined by |policy|. See the comments for
|
||
|
// that enum for details.
|
||
|
GURL SanitizeUrl(const GURL& url, ProxyService::SanitizeUrlPolicy policy) {
|
||
|
DCHECK(url.is_valid());
|
||
|
|
||
|
GURL::Replacements replacements;
|
||
|
replacements.ClearUsername();
|
||
|
replacements.ClearPassword();
|
||
|
replacements.ClearRef();
|
||
|
|
||
|
if (policy == ProxyService::SanitizeUrlPolicy::SAFE &&
|
||
|
url.SchemeIsCryptographic()) {
|
||
|
replacements.ClearPath();
|
||
|
replacements.ClearQuery();
|
||
|
}
|
||
|
|
||
|
return url.ReplaceComponents(replacements);
|
||
|
}
|
||
|
|
||
|
} // namespace
|
||
|
|
||
|
// ProxyService::InitProxyResolver --------------------------------------------
|
||
|
|
||
|
// This glues together two asynchronous steps:
|
||
|
// (1) ProxyScriptDecider -- try to fetch/validate a sequence of PAC scripts
|
||
|
// to figure out what we should configure against.
|
||
|
// (2) Feed the fetched PAC script into the ProxyResolver.
|
||
|
//
|
||
|
// InitProxyResolver is a single-use class which encapsulates cancellation as
|
||
|
// part of its destructor. Start() or StartSkipDecider() should be called just
|
||
|
// once. The instance can be destroyed at any time, and the request will be
|
||
|
// cancelled.
|
||
|
|
||
|
class ProxyService::InitProxyResolver {
|
||
|
public:
|
||
|
InitProxyResolver()
|
||
|
: proxy_resolver_factory_(nullptr),
|
||
|
proxy_resolver_(NULL),
|
||
|
next_state_(STATE_NONE),
|
||
|
quick_check_enabled_(true) {}
|
||
|
|
||
|
~InitProxyResolver() {
|
||
|
// Note that the destruction of ProxyScriptDecider will automatically cancel
|
||
|
// any outstanding work.
|
||
|
}
|
||
|
|
||
|
// Begins initializing the proxy resolver; calls |callback| when done. A
|
||
|
// ProxyResolver instance will be created using |proxy_resolver_factory| and
|
||
|
// returned via |proxy_resolver| if the final result is OK.
|
||
|
int Start(std::unique_ptr<ProxyResolver>* proxy_resolver,
|
||
|
ProxyResolverFactory* proxy_resolver_factory,
|
||
|
ProxyScriptFetcher* proxy_script_fetcher,
|
||
|
DhcpProxyScriptFetcher* dhcp_proxy_script_fetcher,
|
||
|
NetLog* net_log,
|
||
|
const ProxyConfig& config,
|
||
|
TimeDelta wait_delay,
|
||
|
const CompletionCallback& callback) {
|
||
|
DCHECK_EQ(STATE_NONE, next_state_);
|
||
|
proxy_resolver_ = proxy_resolver;
|
||
|
proxy_resolver_factory_ = proxy_resolver_factory;
|
||
|
|
||
|
decider_.reset(new ProxyScriptDecider(
|
||
|
proxy_script_fetcher, dhcp_proxy_script_fetcher, net_log));
|
||
|
decider_->set_quick_check_enabled(quick_check_enabled_);
|
||
|
config_ = config;
|
||
|
wait_delay_ = wait_delay;
|
||
|
callback_ = callback;
|
||
|
|
||
|
next_state_ = STATE_DECIDE_PROXY_SCRIPT;
|
||
|
return DoLoop(OK);
|
||
|
}
|
||
|
|
||
|
// Similar to Start(), however it skips the ProxyScriptDecider stage. Instead
|
||
|
// |effective_config|, |decider_result| and |script_data| will be used as the
|
||
|
// inputs for initializing the ProxyResolver. A ProxyResolver instance will
|
||
|
// be created using |proxy_resolver_factory| and returned via
|
||
|
// |proxy_resolver| if the final result is OK.
|
||
|
int StartSkipDecider(std::unique_ptr<ProxyResolver>* proxy_resolver,
|
||
|
ProxyResolverFactory* proxy_resolver_factory,
|
||
|
const ProxyConfig& effective_config,
|
||
|
int decider_result,
|
||
|
ProxyResolverScriptData* script_data,
|
||
|
const CompletionCallback& callback) {
|
||
|
DCHECK_EQ(STATE_NONE, next_state_);
|
||
|
proxy_resolver_ = proxy_resolver;
|
||
|
proxy_resolver_factory_ = proxy_resolver_factory;
|
||
|
|
||
|
effective_config_ = effective_config;
|
||
|
script_data_ = script_data;
|
||
|
callback_ = callback;
|
||
|
|
||
|
if (decider_result != OK)
|
||
|
return decider_result;
|
||
|
|
||
|
next_state_ = STATE_CREATE_RESOLVER;
|
||
|
return DoLoop(OK);
|
||
|
}
|
||
|
|
||
|
// Returns the proxy configuration that was selected by ProxyScriptDecider.
|
||
|
// Should only be called upon completion of the initialization.
|
||
|
const ProxyConfig& effective_config() const {
|
||
|
DCHECK_EQ(STATE_NONE, next_state_);
|
||
|
return effective_config_;
|
||
|
}
|
||
|
|
||
|
// Returns the PAC script data that was selected by ProxyScriptDecider.
|
||
|
// Should only be called upon completion of the initialization.
|
||
|
const scoped_refptr<ProxyResolverScriptData>& script_data() {
|
||
|
DCHECK_EQ(STATE_NONE, next_state_);
|
||
|
return script_data_;
|
||
|
}
|
||
|
|
||
|
LoadState GetLoadState() const {
|
||
|
if (next_state_ == STATE_DECIDE_PROXY_SCRIPT_COMPLETE) {
|
||
|
// In addition to downloading, this state may also include the stall time
|
||
|
// after network change events (kDelayAfterNetworkChangesMs).
|
||
|
return LOAD_STATE_DOWNLOADING_PROXY_SCRIPT;
|
||
|
}
|
||
|
return LOAD_STATE_RESOLVING_PROXY_FOR_URL;
|
||
|
}
|
||
|
|
||
|
// This must be called before the HostResolver is torn down.
|
||
|
void OnShutdown() {
|
||
|
if (decider_)
|
||
|
decider_->OnShutdown();
|
||
|
}
|
||
|
|
||
|
void set_quick_check_enabled(bool enabled) { quick_check_enabled_ = enabled; }
|
||
|
bool quick_check_enabled() const { return quick_check_enabled_; }
|
||
|
|
||
|
private:
|
||
|
enum State {
|
||
|
STATE_NONE,
|
||
|
STATE_DECIDE_PROXY_SCRIPT,
|
||
|
STATE_DECIDE_PROXY_SCRIPT_COMPLETE,
|
||
|
STATE_CREATE_RESOLVER,
|
||
|
STATE_CREATE_RESOLVER_COMPLETE,
|
||
|
};
|
||
|
|
||
|
int DoLoop(int result) {
|
||
|
DCHECK_NE(next_state_, STATE_NONE);
|
||
|
int rv = result;
|
||
|
do {
|
||
|
State state = next_state_;
|
||
|
next_state_ = STATE_NONE;
|
||
|
switch (state) {
|
||
|
case STATE_DECIDE_PROXY_SCRIPT:
|
||
|
DCHECK_EQ(OK, rv);
|
||
|
rv = DoDecideProxyScript();
|
||
|
break;
|
||
|
case STATE_DECIDE_PROXY_SCRIPT_COMPLETE:
|
||
|
rv = DoDecideProxyScriptComplete(rv);
|
||
|
break;
|
||
|
case STATE_CREATE_RESOLVER:
|
||
|
DCHECK_EQ(OK, rv);
|
||
|
rv = DoCreateResolver();
|
||
|
break;
|
||
|
case STATE_CREATE_RESOLVER_COMPLETE:
|
||
|
rv = DoCreateResolverComplete(rv);
|
||
|
break;
|
||
|
default:
|
||
|
NOTREACHED() << "bad state: " << state;
|
||
|
rv = ERR_UNEXPECTED;
|
||
|
break;
|
||
|
}
|
||
|
} while (rv != ERR_IO_PENDING && next_state_ != STATE_NONE);
|
||
|
return rv;
|
||
|
}
|
||
|
|
||
|
int DoDecideProxyScript() {
|
||
|
next_state_ = STATE_DECIDE_PROXY_SCRIPT_COMPLETE;
|
||
|
|
||
|
return decider_->Start(
|
||
|
config_, wait_delay_, proxy_resolver_factory_->expects_pac_bytes(),
|
||
|
base::Bind(&InitProxyResolver::OnIOCompletion, base::Unretained(this)));
|
||
|
}
|
||
|
|
||
|
int DoDecideProxyScriptComplete(int result) {
|
||
|
if (result != OK)
|
||
|
return result;
|
||
|
|
||
|
effective_config_ = decider_->effective_config();
|
||
|
script_data_ = decider_->script_data();
|
||
|
|
||
|
next_state_ = STATE_CREATE_RESOLVER;
|
||
|
return OK;
|
||
|
}
|
||
|
|
||
|
int DoCreateResolver() {
|
||
|
DCHECK(script_data_.get());
|
||
|
// TODO(eroman): Should log this latency to the NetLog.
|
||
|
next_state_ = STATE_CREATE_RESOLVER_COMPLETE;
|
||
|
return proxy_resolver_factory_->CreateProxyResolver(
|
||
|
script_data_, proxy_resolver_,
|
||
|
base::Bind(&InitProxyResolver::OnIOCompletion, base::Unretained(this)),
|
||
|
&create_resolver_request_);
|
||
|
}
|
||
|
|
||
|
int DoCreateResolverComplete(int result) {
|
||
|
if (result != OK)
|
||
|
proxy_resolver_->reset();
|
||
|
return result;
|
||
|
}
|
||
|
|
||
|
void OnIOCompletion(int result) {
|
||
|
DCHECK_NE(STATE_NONE, next_state_);
|
||
|
int rv = DoLoop(result);
|
||
|
if (rv != ERR_IO_PENDING)
|
||
|
DoCallback(rv);
|
||
|
}
|
||
|
|
||
|
void DoCallback(int result) {
|
||
|
DCHECK_NE(ERR_IO_PENDING, result);
|
||
|
callback_.Run(result);
|
||
|
}
|
||
|
|
||
|
ProxyConfig config_;
|
||
|
ProxyConfig effective_config_;
|
||
|
scoped_refptr<ProxyResolverScriptData> script_data_;
|
||
|
TimeDelta wait_delay_;
|
||
|
std::unique_ptr<ProxyScriptDecider> decider_;
|
||
|
ProxyResolverFactory* proxy_resolver_factory_;
|
||
|
std::unique_ptr<ProxyResolverFactory::Request> create_resolver_request_;
|
||
|
std::unique_ptr<ProxyResolver>* proxy_resolver_;
|
||
|
CompletionCallback callback_;
|
||
|
State next_state_;
|
||
|
bool quick_check_enabled_;
|
||
|
|
||
|
DISALLOW_COPY_AND_ASSIGN(InitProxyResolver);
|
||
|
};
|
||
|
|
||
|
// ProxyService::ProxyScriptDeciderPoller -------------------------------------
|
||
|
|
||
|
// This helper class encapsulates the logic to schedule and run periodic
|
||
|
// background checks to see if the PAC script (or effective proxy configuration)
|
||
|
// has changed. If a change is detected, then the caller will be notified via
|
||
|
// the ChangeCallback.
|
||
|
class ProxyService::ProxyScriptDeciderPoller {
|
||
|
public:
|
||
|
typedef base::Callback<void(int, ProxyResolverScriptData*,
|
||
|
const ProxyConfig&)> ChangeCallback;
|
||
|
|
||
|
// Builds a poller helper, and starts polling for updates. Whenever a change
|
||
|
// is observed, |callback| will be invoked with the details.
|
||
|
//
|
||
|
// |config| specifies the (unresolved) proxy configuration to poll.
|
||
|
// |proxy_resolver_expects_pac_bytes| the type of proxy resolver we expect
|
||
|
// to use the resulting script data with
|
||
|
// (so it can choose the right format).
|
||
|
// |proxy_script_fetcher| this pointer must remain alive throughout our
|
||
|
// lifetime. It is the dependency that will be used
|
||
|
// for downloading proxy scripts.
|
||
|
// |dhcp_proxy_script_fetcher| similar to |proxy_script_fetcher|, but for
|
||
|
// the DHCP dependency.
|
||
|
// |init_net_error| This is the initial network error (possibly success)
|
||
|
// encountered by the first PAC fetch attempt. We use it
|
||
|
// to schedule updates more aggressively if the initial
|
||
|
// fetch resulted in an error.
|
||
|
// |init_script_data| the initial script data from the PAC fetch attempt.
|
||
|
// This is the baseline used to determine when the
|
||
|
// script's contents have changed.
|
||
|
// |net_log| the NetLog to log progress into.
|
||
|
ProxyScriptDeciderPoller(ChangeCallback callback,
|
||
|
const ProxyConfig& config,
|
||
|
bool proxy_resolver_expects_pac_bytes,
|
||
|
ProxyScriptFetcher* proxy_script_fetcher,
|
||
|
DhcpProxyScriptFetcher* dhcp_proxy_script_fetcher,
|
||
|
int init_net_error,
|
||
|
const scoped_refptr<ProxyResolverScriptData>&
|
||
|
init_script_data,
|
||
|
NetLog* net_log)
|
||
|
: change_callback_(callback),
|
||
|
config_(config),
|
||
|
proxy_resolver_expects_pac_bytes_(proxy_resolver_expects_pac_bytes),
|
||
|
proxy_script_fetcher_(proxy_script_fetcher),
|
||
|
dhcp_proxy_script_fetcher_(dhcp_proxy_script_fetcher),
|
||
|
last_error_(init_net_error),
|
||
|
last_script_data_(init_script_data),
|
||
|
last_poll_time_(TimeTicks::Now()),
|
||
|
weak_factory_(this) {
|
||
|
// Set the initial poll delay.
|
||
|
next_poll_mode_ = poll_policy()->GetNextDelay(
|
||
|
last_error_, TimeDelta::FromSeconds(-1), &next_poll_delay_);
|
||
|
TryToStartNextPoll(false);
|
||
|
}
|
||
|
|
||
|
void OnLazyPoll() {
|
||
|
// We have just been notified of network activity. Use this opportunity to
|
||
|
// see if we can start our next poll.
|
||
|
TryToStartNextPoll(true);
|
||
|
}
|
||
|
|
||
|
static const PacPollPolicy* set_policy(const PacPollPolicy* policy) {
|
||
|
const PacPollPolicy* prev = poll_policy_;
|
||
|
poll_policy_ = policy;
|
||
|
return prev;
|
||
|
}
|
||
|
|
||
|
void set_quick_check_enabled(bool enabled) { quick_check_enabled_ = enabled; }
|
||
|
bool quick_check_enabled() const { return quick_check_enabled_; }
|
||
|
|
||
|
private:
|
||
|
// Returns the effective poll policy (the one injected by unit-tests, or the
|
||
|
// default).
|
||
|
const PacPollPolicy* poll_policy() {
|
||
|
if (poll_policy_)
|
||
|
return poll_policy_;
|
||
|
return &default_poll_policy_;
|
||
|
}
|
||
|
|
||
|
void StartPollTimer() {
|
||
|
DCHECK(!decider_.get());
|
||
|
|
||
|
base::ThreadTaskRunnerHandle::Get()->PostDelayedTask(
|
||
|
FROM_HERE, base::Bind(&ProxyScriptDeciderPoller::DoPoll,
|
||
|
weak_factory_.GetWeakPtr()),
|
||
|
next_poll_delay_);
|
||
|
}
|
||
|
|
||
|
void TryToStartNextPoll(bool triggered_by_activity) {
|
||
|
switch (next_poll_mode_) {
|
||
|
case PacPollPolicy::MODE_USE_TIMER:
|
||
|
if (!triggered_by_activity)
|
||
|
StartPollTimer();
|
||
|
break;
|
||
|
|
||
|
case PacPollPolicy::MODE_START_AFTER_ACTIVITY:
|
||
|
if (triggered_by_activity && !decider_.get()) {
|
||
|
TimeDelta elapsed_time = TimeTicks::Now() - last_poll_time_;
|
||
|
if (elapsed_time >= next_poll_delay_)
|
||
|
DoPoll();
|
||
|
}
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
void DoPoll() {
|
||
|
last_poll_time_ = TimeTicks::Now();
|
||
|
|
||
|
// Start the proxy script decider to see if anything has changed.
|
||
|
// TODO(eroman): Pass a proper NetLog rather than NULL.
|
||
|
decider_.reset(new ProxyScriptDecider(
|
||
|
proxy_script_fetcher_, dhcp_proxy_script_fetcher_, NULL));
|
||
|
decider_->set_quick_check_enabled(quick_check_enabled_);
|
||
|
int result = decider_->Start(
|
||
|
config_, TimeDelta(), proxy_resolver_expects_pac_bytes_,
|
||
|
base::Bind(&ProxyScriptDeciderPoller::OnProxyScriptDeciderCompleted,
|
||
|
base::Unretained(this)));
|
||
|
|
||
|
if (result != ERR_IO_PENDING)
|
||
|
OnProxyScriptDeciderCompleted(result);
|
||
|
}
|
||
|
|
||
|
void OnProxyScriptDeciderCompleted(int result) {
|
||
|
if (HasScriptDataChanged(result, decider_->script_data())) {
|
||
|
// Something has changed, we must notify the ProxyService so it can
|
||
|
// re-initialize its ProxyResolver. Note that we post a notification task
|
||
|
// rather than calling it directly -- this is done to avoid an ugly
|
||
|
// destruction sequence, since |this| might be destroyed as a result of
|
||
|
// the notification.
|
||
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
||
|
FROM_HERE,
|
||
|
base::Bind(&ProxyScriptDeciderPoller::NotifyProxyServiceOfChange,
|
||
|
weak_factory_.GetWeakPtr(), result,
|
||
|
decider_->script_data(),
|
||
|
decider_->effective_config()));
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
decider_.reset();
|
||
|
|
||
|
// Decide when the next poll should take place, and possibly start the
|
||
|
// next timer.
|
||
|
next_poll_mode_ = poll_policy()->GetNextDelay(
|
||
|
last_error_, next_poll_delay_, &next_poll_delay_);
|
||
|
TryToStartNextPoll(false);
|
||
|
}
|
||
|
|
||
|
bool HasScriptDataChanged(int result,
|
||
|
const scoped_refptr<ProxyResolverScriptData>& script_data) {
|
||
|
if (result != last_error_) {
|
||
|
// Something changed -- it was failing before and now it succeeded, or
|
||
|
// conversely it succeeded before and now it failed. Or it failed in
|
||
|
// both cases, however the specific failure error codes differ.
|
||
|
return true;
|
||
|
}
|
||
|
|
||
|
if (result != OK) {
|
||
|
// If it failed last time and failed again with the same error code this
|
||
|
// time, then nothing has actually changed.
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
// Otherwise if it succeeded both this time and last time, we need to look
|
||
|
// closer and see if we ended up downloading different content for the PAC
|
||
|
// script.
|
||
|
return !script_data->Equals(last_script_data_.get());
|
||
|
}
|
||
|
|
||
|
void NotifyProxyServiceOfChange(
|
||
|
int result,
|
||
|
const scoped_refptr<ProxyResolverScriptData>& script_data,
|
||
|
const ProxyConfig& effective_config) {
|
||
|
// Note that |this| may be deleted after calling into the ProxyService.
|
||
|
change_callback_.Run(result, script_data.get(), effective_config);
|
||
|
}
|
||
|
|
||
|
ChangeCallback change_callback_;
|
||
|
ProxyConfig config_;
|
||
|
bool proxy_resolver_expects_pac_bytes_;
|
||
|
ProxyScriptFetcher* proxy_script_fetcher_;
|
||
|
DhcpProxyScriptFetcher* dhcp_proxy_script_fetcher_;
|
||
|
|
||
|
int last_error_;
|
||
|
scoped_refptr<ProxyResolverScriptData> last_script_data_;
|
||
|
|
||
|
std::unique_ptr<ProxyScriptDecider> decider_;
|
||
|
TimeDelta next_poll_delay_;
|
||
|
PacPollPolicy::Mode next_poll_mode_;
|
||
|
|
||
|
TimeTicks last_poll_time_;
|
||
|
|
||
|
// Polling policy injected by unit-tests. Otherwise this is NULL and the
|
||
|
// default policy will be used.
|
||
|
static const PacPollPolicy* poll_policy_;
|
||
|
|
||
|
const DefaultPollPolicy default_poll_policy_;
|
||
|
|
||
|
bool quick_check_enabled_;
|
||
|
|
||
|
base::WeakPtrFactory<ProxyScriptDeciderPoller> weak_factory_;
|
||
|
|
||
|
DISALLOW_COPY_AND_ASSIGN(ProxyScriptDeciderPoller);
|
||
|
};
|
||
|
|
||
|
// static
|
||
|
const ProxyService::PacPollPolicy*
|
||
|
ProxyService::ProxyScriptDeciderPoller::poll_policy_ = NULL;
|
||
|
|
||
|
// ProxyService::PacRequest ---------------------------------------------------
|
||
|
|
||
|
class ProxyService::PacRequest
|
||
|
: public base::RefCounted<ProxyService::PacRequest> {
|
||
|
public:
|
||
|
PacRequest(ProxyService* service,
|
||
|
const GURL& url,
|
||
|
const std::string& method,
|
||
|
ProxyDelegate* proxy_delegate,
|
||
|
ProxyInfo* results,
|
||
|
const CompletionCallback& user_callback,
|
||
|
const NetLogWithSource& net_log)
|
||
|
: service_(service),
|
||
|
user_callback_(user_callback),
|
||
|
results_(results),
|
||
|
url_(url),
|
||
|
method_(method),
|
||
|
proxy_delegate_(proxy_delegate),
|
||
|
resolve_job_(nullptr),
|
||
|
config_id_(ProxyConfig::kInvalidConfigID),
|
||
|
config_source_(PROXY_CONFIG_SOURCE_UNKNOWN),
|
||
|
net_log_(net_log),
|
||
|
creation_time_(TimeTicks::Now()) {
|
||
|
DCHECK(!user_callback.is_null());
|
||
|
}
|
||
|
|
||
|
// Starts the resolve proxy request.
|
||
|
int Start() {
|
||
|
DCHECK(!was_cancelled());
|
||
|
DCHECK(!is_started());
|
||
|
|
||
|
DCHECK(service_->config_.is_valid());
|
||
|
|
||
|
config_id_ = service_->config_.id();
|
||
|
config_source_ = service_->config_.source();
|
||
|
|
||
|
return resolver()->GetProxyForURL(
|
||
|
url_, results_,
|
||
|
base::Bind(&PacRequest::QueryComplete, base::Unretained(this)),
|
||
|
&resolve_job_, net_log_);
|
||
|
}
|
||
|
|
||
|
bool is_started() const {
|
||
|
// Note that !! casts to bool. (VS gives a warning otherwise).
|
||
|
return !!resolve_job_.get();
|
||
|
}
|
||
|
|
||
|
void StartAndCompleteCheckingForSynchronous() {
|
||
|
int rv =
|
||
|
service_->TryToCompleteSynchronously(url_, proxy_delegate_, results_);
|
||
|
if (rv == ERR_IO_PENDING)
|
||
|
rv = Start();
|
||
|
if (rv != ERR_IO_PENDING)
|
||
|
QueryComplete(rv);
|
||
|
}
|
||
|
|
||
|
void CancelResolveJob() {
|
||
|
DCHECK(is_started());
|
||
|
// The request may already be running in the resolver.
|
||
|
resolve_job_.reset();
|
||
|
DCHECK(!is_started());
|
||
|
}
|
||
|
|
||
|
void Cancel() {
|
||
|
net_log_.AddEvent(NetLogEventType::CANCELLED);
|
||
|
|
||
|
if (is_started())
|
||
|
CancelResolveJob();
|
||
|
|
||
|
// Mark as cancelled, to prevent accessing this again later.
|
||
|
service_ = NULL;
|
||
|
user_callback_.Reset();
|
||
|
results_ = NULL;
|
||
|
|
||
|
net_log_.EndEvent(NetLogEventType::PROXY_SERVICE);
|
||
|
}
|
||
|
|
||
|
// Returns true if Cancel() has been called.
|
||
|
bool was_cancelled() const {
|
||
|
return user_callback_.is_null();
|
||
|
}
|
||
|
|
||
|
// Helper to call after ProxyResolver completion (both synchronous and
|
||
|
// asynchronous). Fixes up the result that is to be returned to user.
|
||
|
int QueryDidComplete(int result_code) {
|
||
|
DCHECK(!was_cancelled());
|
||
|
|
||
|
// Clear |resolve_job_| so is_started() returns false while
|
||
|
// DidFinishResolvingProxy() runs.
|
||
|
resolve_job_.reset();
|
||
|
|
||
|
// Note that DidFinishResolvingProxy might modify |results_|.
|
||
|
int rv = service_->DidFinishResolvingProxy(url_, method_, proxy_delegate_,
|
||
|
results_, result_code, net_log_);
|
||
|
|
||
|
// Make a note in the results which configuration was in use at the
|
||
|
// time of the resolve.
|
||
|
results_->config_id_ = config_id_;
|
||
|
results_->config_source_ = config_source_;
|
||
|
results_->did_use_pac_script_ = true;
|
||
|
results_->proxy_resolve_start_time_ = creation_time_;
|
||
|
results_->proxy_resolve_end_time_ = TimeTicks::Now();
|
||
|
|
||
|
// Reset the state associated with in-progress-resolve.
|
||
|
config_id_ = ProxyConfig::kInvalidConfigID;
|
||
|
config_source_ = PROXY_CONFIG_SOURCE_UNKNOWN;
|
||
|
|
||
|
return rv;
|
||
|
}
|
||
|
|
||
|
NetLogWithSource* net_log() { return &net_log_; }
|
||
|
|
||
|
LoadState GetLoadState() const {
|
||
|
if (is_started())
|
||
|
return resolve_job_->GetLoadState();
|
||
|
return LOAD_STATE_RESOLVING_PROXY_FOR_URL;
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
friend class base::RefCounted<ProxyService::PacRequest>;
|
||
|
|
||
|
~PacRequest() {}
|
||
|
|
||
|
// Callback for when the ProxyResolver request has completed.
|
||
|
void QueryComplete(int result_code) {
|
||
|
result_code = QueryDidComplete(result_code);
|
||
|
|
||
|
// Remove this completed PacRequest from the service's pending list.
|
||
|
/// (which will probably cause deletion of |this|).
|
||
|
if (!user_callback_.is_null()) {
|
||
|
CompletionCallback callback = user_callback_;
|
||
|
service_->RemovePendingRequest(this);
|
||
|
callback.Run(result_code);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
ProxyResolver* resolver() const { return service_->resolver_.get(); }
|
||
|
|
||
|
// Note that we don't hold a reference to the ProxyService. Outstanding
|
||
|
// requests are cancelled during ~ProxyService, so this is guaranteed
|
||
|
// to be valid throughout our lifetime.
|
||
|
ProxyService* service_;
|
||
|
CompletionCallback user_callback_;
|
||
|
ProxyInfo* results_;
|
||
|
GURL url_;
|
||
|
std::string method_;
|
||
|
ProxyDelegate* proxy_delegate_;
|
||
|
std::unique_ptr<ProxyResolver::Request> resolve_job_;
|
||
|
ProxyConfig::ID config_id_; // The config id when the resolve was started.
|
||
|
ProxyConfigSource config_source_; // The source of proxy settings.
|
||
|
NetLogWithSource net_log_;
|
||
|
// Time when the request was created. Stored here rather than in |results_|
|
||
|
// because the time in |results_| will be cleared.
|
||
|
TimeTicks creation_time_;
|
||
|
};
|
||
|
|
||
|
// ProxyService ---------------------------------------------------------------
|
||
|
|
||
|
ProxyService::ProxyService(
|
||
|
std::unique_ptr<ProxyConfigService> config_service,
|
||
|
std::unique_ptr<ProxyResolverFactory> resolver_factory,
|
||
|
NetLog* net_log)
|
||
|
: resolver_factory_(std::move(resolver_factory)),
|
||
|
next_config_id_(1),
|
||
|
current_state_(STATE_NONE),
|
||
|
net_log_(net_log),
|
||
|
stall_proxy_auto_config_delay_(
|
||
|
TimeDelta::FromMilliseconds(kDelayAfterNetworkChangesMs)),
|
||
|
quick_check_enabled_(true),
|
||
|
sanitize_url_policy_(SanitizeUrlPolicy::SAFE) {
|
||
|
NetworkChangeNotifier::AddIPAddressObserver(this);
|
||
|
NetworkChangeNotifier::AddDNSObserver(this);
|
||
|
ResetConfigService(std::move(config_service));
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyService> ProxyService::CreateUsingSystemProxyResolver(
|
||
|
std::unique_ptr<ProxyConfigService> proxy_config_service,
|
||
|
NetLog* net_log) {
|
||
|
DCHECK(proxy_config_service);
|
||
|
|
||
|
if (!ProxyResolverFactoryForSystem::IsSupported()) {
|
||
|
VLOG(1) << "PAC support disabled because there is no system implementation";
|
||
|
return CreateWithoutProxyResolver(std::move(proxy_config_service), net_log);
|
||
|
}
|
||
|
|
||
|
return std::make_unique<ProxyService>(
|
||
|
std::move(proxy_config_service),
|
||
|
std::make_unique<ProxyResolverFactoryForSystem>(kDefaultNumPacThreads),
|
||
|
net_log);
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyService> ProxyService::CreateWithoutProxyResolver(
|
||
|
std::unique_ptr<ProxyConfigService> proxy_config_service,
|
||
|
NetLog* net_log) {
|
||
|
return std::make_unique<ProxyService>(
|
||
|
std::move(proxy_config_service),
|
||
|
std::make_unique<ProxyResolverFactoryForNullResolver>(), net_log);
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyService> ProxyService::CreateFixed(const ProxyConfig& pc) {
|
||
|
// TODO(eroman): This isn't quite right, won't work if |pc| specifies
|
||
|
// a PAC script.
|
||
|
return CreateUsingSystemProxyResolver(
|
||
|
std::make_unique<ProxyConfigServiceFixed>(pc), NULL);
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyService> ProxyService::CreateFixed(
|
||
|
const std::string& proxy) {
|
||
|
ProxyConfig proxy_config;
|
||
|
proxy_config.proxy_rules().ParseFromString(proxy);
|
||
|
return ProxyService::CreateFixed(proxy_config);
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyService> ProxyService::CreateDirect() {
|
||
|
return CreateDirectWithNetLog(NULL);
|
||
|
}
|
||
|
|
||
|
std::unique_ptr<ProxyService> ProxyService::CreateDirectWithNetLog(
|
||
|
NetLog* net_log) {
|
||
|
// Use direct connections.
|
||
|
return std::make_unique<ProxyService>(
|
||
|
std::make_unique<ProxyConfigServiceDirect>(),
|
||
|
std::make_unique<ProxyResolverFactoryForNullResolver>(), net_log);
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyService> ProxyService::CreateFixedFromPacResult(
|
||
|
const std::string& pac_string) {
|
||
|
// We need the settings to contain an "automatic" setting, otherwise the
|
||
|
// ProxyResolver dependency we give it will never be used.
|
||
|
std::unique_ptr<ProxyConfigService> proxy_config_service(
|
||
|
new ProxyConfigServiceFixed(ProxyConfig::CreateAutoDetect()));
|
||
|
|
||
|
return std::make_unique<ProxyService>(
|
||
|
std::move(proxy_config_service),
|
||
|
std::make_unique<ProxyResolverFactoryForPacResult>(pac_string), nullptr);
|
||
|
}
|
||
|
|
||
|
int ProxyService::ResolveProxy(const GURL& raw_url,
|
||
|
const std::string& method,
|
||
|
ProxyInfo* result,
|
||
|
const CompletionCallback& callback,
|
||
|
PacRequest** pac_request,
|
||
|
ProxyDelegate* proxy_delegate,
|
||
|
const NetLogWithSource& net_log) {
|
||
|
DCHECK(!callback.is_null());
|
||
|
return ResolveProxyHelper(raw_url, method, result, callback, pac_request,
|
||
|
proxy_delegate, net_log);
|
||
|
}
|
||
|
|
||
|
int ProxyService::ResolveProxyHelper(const GURL& raw_url,
|
||
|
const std::string& method,
|
||
|
ProxyInfo* result,
|
||
|
const CompletionCallback& callback,
|
||
|
PacRequest** pac_request,
|
||
|
ProxyDelegate* proxy_delegate,
|
||
|
const NetLogWithSource& net_log) {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
|
||
|
net_log.BeginEvent(NetLogEventType::PROXY_SERVICE);
|
||
|
|
||
|
// Notify our polling-based dependencies that a resolve is taking place.
|
||
|
// This way they can schedule their polls in response to network activity.
|
||
|
config_service_->OnLazyPoll();
|
||
|
if (script_poller_.get())
|
||
|
script_poller_->OnLazyPoll();
|
||
|
|
||
|
if (current_state_ == STATE_NONE)
|
||
|
ApplyProxyConfigIfAvailable();
|
||
|
|
||
|
// Sanitize the URL before passing it on to the proxy resolver (i.e. PAC
|
||
|
// script). The goal is to remove sensitive data (like embedded user names
|
||
|
// and password), and local data (i.e. reference fragment) which does not need
|
||
|
// to be disclosed to the resolver.
|
||
|
GURL url = SanitizeUrl(raw_url, sanitize_url_policy_);
|
||
|
|
||
|
// Check if the request can be completed right away. (This is the case when
|
||
|
// using a direct connection for example).
|
||
|
int rv = TryToCompleteSynchronously(url, proxy_delegate, result);
|
||
|
if (rv != ERR_IO_PENDING) {
|
||
|
rv = DidFinishResolvingProxy(url, method, proxy_delegate, result, rv,
|
||
|
net_log);
|
||
|
return rv;
|
||
|
}
|
||
|
|
||
|
if (callback.is_null())
|
||
|
return ERR_IO_PENDING;
|
||
|
|
||
|
scoped_refptr<PacRequest> req(new PacRequest(
|
||
|
this, url, method, proxy_delegate, result, callback, net_log));
|
||
|
|
||
|
if (current_state_ == STATE_READY) {
|
||
|
// Start the resolve request.
|
||
|
rv = req->Start();
|
||
|
if (rv != ERR_IO_PENDING)
|
||
|
return req->QueryDidComplete(rv);
|
||
|
} else {
|
||
|
req->net_log()->BeginEvent(
|
||
|
NetLogEventType::PROXY_SERVICE_WAITING_FOR_INIT_PAC);
|
||
|
}
|
||
|
|
||
|
DCHECK_EQ(ERR_IO_PENDING, rv);
|
||
|
DCHECK(!ContainsPendingRequest(req.get()));
|
||
|
pending_requests_.insert(req);
|
||
|
|
||
|
// Completion will be notified through |callback|, unless the caller cancels
|
||
|
// the request using |pac_request|.
|
||
|
if (pac_request)
|
||
|
*pac_request = req.get();
|
||
|
return rv; // ERR_IO_PENDING
|
||
|
}
|
||
|
|
||
|
bool ProxyService::TryResolveProxySynchronously(
|
||
|
const GURL& raw_url,
|
||
|
const std::string& method,
|
||
|
ProxyInfo* result,
|
||
|
ProxyDelegate* proxy_delegate,
|
||
|
const NetLogWithSource& net_log) {
|
||
|
CompletionCallback null_callback;
|
||
|
return ResolveProxyHelper(raw_url, method, result, null_callback,
|
||
|
nullptr /* pac_request*/, proxy_delegate,
|
||
|
net_log) == OK;
|
||
|
}
|
||
|
|
||
|
int ProxyService::TryToCompleteSynchronously(const GURL& url,
|
||
|
ProxyDelegate* proxy_delegate,
|
||
|
ProxyInfo* result) {
|
||
|
DCHECK_NE(STATE_NONE, current_state_);
|
||
|
|
||
|
if (current_state_ != STATE_READY)
|
||
|
return ERR_IO_PENDING; // Still initializing.
|
||
|
|
||
|
DCHECK_NE(config_.id(), ProxyConfig::kInvalidConfigID);
|
||
|
|
||
|
// If it was impossible to fetch or parse the PAC script, we cannot complete
|
||
|
// the request here and bail out.
|
||
|
if (permanent_error_ != OK)
|
||
|
return permanent_error_;
|
||
|
|
||
|
if (config_.HasAutomaticSettings())
|
||
|
return ERR_IO_PENDING; // Must submit the request to the proxy resolver.
|
||
|
|
||
|
// Use the manual proxy settings.
|
||
|
config_.proxy_rules().Apply(url, result);
|
||
|
result->config_source_ = config_.source();
|
||
|
result->config_id_ = config_.id();
|
||
|
|
||
|
return OK;
|
||
|
}
|
||
|
|
||
|
ProxyService::~ProxyService() {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
NetworkChangeNotifier::RemoveIPAddressObserver(this);
|
||
|
NetworkChangeNotifier::RemoveDNSObserver(this);
|
||
|
config_service_->RemoveObserver(this);
|
||
|
|
||
|
// Cancel any inprogress requests.
|
||
|
for (PendingRequests::iterator it = pending_requests_.begin();
|
||
|
it != pending_requests_.end();
|
||
|
++it) {
|
||
|
(*it)->Cancel();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
void ProxyService::SuspendAllPendingRequests() {
|
||
|
for (PendingRequests::iterator it = pending_requests_.begin();
|
||
|
it != pending_requests_.end();
|
||
|
++it) {
|
||
|
PacRequest* req = it->get();
|
||
|
if (req->is_started()) {
|
||
|
req->CancelResolveJob();
|
||
|
|
||
|
req->net_log()->BeginEvent(
|
||
|
NetLogEventType::PROXY_SERVICE_WAITING_FOR_INIT_PAC);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
void ProxyService::SetReady() {
|
||
|
DCHECK(!init_proxy_resolver_.get());
|
||
|
current_state_ = STATE_READY;
|
||
|
|
||
|
// Make a copy in case |this| is deleted during the synchronous completion
|
||
|
// of one of the requests. If |this| is deleted then all of the PacRequest
|
||
|
// instances will be Cancel()-ed.
|
||
|
PendingRequests pending_copy = pending_requests_;
|
||
|
|
||
|
for (PendingRequests::iterator it = pending_copy.begin();
|
||
|
it != pending_copy.end();
|
||
|
++it) {
|
||
|
PacRequest* req = it->get();
|
||
|
if (!req->is_started() && !req->was_cancelled()) {
|
||
|
req->net_log()->EndEvent(
|
||
|
NetLogEventType::PROXY_SERVICE_WAITING_FOR_INIT_PAC);
|
||
|
|
||
|
// Note that we re-check for synchronous completion, in case we are
|
||
|
// no longer using a ProxyResolver (can happen if we fell-back to manual).
|
||
|
req->StartAndCompleteCheckingForSynchronous();
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
void ProxyService::ApplyProxyConfigIfAvailable() {
|
||
|
DCHECK_EQ(STATE_NONE, current_state_);
|
||
|
|
||
|
config_service_->OnLazyPoll();
|
||
|
|
||
|
// If we have already fetched the configuration, start applying it.
|
||
|
if (fetched_config_.is_valid()) {
|
||
|
InitializeUsingLastFetchedConfig();
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
// Otherwise we need to first fetch the configuration.
|
||
|
current_state_ = STATE_WAITING_FOR_PROXY_CONFIG;
|
||
|
|
||
|
// Retrieve the current proxy configuration from the ProxyConfigService.
|
||
|
// If a configuration is not available yet, we will get called back later
|
||
|
// by our ProxyConfigService::Observer once it changes.
|
||
|
ProxyConfig config;
|
||
|
ProxyConfigService::ConfigAvailability availability =
|
||
|
config_service_->GetLatestProxyConfig(&config);
|
||
|
if (availability != ProxyConfigService::CONFIG_PENDING)
|
||
|
OnProxyConfigChanged(config, availability);
|
||
|
}
|
||
|
|
||
|
void ProxyService::OnInitProxyResolverComplete(int result) {
|
||
|
DCHECK_EQ(STATE_WAITING_FOR_INIT_PROXY_RESOLVER, current_state_);
|
||
|
DCHECK(init_proxy_resolver_.get());
|
||
|
DCHECK(fetched_config_.HasAutomaticSettings());
|
||
|
config_ = init_proxy_resolver_->effective_config();
|
||
|
|
||
|
// At this point we have decided which proxy settings to use (i.e. which PAC
|
||
|
// script if any). We start up a background poller to periodically revisit
|
||
|
// this decision. If the contents of the PAC script change, or if the
|
||
|
// result of proxy auto-discovery changes, this poller will notice it and
|
||
|
// will trigger a re-initialization using the newly discovered PAC.
|
||
|
script_poller_.reset(new ProxyScriptDeciderPoller(
|
||
|
base::Bind(&ProxyService::InitializeUsingDecidedConfig,
|
||
|
base::Unretained(this)),
|
||
|
fetched_config_, resolver_factory_->expects_pac_bytes(),
|
||
|
proxy_script_fetcher_.get(), dhcp_proxy_script_fetcher_.get(), result,
|
||
|
init_proxy_resolver_->script_data(), NULL));
|
||
|
script_poller_->set_quick_check_enabled(quick_check_enabled_);
|
||
|
|
||
|
init_proxy_resolver_.reset();
|
||
|
|
||
|
if (result != OK) {
|
||
|
if (fetched_config_.pac_mandatory()) {
|
||
|
VLOG(1) << "Failed configuring with mandatory PAC script, blocking all "
|
||
|
"traffic.";
|
||
|
config_ = fetched_config_;
|
||
|
result = ERR_MANDATORY_PROXY_CONFIGURATION_FAILED;
|
||
|
} else {
|
||
|
VLOG(1) << "Failed configuring with PAC script, falling-back to manual "
|
||
|
"proxy servers.";
|
||
|
config_ = fetched_config_;
|
||
|
config_.ClearAutomaticSettings();
|
||
|
result = OK;
|
||
|
}
|
||
|
}
|
||
|
permanent_error_ = result;
|
||
|
|
||
|
// TODO(eroman): Make this ID unique in the case where configuration changed
|
||
|
// due to ProxyScriptDeciderPoller.
|
||
|
config_.set_id(fetched_config_.id());
|
||
|
config_.set_source(fetched_config_.source());
|
||
|
|
||
|
// Resume any requests which we had to defer until the PAC script was
|
||
|
// downloaded.
|
||
|
SetReady();
|
||
|
}
|
||
|
|
||
|
int ProxyService::ReconsiderProxyAfterError(const GURL& url,
|
||
|
const std::string& method,
|
||
|
int net_error,
|
||
|
ProxyInfo* result,
|
||
|
const CompletionCallback& callback,
|
||
|
PacRequest** pac_request,
|
||
|
ProxyDelegate* proxy_delegate,
|
||
|
const NetLogWithSource& net_log) {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
|
||
|
// Check to see if we have a new config since ResolveProxy was called. We
|
||
|
// want to re-run ResolveProxy in two cases: 1) we have a new config, or 2) a
|
||
|
// direct connection failed and we never tried the current config.
|
||
|
|
||
|
DCHECK(result);
|
||
|
bool re_resolve = result->config_id_ != config_.id();
|
||
|
|
||
|
if (re_resolve) {
|
||
|
// If we have a new config or the config was never tried, we delete the
|
||
|
// list of bad proxies and we try again.
|
||
|
proxy_retry_info_.clear();
|
||
|
return ResolveProxy(url, method, result, callback, pac_request,
|
||
|
proxy_delegate, net_log);
|
||
|
}
|
||
|
|
||
|
DCHECK(!result->is_empty());
|
||
|
ProxyServer bad_proxy = result->proxy_server();
|
||
|
|
||
|
// We don't have new proxy settings to try, try to fallback to the next proxy
|
||
|
// in the list.
|
||
|
bool did_fallback = result->Fallback(net_error, net_log);
|
||
|
|
||
|
// Return synchronous failure if there is nothing left to fall-back to.
|
||
|
// TODO(eroman): This is a yucky API, clean it up.
|
||
|
return did_fallback ? OK : ERR_FAILED;
|
||
|
}
|
||
|
|
||
|
bool ProxyService::MarkProxiesAsBadUntil(
|
||
|
const ProxyInfo& result,
|
||
|
base::TimeDelta retry_delay,
|
||
|
const std::vector<ProxyServer>& additional_bad_proxies,
|
||
|
const NetLogWithSource& net_log) {
|
||
|
result.proxy_list_.UpdateRetryInfoOnFallback(&proxy_retry_info_, retry_delay,
|
||
|
false, additional_bad_proxies,
|
||
|
OK, net_log);
|
||
|
return result.proxy_list_.size() > (additional_bad_proxies.size() + 1);
|
||
|
}
|
||
|
|
||
|
void ProxyService::ReportSuccess(const ProxyInfo& result,
|
||
|
ProxyDelegate* proxy_delegate) {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
|
||
|
const ProxyRetryInfoMap& new_retry_info = result.proxy_retry_info();
|
||
|
if (new_retry_info.empty())
|
||
|
return;
|
||
|
|
||
|
for (ProxyRetryInfoMap::const_iterator iter = new_retry_info.begin();
|
||
|
iter != new_retry_info.end(); ++iter) {
|
||
|
ProxyRetryInfoMap::iterator existing = proxy_retry_info_.find(iter->first);
|
||
|
if (existing == proxy_retry_info_.end()) {
|
||
|
proxy_retry_info_[iter->first] = iter->second;
|
||
|
if (proxy_delegate) {
|
||
|
const ProxyServer& bad_proxy =
|
||
|
ProxyServer::FromURI(iter->first, ProxyServer::SCHEME_HTTP);
|
||
|
const ProxyRetryInfo& proxy_retry_info = iter->second;
|
||
|
proxy_delegate->OnFallback(bad_proxy, proxy_retry_info.net_error);
|
||
|
}
|
||
|
}
|
||
|
else if (existing->second.bad_until < iter->second.bad_until)
|
||
|
existing->second.bad_until = iter->second.bad_until;
|
||
|
}
|
||
|
if (net_log_) {
|
||
|
net_log_->AddGlobalEntry(
|
||
|
NetLogEventType::BAD_PROXY_LIST_REPORTED,
|
||
|
base::Bind(&NetLogBadProxyListCallback, &new_retry_info));
|
||
|
}
|
||
|
}
|
||
|
|
||
|
void ProxyService::CancelPacRequest(PacRequest* req) {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
DCHECK(req);
|
||
|
req->Cancel();
|
||
|
RemovePendingRequest(req);
|
||
|
}
|
||
|
|
||
|
LoadState ProxyService::GetLoadState(const PacRequest* req) const {
|
||
|
CHECK(req);
|
||
|
if (current_state_ == STATE_WAITING_FOR_INIT_PROXY_RESOLVER)
|
||
|
return init_proxy_resolver_->GetLoadState();
|
||
|
return req->GetLoadState();
|
||
|
}
|
||
|
|
||
|
bool ProxyService::ContainsPendingRequest(PacRequest* req) {
|
||
|
return pending_requests_.count(req) == 1;
|
||
|
}
|
||
|
|
||
|
void ProxyService::RemovePendingRequest(PacRequest* req) {
|
||
|
DCHECK(ContainsPendingRequest(req));
|
||
|
pending_requests_.erase(req);
|
||
|
}
|
||
|
|
||
|
int ProxyService::DidFinishResolvingProxy(const GURL& url,
|
||
|
const std::string& method,
|
||
|
ProxyDelegate* proxy_delegate,
|
||
|
ProxyInfo* result,
|
||
|
int result_code,
|
||
|
const NetLogWithSource& net_log) {
|
||
|
// Log the result of the proxy resolution.
|
||
|
if (result_code == OK) {
|
||
|
// Allow the proxy delegate to interpose on the resolution decision,
|
||
|
// possibly modifying the ProxyInfo.
|
||
|
if (proxy_delegate)
|
||
|
proxy_delegate->OnResolveProxy(url, method, proxy_retry_info_, result);
|
||
|
|
||
|
net_log.AddEvent(NetLogEventType::PROXY_SERVICE_RESOLVED_PROXY_LIST,
|
||
|
base::Bind(&NetLogFinishedResolvingProxyCallback, result));
|
||
|
|
||
|
// This check is done to only log the NetLog event when necessary, it's
|
||
|
// not a performance optimization.
|
||
|
if (!proxy_retry_info_.empty()) {
|
||
|
result->DeprioritizeBadProxies(proxy_retry_info_);
|
||
|
net_log.AddEvent(
|
||
|
NetLogEventType::PROXY_SERVICE_DEPRIORITIZED_BAD_PROXIES,
|
||
|
base::Bind(&NetLogFinishedResolvingProxyCallback, result));
|
||
|
}
|
||
|
} else {
|
||
|
net_log.AddEventWithNetErrorCode(
|
||
|
NetLogEventType::PROXY_SERVICE_RESOLVED_PROXY_LIST, result_code);
|
||
|
|
||
|
bool reset_config = result_code == ERR_PAC_SCRIPT_TERMINATED;
|
||
|
if (!config_.pac_mandatory()) {
|
||
|
// Fall-back to direct when the proxy resolver fails. This corresponds
|
||
|
// with a javascript runtime error in the PAC script.
|
||
|
//
|
||
|
// This implicit fall-back to direct matches Firefox 3.5 and
|
||
|
// Internet Explorer 8. For more information, see:
|
||
|
//
|
||
|
// http://www.chromium.org/developers/design-documents/proxy-settings-fallback
|
||
|
result->UseDirect();
|
||
|
result_code = OK;
|
||
|
|
||
|
// Allow the proxy delegate to interpose on the resolution decision,
|
||
|
// possibly modifying the ProxyInfo.
|
||
|
if (proxy_delegate)
|
||
|
proxy_delegate->OnResolveProxy(url, method, proxy_retry_info_, result);
|
||
|
} else {
|
||
|
result_code = ERR_MANDATORY_PROXY_CONFIGURATION_FAILED;
|
||
|
}
|
||
|
if (reset_config) {
|
||
|
ResetProxyConfig(false);
|
||
|
// If the ProxyResolver crashed, force it to be re-initialized for the
|
||
|
// next request by resetting the proxy config. If there are other pending
|
||
|
// requests, trigger the recreation immediately so those requests retry.
|
||
|
if (pending_requests_.size() > 1)
|
||
|
ApplyProxyConfigIfAvailable();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
net_log.EndEvent(NetLogEventType::PROXY_SERVICE);
|
||
|
return result_code;
|
||
|
}
|
||
|
|
||
|
void ProxyService::SetProxyScriptFetchers(
|
||
|
std::unique_ptr<ProxyScriptFetcher> proxy_script_fetcher,
|
||
|
std::unique_ptr<DhcpProxyScriptFetcher> dhcp_proxy_script_fetcher) {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
State previous_state = ResetProxyConfig(false);
|
||
|
proxy_script_fetcher_ = std::move(proxy_script_fetcher);
|
||
|
dhcp_proxy_script_fetcher_ = std::move(dhcp_proxy_script_fetcher);
|
||
|
if (previous_state != STATE_NONE)
|
||
|
ApplyProxyConfigIfAvailable();
|
||
|
}
|
||
|
|
||
|
void ProxyService::OnShutdown() {
|
||
|
// Order here does not matter for correctness. |init_proxy_resolver_| is first
|
||
|
// because shutting it down also cancels its requests using the fetcher.
|
||
|
if (init_proxy_resolver_)
|
||
|
init_proxy_resolver_->OnShutdown();
|
||
|
if (proxy_script_fetcher_)
|
||
|
proxy_script_fetcher_->OnShutdown();
|
||
|
if (dhcp_proxy_script_fetcher_)
|
||
|
dhcp_proxy_script_fetcher_->OnShutdown();
|
||
|
}
|
||
|
|
||
|
ProxyScriptFetcher* ProxyService::GetProxyScriptFetcher() const {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
return proxy_script_fetcher_.get();
|
||
|
}
|
||
|
|
||
|
ProxyService::State ProxyService::ResetProxyConfig(bool reset_fetched_config) {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
State previous_state = current_state_;
|
||
|
|
||
|
permanent_error_ = OK;
|
||
|
proxy_retry_info_.clear();
|
||
|
script_poller_.reset();
|
||
|
init_proxy_resolver_.reset();
|
||
|
SuspendAllPendingRequests();
|
||
|
resolver_.reset();
|
||
|
config_ = ProxyConfig();
|
||
|
if (reset_fetched_config)
|
||
|
fetched_config_ = ProxyConfig();
|
||
|
current_state_ = STATE_NONE;
|
||
|
|
||
|
return previous_state;
|
||
|
}
|
||
|
|
||
|
void ProxyService::ResetConfigService(
|
||
|
std::unique_ptr<ProxyConfigService> new_proxy_config_service) {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
State previous_state = ResetProxyConfig(true);
|
||
|
|
||
|
// Release the old configuration service.
|
||
|
if (config_service_.get())
|
||
|
config_service_->RemoveObserver(this);
|
||
|
|
||
|
// Set the new configuration service.
|
||
|
config_service_ = std::move(new_proxy_config_service);
|
||
|
config_service_->AddObserver(this);
|
||
|
|
||
|
if (previous_state != STATE_NONE)
|
||
|
ApplyProxyConfigIfAvailable();
|
||
|
}
|
||
|
|
||
|
void ProxyService::ForceReloadProxyConfig() {
|
||
|
DCHECK_CALLED_ON_VALID_THREAD(thread_checker_);
|
||
|
ResetProxyConfig(false);
|
||
|
ApplyProxyConfigIfAvailable();
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyConfigService>
|
||
|
ProxyService::CreateSystemProxyConfigService(
|
||
|
const scoped_refptr<base::SequencedTaskRunner>& io_task_runner) {
|
||
|
#if defined(OS_WIN)
|
||
|
return std::make_unique<ProxyConfigServiceWin>();
|
||
|
#elif defined(OS_IOS)
|
||
|
return std::make_unique<ProxyConfigServiceIOS>();
|
||
|
#elif defined(OS_MACOSX)
|
||
|
return std::make_unique<ProxyConfigServiceMac>(io_task_runner);
|
||
|
#elif defined(OS_CHROMEOS)
|
||
|
LOG(ERROR) << "ProxyConfigService for ChromeOS should be created in "
|
||
|
<< "profile_io_data.cc::CreateProxyConfigService and this should "
|
||
|
<< "be used only for examples.";
|
||
|
return std::make_unique<UnsetProxyConfigService>();
|
||
|
#elif defined(OS_LINUX)
|
||
|
std::unique_ptr<ProxyConfigServiceLinux> linux_config_service(
|
||
|
new ProxyConfigServiceLinux());
|
||
|
|
||
|
// Assume we got called on the thread that runs the default glib
|
||
|
// main loop, so the current thread is where we should be running
|
||
|
// gconf calls from.
|
||
|
scoped_refptr<base::SingleThreadTaskRunner> glib_thread_task_runner =
|
||
|
base::ThreadTaskRunnerHandle::Get();
|
||
|
|
||
|
// Synchronously fetch the current proxy config (since we are running on
|
||
|
// glib_default_loop). Additionally register for notifications (delivered in
|
||
|
// either |glib_default_loop| or an internal sequenced task runner) to
|
||
|
// keep us updated when the proxy config changes.
|
||
|
linux_config_service->SetupAndFetchInitialConfig(glib_thread_task_runner,
|
||
|
io_task_runner);
|
||
|
|
||
|
return std::move(linux_config_service);
|
||
|
#elif defined(OS_ANDROID)
|
||
|
return std::make_unique<ProxyConfigServiceAndroid>(
|
||
|
io_task_runner, base::ThreadTaskRunnerHandle::Get());
|
||
|
#else
|
||
|
LOG(WARNING) << "Failed to choose a system proxy settings fetcher "
|
||
|
"for this platform.";
|
||
|
return std::make_unique<ProxyConfigServiceDirect>();
|
||
|
#endif
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
const ProxyService::PacPollPolicy* ProxyService::set_pac_script_poll_policy(
|
||
|
const PacPollPolicy* policy) {
|
||
|
return ProxyScriptDeciderPoller::set_policy(policy);
|
||
|
}
|
||
|
|
||
|
// static
|
||
|
std::unique_ptr<ProxyService::PacPollPolicy>
|
||
|
ProxyService::CreateDefaultPacPollPolicy() {
|
||
|
return std::unique_ptr<PacPollPolicy>(new DefaultPollPolicy());
|
||
|
}
|
||
|
|
||
|
void ProxyService::OnProxyConfigChanged(
|
||
|
const ProxyConfig& config,
|
||
|
ProxyConfigService::ConfigAvailability availability) {
|
||
|
// Retrieve the current proxy configuration from the ProxyConfigService.
|
||
|
// If a configuration is not available yet, we will get called back later
|
||
|
// by our ProxyConfigService::Observer once it changes.
|
||
|
ProxyConfig effective_config;
|
||
|
switch (availability) {
|
||
|
case ProxyConfigService::CONFIG_PENDING:
|
||
|
// ProxyConfigService implementors should never pass CONFIG_PENDING.
|
||
|
NOTREACHED() << "Proxy config change with CONFIG_PENDING availability!";
|
||
|
return;
|
||
|
case ProxyConfigService::CONFIG_VALID:
|
||
|
effective_config = config;
|
||
|
break;
|
||
|
case ProxyConfigService::CONFIG_UNSET:
|
||
|
effective_config = ProxyConfig::CreateDirect();
|
||
|
break;
|
||
|
}
|
||
|
|
||
|
// Emit the proxy settings change to the NetLog stream.
|
||
|
if (net_log_) {
|
||
|
net_log_->AddGlobalEntry(NetLogEventType::PROXY_CONFIG_CHANGED,
|
||
|
base::Bind(&NetLogProxyConfigChangedCallback,
|
||
|
&fetched_config_, &effective_config));
|
||
|
}
|
||
|
|
||
|
// Set the new configuration as the most recently fetched one.
|
||
|
fetched_config_ = effective_config;
|
||
|
fetched_config_.set_id(1); // Needed for a later DCHECK of is_valid().
|
||
|
|
||
|
InitializeUsingLastFetchedConfig();
|
||
|
}
|
||
|
|
||
|
void ProxyService::InitializeUsingLastFetchedConfig() {
|
||
|
ResetProxyConfig(false);
|
||
|
|
||
|
DCHECK(fetched_config_.is_valid());
|
||
|
|
||
|
// Increment the ID to reflect that the config has changed.
|
||
|
fetched_config_.set_id(next_config_id_++);
|
||
|
|
||
|
if (!fetched_config_.HasAutomaticSettings()) {
|
||
|
config_ = fetched_config_;
|
||
|
SetReady();
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
// Start downloading + testing the PAC scripts for this new configuration.
|
||
|
current_state_ = STATE_WAITING_FOR_INIT_PROXY_RESOLVER;
|
||
|
|
||
|
// If we changed networks recently, we should delay running proxy auto-config.
|
||
|
TimeDelta wait_delay =
|
||
|
stall_proxy_autoconfig_until_ - TimeTicks::Now();
|
||
|
|
||
|
init_proxy_resolver_.reset(new InitProxyResolver());
|
||
|
init_proxy_resolver_->set_quick_check_enabled(quick_check_enabled_);
|
||
|
int rv = init_proxy_resolver_->Start(
|
||
|
&resolver_, resolver_factory_.get(), proxy_script_fetcher_.get(),
|
||
|
dhcp_proxy_script_fetcher_.get(), net_log_, fetched_config_, wait_delay,
|
||
|
base::Bind(&ProxyService::OnInitProxyResolverComplete,
|
||
|
base::Unretained(this)));
|
||
|
|
||
|
if (rv != ERR_IO_PENDING)
|
||
|
OnInitProxyResolverComplete(rv);
|
||
|
}
|
||
|
|
||
|
void ProxyService::InitializeUsingDecidedConfig(
|
||
|
int decider_result,
|
||
|
ProxyResolverScriptData* script_data,
|
||
|
const ProxyConfig& effective_config) {
|
||
|
DCHECK(fetched_config_.is_valid());
|
||
|
DCHECK(fetched_config_.HasAutomaticSettings());
|
||
|
|
||
|
ResetProxyConfig(false);
|
||
|
|
||
|
current_state_ = STATE_WAITING_FOR_INIT_PROXY_RESOLVER;
|
||
|
|
||
|
init_proxy_resolver_.reset(new InitProxyResolver());
|
||
|
int rv = init_proxy_resolver_->StartSkipDecider(
|
||
|
&resolver_, resolver_factory_.get(), effective_config, decider_result,
|
||
|
script_data, base::Bind(&ProxyService::OnInitProxyResolverComplete,
|
||
|
base::Unretained(this)));
|
||
|
|
||
|
if (rv != ERR_IO_PENDING)
|
||
|
OnInitProxyResolverComplete(rv);
|
||
|
}
|
||
|
|
||
|
void ProxyService::OnIPAddressChanged() {
|
||
|
// See the comment block by |kDelayAfterNetworkChangesMs| for info.
|
||
|
stall_proxy_autoconfig_until_ =
|
||
|
TimeTicks::Now() + stall_proxy_auto_config_delay_;
|
||
|
|
||
|
State previous_state = ResetProxyConfig(false);
|
||
|
if (previous_state != STATE_NONE)
|
||
|
ApplyProxyConfigIfAvailable();
|
||
|
}
|
||
|
|
||
|
void ProxyService::OnDNSChanged() {
|
||
|
OnIPAddressChanged();
|
||
|
}
|
||
|
|
||
|
} // namespace net
|