mirror of
https://github.com/klzgrad/naiveproxy.git
synced 2024-11-28 08:16:09 +03:00
676 lines
25 KiB
C++
676 lines
25 KiB
C++
// Copyright 2016 The Chromium Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file.
|
|
|
|
#include "base/task_scheduler/task_tracker.h"
|
|
|
|
#include <limits>
|
|
#include <string>
|
|
|
|
#include "base/callback.h"
|
|
#include "base/debug/task_annotator.h"
|
|
#include "base/json/json_writer.h"
|
|
#include "base/memory/ptr_util.h"
|
|
#include "base/metrics/histogram_macros.h"
|
|
#include "base/sequence_token.h"
|
|
#include "base/synchronization/condition_variable.h"
|
|
#include "base/task_scheduler/scoped_set_task_priority_for_current_thread.h"
|
|
#include "base/threading/sequence_local_storage_map.h"
|
|
#include "base/threading/sequenced_task_runner_handle.h"
|
|
#include "base/threading/thread_restrictions.h"
|
|
#include "base/threading/thread_task_runner_handle.h"
|
|
#include "base/time/time.h"
|
|
#include "base/trace_event/trace_event.h"
|
|
#include "base/values.h"
|
|
|
|
namespace base {
|
|
namespace internal {
|
|
|
|
namespace {
|
|
|
|
constexpr char kParallelExecutionMode[] = "parallel";
|
|
constexpr char kSequencedExecutionMode[] = "sequenced";
|
|
constexpr char kSingleThreadExecutionMode[] = "single thread";
|
|
|
|
// An immutable copy of a scheduler task's info required by tracing.
|
|
class TaskTracingInfo : public trace_event::ConvertableToTraceFormat {
|
|
public:
|
|
TaskTracingInfo(const TaskTraits& task_traits,
|
|
const char* execution_mode,
|
|
const SequenceToken& sequence_token)
|
|
: task_traits_(task_traits),
|
|
execution_mode_(execution_mode),
|
|
sequence_token_(sequence_token) {}
|
|
|
|
// trace_event::ConvertableToTraceFormat implementation.
|
|
void AppendAsTraceFormat(std::string* out) const override;
|
|
|
|
private:
|
|
const TaskTraits task_traits_;
|
|
const char* const execution_mode_;
|
|
const SequenceToken sequence_token_;
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(TaskTracingInfo);
|
|
};
|
|
|
|
void TaskTracingInfo::AppendAsTraceFormat(std::string* out) const {
|
|
DictionaryValue dict;
|
|
|
|
dict.SetString("task_priority",
|
|
base::TaskPriorityToString(task_traits_.priority()));
|
|
dict.SetString("execution_mode", execution_mode_);
|
|
if (execution_mode_ != kParallelExecutionMode)
|
|
dict.SetInteger("sequence_token", sequence_token_.ToInternalValue());
|
|
|
|
std::string tmp;
|
|
JSONWriter::Write(dict, &tmp);
|
|
out->append(tmp);
|
|
}
|
|
|
|
const char kQueueFunctionName[] = "base::PostTask";
|
|
|
|
// This name conveys that a Task is run by the task scheduler without revealing
|
|
// its implementation details.
|
|
const char kRunFunctionName[] = "TaskSchedulerRunTask";
|
|
|
|
HistogramBase* GetTaskLatencyHistogram(const char* suffix) {
|
|
// Mimics the UMA_HISTOGRAM_TIMES macro except we don't specify bounds with
|
|
// TimeDeltas as FactoryTimeGet assumes millisecond granularity. The minimums
|
|
// and maximums were chosen to place the 1ms mark at around the 70% range
|
|
// coverage for buckets giving us good info for tasks that have a latency
|
|
// below 1ms (most of them) and enough info to assess how bad the latency is
|
|
// for tasks that exceed this threshold.
|
|
return Histogram::FactoryGet(
|
|
std::string("TaskScheduler.TaskLatencyMicroseconds.") + suffix, 1, 20000,
|
|
50, HistogramBase::kUmaTargetedHistogramFlag);
|
|
}
|
|
|
|
// Upper bound for the
|
|
// TaskScheduler.BlockShutdownTasksPostedDuringShutdown histogram.
|
|
const HistogramBase::Sample kMaxBlockShutdownTasksPostedDuringShutdown = 1000;
|
|
|
|
void RecordNumBlockShutdownTasksPostedDuringShutdown(
|
|
HistogramBase::Sample value) {
|
|
UMA_HISTOGRAM_CUSTOM_COUNTS(
|
|
"TaskScheduler.BlockShutdownTasksPostedDuringShutdown", value, 1,
|
|
kMaxBlockShutdownTasksPostedDuringShutdown, 50);
|
|
}
|
|
|
|
} // namespace
|
|
|
|
// Atomic internal state used by TaskTracker. Sequential consistency shouldn't
|
|
// be assumed from these calls (i.e. a thread reading
|
|
// |HasShutdownStarted() == true| isn't guaranteed to see all writes made before
|
|
// |StartShutdown()| on the thread that invoked it).
|
|
class TaskTracker::State {
|
|
public:
|
|
State() = default;
|
|
|
|
// Sets a flag indicating that shutdown has started. Returns true if there are
|
|
// tasks blocking shutdown. Can only be called once.
|
|
bool StartShutdown() {
|
|
const auto new_value =
|
|
subtle::NoBarrier_AtomicIncrement(&bits_, kShutdownHasStartedMask);
|
|
|
|
// Check that the "shutdown has started" bit isn't zero. This would happen
|
|
// if it was incremented twice.
|
|
DCHECK(new_value & kShutdownHasStartedMask);
|
|
|
|
const auto num_tasks_blocking_shutdown =
|
|
new_value >> kNumTasksBlockingShutdownBitOffset;
|
|
return num_tasks_blocking_shutdown != 0;
|
|
}
|
|
|
|
// Returns true if shutdown has started.
|
|
bool HasShutdownStarted() const {
|
|
return subtle::NoBarrier_Load(&bits_) & kShutdownHasStartedMask;
|
|
}
|
|
|
|
// Returns true if there are tasks blocking shutdown.
|
|
bool AreTasksBlockingShutdown() const {
|
|
const auto num_tasks_blocking_shutdown =
|
|
subtle::NoBarrier_Load(&bits_) >> kNumTasksBlockingShutdownBitOffset;
|
|
DCHECK_GE(num_tasks_blocking_shutdown, 0);
|
|
return num_tasks_blocking_shutdown != 0;
|
|
}
|
|
|
|
// Increments the number of tasks blocking shutdown. Returns true if shutdown
|
|
// has started.
|
|
bool IncrementNumTasksBlockingShutdown() {
|
|
#if DCHECK_IS_ON()
|
|
// Verify that no overflow will occur.
|
|
const auto num_tasks_blocking_shutdown =
|
|
subtle::NoBarrier_Load(&bits_) >> kNumTasksBlockingShutdownBitOffset;
|
|
DCHECK_LT(num_tasks_blocking_shutdown,
|
|
std::numeric_limits<subtle::Atomic32>::max() -
|
|
kNumTasksBlockingShutdownIncrement);
|
|
#endif
|
|
|
|
const auto new_bits = subtle::NoBarrier_AtomicIncrement(
|
|
&bits_, kNumTasksBlockingShutdownIncrement);
|
|
return new_bits & kShutdownHasStartedMask;
|
|
}
|
|
|
|
// Decrements the number of tasks blocking shutdown. Returns true if shutdown
|
|
// has started and the number of tasks blocking shutdown becomes zero.
|
|
bool DecrementNumTasksBlockingShutdown() {
|
|
const auto new_bits = subtle::NoBarrier_AtomicIncrement(
|
|
&bits_, -kNumTasksBlockingShutdownIncrement);
|
|
const bool shutdown_has_started = new_bits & kShutdownHasStartedMask;
|
|
const auto num_tasks_blocking_shutdown =
|
|
new_bits >> kNumTasksBlockingShutdownBitOffset;
|
|
DCHECK_GE(num_tasks_blocking_shutdown, 0);
|
|
return shutdown_has_started && num_tasks_blocking_shutdown == 0;
|
|
}
|
|
|
|
private:
|
|
static constexpr subtle::Atomic32 kShutdownHasStartedMask = 1;
|
|
static constexpr subtle::Atomic32 kNumTasksBlockingShutdownBitOffset = 1;
|
|
static constexpr subtle::Atomic32 kNumTasksBlockingShutdownIncrement =
|
|
1 << kNumTasksBlockingShutdownBitOffset;
|
|
|
|
// The LSB indicates whether shutdown has started. The other bits count the
|
|
// number of tasks blocking shutdown.
|
|
// No barriers are required to read/write |bits_| as this class is only used
|
|
// as an atomic state checker, it doesn't provide sequential consistency
|
|
// guarantees w.r.t. external state. Sequencing of the TaskTracker::State
|
|
// operations themselves is guaranteed by the AtomicIncrement RMW (read-
|
|
// modify-write) semantics however. For example, if two threads are racing to
|
|
// call IncrementNumTasksBlockingShutdown() and StartShutdown() respectively,
|
|
// either the first thread will win and the StartShutdown() call will see the
|
|
// blocking task or the second thread will win and
|
|
// IncrementNumTasksBlockingShutdown() will know that shutdown has started.
|
|
subtle::Atomic32 bits_ = 0;
|
|
|
|
DISALLOW_COPY_AND_ASSIGN(State);
|
|
};
|
|
|
|
struct TaskTracker::PreemptedBackgroundSequence {
|
|
PreemptedBackgroundSequence() = default;
|
|
PreemptedBackgroundSequence(scoped_refptr<Sequence> sequence_in,
|
|
TimeTicks next_task_sequenced_time_in,
|
|
CanScheduleSequenceObserver* observer_in)
|
|
: sequence(std::move(sequence_in)),
|
|
next_task_sequenced_time(next_task_sequenced_time_in),
|
|
observer(observer_in) {}
|
|
PreemptedBackgroundSequence(PreemptedBackgroundSequence&& other) = default;
|
|
~PreemptedBackgroundSequence() = default;
|
|
PreemptedBackgroundSequence& operator=(PreemptedBackgroundSequence&& other) =
|
|
default;
|
|
bool operator<(const PreemptedBackgroundSequence& other) const {
|
|
return next_task_sequenced_time < other.next_task_sequenced_time;
|
|
}
|
|
bool operator>(const PreemptedBackgroundSequence& other) const {
|
|
return next_task_sequenced_time > other.next_task_sequenced_time;
|
|
}
|
|
|
|
// A background sequence waiting to be scheduled.
|
|
scoped_refptr<Sequence> sequence;
|
|
|
|
// The sequenced time of the next task in |sequence|.
|
|
TimeTicks next_task_sequenced_time;
|
|
|
|
// An observer to notify when |sequence| can be scheduled.
|
|
CanScheduleSequenceObserver* observer = nullptr;
|
|
};
|
|
|
|
TaskTracker::TaskTracker(int max_num_scheduled_background_sequences)
|
|
: state_(new State),
|
|
flush_cv_(flush_lock_.CreateConditionVariable()),
|
|
shutdown_lock_(&flush_lock_),
|
|
max_num_scheduled_background_sequences_(
|
|
max_num_scheduled_background_sequences),
|
|
task_latency_histograms_{
|
|
{GetTaskLatencyHistogram("BackgroundTaskPriority"),
|
|
GetTaskLatencyHistogram("BackgroundTaskPriority.MayBlock")},
|
|
{GetTaskLatencyHistogram("UserVisibleTaskPriority"),
|
|
GetTaskLatencyHistogram("UserVisibleTaskPriority.MayBlock")},
|
|
{GetTaskLatencyHistogram("UserBlockingTaskPriority"),
|
|
GetTaskLatencyHistogram("UserBlockingTaskPriority.MayBlock")}} {
|
|
// Confirm that all |task_latency_histograms_| have been initialized above.
|
|
DCHECK(*(&task_latency_histograms_[static_cast<int>(TaskPriority::HIGHEST) +
|
|
1][0] -
|
|
1));
|
|
}
|
|
|
|
TaskTracker::~TaskTracker() = default;
|
|
|
|
void TaskTracker::Shutdown() {
|
|
PerformShutdown();
|
|
DCHECK(IsShutdownComplete());
|
|
|
|
// Unblock Flush() when shutdown completes.
|
|
AutoSchedulerLock auto_lock(flush_lock_);
|
|
flush_cv_->Signal();
|
|
}
|
|
|
|
void TaskTracker::Flush() {
|
|
AutoSchedulerLock auto_lock(flush_lock_);
|
|
while (subtle::Acquire_Load(&num_incomplete_undelayed_tasks_) != 0 &&
|
|
!IsShutdownComplete()) {
|
|
flush_cv_->Wait();
|
|
}
|
|
}
|
|
|
|
bool TaskTracker::WillPostTask(const Task* task) {
|
|
DCHECK(task);
|
|
|
|
if (!BeforePostTask(task->traits.shutdown_behavior()))
|
|
return false;
|
|
|
|
if (task->delayed_run_time.is_null())
|
|
subtle::NoBarrier_AtomicIncrement(&num_incomplete_undelayed_tasks_, 1);
|
|
|
|
debug::TaskAnnotator task_annotator;
|
|
task_annotator.DidQueueTask(kQueueFunctionName, *task);
|
|
|
|
return true;
|
|
}
|
|
|
|
scoped_refptr<Sequence> TaskTracker::WillScheduleSequence(
|
|
scoped_refptr<Sequence> sequence,
|
|
CanScheduleSequenceObserver* observer) {
|
|
const SequenceSortKey sort_key = sequence->GetSortKey();
|
|
|
|
// A foreground sequence can always be scheduled.
|
|
if (sort_key.priority() != TaskPriority::BACKGROUND)
|
|
return sequence;
|
|
|
|
// It is convenient not to have to specify an observer when scheduling
|
|
// foreground sequences in tests.
|
|
DCHECK(observer);
|
|
|
|
AutoSchedulerLock auto_lock(background_lock_);
|
|
|
|
if (num_scheduled_background_sequences_ <
|
|
max_num_scheduled_background_sequences_) {
|
|
++num_scheduled_background_sequences_;
|
|
return sequence;
|
|
}
|
|
|
|
preempted_background_sequences_.emplace(
|
|
std::move(sequence), sort_key.next_task_sequenced_time(), observer);
|
|
return nullptr;
|
|
}
|
|
|
|
scoped_refptr<Sequence> TaskTracker::RunNextTask(
|
|
scoped_refptr<Sequence> sequence,
|
|
CanScheduleSequenceObserver* observer) {
|
|
DCHECK(sequence);
|
|
|
|
// Run the next task in |sequence|.
|
|
std::unique_ptr<Task> task = sequence->TakeTask();
|
|
DCHECK(task);
|
|
|
|
const TaskShutdownBehavior shutdown_behavior =
|
|
task->traits.shutdown_behavior();
|
|
const TaskPriority task_priority = task->traits.priority();
|
|
const bool can_run_task = BeforeRunTask(shutdown_behavior);
|
|
const bool is_delayed = !task->delayed_run_time.is_null();
|
|
|
|
RunOrSkipTask(std::move(task), sequence.get(), can_run_task);
|
|
if (can_run_task)
|
|
AfterRunTask(shutdown_behavior);
|
|
|
|
if (!is_delayed)
|
|
DecrementNumIncompleteUndelayedTasks();
|
|
|
|
const bool sequence_is_empty_after_pop = sequence->Pop();
|
|
|
|
// Never reschedule a Sequence emptied by Pop(). The contract is such that
|
|
// next poster to make it non-empty is responsible to schedule it.
|
|
if (sequence_is_empty_after_pop)
|
|
sequence = nullptr;
|
|
|
|
if (task_priority == TaskPriority::BACKGROUND) {
|
|
// Allow |sequence| to be rescheduled only if its next task is set to run
|
|
// earlier than the earliest currently preempted sequence
|
|
return ManageBackgroundSequencesAfterRunningTask(std::move(sequence),
|
|
observer);
|
|
}
|
|
|
|
return sequence;
|
|
}
|
|
|
|
bool TaskTracker::HasShutdownStarted() const {
|
|
return state_->HasShutdownStarted();
|
|
}
|
|
|
|
bool TaskTracker::IsShutdownComplete() const {
|
|
AutoSchedulerLock auto_lock(shutdown_lock_);
|
|
return shutdown_event_ && shutdown_event_->IsSignaled();
|
|
}
|
|
|
|
void TaskTracker::SetHasShutdownStartedForTesting() {
|
|
AutoSchedulerLock auto_lock(shutdown_lock_);
|
|
|
|
// Create a dummy |shutdown_event_| to satisfy TaskTracker's expectation of
|
|
// its existence during shutdown (e.g. in OnBlockingShutdownTasksComplete()).
|
|
shutdown_event_.reset(
|
|
new WaitableEvent(WaitableEvent::ResetPolicy::MANUAL,
|
|
WaitableEvent::InitialState::NOT_SIGNALED));
|
|
|
|
state_->StartShutdown();
|
|
}
|
|
|
|
void TaskTracker::RunOrSkipTask(std::unique_ptr<Task> task,
|
|
Sequence* sequence,
|
|
bool can_run_task) {
|
|
RecordTaskLatencyHistogram(task.get());
|
|
|
|
const bool previous_singleton_allowed =
|
|
ThreadRestrictions::SetSingletonAllowed(
|
|
task->traits.shutdown_behavior() !=
|
|
TaskShutdownBehavior::CONTINUE_ON_SHUTDOWN);
|
|
const bool previous_io_allowed =
|
|
ThreadRestrictions::SetIOAllowed(task->traits.may_block());
|
|
const bool previous_wait_allowed = ThreadRestrictions::SetWaitAllowed(
|
|
task->traits.with_base_sync_primitives());
|
|
|
|
{
|
|
const SequenceToken& sequence_token = sequence->token();
|
|
DCHECK(sequence_token.IsValid());
|
|
ScopedSetSequenceTokenForCurrentThread
|
|
scoped_set_sequence_token_for_current_thread(sequence_token);
|
|
ScopedSetTaskPriorityForCurrentThread
|
|
scoped_set_task_priority_for_current_thread(task->traits.priority());
|
|
ScopedSetSequenceLocalStorageMapForCurrentThread
|
|
scoped_set_sequence_local_storage_map_for_current_thread(
|
|
sequence->sequence_local_storage());
|
|
|
|
// Set up TaskRunnerHandle as expected for the scope of the task.
|
|
std::unique_ptr<SequencedTaskRunnerHandle> sequenced_task_runner_handle;
|
|
std::unique_ptr<ThreadTaskRunnerHandle> single_thread_task_runner_handle;
|
|
DCHECK(!task->sequenced_task_runner_ref ||
|
|
!task->single_thread_task_runner_ref);
|
|
if (task->sequenced_task_runner_ref) {
|
|
sequenced_task_runner_handle.reset(
|
|
new SequencedTaskRunnerHandle(task->sequenced_task_runner_ref));
|
|
} else if (task->single_thread_task_runner_ref) {
|
|
single_thread_task_runner_handle.reset(
|
|
new ThreadTaskRunnerHandle(task->single_thread_task_runner_ref));
|
|
}
|
|
|
|
if (can_run_task) {
|
|
TRACE_TASK_EXECUTION(kRunFunctionName, *task);
|
|
|
|
const char* const execution_mode =
|
|
task->single_thread_task_runner_ref
|
|
? kSingleThreadExecutionMode
|
|
: (task->sequenced_task_runner_ref ? kSequencedExecutionMode
|
|
: kParallelExecutionMode);
|
|
// TODO(gab): In a better world this would be tacked on as an extra arg
|
|
// to the trace event generated above. This is not possible however until
|
|
// http://crbug.com/652692 is resolved.
|
|
TRACE_EVENT1("task_scheduler", "TaskTracker::RunTask", "task_info",
|
|
std::make_unique<TaskTracingInfo>(
|
|
task->traits, execution_mode, sequence_token));
|
|
|
|
debug::TaskAnnotator().RunTask(kQueueFunctionName, task.get());
|
|
}
|
|
|
|
task.reset();
|
|
}
|
|
|
|
ThreadRestrictions::SetWaitAllowed(previous_wait_allowed);
|
|
ThreadRestrictions::SetIOAllowed(previous_io_allowed);
|
|
ThreadRestrictions::SetSingletonAllowed(previous_singleton_allowed);
|
|
}
|
|
|
|
void TaskTracker::PerformShutdown() {
|
|
{
|
|
AutoSchedulerLock auto_lock(shutdown_lock_);
|
|
|
|
// This method can only be called once.
|
|
DCHECK(!shutdown_event_);
|
|
DCHECK(!num_block_shutdown_tasks_posted_during_shutdown_);
|
|
DCHECK(!state_->HasShutdownStarted());
|
|
|
|
shutdown_event_.reset(
|
|
new WaitableEvent(WaitableEvent::ResetPolicy::MANUAL,
|
|
WaitableEvent::InitialState::NOT_SIGNALED));
|
|
|
|
const bool tasks_are_blocking_shutdown = state_->StartShutdown();
|
|
|
|
// From now, if a thread causes the number of tasks blocking shutdown to
|
|
// become zero, it will call OnBlockingShutdownTasksComplete().
|
|
|
|
if (!tasks_are_blocking_shutdown) {
|
|
// If another thread posts a BLOCK_SHUTDOWN task at this moment, it will
|
|
// block until this method releases |shutdown_lock_|. Then, it will fail
|
|
// DCHECK(!shutdown_event_->IsSignaled()). This is the desired behavior
|
|
// because posting a BLOCK_SHUTDOWN task when TaskTracker::Shutdown() has
|
|
// started and no tasks are blocking shutdown isn't allowed.
|
|
shutdown_event_->Signal();
|
|
return;
|
|
}
|
|
}
|
|
|
|
// It is safe to access |shutdown_event_| without holding |lock_| because the
|
|
// pointer never changes after being set above.
|
|
{
|
|
base::ThreadRestrictions::ScopedAllowWait allow_wait;
|
|
shutdown_event_->Wait();
|
|
}
|
|
|
|
{
|
|
AutoSchedulerLock auto_lock(shutdown_lock_);
|
|
|
|
// Record TaskScheduler.BlockShutdownTasksPostedDuringShutdown if less than
|
|
// |kMaxBlockShutdownTasksPostedDuringShutdown| BLOCK_SHUTDOWN tasks were
|
|
// posted during shutdown. Otherwise, the histogram has already been
|
|
// recorded in BeforePostTask().
|
|
if (num_block_shutdown_tasks_posted_during_shutdown_ <
|
|
kMaxBlockShutdownTasksPostedDuringShutdown) {
|
|
RecordNumBlockShutdownTasksPostedDuringShutdown(
|
|
num_block_shutdown_tasks_posted_during_shutdown_);
|
|
}
|
|
}
|
|
}
|
|
|
|
#if DCHECK_IS_ON()
|
|
bool TaskTracker::IsPostingBlockShutdownTaskAfterShutdownAllowed() {
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
int TaskTracker::GetNumIncompleteUndelayedTasksForTesting() const {
|
|
return subtle::NoBarrier_Load(&num_incomplete_undelayed_tasks_);
|
|
}
|
|
|
|
bool TaskTracker::BeforePostTask(TaskShutdownBehavior shutdown_behavior) {
|
|
if (shutdown_behavior == TaskShutdownBehavior::BLOCK_SHUTDOWN) {
|
|
// BLOCK_SHUTDOWN tasks block shutdown between the moment they are posted
|
|
// and the moment they complete their execution.
|
|
const bool shutdown_started = state_->IncrementNumTasksBlockingShutdown();
|
|
|
|
if (shutdown_started) {
|
|
AutoSchedulerLock auto_lock(shutdown_lock_);
|
|
|
|
// A BLOCK_SHUTDOWN task posted after shutdown has completed is an
|
|
// ordering bug. This aims to catch those early.
|
|
DCHECK(shutdown_event_);
|
|
if (shutdown_event_->IsSignaled()) {
|
|
#if DCHECK_IS_ON()
|
|
// clang-format off
|
|
// TODO(robliao): http://crbug.com/698140. Since the service thread
|
|
// doesn't stop processing its own tasks at shutdown, we may still
|
|
// attempt to post a BLOCK_SHUTDOWN task in response to a
|
|
// FileDescriptorWatcher. Same is true for FilePathWatcher
|
|
// (http://crbug.com/728235). Until it's possible for such services to
|
|
// post to non-BLOCK_SHUTDOWN sequences which are themselves funneled to
|
|
// the main execution sequence (a future plan for the post_task.h API),
|
|
// this DCHECK will be flaky and must be disabled.
|
|
// DCHECK(IsPostingBlockShutdownTaskAfterShutdownAllowed());
|
|
// clang-format on
|
|
#endif
|
|
state_->DecrementNumTasksBlockingShutdown();
|
|
return false;
|
|
}
|
|
|
|
++num_block_shutdown_tasks_posted_during_shutdown_;
|
|
|
|
if (num_block_shutdown_tasks_posted_during_shutdown_ ==
|
|
kMaxBlockShutdownTasksPostedDuringShutdown) {
|
|
// Record the TaskScheduler.BlockShutdownTasksPostedDuringShutdown
|
|
// histogram as soon as its upper bound is hit. That way, a value will
|
|
// be recorded even if an infinite number of BLOCK_SHUTDOWN tasks are
|
|
// posted, preventing shutdown to complete.
|
|
RecordNumBlockShutdownTasksPostedDuringShutdown(
|
|
num_block_shutdown_tasks_posted_during_shutdown_);
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
// A non BLOCK_SHUTDOWN task is allowed to be posted iff shutdown hasn't
|
|
// started.
|
|
return !state_->HasShutdownStarted();
|
|
}
|
|
|
|
bool TaskTracker::BeforeRunTask(TaskShutdownBehavior shutdown_behavior) {
|
|
switch (shutdown_behavior) {
|
|
case TaskShutdownBehavior::BLOCK_SHUTDOWN: {
|
|
// The number of tasks blocking shutdown has been incremented when the
|
|
// task was posted.
|
|
DCHECK(state_->AreTasksBlockingShutdown());
|
|
|
|
// Trying to run a BLOCK_SHUTDOWN task after shutdown has completed is
|
|
// unexpected as it either shouldn't have been posted if shutdown
|
|
// completed or should be blocking shutdown if it was posted before it
|
|
// did.
|
|
DCHECK(!state_->HasShutdownStarted() || !IsShutdownComplete());
|
|
|
|
return true;
|
|
}
|
|
|
|
case TaskShutdownBehavior::SKIP_ON_SHUTDOWN: {
|
|
// SKIP_ON_SHUTDOWN tasks block shutdown while they are running.
|
|
const bool shutdown_started = state_->IncrementNumTasksBlockingShutdown();
|
|
|
|
if (shutdown_started) {
|
|
// The SKIP_ON_SHUTDOWN task isn't allowed to run during shutdown.
|
|
// Decrement the number of tasks blocking shutdown that was wrongly
|
|
// incremented.
|
|
const bool shutdown_started_and_no_tasks_block_shutdown =
|
|
state_->DecrementNumTasksBlockingShutdown();
|
|
if (shutdown_started_and_no_tasks_block_shutdown)
|
|
OnBlockingShutdownTasksComplete();
|
|
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
case TaskShutdownBehavior::CONTINUE_ON_SHUTDOWN: {
|
|
return !state_->HasShutdownStarted();
|
|
}
|
|
}
|
|
|
|
NOTREACHED();
|
|
return false;
|
|
}
|
|
|
|
void TaskTracker::AfterRunTask(TaskShutdownBehavior shutdown_behavior) {
|
|
if (shutdown_behavior == TaskShutdownBehavior::BLOCK_SHUTDOWN ||
|
|
shutdown_behavior == TaskShutdownBehavior::SKIP_ON_SHUTDOWN) {
|
|
const bool shutdown_started_and_no_tasks_block_shutdown =
|
|
state_->DecrementNumTasksBlockingShutdown();
|
|
if (shutdown_started_and_no_tasks_block_shutdown)
|
|
OnBlockingShutdownTasksComplete();
|
|
}
|
|
}
|
|
|
|
void TaskTracker::OnBlockingShutdownTasksComplete() {
|
|
AutoSchedulerLock auto_lock(shutdown_lock_);
|
|
|
|
// This method can only be called after shutdown has started.
|
|
DCHECK(state_->HasShutdownStarted());
|
|
DCHECK(shutdown_event_);
|
|
|
|
shutdown_event_->Signal();
|
|
}
|
|
|
|
void TaskTracker::DecrementNumIncompleteUndelayedTasks() {
|
|
const auto new_num_incomplete_undelayed_tasks =
|
|
subtle::Barrier_AtomicIncrement(&num_incomplete_undelayed_tasks_, -1);
|
|
DCHECK_GE(new_num_incomplete_undelayed_tasks, 0);
|
|
if (new_num_incomplete_undelayed_tasks == 0) {
|
|
AutoSchedulerLock auto_lock(flush_lock_);
|
|
flush_cv_->Signal();
|
|
}
|
|
}
|
|
|
|
scoped_refptr<Sequence> TaskTracker::ManageBackgroundSequencesAfterRunningTask(
|
|
scoped_refptr<Sequence> just_ran_sequence,
|
|
CanScheduleSequenceObserver* observer) {
|
|
const TimeTicks next_task_sequenced_time =
|
|
just_ran_sequence
|
|
? just_ran_sequence->GetSortKey().next_task_sequenced_time()
|
|
: TimeTicks();
|
|
PreemptedBackgroundSequence background_sequence_to_schedule;
|
|
|
|
{
|
|
AutoSchedulerLock auto_lock(background_lock_);
|
|
|
|
DCHECK(preempted_background_sequences_.empty() ||
|
|
num_scheduled_background_sequences_ ==
|
|
max_num_scheduled_background_sequences_);
|
|
--num_scheduled_background_sequences_;
|
|
|
|
if (just_ran_sequence) {
|
|
if (preempted_background_sequences_.empty() ||
|
|
preempted_background_sequences_.top().next_task_sequenced_time >
|
|
next_task_sequenced_time) {
|
|
++num_scheduled_background_sequences_;
|
|
return just_ran_sequence;
|
|
}
|
|
|
|
preempted_background_sequences_.emplace(
|
|
std::move(just_ran_sequence), next_task_sequenced_time, observer);
|
|
}
|
|
|
|
DCHECK(preempted_background_sequences_.empty() ||
|
|
num_scheduled_background_sequences_ ==
|
|
max_num_scheduled_background_sequences_ - 1);
|
|
|
|
if (!preempted_background_sequences_.empty()) {
|
|
// The const_cast on top is okay since the PreemptedBackgroundSequence is
|
|
// transactionnaly being popped from |preempted_background_sequences_|
|
|
// right after and the move doesn't alter the sort order (a requirement
|
|
// for the Windows STL's consistency debug-checks for
|
|
// std::priority_queue::top()).
|
|
++num_scheduled_background_sequences_;
|
|
background_sequence_to_schedule =
|
|
std::move(const_cast<PreemptedBackgroundSequence&>(
|
|
preempted_background_sequences_.top()));
|
|
preempted_background_sequences_.pop();
|
|
}
|
|
}
|
|
|
|
// |background_sequence_to_schedule.sequence| may be null if there was no
|
|
// preempted background sequence.
|
|
if (background_sequence_to_schedule.sequence) {
|
|
DCHECK(background_sequence_to_schedule.observer);
|
|
background_sequence_to_schedule.observer->OnCanScheduleSequence(
|
|
std::move(background_sequence_to_schedule.sequence));
|
|
}
|
|
|
|
return nullptr;
|
|
}
|
|
|
|
void TaskTracker::RecordTaskLatencyHistogram(Task* task) {
|
|
const TimeDelta task_latency = TimeTicks::Now() - task->sequenced_time;
|
|
task_latency_histograms_[static_cast<int>(task->traits.priority())]
|
|
[task->traits.may_block() ||
|
|
task->traits.with_base_sync_primitives()
|
|
? 1
|
|
: 0]
|
|
->Add(task_latency.InMicroseconds());
|
|
}
|
|
|
|
} // namespace internal
|
|
} // namespace base
|