mirror of
https://github.com/klzgrad/naiveproxy.git
synced 2024-11-28 16:26:10 +03:00
1589 lines
61 KiB
C++
1589 lines
61 KiB
C++
// Copyright (c) 2013 The Chromium Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file.
|
|
|
|
#include "net/disk_cache/simple/simple_entry_impl.h"
|
|
|
|
#include <algorithm>
|
|
#include <cstring>
|
|
#include <limits>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
#include "base/bind.h"
|
|
#include "base/bind_helpers.h"
|
|
#include "base/callback.h"
|
|
#include "base/location.h"
|
|
#include "base/logging.h"
|
|
#include "base/single_thread_task_runner.h"
|
|
#include "base/task_runner.h"
|
|
#include "base/task_runner_util.h"
|
|
#include "base/threading/thread_task_runner_handle.h"
|
|
#include "base/time/time.h"
|
|
#include "base/trace_event/memory_usage_estimator.h"
|
|
#include "net/base/io_buffer.h"
|
|
#include "net/base/net_errors.h"
|
|
#include "net/base/prioritized_task_runner.h"
|
|
#include "net/disk_cache/backend_cleanup_tracker.h"
|
|
#include "net/disk_cache/net_log_parameters.h"
|
|
#include "net/disk_cache/simple/simple_backend_impl.h"
|
|
#include "net/disk_cache/simple/simple_histogram_enums.h"
|
|
#include "net/disk_cache/simple/simple_histogram_macros.h"
|
|
#include "net/disk_cache/simple/simple_index.h"
|
|
#include "net/disk_cache/simple/simple_net_log_parameters.h"
|
|
#include "net/disk_cache/simple/simple_synchronous_entry.h"
|
|
#include "net/disk_cache/simple/simple_util.h"
|
|
#include "net/log/net_log.h"
|
|
#include "net/log/net_log_source_type.h"
|
|
#include "third_party/zlib/zlib.h"
|
|
|
|
namespace disk_cache {
|
|
namespace {
|
|
|
|
// An entry can store sparse data taking up to 1 / kMaxSparseDataSizeDivisor of
|
|
// the cache.
|
|
const int64_t kMaxSparseDataSizeDivisor = 10;
|
|
|
|
// Used in histograms, please only add entries at the end.
|
|
enum SimpleEntryWriteResult {
|
|
SIMPLE_ENTRY_WRITE_RESULT_SUCCESS = 0,
|
|
SIMPLE_ENTRY_WRITE_RESULT_INVALID_ARGUMENT = 1,
|
|
SIMPLE_ENTRY_WRITE_RESULT_OVER_MAX_SIZE = 2,
|
|
SIMPLE_ENTRY_WRITE_RESULT_BAD_STATE = 3,
|
|
SIMPLE_ENTRY_WRITE_RESULT_SYNC_WRITE_FAILURE = 4,
|
|
SIMPLE_ENTRY_WRITE_RESULT_FAST_EMPTY_RETURN = 5,
|
|
SIMPLE_ENTRY_WRITE_RESULT_MAX = 6,
|
|
};
|
|
|
|
void RecordReadResult(net::CacheType cache_type, SimpleReadResult result) {
|
|
SIMPLE_CACHE_UMA(ENUMERATION,
|
|
"ReadResult", cache_type, result, READ_RESULT_MAX);
|
|
}
|
|
|
|
void RecordWriteResult(net::CacheType cache_type,
|
|
SimpleEntryWriteResult result) {
|
|
SIMPLE_CACHE_UMA(ENUMERATION, "WriteResult2", cache_type, result,
|
|
SIMPLE_ENTRY_WRITE_RESULT_MAX);
|
|
}
|
|
|
|
void RecordHeaderSize(net::CacheType cache_type, int size) {
|
|
SIMPLE_CACHE_UMA(COUNTS_10000, "HeaderSize", cache_type, size);
|
|
}
|
|
|
|
int g_open_entry_count = 0;
|
|
|
|
void AdjustOpenEntryCountBy(net::CacheType cache_type, int offset) {
|
|
g_open_entry_count += offset;
|
|
SIMPLE_CACHE_UMA(COUNTS_10000,
|
|
"GlobalOpenEntryCount", cache_type, g_open_entry_count);
|
|
}
|
|
|
|
void InvokeCallbackIfBackendIsAlive(
|
|
const base::WeakPtr<SimpleBackendImpl>& backend,
|
|
net::CompletionOnceCallback completion_callback,
|
|
int result) {
|
|
DCHECK(!completion_callback.is_null());
|
|
if (!backend.get())
|
|
return;
|
|
std::move(completion_callback).Run(result);
|
|
}
|
|
|
|
// If |sync_possible| is false, and callback is available, posts rv to it and
|
|
// return net::ERR_IO_PENDING; otherwise just passes through rv.
|
|
int PostToCallbackIfNeeded(bool sync_possible,
|
|
net::CompletionOnceCallback callback,
|
|
int rv) {
|
|
if (!sync_possible && !callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(callback), rv));
|
|
return net::ERR_IO_PENDING;
|
|
} else {
|
|
return rv;
|
|
}
|
|
}
|
|
|
|
} // namespace
|
|
|
|
using base::Closure;
|
|
using base::OnceClosure;
|
|
using base::FilePath;
|
|
using base::Time;
|
|
using base::TaskRunner;
|
|
|
|
// Static function called by base::trace_event::EstimateMemoryUsage() to
|
|
// estimate the memory of SimpleEntryOperation.
|
|
// This needs to be in disk_cache namespace.
|
|
size_t EstimateMemoryUsage(const SimpleEntryOperation& op) {
|
|
return 0;
|
|
}
|
|
|
|
// A helper class to insure that RunNextOperationIfNeeded() is called when
|
|
// exiting the current stack frame.
|
|
class SimpleEntryImpl::ScopedOperationRunner {
|
|
public:
|
|
explicit ScopedOperationRunner(SimpleEntryImpl* entry) : entry_(entry) {
|
|
}
|
|
|
|
~ScopedOperationRunner() {
|
|
entry_->RunNextOperationIfNeeded();
|
|
}
|
|
|
|
private:
|
|
SimpleEntryImpl* const entry_;
|
|
};
|
|
|
|
SimpleEntryImpl::ActiveEntryProxy::~ActiveEntryProxy() = default;
|
|
|
|
SimpleEntryImpl::SimpleEntryImpl(
|
|
net::CacheType cache_type,
|
|
const FilePath& path,
|
|
scoped_refptr<BackendCleanupTracker> cleanup_tracker,
|
|
const uint64_t entry_hash,
|
|
OperationsMode operations_mode,
|
|
SimpleBackendImpl* backend,
|
|
SimpleFileTracker* file_tracker,
|
|
net::NetLog* net_log,
|
|
uint32_t entry_priority)
|
|
: cleanup_tracker_(std::move(cleanup_tracker)),
|
|
backend_(backend->AsWeakPtr()),
|
|
file_tracker_(file_tracker),
|
|
cache_type_(cache_type),
|
|
path_(path),
|
|
entry_hash_(entry_hash),
|
|
use_optimistic_operations_(operations_mode == OPTIMISTIC_OPERATIONS),
|
|
is_initial_stream1_read_(true),
|
|
last_used_(Time::Now()),
|
|
last_modified_(last_used_),
|
|
sparse_data_size_(0),
|
|
open_count_(0),
|
|
doom_state_(DOOM_NONE),
|
|
optimistic_create_pending_doom_state_(CREATE_NORMAL),
|
|
state_(STATE_UNINITIALIZED),
|
|
synchronous_entry_(NULL),
|
|
prioritized_task_runner_(backend_->prioritized_task_runner()),
|
|
net_log_(
|
|
net::NetLogWithSource::Make(net_log,
|
|
net::NetLogSourceType::DISK_CACHE_ENTRY)),
|
|
stream_0_data_(base::MakeRefCounted<net::GrowableIOBuffer>()),
|
|
entry_priority_(entry_priority) {
|
|
static_assert(arraysize(data_size_) == arraysize(crc32s_end_offset_),
|
|
"arrays should be the same size");
|
|
static_assert(arraysize(data_size_) == arraysize(crc32s_),
|
|
"arrays should be the same size");
|
|
static_assert(arraysize(data_size_) == arraysize(have_written_),
|
|
"arrays should be the same size");
|
|
static_assert(arraysize(data_size_) == arraysize(crc_check_state_),
|
|
"arrays should be the same size");
|
|
ResetEntry();
|
|
net_log_.BeginEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY,
|
|
CreateNetLogSimpleEntryConstructionCallback(this));
|
|
}
|
|
|
|
void SimpleEntryImpl::SetActiveEntryProxy(
|
|
std::unique_ptr<ActiveEntryProxy> active_entry_proxy) {
|
|
DCHECK(!active_entry_proxy_);
|
|
active_entry_proxy_ = std::move(active_entry_proxy);
|
|
}
|
|
|
|
int SimpleEntryImpl::OpenEntry(Entry** out_entry,
|
|
net::CompletionOnceCallback callback) {
|
|
DCHECK(backend_.get());
|
|
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_OPEN_CALL);
|
|
|
|
bool have_index = backend_->index()->initialized();
|
|
// This enumeration is used in histograms, add entries only at end.
|
|
enum OpenEntryIndexEnum {
|
|
INDEX_NOEXIST = 0,
|
|
INDEX_MISS = 1,
|
|
INDEX_HIT = 2,
|
|
INDEX_MAX = 3,
|
|
};
|
|
OpenEntryIndexEnum open_entry_index_enum = INDEX_NOEXIST;
|
|
if (have_index) {
|
|
if (backend_->index()->Has(entry_hash_))
|
|
open_entry_index_enum = INDEX_HIT;
|
|
else
|
|
open_entry_index_enum = INDEX_MISS;
|
|
}
|
|
SIMPLE_CACHE_UMA(ENUMERATION,
|
|
"OpenEntryIndexState", cache_type_,
|
|
open_entry_index_enum, INDEX_MAX);
|
|
|
|
// If entry is not known to the index, initiate fast failover to the network.
|
|
if (open_entry_index_enum == INDEX_MISS) {
|
|
net_log_.AddEventWithNetErrorCode(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_OPEN_END, net::ERR_FAILED);
|
|
return net::ERR_FAILED;
|
|
}
|
|
|
|
pending_operations_.push(SimpleEntryOperation::OpenOperation(
|
|
this, have_index, std::move(callback), out_entry));
|
|
RunNextOperationIfNeeded();
|
|
return net::ERR_IO_PENDING;
|
|
}
|
|
|
|
int SimpleEntryImpl::CreateEntry(Entry** out_entry,
|
|
net::CompletionOnceCallback callback) {
|
|
DCHECK(backend_.get());
|
|
DCHECK_EQ(entry_hash_, simple_util::GetEntryHashKey(key_));
|
|
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_CREATE_CALL);
|
|
|
|
bool have_index = backend_->index()->initialized();
|
|
int ret_value = net::ERR_FAILED;
|
|
if (use_optimistic_operations_ &&
|
|
state_ == STATE_UNINITIALIZED && pending_operations_.size() == 0) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_CREATE_OPTIMISTIC);
|
|
|
|
ReturnEntryToCaller(out_entry);
|
|
pending_operations_.push(SimpleEntryOperation::CreateOperation(
|
|
this, have_index, CompletionOnceCallback(),
|
|
static_cast<Entry**>(NULL)));
|
|
ret_value = net::OK;
|
|
|
|
// If we are optimistically returning before a preceeding doom, we need to
|
|
// wait for that IO, about which we will be notified externally.
|
|
if (optimistic_create_pending_doom_state_ != CREATE_NORMAL) {
|
|
DCHECK_EQ(CREATE_OPTIMISTIC_PENDING_DOOM,
|
|
optimistic_create_pending_doom_state_);
|
|
state_ = STATE_IO_PENDING;
|
|
}
|
|
} else {
|
|
pending_operations_.push(SimpleEntryOperation::CreateOperation(
|
|
this, have_index, std::move(callback), out_entry));
|
|
ret_value = net::ERR_IO_PENDING;
|
|
}
|
|
|
|
// We insert the entry in the index before creating the entry files in the
|
|
// SimpleSynchronousEntry, because this way the worst scenario is when we
|
|
// have the entry in the index but we don't have the created files yet, this
|
|
// way we never leak files. CreationOperationComplete will remove the entry
|
|
// from the index if the creation fails.
|
|
backend_->index()->Insert(entry_hash_);
|
|
|
|
RunNextOperationIfNeeded();
|
|
return ret_value;
|
|
}
|
|
|
|
int SimpleEntryImpl::DoomEntry(net::CompletionOnceCallback callback) {
|
|
if (doom_state_ != DOOM_NONE)
|
|
return net::OK;
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_DOOM_CALL);
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_DOOM_BEGIN);
|
|
|
|
MarkAsDoomed(DOOM_QUEUED);
|
|
if (backend_.get()) {
|
|
if (optimistic_create_pending_doom_state_ == CREATE_NORMAL) {
|
|
backend_->OnDoomStart(entry_hash_);
|
|
} else {
|
|
DCHECK_EQ(STATE_IO_PENDING, state_);
|
|
DCHECK_EQ(CREATE_OPTIMISTIC_PENDING_DOOM,
|
|
optimistic_create_pending_doom_state_);
|
|
// If we are in this state, we went ahead with making the entry even
|
|
// though the backend was already keeping track of a doom, so it can't
|
|
// keep track of ours. So we delay notifying it until
|
|
// NotifyDoomBeforeCreateComplete is called. Since this path is invoked
|
|
// only when the queue of post-doom callbacks was previously empty, while
|
|
// the CompletionOnceCallback for the op is posted,
|
|
// NotifyDoomBeforeCreateComplete() will be the first thing running after
|
|
// the previous doom completes, so at that point we can immediately grab
|
|
// a spot in entries_pending_doom_.
|
|
optimistic_create_pending_doom_state_ =
|
|
CREATE_OPTIMISTIC_PENDING_DOOM_FOLLOWED_BY_DOOM;
|
|
}
|
|
}
|
|
pending_operations_.push(
|
|
SimpleEntryOperation::DoomOperation(this, std::move(callback)));
|
|
RunNextOperationIfNeeded();
|
|
return net::ERR_IO_PENDING;
|
|
}
|
|
|
|
void SimpleEntryImpl::SetCreatePendingDoom() {
|
|
DCHECK_EQ(CREATE_NORMAL, optimistic_create_pending_doom_state_);
|
|
optimistic_create_pending_doom_state_ = CREATE_OPTIMISTIC_PENDING_DOOM;
|
|
}
|
|
|
|
void SimpleEntryImpl::NotifyDoomBeforeCreateComplete() {
|
|
DCHECK_EQ(STATE_IO_PENDING, state_);
|
|
DCHECK_NE(CREATE_NORMAL, optimistic_create_pending_doom_state_);
|
|
if (backend_.get() && optimistic_create_pending_doom_state_ ==
|
|
CREATE_OPTIMISTIC_PENDING_DOOM_FOLLOWED_BY_DOOM)
|
|
backend_->OnDoomStart(entry_hash_);
|
|
|
|
state_ = STATE_UNINITIALIZED;
|
|
optimistic_create_pending_doom_state_ = CREATE_NORMAL;
|
|
RunNextOperationIfNeeded();
|
|
}
|
|
|
|
void SimpleEntryImpl::SetKey(const std::string& key) {
|
|
key_ = key;
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_SET_KEY,
|
|
net::NetLog::StringCallback("key", &key));
|
|
}
|
|
|
|
void SimpleEntryImpl::Doom() {
|
|
DoomEntry(CompletionOnceCallback());
|
|
}
|
|
|
|
void SimpleEntryImpl::Close() {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK_LT(0, open_count_);
|
|
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_CLOSE_CALL);
|
|
|
|
if (--open_count_ > 0) {
|
|
DCHECK(!HasOneRef());
|
|
Release(); // Balanced in ReturnEntryToCaller().
|
|
return;
|
|
}
|
|
|
|
pending_operations_.push(SimpleEntryOperation::CloseOperation(this));
|
|
DCHECK(!HasOneRef());
|
|
Release(); // Balanced in ReturnEntryToCaller().
|
|
RunNextOperationIfNeeded();
|
|
}
|
|
|
|
std::string SimpleEntryImpl::GetKey() const {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
return key_;
|
|
}
|
|
|
|
Time SimpleEntryImpl::GetLastUsed() const {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
return last_used_;
|
|
}
|
|
|
|
Time SimpleEntryImpl::GetLastModified() const {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
return last_modified_;
|
|
}
|
|
|
|
int32_t SimpleEntryImpl::GetDataSize(int stream_index) const {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK_LE(0, data_size_[stream_index]);
|
|
return data_size_[stream_index];
|
|
}
|
|
|
|
int SimpleEntryImpl::ReadData(int stream_index,
|
|
int offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_CALL,
|
|
CreateNetLogReadWriteDataCallback(stream_index, offset,
|
|
buf_len, false));
|
|
}
|
|
|
|
if (stream_index < 0 || stream_index >= kSimpleEntryStreamCount ||
|
|
buf_len < 0) {
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_END,
|
|
CreateNetLogReadWriteCompleteCallback(net::ERR_INVALID_ARGUMENT));
|
|
}
|
|
|
|
RecordReadResult(cache_type_, READ_RESULT_INVALID_ARGUMENT);
|
|
return net::ERR_INVALID_ARGUMENT;
|
|
}
|
|
|
|
// If this is the only operation, bypass the queue, and also see if there is
|
|
// in-memory data to handle it synchronously. In principle, multiple reads can
|
|
// be parallelized, but past studies have shown that parallelizable ones
|
|
// happen <1% of the time, so it's probably not worth the effort.
|
|
bool alone_in_queue =
|
|
pending_operations_.size() == 0 && state_ == STATE_READY;
|
|
|
|
if (alone_in_queue) {
|
|
return ReadDataInternal(/*sync_possible = */ true, stream_index, offset,
|
|
buf, buf_len, std::move(callback));
|
|
}
|
|
|
|
pending_operations_.push(SimpleEntryOperation::ReadOperation(
|
|
this, stream_index, offset, buf_len, buf, std::move(callback)));
|
|
RunNextOperationIfNeeded();
|
|
return net::ERR_IO_PENDING;
|
|
}
|
|
|
|
int SimpleEntryImpl::WriteData(int stream_index,
|
|
int offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
CompletionOnceCallback callback,
|
|
bool truncate) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_CALL,
|
|
CreateNetLogReadWriteDataCallback(stream_index, offset,
|
|
buf_len, truncate));
|
|
}
|
|
|
|
if (stream_index < 0 || stream_index >= kSimpleEntryStreamCount ||
|
|
offset < 0 || buf_len < 0) {
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_END,
|
|
CreateNetLogReadWriteCompleteCallback(net::ERR_INVALID_ARGUMENT));
|
|
}
|
|
RecordWriteResult(cache_type_, SIMPLE_ENTRY_WRITE_RESULT_INVALID_ARGUMENT);
|
|
return net::ERR_INVALID_ARGUMENT;
|
|
}
|
|
if (backend_.get() && offset + buf_len > backend_->GetMaxFileSize()) {
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_END,
|
|
CreateNetLogReadWriteCompleteCallback(net::ERR_FAILED));
|
|
}
|
|
RecordWriteResult(cache_type_, SIMPLE_ENTRY_WRITE_RESULT_OVER_MAX_SIZE);
|
|
return net::ERR_FAILED;
|
|
}
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
// Stream 0 data is kept in memory, so can be written immediatly if there are
|
|
// no IO operations pending.
|
|
if (stream_index == 0 && state_ == STATE_READY &&
|
|
pending_operations_.size() == 0)
|
|
return SetStream0Data(buf, offset, buf_len, truncate);
|
|
|
|
// We can only do optimistic Write if there is no pending operations, so
|
|
// that we are sure that the next call to RunNextOperationIfNeeded will
|
|
// actually run the write operation that sets the stream size. It also
|
|
// prevents from previous possibly-conflicting writes that could be stacked
|
|
// in the |pending_operations_|. We could optimize this for when we have
|
|
// only read operations enqueued, but past studies have shown that that such
|
|
// parallelizable cases are very rare.
|
|
const bool optimistic =
|
|
(use_optimistic_operations_ && state_ == STATE_READY &&
|
|
pending_operations_.size() == 0);
|
|
CompletionOnceCallback op_callback;
|
|
scoped_refptr<net::IOBuffer> op_buf;
|
|
int ret_value = net::ERR_FAILED;
|
|
if (!optimistic) {
|
|
op_buf = buf;
|
|
op_callback = std::move(callback);
|
|
ret_value = net::ERR_IO_PENDING;
|
|
} else {
|
|
// TODO(morlovich,pasko): For performance, don't use a copy of an IOBuffer
|
|
// here to avoid paying the price of the RefCountedThreadSafe atomic
|
|
// operations.
|
|
if (buf) {
|
|
op_buf = base::MakeRefCounted<IOBuffer>(buf_len);
|
|
memcpy(op_buf->data(), buf->data(), buf_len);
|
|
}
|
|
op_callback = CompletionOnceCallback();
|
|
ret_value = buf_len;
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_OPTIMISTIC,
|
|
CreateNetLogReadWriteCompleteCallback(buf_len));
|
|
}
|
|
}
|
|
|
|
pending_operations_.push(SimpleEntryOperation::WriteOperation(
|
|
this, stream_index, offset, buf_len, op_buf.get(), truncate, optimistic,
|
|
std::move(op_callback)));
|
|
return ret_value;
|
|
}
|
|
|
|
int SimpleEntryImpl::ReadSparseData(int64_t offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_SPARSE_CALL,
|
|
CreateNetLogSparseOperationCallback(offset, buf_len));
|
|
}
|
|
|
|
ScopedOperationRunner operation_runner(this);
|
|
pending_operations_.push(SimpleEntryOperation::ReadSparseOperation(
|
|
this, offset, buf_len, buf, std::move(callback)));
|
|
return net::ERR_IO_PENDING;
|
|
}
|
|
|
|
int SimpleEntryImpl::WriteSparseData(int64_t offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_SPARSE_CALL,
|
|
CreateNetLogSparseOperationCallback(offset, buf_len));
|
|
}
|
|
|
|
ScopedOperationRunner operation_runner(this);
|
|
pending_operations_.push(SimpleEntryOperation::WriteSparseOperation(
|
|
this, offset, buf_len, buf, std::move(callback)));
|
|
return net::ERR_IO_PENDING;
|
|
}
|
|
|
|
int SimpleEntryImpl::GetAvailableRange(int64_t offset,
|
|
int len,
|
|
int64_t* start,
|
|
CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
|
|
ScopedOperationRunner operation_runner(this);
|
|
pending_operations_.push(SimpleEntryOperation::GetAvailableRangeOperation(
|
|
this, offset, len, start, std::move(callback)));
|
|
return net::ERR_IO_PENDING;
|
|
}
|
|
|
|
bool SimpleEntryImpl::CouldBeSparse() const {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
// TODO(morlovich): Actually check.
|
|
return true;
|
|
}
|
|
|
|
void SimpleEntryImpl::CancelSparseIO() {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
// The Simple Cache does not return distinct objects for the same non-doomed
|
|
// entry, so there's no need to coordinate which object is performing sparse
|
|
// I/O. Therefore, CancelSparseIO and ReadyForSparseIO succeed instantly.
|
|
}
|
|
|
|
int SimpleEntryImpl::ReadyForSparseIO(CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
// The simple Cache does not return distinct objects for the same non-doomed
|
|
// entry, so there's no need to coordinate which object is performing sparse
|
|
// I/O. Therefore, CancelSparseIO and ReadyForSparseIO succeed instantly.
|
|
return net::OK;
|
|
}
|
|
|
|
void SimpleEntryImpl::SetLastUsedTimeForTest(base::Time time) {
|
|
// Note that we do not update |last_used_| here as it gets overwritten
|
|
// by UpdateDataFromEntryStat with the current time. It would be more involved
|
|
// to make that value stick and is not needed by the current tests.
|
|
backend_->index()->SetLastUsedTimeForTest(entry_hash_, time);
|
|
}
|
|
|
|
size_t SimpleEntryImpl::EstimateMemoryUsage() const {
|
|
// TODO(xunjieli): crbug.com/669108. It'd be nice to have the rest of |entry|
|
|
// measured, but the ownership of SimpleSynchronousEntry isn't straightforward
|
|
return sizeof(SimpleSynchronousEntry) +
|
|
base::trace_event::EstimateMemoryUsage(pending_operations_) +
|
|
(stream_0_data_ ? stream_0_data_->capacity() : 0) +
|
|
(stream_1_prefetch_data_ ? stream_1_prefetch_data_->capacity() : 0);
|
|
}
|
|
|
|
void SimpleEntryImpl::SetPriority(uint32_t entry_priority) {
|
|
entry_priority_ = entry_priority;
|
|
}
|
|
|
|
SimpleEntryImpl::~SimpleEntryImpl() {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK_EQ(0U, pending_operations_.size());
|
|
|
|
// STATE_IO_PENDING is possible here in one corner case: the entry had
|
|
// dispatched the final Close() operation to SimpleSynchronousEntry, and the
|
|
// only thing keeping |this| alive were the callbacks for that
|
|
// PostTaskAndReply. If at that point the message loop is shut down, all
|
|
// outstanding tasks get destroyed, dropping the last reference without
|
|
// CloseOperationComplete ever getting to run to exit from IO_PENDING.
|
|
DCHECK(state_ == STATE_UNINITIALIZED || state_ == STATE_FAILURE ||
|
|
state_ == STATE_IO_PENDING);
|
|
DCHECK(!synchronous_entry_);
|
|
net_log_.EndEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY);
|
|
}
|
|
|
|
void SimpleEntryImpl::PostClientCallback(net::CompletionOnceCallback callback,
|
|
int result) {
|
|
if (callback.is_null())
|
|
return;
|
|
// Note that the callback is posted rather than directly invoked to avoid
|
|
// reentrancy issues.
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(&InvokeCallbackIfBackendIsAlive, backend_,
|
|
std::move(callback), result));
|
|
}
|
|
|
|
void SimpleEntryImpl::ResetEntry() {
|
|
// If we're doomed, we can't really do anything else with the entry, since
|
|
// we no longer own the name and are disconnected from the active entry table.
|
|
// We preserve doom_state_ accross this entry for this same reason.
|
|
state_ = doom_state_ == DOOM_COMPLETED ? STATE_FAILURE : STATE_UNINITIALIZED;
|
|
std::memset(crc32s_end_offset_, 0, sizeof(crc32s_end_offset_));
|
|
std::memset(crc32s_, 0, sizeof(crc32s_));
|
|
std::memset(have_written_, 0, sizeof(have_written_));
|
|
std::memset(data_size_, 0, sizeof(data_size_));
|
|
for (size_t i = 0; i < arraysize(crc_check_state_); ++i) {
|
|
crc_check_state_[i] = CRC_CHECK_NEVER_READ_AT_ALL;
|
|
}
|
|
}
|
|
|
|
void SimpleEntryImpl::ReturnEntryToCaller(Entry** out_entry) {
|
|
DCHECK(out_entry);
|
|
++open_count_;
|
|
AddRef(); // Balanced in Close()
|
|
if (!backend_.get()) {
|
|
// This method can be called when an asynchronous operation completed.
|
|
// If the backend no longer exists, the callback won't be invoked, and so we
|
|
// must close ourselves to avoid leaking. As well, there's no guarantee the
|
|
// client-provided pointer (|out_entry|) hasn't been freed, and no point
|
|
// dereferencing it, either.
|
|
Close();
|
|
return;
|
|
}
|
|
*out_entry = this;
|
|
}
|
|
|
|
void SimpleEntryImpl::MarkAsDoomed(DoomState new_state) {
|
|
DCHECK_NE(DOOM_NONE, new_state);
|
|
doom_state_ = new_state;
|
|
if (!backend_.get())
|
|
return;
|
|
backend_->index()->Remove(entry_hash_);
|
|
active_entry_proxy_.reset();
|
|
}
|
|
|
|
void SimpleEntryImpl::RunNextOperationIfNeeded() {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
SIMPLE_CACHE_UMA(CUSTOM_COUNTS,
|
|
"EntryOperationsPending", cache_type_,
|
|
pending_operations_.size(), 0, 100, 20);
|
|
if (!pending_operations_.empty() && state_ != STATE_IO_PENDING) {
|
|
SimpleEntryOperation operation = std::move(pending_operations_.front());
|
|
pending_operations_.pop();
|
|
switch (operation.type()) {
|
|
case SimpleEntryOperation::TYPE_OPEN:
|
|
OpenEntryInternal(operation.have_index(), operation.ReleaseCallback(),
|
|
operation.out_entry());
|
|
break;
|
|
case SimpleEntryOperation::TYPE_CREATE:
|
|
CreateEntryInternal(operation.have_index(), operation.ReleaseCallback(),
|
|
operation.out_entry());
|
|
break;
|
|
case SimpleEntryOperation::TYPE_CLOSE:
|
|
CloseInternal();
|
|
break;
|
|
case SimpleEntryOperation::TYPE_READ:
|
|
ReadDataInternal(/* sync_possible= */ false, operation.index(),
|
|
operation.offset(), operation.buf(),
|
|
operation.length(), operation.ReleaseCallback());
|
|
break;
|
|
case SimpleEntryOperation::TYPE_WRITE:
|
|
WriteDataInternal(operation.index(), operation.offset(),
|
|
operation.buf(), operation.length(),
|
|
operation.ReleaseCallback(), operation.truncate());
|
|
break;
|
|
case SimpleEntryOperation::TYPE_READ_SPARSE:
|
|
ReadSparseDataInternal(operation.sparse_offset(), operation.buf(),
|
|
operation.length(), operation.ReleaseCallback());
|
|
break;
|
|
case SimpleEntryOperation::TYPE_WRITE_SPARSE:
|
|
WriteSparseDataInternal(operation.sparse_offset(), operation.buf(),
|
|
operation.length(),
|
|
operation.ReleaseCallback());
|
|
break;
|
|
case SimpleEntryOperation::TYPE_GET_AVAILABLE_RANGE:
|
|
GetAvailableRangeInternal(operation.sparse_offset(), operation.length(),
|
|
operation.out_start(),
|
|
operation.ReleaseCallback());
|
|
break;
|
|
case SimpleEntryOperation::TYPE_DOOM:
|
|
DoomEntryInternal(operation.ReleaseCallback());
|
|
break;
|
|
default:
|
|
NOTREACHED();
|
|
}
|
|
// |this| may have been deleted.
|
|
}
|
|
}
|
|
|
|
void SimpleEntryImpl::OpenEntryInternal(bool have_index,
|
|
net::CompletionOnceCallback callback,
|
|
Entry** out_entry) {
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_OPEN_BEGIN);
|
|
|
|
if (state_ == STATE_READY) {
|
|
ReturnEntryToCaller(out_entry);
|
|
PostClientCallback(std::move(callback), net::OK);
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_OPEN_END,
|
|
CreateNetLogSimpleEntryCreationCallback(this, net::OK));
|
|
return;
|
|
}
|
|
if (state_ == STATE_FAILURE) {
|
|
PostClientCallback(std::move(callback), net::ERR_FAILED);
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_OPEN_END,
|
|
CreateNetLogSimpleEntryCreationCallback(this, net::ERR_FAILED));
|
|
return;
|
|
}
|
|
|
|
DCHECK_EQ(STATE_UNINITIALIZED, state_);
|
|
DCHECK(!synchronous_entry_);
|
|
state_ = STATE_IO_PENDING;
|
|
const base::TimeTicks start_time = base::TimeTicks::Now();
|
|
std::unique_ptr<SimpleEntryCreationResults> results(
|
|
new SimpleEntryCreationResults(SimpleEntryStat(
|
|
last_used_, last_modified_, data_size_, sparse_data_size_)));
|
|
|
|
base::OnceClosure task = base::BindOnce(
|
|
&SimpleSynchronousEntry::OpenEntry, cache_type_, path_, key_, entry_hash_,
|
|
have_index, start_time, file_tracker_, results.get());
|
|
|
|
base::OnceClosure reply = base::BindOnce(
|
|
&SimpleEntryImpl::CreationOperationComplete, this, std::move(callback),
|
|
start_time, base::Passed(&results), out_entry,
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_OPEN_END);
|
|
|
|
prioritized_task_runner_->PostTaskAndReply(FROM_HERE, std::move(task),
|
|
std::move(reply), entry_priority_);
|
|
}
|
|
|
|
void SimpleEntryImpl::CreateEntryInternal(bool have_index,
|
|
net::CompletionOnceCallback callback,
|
|
Entry** out_entry) {
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_CREATE_BEGIN);
|
|
|
|
if (state_ != STATE_UNINITIALIZED) {
|
|
// There is already an active normal entry.
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_CREATE_END,
|
|
CreateNetLogSimpleEntryCreationCallback(this, net::ERR_FAILED));
|
|
PostClientCallback(std::move(callback), net::ERR_FAILED);
|
|
return;
|
|
}
|
|
DCHECK_EQ(STATE_UNINITIALIZED, state_);
|
|
DCHECK(!synchronous_entry_);
|
|
|
|
state_ = STATE_IO_PENDING;
|
|
|
|
// Since we don't know the correct values for |last_used_| and
|
|
// |last_modified_| yet, we make this approximation.
|
|
last_used_ = last_modified_ = base::Time::Now();
|
|
|
|
// If creation succeeds, we should mark all streams to be saved on close.
|
|
for (int i = 0; i < kSimpleEntryStreamCount; ++i)
|
|
have_written_[i] = true;
|
|
|
|
const base::TimeTicks start_time = base::TimeTicks::Now();
|
|
std::unique_ptr<SimpleEntryCreationResults> results(
|
|
new SimpleEntryCreationResults(SimpleEntryStat(
|
|
last_used_, last_modified_, data_size_, sparse_data_size_)));
|
|
|
|
OnceClosure task = base::BindOnce(
|
|
&SimpleSynchronousEntry::CreateEntry, cache_type_, path_, key_,
|
|
entry_hash_, have_index, start_time, file_tracker_, results.get());
|
|
OnceClosure reply = base::BindOnce(
|
|
&SimpleEntryImpl::CreationOperationComplete, this, std::move(callback),
|
|
start_time, base::Passed(&results), out_entry,
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_CREATE_END);
|
|
prioritized_task_runner_->PostTaskAndReply(FROM_HERE, std::move(task),
|
|
std::move(reply), entry_priority_);
|
|
}
|
|
|
|
void SimpleEntryImpl::CloseInternal() {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
typedef SimpleSynchronousEntry::CRCRecord CRCRecord;
|
|
std::unique_ptr<std::vector<CRCRecord>> crc32s_to_write(
|
|
new std::vector<CRCRecord>());
|
|
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_CLOSE_BEGIN);
|
|
|
|
if (state_ == STATE_READY) {
|
|
DCHECK(synchronous_entry_);
|
|
state_ = STATE_IO_PENDING;
|
|
for (int i = 0; i < kSimpleEntryStreamCount; ++i) {
|
|
if (have_written_[i]) {
|
|
if (GetDataSize(i) == crc32s_end_offset_[i]) {
|
|
int32_t crc = GetDataSize(i) == 0 ? crc32(0, Z_NULL, 0) : crc32s_[i];
|
|
crc32s_to_write->push_back(CRCRecord(i, true, crc));
|
|
} else {
|
|
crc32s_to_write->push_back(CRCRecord(i, false, 0));
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
DCHECK(STATE_UNINITIALIZED == state_ || STATE_FAILURE == state_);
|
|
}
|
|
|
|
if (synchronous_entry_) {
|
|
Closure task = base::Bind(
|
|
&SimpleSynchronousEntry::Close, base::Unretained(synchronous_entry_),
|
|
SimpleEntryStat(last_used_, last_modified_, data_size_,
|
|
sparse_data_size_),
|
|
base::Passed(&crc32s_to_write), base::RetainedRef(stream_0_data_));
|
|
Closure reply = base::Bind(&SimpleEntryImpl::CloseOperationComplete, this);
|
|
synchronous_entry_ = NULL;
|
|
prioritized_task_runner_->PostTaskAndReply(
|
|
FROM_HERE, std::move(task), std::move(reply), entry_priority_);
|
|
|
|
for (int i = 0; i < kSimpleEntryStreamCount; ++i) {
|
|
if (!have_written_[i]) {
|
|
SIMPLE_CACHE_UMA(ENUMERATION,
|
|
"CheckCRCResult", cache_type_,
|
|
crc_check_state_[i], CRC_CHECK_MAX);
|
|
}
|
|
}
|
|
} else {
|
|
CloseOperationComplete();
|
|
}
|
|
}
|
|
|
|
int SimpleEntryImpl::ReadDataInternal(bool sync_possible,
|
|
int stream_index,
|
|
int offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
net::CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_BEGIN,
|
|
CreateNetLogReadWriteDataCallback(stream_index, offset,
|
|
buf_len, false));
|
|
}
|
|
|
|
if (state_ == STATE_FAILURE || state_ == STATE_UNINITIALIZED) {
|
|
RecordReadResult(cache_type_, READ_RESULT_BAD_STATE);
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_END,
|
|
CreateNetLogReadWriteCompleteCallback(net::ERR_FAILED));
|
|
}
|
|
// Note that the API states that client-provided callbacks for entry-level
|
|
// (i.e. non-backend) operations (e.g. read, write) are invoked even if
|
|
// the backend was already destroyed.
|
|
return PostToCallbackIfNeeded(sync_possible, std::move(callback),
|
|
net::ERR_FAILED);
|
|
}
|
|
DCHECK_EQ(STATE_READY, state_);
|
|
if (offset >= GetDataSize(stream_index) || offset < 0 || !buf_len) {
|
|
RecordReadResult(cache_type_, sync_possible
|
|
? READ_RESULT_NONBLOCK_EMPTY_RETURN
|
|
: READ_RESULT_FAST_EMPTY_RETURN);
|
|
// If there is nothing to read, we bail out before setting state_ to
|
|
// STATE_IO_PENDING (so ScopedOperationRunner might start us on next op
|
|
// here).
|
|
return PostToCallbackIfNeeded(sync_possible, std::move(callback), 0);
|
|
}
|
|
|
|
buf_len = std::min(buf_len, GetDataSize(stream_index) - offset);
|
|
|
|
// Since stream 0 data is kept in memory, it is read immediately.
|
|
if (stream_index == 0) {
|
|
int rv = ReadFromBuffer(stream_0_data_.get(), offset, buf_len, buf);
|
|
return PostToCallbackIfNeeded(sync_possible, std::move(callback), rv);
|
|
}
|
|
|
|
// Sometimes we can read in-ram prefetched stream 1 data immediately, too.
|
|
if (stream_index == 1) {
|
|
if (is_initial_stream1_read_) {
|
|
SIMPLE_CACHE_UMA(BOOLEAN, "ReadStream1FromPrefetched", cache_type_,
|
|
stream_1_prefetch_data_ != nullptr);
|
|
}
|
|
is_initial_stream1_read_ = false;
|
|
|
|
if (stream_1_prefetch_data_) {
|
|
int rv =
|
|
ReadFromBuffer(stream_1_prefetch_data_.get(), offset, buf_len, buf);
|
|
return PostToCallbackIfNeeded(sync_possible, std::move(callback), rv);
|
|
}
|
|
}
|
|
|
|
state_ = STATE_IO_PENDING;
|
|
if (doom_state_ == DOOM_NONE && backend_.get())
|
|
backend_->index()->UseIfExists(entry_hash_);
|
|
|
|
SimpleSynchronousEntry::ReadRequest read_req(stream_index, offset, buf_len);
|
|
// Figure out if we should be computing the checksum for this read,
|
|
// and whether we should be verifying it, too.
|
|
if (crc32s_end_offset_[stream_index] == offset) {
|
|
read_req.request_update_crc = true;
|
|
read_req.previous_crc32 =
|
|
offset == 0 ? crc32(0, Z_NULL, 0) : crc32s_[stream_index];
|
|
|
|
// We can't verify the checksum if we already overwrote part of the file.
|
|
// (It may still make sense to compute it if the overwritten area and the
|
|
// about-to-read-in area are adjoint).
|
|
read_req.request_verify_crc = !have_written_[stream_index];
|
|
}
|
|
|
|
std::unique_ptr<SimpleSynchronousEntry::ReadResult> result =
|
|
std::make_unique<SimpleSynchronousEntry::ReadResult>();
|
|
std::unique_ptr<SimpleEntryStat> entry_stat(new SimpleEntryStat(
|
|
last_used_, last_modified_, data_size_, sparse_data_size_));
|
|
OnceClosure task = base::BindOnce(
|
|
&SimpleSynchronousEntry::ReadData, base::Unretained(synchronous_entry_),
|
|
read_req, entry_stat.get(), base::RetainedRef(buf), result.get());
|
|
OnceClosure reply = base::BindOnce(
|
|
&SimpleEntryImpl::ReadOperationComplete, this, stream_index, offset,
|
|
std::move(callback), base::Passed(&entry_stat), base::Passed(&result));
|
|
prioritized_task_runner_->PostTaskAndReply(FROM_HERE, std::move(task),
|
|
std::move(reply), entry_priority_);
|
|
return net::ERR_IO_PENDING;
|
|
}
|
|
|
|
void SimpleEntryImpl::WriteDataInternal(int stream_index,
|
|
int offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
net::CompletionOnceCallback callback,
|
|
bool truncate) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_BEGIN,
|
|
CreateNetLogReadWriteDataCallback(stream_index, offset,
|
|
buf_len, truncate));
|
|
}
|
|
|
|
if (state_ == STATE_FAILURE || state_ == STATE_UNINITIALIZED) {
|
|
RecordWriteResult(cache_type_, SIMPLE_ENTRY_WRITE_RESULT_BAD_STATE);
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_END,
|
|
CreateNetLogReadWriteCompleteCallback(net::ERR_FAILED));
|
|
}
|
|
if (!callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(callback), net::ERR_FAILED));
|
|
}
|
|
// |this| may be destroyed after return here.
|
|
return;
|
|
}
|
|
|
|
DCHECK_EQ(STATE_READY, state_);
|
|
|
|
// Since stream 0 data is kept in memory, it will be written immediatly.
|
|
if (stream_index == 0) {
|
|
int ret_value = SetStream0Data(buf, offset, buf_len, truncate);
|
|
if (!callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(callback), ret_value));
|
|
}
|
|
return;
|
|
}
|
|
|
|
// Ignore zero-length writes that do not change the file size.
|
|
if (buf_len == 0) {
|
|
int32_t data_size = data_size_[stream_index];
|
|
if (truncate ? (offset == data_size) : (offset <= data_size)) {
|
|
RecordWriteResult(cache_type_,
|
|
SIMPLE_ENTRY_WRITE_RESULT_FAST_EMPTY_RETURN);
|
|
if (!callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(callback), 0));
|
|
}
|
|
return;
|
|
}
|
|
}
|
|
state_ = STATE_IO_PENDING;
|
|
if (doom_state_ == DOOM_NONE && backend_.get())
|
|
backend_->index()->UseIfExists(entry_hash_);
|
|
|
|
// Any stream 1 write invalidates the prefetched data.
|
|
if (stream_index == 1)
|
|
stream_1_prefetch_data_ = nullptr;
|
|
|
|
bool request_update_crc = false;
|
|
uint32_t initial_crc = 0;
|
|
|
|
if (offset < crc32s_end_offset_[stream_index]) {
|
|
// If a range for which the crc32 was already computed is rewritten, the
|
|
// computation of the crc32 need to start from 0 again.
|
|
crc32s_end_offset_[stream_index] = 0;
|
|
}
|
|
|
|
if (crc32s_end_offset_[stream_index] == offset) {
|
|
request_update_crc = true;
|
|
initial_crc = (offset != 0) ? crc32s_[stream_index] : crc32(0, Z_NULL, 0);
|
|
}
|
|
|
|
// |entry_stat| needs to be initialized before modifying |data_size_|.
|
|
std::unique_ptr<SimpleEntryStat> entry_stat(new SimpleEntryStat(
|
|
last_used_, last_modified_, data_size_, sparse_data_size_));
|
|
if (truncate) {
|
|
data_size_[stream_index] = offset + buf_len;
|
|
} else {
|
|
data_size_[stream_index] = std::max(offset + buf_len,
|
|
GetDataSize(stream_index));
|
|
}
|
|
|
|
std::unique_ptr<SimpleSynchronousEntry::WriteResult> write_result =
|
|
std::make_unique<SimpleSynchronousEntry::WriteResult>();
|
|
|
|
// Since we don't know the correct values for |last_used_| and
|
|
// |last_modified_| yet, we make this approximation.
|
|
last_used_ = last_modified_ = base::Time::Now();
|
|
|
|
have_written_[stream_index] = true;
|
|
// Writing on stream 1 affects the placement of stream 0 in the file, the EOF
|
|
// record will have to be rewritten.
|
|
if (stream_index == 1)
|
|
have_written_[0] = true;
|
|
|
|
// Retain a reference to |buf| in |reply| instead of |task|, so that we can
|
|
// reduce cross thread malloc/free pairs. The cross thread malloc/free pair
|
|
// increases the apparent memory usage due to the thread cached free list.
|
|
// TODO(morlovich): Remove the doom_state_ argument to WriteData, since with
|
|
// renaming rather than delete, creating a new stream 2 of doomed entry will
|
|
// just work.
|
|
OnceClosure task = base::BindOnce(
|
|
&SimpleSynchronousEntry::WriteData, base::Unretained(synchronous_entry_),
|
|
SimpleSynchronousEntry::WriteRequest(
|
|
stream_index, offset, buf_len, initial_crc, truncate,
|
|
doom_state_ != DOOM_NONE, request_update_crc),
|
|
base::Unretained(buf), entry_stat.get(), write_result.get());
|
|
OnceClosure reply = base::BindOnce(
|
|
&SimpleEntryImpl::WriteOperationComplete, this, stream_index,
|
|
std::move(callback), base::Passed(&entry_stat),
|
|
base::Passed(&write_result), base::RetainedRef(buf));
|
|
prioritized_task_runner_->PostTaskAndReply(FROM_HERE, std::move(task),
|
|
std::move(reply), entry_priority_);
|
|
}
|
|
|
|
void SimpleEntryImpl::ReadSparseDataInternal(
|
|
int64_t sparse_offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
net::CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_SPARSE_BEGIN,
|
|
CreateNetLogSparseOperationCallback(sparse_offset, buf_len));
|
|
}
|
|
|
|
if (state_ == STATE_FAILURE || state_ == STATE_UNINITIALIZED) {
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_SPARSE_END,
|
|
CreateNetLogReadWriteCompleteCallback(net::ERR_FAILED));
|
|
}
|
|
if (!callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(callback), net::ERR_FAILED));
|
|
}
|
|
// |this| may be destroyed after return here.
|
|
return;
|
|
}
|
|
|
|
DCHECK_EQ(STATE_READY, state_);
|
|
state_ = STATE_IO_PENDING;
|
|
|
|
std::unique_ptr<int> result(new int());
|
|
std::unique_ptr<base::Time> last_used(new base::Time());
|
|
OnceClosure task = base::BindOnce(
|
|
&SimpleSynchronousEntry::ReadSparseData,
|
|
base::Unretained(synchronous_entry_),
|
|
SimpleSynchronousEntry::SparseRequest(sparse_offset, buf_len),
|
|
base::RetainedRef(buf), last_used.get(), result.get());
|
|
OnceClosure reply = base::BindOnce(
|
|
&SimpleEntryImpl::ReadSparseOperationComplete, this, std::move(callback),
|
|
base::Passed(&last_used), base::Passed(&result));
|
|
prioritized_task_runner_->PostTaskAndReply(FROM_HERE, std::move(task),
|
|
std::move(reply), entry_priority_);
|
|
}
|
|
|
|
void SimpleEntryImpl::WriteSparseDataInternal(
|
|
int64_t sparse_offset,
|
|
net::IOBuffer* buf,
|
|
int buf_len,
|
|
net::CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_SPARSE_BEGIN,
|
|
CreateNetLogSparseOperationCallback(sparse_offset, buf_len));
|
|
}
|
|
|
|
if (state_ == STATE_FAILURE || state_ == STATE_UNINITIALIZED) {
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(
|
|
net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_SPARSE_END,
|
|
CreateNetLogReadWriteCompleteCallback(net::ERR_FAILED));
|
|
}
|
|
if (!callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(callback), net::ERR_FAILED));
|
|
}
|
|
// |this| may be destroyed after return here.
|
|
return;
|
|
}
|
|
|
|
DCHECK_EQ(STATE_READY, state_);
|
|
state_ = STATE_IO_PENDING;
|
|
|
|
uint64_t max_sparse_data_size = std::numeric_limits<int64_t>::max();
|
|
if (backend_.get()) {
|
|
uint64_t max_cache_size = backend_->index()->max_size();
|
|
max_sparse_data_size = max_cache_size / kMaxSparseDataSizeDivisor;
|
|
}
|
|
|
|
std::unique_ptr<SimpleEntryStat> entry_stat(new SimpleEntryStat(
|
|
last_used_, last_modified_, data_size_, sparse_data_size_));
|
|
|
|
last_used_ = last_modified_ = base::Time::Now();
|
|
|
|
std::unique_ptr<int> result(new int());
|
|
OnceClosure task = base::BindOnce(
|
|
&SimpleSynchronousEntry::WriteSparseData,
|
|
base::Unretained(synchronous_entry_),
|
|
SimpleSynchronousEntry::SparseRequest(sparse_offset, buf_len),
|
|
base::RetainedRef(buf), max_sparse_data_size, entry_stat.get(),
|
|
result.get());
|
|
OnceClosure reply = base::BindOnce(
|
|
&SimpleEntryImpl::WriteSparseOperationComplete, this, std::move(callback),
|
|
base::Passed(&entry_stat), base::Passed(&result));
|
|
prioritized_task_runner_->PostTaskAndReply(FROM_HERE, std::move(task),
|
|
std::move(reply), entry_priority_);
|
|
}
|
|
|
|
void SimpleEntryImpl::GetAvailableRangeInternal(
|
|
int64_t sparse_offset,
|
|
int len,
|
|
int64_t* out_start,
|
|
net::CompletionOnceCallback callback) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
ScopedOperationRunner operation_runner(this);
|
|
|
|
if (state_ == STATE_FAILURE || state_ == STATE_UNINITIALIZED) {
|
|
if (!callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(callback), net::ERR_FAILED));
|
|
}
|
|
// |this| may be destroyed after return here.
|
|
return;
|
|
}
|
|
|
|
DCHECK_EQ(STATE_READY, state_);
|
|
state_ = STATE_IO_PENDING;
|
|
|
|
std::unique_ptr<int> result(new int());
|
|
OnceClosure task =
|
|
base::BindOnce(&SimpleSynchronousEntry::GetAvailableRange,
|
|
base::Unretained(synchronous_entry_),
|
|
SimpleSynchronousEntry::SparseRequest(sparse_offset, len),
|
|
out_start, result.get());
|
|
OnceClosure reply =
|
|
base::BindOnce(&SimpleEntryImpl::GetAvailableRangeOperationComplete, this,
|
|
std::move(callback), base::Passed(&result));
|
|
prioritized_task_runner_->PostTaskAndReply(FROM_HERE, std::move(task),
|
|
std::move(reply), entry_priority_);
|
|
}
|
|
|
|
void SimpleEntryImpl::DoomEntryInternal(net::CompletionOnceCallback callback) {
|
|
if (doom_state_ == DOOM_COMPLETED) {
|
|
// During the time we were sitting on a queue, some operation failed
|
|
// and cleaned our files up, so we don't have to do anything.
|
|
DoomOperationComplete(std::move(callback), state_, net::OK);
|
|
return;
|
|
}
|
|
|
|
if (!backend_) {
|
|
// If there's no backend, we want to truncate the files rather than delete
|
|
// or rename them. Either op will update the entry directory's mtime, which
|
|
// will likely force a full index rebuild on the next startup; this is
|
|
// clearly an undesirable cost. Instead, the lesser evil is to set the entry
|
|
// files to length zero, leaving the invalid entry in the index. On the next
|
|
// attempt to open the entry, it will fail asynchronously (since the magic
|
|
// numbers will not be found), and the files will actually be removed.
|
|
// Since there is no backend, new entries to conflict with us also can't be
|
|
// created.
|
|
prioritized_task_runner_->PostTaskAndReplyWithResult(
|
|
FROM_HERE,
|
|
base::BindOnce(&SimpleSynchronousEntry::TruncateEntryFiles, path_,
|
|
entry_hash_),
|
|
base::BindOnce(&SimpleEntryImpl::DoomOperationComplete, this,
|
|
std::move(callback),
|
|
// Return to STATE_FAILURE after dooming, since no
|
|
// operation can succeed on the truncated entry files.
|
|
STATE_FAILURE),
|
|
entry_priority_);
|
|
state_ = STATE_IO_PENDING;
|
|
return;
|
|
}
|
|
|
|
if (synchronous_entry_) {
|
|
// If there is a backing object, we have to go through its instance methods,
|
|
// so that it can rename itself and keep track of the altenative name.
|
|
prioritized_task_runner_->PostTaskAndReplyWithResult(
|
|
FROM_HERE,
|
|
base::BindOnce(&SimpleSynchronousEntry::Doom,
|
|
base::Unretained(synchronous_entry_)),
|
|
base::BindOnce(&SimpleEntryImpl::DoomOperationComplete, this,
|
|
std::move(callback), state_),
|
|
entry_priority_);
|
|
} else {
|
|
DCHECK_EQ(STATE_UNINITIALIZED, state_);
|
|
// If nothing is open, we can just delete the files. We know they have the
|
|
// base names, since if we ever renamed them our doom_state_ would be
|
|
// DOOM_COMPLETED, and we would exit at function entry.
|
|
prioritized_task_runner_->PostTaskAndReplyWithResult(
|
|
FROM_HERE,
|
|
base::BindOnce(&SimpleSynchronousEntry::DeleteEntryFiles, path_,
|
|
cache_type_, entry_hash_),
|
|
base::BindOnce(&SimpleEntryImpl::DoomOperationComplete, this,
|
|
std::move(callback), state_),
|
|
entry_priority_);
|
|
}
|
|
state_ = STATE_IO_PENDING;
|
|
}
|
|
|
|
void SimpleEntryImpl::CreationOperationComplete(
|
|
net::CompletionOnceCallback completion_callback,
|
|
const base::TimeTicks& start_time,
|
|
std::unique_ptr<SimpleEntryCreationResults> in_results,
|
|
Entry** out_entry,
|
|
net::NetLogEventType end_event_type) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK_EQ(state_, STATE_IO_PENDING);
|
|
DCHECK(in_results);
|
|
ScopedOperationRunner operation_runner(this);
|
|
SIMPLE_CACHE_UMA(BOOLEAN,
|
|
"EntryCreationResult", cache_type_,
|
|
in_results->result == net::OK);
|
|
if (in_results->result != net::OK) {
|
|
if (in_results->result != net::ERR_FILE_EXISTS) {
|
|
// Here we keep index up-to-date, but don't remove ourselves from active
|
|
// entries since we may have queued operations, and it would be
|
|
// problematic to run further Creates, Opens, or Dooms if we are not
|
|
// the active entry. We can only do this because OpenEntryInternal
|
|
// and CreateEntryInternal have to start from STATE_UNINITIALIZED, so
|
|
// nothing else is going on which may be confused.
|
|
if (backend_)
|
|
backend_->index()->Remove(entry_hash_);
|
|
}
|
|
|
|
net_log_.AddEventWithNetErrorCode(end_event_type, net::ERR_FAILED);
|
|
PostClientCallback(std::move(completion_callback), net::ERR_FAILED);
|
|
ResetEntry();
|
|
return;
|
|
}
|
|
|
|
// Make sure to keep the index up-to-date. We likely already did this when
|
|
// CreateEntry was called, but it's possible we were sitting on a queue
|
|
// after an op that removed us.
|
|
if (backend_ && doom_state_ == DOOM_NONE)
|
|
backend_->index()->Insert(entry_hash_);
|
|
|
|
// If out_entry is NULL, it means we already called ReturnEntryToCaller from
|
|
// the optimistic Create case.
|
|
if (out_entry)
|
|
ReturnEntryToCaller(out_entry);
|
|
|
|
state_ = STATE_READY;
|
|
synchronous_entry_ = in_results->sync_entry;
|
|
|
|
// Copy over any pre-fetched data and its CRCs.
|
|
for (int stream = 0; stream < 2; ++stream) {
|
|
const SimpleStreamPrefetchData& prefetched =
|
|
in_results->stream_prefetch_data[stream];
|
|
if (prefetched.data.get()) {
|
|
if (stream == 0)
|
|
stream_0_data_ = prefetched.data;
|
|
else
|
|
stream_1_prefetch_data_ = prefetched.data;
|
|
|
|
// The crc was read in SimpleSynchronousEntry.
|
|
crc_check_state_[stream] = CRC_CHECK_DONE;
|
|
crc32s_[stream] = prefetched.stream_crc32;
|
|
crc32s_end_offset_[stream] = in_results->entry_stat.data_size(stream);
|
|
}
|
|
}
|
|
|
|
// If this entry was opened by hash, key_ could still be empty. If so, update
|
|
// it with the key read from the synchronous entry.
|
|
if (key_.empty()) {
|
|
SetKey(synchronous_entry_->key());
|
|
} else {
|
|
// This should only be triggered when creating an entry. In the open case
|
|
// the key is either copied from the arguments to open, or checked
|
|
// in the synchronous entry.
|
|
DCHECK_EQ(key_, synchronous_entry_->key());
|
|
}
|
|
UpdateDataFromEntryStat(in_results->entry_stat);
|
|
SIMPLE_CACHE_UMA(TIMES,
|
|
"EntryCreationTime", cache_type_,
|
|
(base::TimeTicks::Now() - start_time));
|
|
AdjustOpenEntryCountBy(cache_type_, 1);
|
|
|
|
net_log_.AddEvent(end_event_type);
|
|
PostClientCallback(std::move(completion_callback), net::OK);
|
|
}
|
|
|
|
void SimpleEntryImpl::EntryOperationComplete(
|
|
net::CompletionOnceCallback completion_callback,
|
|
const SimpleEntryStat& entry_stat,
|
|
int result) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK(synchronous_entry_);
|
|
DCHECK_EQ(STATE_IO_PENDING, state_);
|
|
if (result < 0) {
|
|
state_ = STATE_FAILURE;
|
|
MarkAsDoomed(DOOM_COMPLETED);
|
|
} else {
|
|
state_ = STATE_READY;
|
|
UpdateDataFromEntryStat(entry_stat);
|
|
}
|
|
|
|
if (!completion_callback.is_null()) {
|
|
base::ThreadTaskRunnerHandle::Get()->PostTask(
|
|
FROM_HERE, base::BindOnce(std::move(completion_callback), result));
|
|
}
|
|
RunNextOperationIfNeeded();
|
|
}
|
|
|
|
void SimpleEntryImpl::ReadOperationComplete(
|
|
int stream_index,
|
|
int offset,
|
|
net::CompletionOnceCallback completion_callback,
|
|
std::unique_ptr<SimpleEntryStat> entry_stat,
|
|
std::unique_ptr<SimpleSynchronousEntry::ReadResult> read_result) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK(synchronous_entry_);
|
|
DCHECK_EQ(STATE_IO_PENDING, state_);
|
|
DCHECK(read_result);
|
|
int result = read_result->result;
|
|
|
|
if (result > 0 &&
|
|
crc_check_state_[stream_index] == CRC_CHECK_NEVER_READ_AT_ALL) {
|
|
crc_check_state_[stream_index] = CRC_CHECK_NEVER_READ_TO_END;
|
|
}
|
|
|
|
if (read_result->crc_updated) {
|
|
if (result > 0) {
|
|
DCHECK_EQ(crc32s_end_offset_[stream_index], offset);
|
|
crc32s_end_offset_[stream_index] += result;
|
|
crc32s_[stream_index] = read_result->updated_crc32;
|
|
}
|
|
|
|
if (read_result->crc_performed_verify)
|
|
crc_check_state_[stream_index] = CRC_CHECK_DONE;
|
|
}
|
|
|
|
if (result < 0) {
|
|
crc32s_end_offset_[stream_index] = 0;
|
|
} else {
|
|
if (crc_check_state_[stream_index] == CRC_CHECK_NEVER_READ_TO_END &&
|
|
offset + result == GetDataSize(stream_index)) {
|
|
crc_check_state_[stream_index] = CRC_CHECK_NOT_DONE;
|
|
}
|
|
}
|
|
RecordReadResultConsideringChecksum(read_result);
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_END,
|
|
CreateNetLogReadWriteCompleteCallback(result));
|
|
}
|
|
|
|
EntryOperationComplete(std::move(completion_callback), *entry_stat, result);
|
|
}
|
|
|
|
void SimpleEntryImpl::WriteOperationComplete(
|
|
int stream_index,
|
|
net::CompletionOnceCallback completion_callback,
|
|
std::unique_ptr<SimpleEntryStat> entry_stat,
|
|
std::unique_ptr<SimpleSynchronousEntry::WriteResult> write_result,
|
|
net::IOBuffer* buf) {
|
|
int result = write_result->result;
|
|
if (result >= 0)
|
|
RecordWriteResult(cache_type_, SIMPLE_ENTRY_WRITE_RESULT_SUCCESS);
|
|
else
|
|
RecordWriteResult(cache_type_,
|
|
SIMPLE_ENTRY_WRITE_RESULT_SYNC_WRITE_FAILURE);
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_END,
|
|
CreateNetLogReadWriteCompleteCallback(result));
|
|
}
|
|
|
|
if (result < 0)
|
|
crc32s_end_offset_[stream_index] = 0;
|
|
|
|
if (result > 0 && write_result->crc_updated) {
|
|
crc32s_end_offset_[stream_index] += result;
|
|
crc32s_[stream_index] = write_result->updated_crc32;
|
|
}
|
|
|
|
EntryOperationComplete(std::move(completion_callback), *entry_stat, result);
|
|
}
|
|
|
|
void SimpleEntryImpl::ReadSparseOperationComplete(
|
|
net::CompletionOnceCallback completion_callback,
|
|
std::unique_ptr<base::Time> last_used,
|
|
std::unique_ptr<int> result) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK(synchronous_entry_);
|
|
DCHECK(result);
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_READ_SPARSE_END,
|
|
CreateNetLogReadWriteCompleteCallback(*result));
|
|
}
|
|
|
|
SimpleEntryStat entry_stat(*last_used, last_modified_, data_size_,
|
|
sparse_data_size_);
|
|
EntryOperationComplete(std::move(completion_callback), entry_stat, *result);
|
|
}
|
|
|
|
void SimpleEntryImpl::WriteSparseOperationComplete(
|
|
net::CompletionOnceCallback completion_callback,
|
|
std::unique_ptr<SimpleEntryStat> entry_stat,
|
|
std::unique_ptr<int> result) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK(synchronous_entry_);
|
|
DCHECK(result);
|
|
|
|
if (net_log_.IsCapturing()) {
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_WRITE_SPARSE_END,
|
|
CreateNetLogReadWriteCompleteCallback(*result));
|
|
}
|
|
|
|
EntryOperationComplete(std::move(completion_callback), *entry_stat, *result);
|
|
}
|
|
|
|
void SimpleEntryImpl::GetAvailableRangeOperationComplete(
|
|
net::CompletionOnceCallback completion_callback,
|
|
std::unique_ptr<int> result) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK(synchronous_entry_);
|
|
DCHECK(result);
|
|
|
|
SimpleEntryStat entry_stat(last_used_, last_modified_, data_size_,
|
|
sparse_data_size_);
|
|
EntryOperationComplete(std::move(completion_callback), entry_stat, *result);
|
|
}
|
|
|
|
void SimpleEntryImpl::DoomOperationComplete(
|
|
net::CompletionOnceCallback callback,
|
|
State state_to_restore,
|
|
int result) {
|
|
state_ = state_to_restore;
|
|
doom_state_ = DOOM_COMPLETED;
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_DOOM_END);
|
|
PostClientCallback(std::move(callback), result);
|
|
RunNextOperationIfNeeded();
|
|
if (backend_)
|
|
backend_->OnDoomComplete(entry_hash_);
|
|
}
|
|
|
|
void SimpleEntryImpl::RecordReadResultConsideringChecksum(
|
|
const std::unique_ptr<SimpleSynchronousEntry::ReadResult>& read_result)
|
|
const {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK(synchronous_entry_);
|
|
DCHECK_EQ(STATE_IO_PENDING, state_);
|
|
|
|
if (read_result->result >= 0) {
|
|
RecordReadResult(cache_type_, READ_RESULT_SUCCESS);
|
|
} else {
|
|
if (read_result->crc_updated && read_result->crc_performed_verify &&
|
|
!read_result->crc_verify_ok)
|
|
RecordReadResult(cache_type_, READ_RESULT_SYNC_CHECKSUM_FAILURE);
|
|
else
|
|
RecordReadResult(cache_type_, READ_RESULT_SYNC_READ_FAILURE);
|
|
}
|
|
}
|
|
|
|
void SimpleEntryImpl::CloseOperationComplete() {
|
|
DCHECK(!synchronous_entry_);
|
|
DCHECK_EQ(0, open_count_);
|
|
DCHECK(STATE_IO_PENDING == state_ || STATE_FAILURE == state_ ||
|
|
STATE_UNINITIALIZED == state_);
|
|
net_log_.AddEvent(net::NetLogEventType::SIMPLE_CACHE_ENTRY_CLOSE_END);
|
|
AdjustOpenEntryCountBy(cache_type_, -1);
|
|
ResetEntry();
|
|
RunNextOperationIfNeeded();
|
|
}
|
|
|
|
void SimpleEntryImpl::UpdateDataFromEntryStat(
|
|
const SimpleEntryStat& entry_stat) {
|
|
DCHECK(io_thread_checker_.CalledOnValidThread());
|
|
DCHECK(synchronous_entry_);
|
|
DCHECK_EQ(STATE_READY, state_);
|
|
|
|
last_used_ = entry_stat.last_used();
|
|
last_modified_ = entry_stat.last_modified();
|
|
for (int i = 0; i < kSimpleEntryStreamCount; ++i) {
|
|
data_size_[i] = entry_stat.data_size(i);
|
|
}
|
|
sparse_data_size_ = entry_stat.sparse_data_size();
|
|
if (doom_state_ == DOOM_NONE && backend_.get()) {
|
|
backend_->index()->UpdateEntrySize(
|
|
entry_hash_, base::checked_cast<uint32_t>(GetDiskUsage()));
|
|
}
|
|
}
|
|
|
|
int64_t SimpleEntryImpl::GetDiskUsage() const {
|
|
int64_t file_size = 0;
|
|
for (int i = 0; i < kSimpleEntryStreamCount; ++i) {
|
|
file_size +=
|
|
simple_util::GetFileSizeFromDataSize(key_.size(), data_size_[i]);
|
|
}
|
|
file_size += sparse_data_size_;
|
|
return file_size;
|
|
}
|
|
|
|
int SimpleEntryImpl::ReadFromBuffer(net::GrowableIOBuffer* in_buf,
|
|
int offset,
|
|
int buf_len,
|
|
net::IOBuffer* out_buf) {
|
|
DCHECK_GE(buf_len, 0);
|
|
|
|
memcpy(out_buf->data(), in_buf->data() + offset, buf_len);
|
|
UpdateDataFromEntryStat(SimpleEntryStat(base::Time::Now(), last_modified_,
|
|
data_size_, sparse_data_size_));
|
|
RecordReadResult(cache_type_, READ_RESULT_SUCCESS);
|
|
return buf_len;
|
|
}
|
|
|
|
int SimpleEntryImpl::SetStream0Data(net::IOBuffer* buf,
|
|
int offset,
|
|
int buf_len,
|
|
bool truncate) {
|
|
// Currently, stream 0 is only used for HTTP headers, and always writes them
|
|
// with a single, truncating write. Detect these writes and record the size
|
|
// changes of the headers. Also, support writes to stream 0 that have
|
|
// different access patterns, as required by the API contract.
|
|
// All other clients of the Simple Cache are encouraged to use stream 1.
|
|
have_written_[0] = true;
|
|
int data_size = GetDataSize(0);
|
|
if (offset == 0 && truncate) {
|
|
stream_0_data_->SetCapacity(buf_len);
|
|
memcpy(stream_0_data_->data(), buf->data(), buf_len);
|
|
data_size_[0] = buf_len;
|
|
} else {
|
|
const int buffer_size =
|
|
truncate ? offset + buf_len : std::max(offset + buf_len, data_size);
|
|
stream_0_data_->SetCapacity(buffer_size);
|
|
// If |stream_0_data_| was extended, the extension until offset needs to be
|
|
// zero-filled.
|
|
const int fill_size = offset <= data_size ? 0 : offset - data_size;
|
|
if (fill_size > 0)
|
|
memset(stream_0_data_->data() + data_size, 0, fill_size);
|
|
if (buf)
|
|
memcpy(stream_0_data_->data() + offset, buf->data(), buf_len);
|
|
data_size_[0] = buffer_size;
|
|
}
|
|
RecordHeaderSize(cache_type_, data_size_[0]);
|
|
base::Time modification_time = base::Time::Now();
|
|
|
|
// Reset checksum; SimpleSynchronousEntry::Close will compute it for us,
|
|
// and do it off the I/O thread.
|
|
crc32s_end_offset_[0] = 0;
|
|
|
|
UpdateDataFromEntryStat(
|
|
SimpleEntryStat(modification_time, modification_time, data_size_,
|
|
sparse_data_size_));
|
|
RecordWriteResult(cache_type_, SIMPLE_ENTRY_WRITE_RESULT_SUCCESS);
|
|
return buf_len;
|
|
}
|
|
|
|
} // namespace disk_cache
|