mirror of
https://github.com/klzgrad/naiveproxy.git
synced 2024-11-24 14:26:09 +03:00
89 lines
2.8 KiB
C++
89 lines
2.8 KiB
C++
|
// Copyright 2015 The Chromium Authors. All rights reserved.
|
||
|
// Use of this source code is governed by a BSD-style license that can be
|
||
|
// found in the LICENSE file.
|
||
|
|
||
|
#include "base/trace_event/heap_profiler_allocation_context.h"
|
||
|
|
||
|
#include <cstring>
|
||
|
|
||
|
#include "base/hash.h"
|
||
|
#include "base/macros.h"
|
||
|
|
||
|
namespace base {
|
||
|
namespace trace_event {
|
||
|
|
||
|
bool operator < (const StackFrame& lhs, const StackFrame& rhs) {
|
||
|
return lhs.value < rhs.value;
|
||
|
}
|
||
|
|
||
|
bool operator == (const StackFrame& lhs, const StackFrame& rhs) {
|
||
|
return lhs.value == rhs.value;
|
||
|
}
|
||
|
|
||
|
bool operator != (const StackFrame& lhs, const StackFrame& rhs) {
|
||
|
return !(lhs.value == rhs.value);
|
||
|
}
|
||
|
|
||
|
Backtrace::Backtrace() = default;
|
||
|
|
||
|
bool operator==(const Backtrace& lhs, const Backtrace& rhs) {
|
||
|
if (lhs.frame_count != rhs.frame_count) return false;
|
||
|
return std::equal(lhs.frames, lhs.frames + lhs.frame_count, rhs.frames);
|
||
|
}
|
||
|
|
||
|
bool operator!=(const Backtrace& lhs, const Backtrace& rhs) {
|
||
|
return !(lhs == rhs);
|
||
|
}
|
||
|
|
||
|
AllocationContext::AllocationContext(): type_name(nullptr) {}
|
||
|
|
||
|
AllocationContext::AllocationContext(const Backtrace& backtrace,
|
||
|
const char* type_name)
|
||
|
: backtrace(backtrace), type_name(type_name) {}
|
||
|
|
||
|
bool operator==(const AllocationContext& lhs, const AllocationContext& rhs) {
|
||
|
return (lhs.backtrace == rhs.backtrace) && (lhs.type_name == rhs.type_name);
|
||
|
}
|
||
|
|
||
|
bool operator!=(const AllocationContext& lhs, const AllocationContext& rhs) {
|
||
|
return !(lhs == rhs);
|
||
|
}
|
||
|
|
||
|
} // namespace trace_event
|
||
|
} // namespace base
|
||
|
|
||
|
namespace std {
|
||
|
|
||
|
using base::trace_event::AllocationContext;
|
||
|
using base::trace_event::Backtrace;
|
||
|
using base::trace_event::StackFrame;
|
||
|
|
||
|
size_t hash<StackFrame>::operator()(const StackFrame& frame) const {
|
||
|
return hash<const void*>()(frame.value);
|
||
|
}
|
||
|
|
||
|
size_t hash<Backtrace>::operator()(const Backtrace& backtrace) const {
|
||
|
const void* values[Backtrace::kMaxFrameCount];
|
||
|
for (size_t i = 0; i != backtrace.frame_count; ++i) {
|
||
|
values[i] = backtrace.frames[i].value;
|
||
|
}
|
||
|
return base::PersistentHash(values, backtrace.frame_count * sizeof(*values));
|
||
|
}
|
||
|
|
||
|
size_t hash<AllocationContext>::operator()(const AllocationContext& ctx) const {
|
||
|
size_t backtrace_hash = hash<Backtrace>()(ctx.backtrace);
|
||
|
|
||
|
// Multiplicative hash from [Knuth 1998]. Works best if |size_t| is 32 bits,
|
||
|
// because the magic number is a prime very close to 2^32 / golden ratio, but
|
||
|
// will still redistribute keys bijectively on 64-bit architectures because
|
||
|
// the magic number is coprime to 2^64.
|
||
|
size_t type_hash = reinterpret_cast<size_t>(ctx.type_name) * 2654435761;
|
||
|
|
||
|
// Multiply one side to break the commutativity of +. Multiplication with a
|
||
|
// number coprime to |numeric_limits<size_t>::max() + 1| is bijective so
|
||
|
// randomness is preserved.
|
||
|
return (backtrace_hash * 3) + type_hash;
|
||
|
}
|
||
|
|
||
|
} // namespace std
|