mirror of
https://github.com/klzgrad/naiveproxy.git
synced 2024-11-24 06:16:30 +03:00
317 lines
13 KiB
Markdown
317 lines
13 KiB
Markdown
|
# base/containers library
|
|||
|
|
|||
|
## What goes here
|
|||
|
|
|||
|
This directory contains some STL-like containers.
|
|||
|
|
|||
|
Things should be moved here that are generally applicable across the code base.
|
|||
|
Don't add things here just because you need them in one place and think others
|
|||
|
may someday want something similar. You can put specialized containers in
|
|||
|
your component's directory and we can promote them here later if we feel there
|
|||
|
is broad applicability.
|
|||
|
|
|||
|
### Design and naming
|
|||
|
|
|||
|
Containers should adhere as closely to STL as possible. Functions and behaviors
|
|||
|
not present in STL should only be added when they are related to the specific
|
|||
|
data structure implemented by the container.
|
|||
|
|
|||
|
For STL-like containers our policy is that they should use STL-like naming even
|
|||
|
when it may conflict with the style guide. So functions and class names should
|
|||
|
be lower case with underscores. Non-STL-like classes and functions should use
|
|||
|
Google naming. Be sure to use the base namespace.
|
|||
|
|
|||
|
## Map and set selection
|
|||
|
|
|||
|
### Usage advice
|
|||
|
|
|||
|
* Generally avoid **std::unordered\_set** and **std::unordered\_map**. In the
|
|||
|
common case, query performance is unlikely to be sufficiently higher than
|
|||
|
std::map to make a difference, insert performance is slightly worse, and
|
|||
|
the memory overhead is high. This makes sense mostly for large tables where
|
|||
|
you expect a lot of lookups.
|
|||
|
|
|||
|
* Most maps and sets in Chrome are small and contain objects that can be
|
|||
|
moved efficiently. In this case, consider **base::flat\_map** and
|
|||
|
**base::flat\_set**. You need to be aware of the maximum expected size of
|
|||
|
the container since individual inserts and deletes are O(n), giving O(n^2)
|
|||
|
construction time for the entire map. But because it avoids mallocs in most
|
|||
|
cases, inserts are better or comparable to other containers even for
|
|||
|
several dozen items, and efficiently-moved types are unlikely to have
|
|||
|
performance problems for most cases until you have hundreds of items. If
|
|||
|
your container can be constructed in one shot, the constructor from vector
|
|||
|
gives O(n log n) construction times and it should be strictly better than
|
|||
|
a std::map.
|
|||
|
|
|||
|
* **base::small\_map** has better runtime memory usage without the poor
|
|||
|
mutation performance of large containers that base::flat\_map has. But this
|
|||
|
advantage is partially offset by additional code size. Prefer in cases
|
|||
|
where you make<6B>many objects so that the code/heap tradeoff is good.
|
|||
|
|
|||
|
* Use **std::map** and **std::set** if you can't decide. Even if they're not
|
|||
|
great, they're unlikely to be bad or surprising.
|
|||
|
|
|||
|
### Map and set details
|
|||
|
|
|||
|
Sizes are on 64-bit platforms. Stable iterators aren't invalidated when the
|
|||
|
container is mutated.
|
|||
|
|
|||
|
| Container | Empty size | Per-item overhead | Stable iterators? |
|
|||
|
|:---------------------------------------- |:--------------------- |:----------------- |:----------------- |
|
|||
|
| std::map, std::set | 16 bytes | 32 bytes | Yes |
|
|||
|
| std::unordered\_map, std::unordered\_set | 128 bytes | 16-24 bytes | No |
|
|||
|
| base::flat\_map and base::flat\_set | 24 bytes | 0 (see notes) | No |
|
|||
|
| base::small\_map | 24 bytes (see notes) | 32 bytes | No |
|
|||
|
|
|||
|
**Takeaways:** std::unordered\_map and std::unordered\_map have high
|
|||
|
overhead for small container sizes, prefer these only for larger workloads.
|
|||
|
|
|||
|
Code size comparisons for a block of code (see appendix) on Windows using
|
|||
|
strings as keys.
|
|||
|
|
|||
|
| Container | Code size |
|
|||
|
|:------------------- |:---------- |
|
|||
|
| std::unordered\_map | 1646 bytes |
|
|||
|
| std::map | 1759 bytes |
|
|||
|
| base::flat\_map | 1872 bytes |
|
|||
|
| base::small\_map | 2410 bytes |
|
|||
|
|
|||
|
**Takeaways:** base::small\_map generates more code because of the inlining of
|
|||
|
both brute-force and red-black tree searching. This makes it less attractive
|
|||
|
for random one-off uses. But if your code is called frequently, the runtime
|
|||
|
memory benefits will be more important. The code sizes of the other maps are
|
|||
|
close enough it's not worth worrying about.
|
|||
|
|
|||
|
### std::map and std::set
|
|||
|
|
|||
|
A red-black tree. Each inserted item requires the memory allocation of a node
|
|||
|
on the heap. Each node contains a left pointer, a right pointer, a parent
|
|||
|
pointer, and a "color" for the red-black tree (32-bytes per item on 64-bits).
|
|||
|
|
|||
|
### std::unordered\_map and std::unordered\_set
|
|||
|
|
|||
|
A hash table. Implemented on Windows as a std::vector + std::list and in libc++
|
|||
|
as the equivalent of a std::vector + a std::forward\_list. Both implementations
|
|||
|
allocate an 8-entry hash table (containing iterators into the list) on
|
|||
|
initialization, and grow to 64 entries once 8 items are inserted. Above 64
|
|||
|
items, the size doubles every time the load factor exceeds 1.
|
|||
|
|
|||
|
The empty size is sizeof(std::unordered\_map) = 64 +
|
|||
|
the initial hash table size which is 8 pointers. The per-item overhead in the
|
|||
|
table above counts the list node (2 pointers on Windows, 1 pointer in libc++),
|
|||
|
plus amortizes the hash table assuming a 0.5 load factor on average.
|
|||
|
|
|||
|
In a microbenchmark on Windows, inserts of 1M integers into a
|
|||
|
std::unordered\_set took 1.07x the time of std::set, and queries took 0.67x the
|
|||
|
time of std::set. For a typical 4-entry set (the statistical mode of map sizes
|
|||
|
in the browser), query performance is identical to std::set and base::flat\_set.
|
|||
|
On ARM, unordered\_set performance can be worse because integer division to
|
|||
|
compute the bucket is slow, and a few "less than" operations can be faster than
|
|||
|
computing a hash depending on the key type. The takeaway is that you should not
|
|||
|
default to using unordered maps because "they're faster."
|
|||
|
|
|||
|
### base::flat\_map and base::flat\_set
|
|||
|
|
|||
|
A sorted std::vector. Seached via binary search, inserts in the middle require
|
|||
|
moving elements to make room. Good cache locality. For large objects and large
|
|||
|
set sizes, std::vector's doubling-when-full strategy can waste memory.
|
|||
|
|
|||
|
Supports efficient construction from a vector of items which avoids the O(n^2)
|
|||
|
insertion time of each element separately.
|
|||
|
|
|||
|
The per-item overhead will depend on the underlying std::vector's reallocation
|
|||
|
strategy and the memory access pattern. Assuming items are being linearly added,
|
|||
|
one would expect it to be 3/4 full, so per-item overhead will be 0.25 *
|
|||
|
sizeof(T).
|
|||
|
|
|||
|
|
|||
|
flat\_set/flat\_map support a notion of transparent comparisons. Therefore you
|
|||
|
can, for example, lookup base::StringPiece in a set of std::strings without
|
|||
|
constructing a temporary std::string. This functionality is based on C++14
|
|||
|
extensions to std::set/std::map interface.
|
|||
|
|
|||
|
You can find more information about transparent comparisons here:
|
|||
|
http://en.cppreference.com/w/cpp/utility/functional/less_void
|
|||
|
|
|||
|
Example, smart pointer set:
|
|||
|
|
|||
|
```cpp
|
|||
|
// Define a custom comparator.
|
|||
|
struct UniquePtrComparator {
|
|||
|
// Mark your comparison as transparent.
|
|||
|
using is_transparent = int;
|
|||
|
|
|||
|
template <typename T>
|
|||
|
bool operator()(const std::unique_ptr<T>& lhs,
|
|||
|
const std::unique_ptr<T>& rhs) const {
|
|||
|
return lhs < rhs;
|
|||
|
}
|
|||
|
|
|||
|
template <typename T>
|
|||
|
bool operator()(const T* lhs, const std::unique_ptr<T>& rhs) const {
|
|||
|
return lhs < rhs.get();
|
|||
|
}
|
|||
|
|
|||
|
template <typename T>
|
|||
|
bool operator()(const std::unique_ptr<T>& lhs, const T* rhs) const {
|
|||
|
return lhs.get() < rhs;
|
|||
|
}
|
|||
|
};
|
|||
|
|
|||
|
// Declare a typedef.
|
|||
|
template <typename T>
|
|||
|
using UniquePtrSet = base::flat_set<std::unique_ptr<T>, UniquePtrComparator>;
|
|||
|
|
|||
|
// ...
|
|||
|
// Collect data.
|
|||
|
std::vector<std::unique_ptr<int>> ptr_vec;
|
|||
|
ptr_vec.reserve(5);
|
|||
|
std::generate_n(std::back_inserter(ptr_vec), 5, []{
|
|||
|
return std::make_unique<int>(0);
|
|||
|
});
|
|||
|
|
|||
|
// Construct a set.
|
|||
|
UniquePtrSet<int> ptr_set(std::move(ptr_vec), base::KEEP_FIRST_OF_DUPES);
|
|||
|
|
|||
|
// Use raw pointers to lookup keys.
|
|||
|
int* ptr = ptr_set.begin()->get();
|
|||
|
EXPECT_TRUE(ptr_set.find(ptr) == ptr_set.begin());
|
|||
|
```
|
|||
|
|
|||
|
Example flat_map<std\::string, int>:
|
|||
|
|
|||
|
```cpp
|
|||
|
base::flat_map<std::string, int> str_to_int({{"a", 1}, {"c", 2},{"b", 2}},
|
|||
|
base::KEEP_FIRST_OF_DUPES);
|
|||
|
|
|||
|
// Does not construct temporary strings.
|
|||
|
str_to_int.find("c")->second = 3;
|
|||
|
str_to_int.erase("c");
|
|||
|
EXPECT_EQ(str_to_int.end(), str_to_int.find("c")->second);
|
|||
|
|
|||
|
// NOTE: This does construct a temporary string. This happens since if the
|
|||
|
// item is not in the container, then it needs to be constructed, which is
|
|||
|
// something that transparent comparators don't have to guarantee.
|
|||
|
str_to_int["c"] = 3;
|
|||
|
```
|
|||
|
|
|||
|
### base::small\_map
|
|||
|
|
|||
|
A small inline buffer that is brute-force searched that overflows into a full
|
|||
|
std::map or std::unordered\_map. This gives the memory benefit of
|
|||
|
base::flat\_map for small data sizes without the degenerate insertion
|
|||
|
performance for large container sizes.
|
|||
|
|
|||
|
Since instantiations require both code for a std::map and a brute-force search
|
|||
|
of the inline container, plus a fancy iterator to cover both cases, code size
|
|||
|
is larger.
|
|||
|
|
|||
|
The initial size in the above table is assuming a very small inline table. The
|
|||
|
actual size will be sizeof(int) + min(sizeof(std::map), sizeof(T) *
|
|||
|
inline\_size).
|
|||
|
|
|||
|
# Deque
|
|||
|
|
|||
|
### Usage advice
|
|||
|
|
|||
|
Chromium code should always use `base::circular_deque` or `base::queue` in
|
|||
|
preference to `std::deque` or `std::queue` due to memory usage and platform
|
|||
|
variation.
|
|||
|
|
|||
|
The `base::circular_deque` implementation (and the `base::queue` which uses it)
|
|||
|
provide performance consistent across platforms that better matches most
|
|||
|
programmer's expectations on performance (it doesn't waste as much space as
|
|||
|
libc++ and doesn't do as many heap allocations as MSVC). It also generates less
|
|||
|
code tham `std::queue`: using it across the code base saves several hundred
|
|||
|
kilobytes.
|
|||
|
|
|||
|
Since `base::deque` does not have stable iterators and it will move the objects
|
|||
|
it contains, it may not be appropriate for all uses. If you need these,
|
|||
|
consider using a `std::list` which will provide constant time insert and erase.
|
|||
|
|
|||
|
### std::deque and std::queue
|
|||
|
|
|||
|
The implementation of `std::deque` varies considerably which makes it hard to
|
|||
|
reason about. All implementations use a sequence of data blocks referenced by
|
|||
|
an array of pointers. The standard guarantees random access, amortized
|
|||
|
constant operations at the ends, and linear mutations in the middle.
|
|||
|
|
|||
|
In Microsoft's implementation, each block is the smaller of 16 bytes or the
|
|||
|
size of the contained element. This means in practice that every expansion of
|
|||
|
the deque of non-trivial classes requires a heap allocation. libc++ (on Android
|
|||
|
and Mac) uses 4K blocks which elimiates the problem of many heap allocations,
|
|||
|
but generally wastes a large amount of space (an Android analysis revealed more
|
|||
|
than 2.5MB wasted space from deque alone, resulting in some optimizations).
|
|||
|
libstdc++ uses an intermediate-size 512 byte buffer.
|
|||
|
|
|||
|
Microsoft's implementation never shrinks the deque capacity, so the capacity
|
|||
|
will always be the maximum number of elements ever contained. libstdc++
|
|||
|
deallocates blocks as they are freed. libc++ keeps up to two empty blocks.
|
|||
|
|
|||
|
### base::circular_deque and base::queue
|
|||
|
|
|||
|
A deque implemented as a circular buffer in an array. The underlying array will
|
|||
|
grow like a `std::vector` while the beginning and end of the deque will move
|
|||
|
around. The items will wrap around the underlying buffer so the storage will
|
|||
|
not be contiguous, but fast random access iterators are still possible.
|
|||
|
|
|||
|
When the underlying buffer is filled, it will be reallocated and the constents
|
|||
|
moved (like a `std::vector`). The underlying buffer will be shrunk if there is
|
|||
|
too much wasted space (_unlike_ a `std::vector`). As a result, iterators are
|
|||
|
not stable across mutations.
|
|||
|
|
|||
|
# Stack
|
|||
|
|
|||
|
`std::stack` is like `std::queue` in that it is a wrapper around an underlying
|
|||
|
container. The default container is `std::deque` so everything from the deque
|
|||
|
section applies.
|
|||
|
|
|||
|
Chromium provides `base/containers/stack.h` which defines `base::stack` that
|
|||
|
should be used in preference to std::stack. This changes the underlying
|
|||
|
container to `base::circular_deque`. The result will be very similar to
|
|||
|
manually specifying a `std::vector` for the underlying implementation except
|
|||
|
that the storage will shrink when it gets too empty (vector will never
|
|||
|
reallocate to a smaller size).
|
|||
|
|
|||
|
Watch out: with some stack usage patterns it's easy to depend on unstable
|
|||
|
behavior:
|
|||
|
|
|||
|
```cpp
|
|||
|
base::stack<Foo> stack;
|
|||
|
for (...) {
|
|||
|
Foo& current = stack.top();
|
|||
|
DoStuff(); // May call stack.push(), say if writing a parser.
|
|||
|
current.done = true; // Current may reference deleted item!
|
|||
|
}
|
|||
|
```
|
|||
|
|
|||
|
## Appendix
|
|||
|
|
|||
|
### Code for map code size comparison
|
|||
|
|
|||
|
This just calls insert and query a number of times, with printfs that prevent
|
|||
|
things from being dead-code eliminated.
|
|||
|
|
|||
|
```cpp
|
|||
|
TEST(Foo, Bar) {
|
|||
|
base::small_map<std::map<std::string, Flubber>> foo;
|
|||
|
foo.insert(std::make_pair("foo", Flubber(8, "bar")));
|
|||
|
foo.insert(std::make_pair("bar", Flubber(8, "bar")));
|
|||
|
foo.insert(std::make_pair("foo1", Flubber(8, "bar")));
|
|||
|
foo.insert(std::make_pair("bar1", Flubber(8, "bar")));
|
|||
|
foo.insert(std::make_pair("foo", Flubber(8, "bar")));
|
|||
|
foo.insert(std::make_pair("bar", Flubber(8, "bar")));
|
|||
|
auto found = foo.find("asdf");
|
|||
|
printf("Found is %d\n", (int)(found == foo.end()));
|
|||
|
found = foo.find("foo");
|
|||
|
printf("Found is %d\n", (int)(found == foo.end()));
|
|||
|
found = foo.find("bar");
|
|||
|
printf("Found is %d\n", (int)(found == foo.end()));
|
|||
|
found = foo.find("asdfhf");
|
|||
|
printf("Found is %d\n", (int)(found == foo.end()));
|
|||
|
found = foo.find("bar1");
|
|||
|
printf("Found is %d\n", (int)(found == foo.end()));
|
|||
|
}
|
|||
|
```
|
|||
|
|