Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(65)

Side by Side Diff: third_party/protobuf/src/google/protobuf/arena.cc

Issue 1842653006: Update //third_party/protobuf to version 3. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: merge Created 4 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
(Empty)
1 // Protocol Buffers - Google's data interchange format
2 // Copyright 2008 Google Inc. All rights reserved.
3 // https://developers.google.com/protocol-buffers/
4 //
5 // Redistribution and use in source and binary forms, with or without
6 // modification, are permitted provided that the following conditions are
7 // met:
8 //
9 // * Redistributions of source code must retain the above copyright
10 // notice, this list of conditions and the following disclaimer.
11 // * Redistributions in binary form must reproduce the above
12 // copyright notice, this list of conditions and the following disclaimer
13 // in the documentation and/or other materials provided with the
14 // distribution.
15 // * Neither the name of Google Inc. nor the names of its
16 // contributors may be used to endorse or promote products derived from
17 // this software without specific prior written permission.
18 //
19 // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
20 // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
21 // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
22 // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
23 // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
24 // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
25 // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
26 // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
27 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
28 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
29 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30
31 #include <google/protobuf/arena.h>
32
33 #ifdef ADDRESS_SANITIZER
34 #include <sanitizer/asan_interface.h>
35 #endif
36
37 namespace google {
38 namespace protobuf {
39
40 google::protobuf::internal::SequenceNumber Arena::lifecycle_id_generator_;
41 #if defined(GOOGLE_PROTOBUF_NO_THREADLOCAL)
42 Arena::ThreadCache& Arena::thread_cache() {
43 static internal::ThreadLocalStorage<ThreadCache>* thread_cache_ =
44 new internal::ThreadLocalStorage<ThreadCache>();
45 return *thread_cache_->Get();
46 }
47 #elif defined(PROTOBUF_USE_DLLS)
48 Arena::ThreadCache& Arena::thread_cache() {
49 static GOOGLE_THREAD_LOCAL ThreadCache thread_cache_ = { -1, NULL };
50 return thread_cache_;
51 }
52 #else
53 GOOGLE_THREAD_LOCAL Arena::ThreadCache Arena::thread_cache_ = { -1, NULL };
54 #endif
55
56 void Arena::Init() {
57 lifecycle_id_ = lifecycle_id_generator_.GetNext();
58 blocks_ = 0;
59 hint_ = 0;
60 owns_first_block_ = true;
61 cleanup_list_ = 0;
62
63 if (options_.initial_block != NULL && options_.initial_block_size > 0) {
64 GOOGLE_CHECK_GE(options_.initial_block_size, sizeof(Block))
65 << ": Initial block size too small for header.";
66
67 // Add first unowned block to list.
68 Block* first_block = reinterpret_cast<Block*>(options_.initial_block);
69 first_block->size = options_.initial_block_size;
70 first_block->pos = kHeaderSize;
71 first_block->next = NULL;
72 // Thread which calls Init() owns the first block. This allows the
73 // single-threaded case to allocate on the first block without taking any
74 // locks.
75 first_block->owner = &thread_cache();
76 SetThreadCacheBlock(first_block);
77 AddBlockInternal(first_block);
78 owns_first_block_ = false;
79 }
80
81 // Call the initialization hook
82 if (options_.on_arena_init != NULL) {
83 hooks_cookie_ = options_.on_arena_init(this);
84 } else {
85 hooks_cookie_ = NULL;
86 }
87 }
88
89 Arena::~Arena() {
90 uint64 space_allocated = ResetInternal();
91
92 // Call the destruction hook
93 if (options_.on_arena_destruction != NULL) {
94 options_.on_arena_destruction(this, hooks_cookie_, space_allocated);
95 }
96 }
97
98 uint64 Arena::Reset() {
99 // Invalidate any ThreadCaches pointing to any blocks we just destroyed.
100 lifecycle_id_ = lifecycle_id_generator_.GetNext();
101 return ResetInternal();
102 }
103
104 uint64 Arena::ResetInternal() {
105 CleanupList();
106 uint64 space_allocated = FreeBlocks();
107
108 // Call the reset hook
109 if (options_.on_arena_reset != NULL) {
110 options_.on_arena_reset(this, hooks_cookie_, space_allocated);
111 }
112
113 return space_allocated;
114 }
115
116 Arena::Block* Arena::NewBlock(void* me, Block* my_last_block, size_t n,
117 size_t start_block_size, size_t max_block_size) {
118 size_t size;
119 if (my_last_block != NULL) {
120 // Double the current block size, up to a limit.
121 size = 2 * (my_last_block->size);
122 if (size > max_block_size) size = max_block_size;
123 } else {
124 size = start_block_size;
125 }
126 if (n > size - kHeaderSize) {
127 // TODO(sanjay): Check if n + kHeaderSize would overflow
128 size = kHeaderSize + n;
129 }
130
131 Block* b = reinterpret_cast<Block*>(options_.block_alloc(size));
132 b->pos = kHeaderSize + n;
133 b->size = size;
134 if (b->avail() == 0) {
135 // Do not attempt to reuse this block.
136 b->owner = NULL;
137 } else {
138 b->owner = me;
139 }
140 #ifdef ADDRESS_SANITIZER
141 // Poison the rest of the block for ASAN. It was unpoisoned by the underlying
142 // malloc but it's not yet usable until we return it as part of an allocation.
143 ASAN_POISON_MEMORY_REGION(
144 reinterpret_cast<char*>(b) + b->pos, b->size - b->pos);
145 #endif
146 return b;
147 }
148
149 void Arena::AddBlock(Block* b) {
150 MutexLock l(&blocks_lock_);
151 AddBlockInternal(b);
152 }
153
154 void Arena::AddBlockInternal(Block* b) {
155 b->next = reinterpret_cast<Block*>(google::protobuf::internal::NoBarrier_Load( &blocks_));
156 google::protobuf::internal::Release_Store(&blocks_, reinterpret_cast<google::p rotobuf::internal::AtomicWord>(b));
157 if (b->avail() != 0) {
158 // Direct future allocations to this block.
159 google::protobuf::internal::Release_Store(&hint_, reinterpret_cast<google::p rotobuf::internal::AtomicWord>(b));
160 }
161 }
162
163 void Arena::AddListNode(void* elem, void (*cleanup)(void*)) {
164 Node* node = reinterpret_cast<Node*>(AllocateAligned(sizeof(Node)));
165 node->elem = elem;
166 node->cleanup = cleanup;
167 node->next = reinterpret_cast<Node*>(
168 google::protobuf::internal::NoBarrier_AtomicExchange(&cleanup_list_,
169 reinterpret_cast<google::protobuf::internal::AtomicWord>(node)));
170 }
171
172 void* Arena::AllocateAligned(const std::type_info* allocated, size_t n) {
173 // Align n to next multiple of 8 (from Hacker's Delight, Chapter 3.)
174 n = (n + 7) & -8;
175
176 // Monitor allocation if needed.
177 if (GOOGLE_PREDICT_FALSE(hooks_cookie_ != NULL) &&
178 options_.on_arena_allocation != NULL) {
179 options_.on_arena_allocation(allocated, n, hooks_cookie_);
180 }
181
182 // If this thread already owns a block in this arena then try to use that.
183 // This fast path optimizes the case where multiple threads allocate from the
184 // same arena.
185 if (thread_cache().last_lifecycle_id_seen == lifecycle_id_ &&
186 thread_cache().last_block_used_ != NULL) {
187 if (thread_cache().last_block_used_->avail() < n) {
188 return SlowAlloc(n);
189 }
190 return AllocFromBlock(thread_cache().last_block_used_, n);
191 }
192
193 // Check whether we own the last accessed block on this arena.
194 // This fast path optimizes the case where a single thread uses multiple
195 // arenas.
196 void* me = &thread_cache();
197 Block* b = reinterpret_cast<Block*>(google::protobuf::internal::Acquire_Load(& hint_));
198 if (!b || b->owner != me || b->avail() < n) {
199 return SlowAlloc(n);
200 }
201 return AllocFromBlock(b, n);
202 }
203
204 void* Arena::AllocFromBlock(Block* b, size_t n) {
205 size_t p = b->pos;
206 b->pos = p + n;
207 #ifdef ADDRESS_SANITIZER
208 ASAN_UNPOISON_MEMORY_REGION(reinterpret_cast<char*>(b) + p, n);
209 #endif
210 return reinterpret_cast<char*>(b) + p;
211 }
212
213 void* Arena::SlowAlloc(size_t n) {
214 void* me = &thread_cache();
215 Block* b = FindBlock(me); // Find block owned by me.
216 // See if allocation fits in my latest block.
217 if (b != NULL && b->avail() >= n) {
218 SetThreadCacheBlock(b);
219 google::protobuf::internal::NoBarrier_Store(&hint_, reinterpret_cast<google: :protobuf::internal::AtomicWord>(b));
220 return AllocFromBlock(b, n);
221 }
222 b = NewBlock(me, b, n, options_.start_block_size, options_.max_block_size);
223 AddBlock(b);
224 if (b->owner == me) { // If this block can be reused (see NewBlock()).
225 SetThreadCacheBlock(b);
226 }
227 return reinterpret_cast<char*>(b) + kHeaderSize;
228 }
229
230 uint64 Arena::SpaceAllocated() const {
231 uint64 space_allocated = 0;
232 Block* b = reinterpret_cast<Block*>(google::protobuf::internal::NoBarrier_Load (&blocks_));
233 while (b != NULL) {
234 space_allocated += (b->size);
235 b = b->next;
236 }
237 return space_allocated;
238 }
239
240 uint64 Arena::SpaceUsed() const {
241 uint64 space_used = 0;
242 Block* b = reinterpret_cast<Block*>(google::protobuf::internal::NoBarrier_Load (&blocks_));
243 while (b != NULL) {
244 space_used += (b->pos - kHeaderSize);
245 b = b->next;
246 }
247 return space_used;
248 }
249
250 uint64 Arena::FreeBlocks() {
251 uint64 space_allocated = 0;
252 Block* b = reinterpret_cast<Block*>(google::protobuf::internal::NoBarrier_Load (&blocks_));
253 Block* first_block = NULL;
254 while (b != NULL) {
255 space_allocated += (b->size);
256 Block* next = b->next;
257 if (next != NULL) {
258 options_.block_dealloc(b, b->size);
259 } else {
260 if (owns_first_block_) {
261 options_.block_dealloc(b, b->size);
262 } else {
263 // User passed in the first block, skip free'ing the memory.
264 first_block = b;
265 }
266 }
267 b = next;
268 }
269 blocks_ = 0;
270 hint_ = 0;
271 if (!owns_first_block_) {
272 // Make the first block that was passed in through ArenaOptions
273 // available for reuse.
274 first_block->pos = kHeaderSize;
275 // Thread which calls Reset() owns the first block. This allows the
276 // single-threaded case to allocate on the first block without taking any
277 // locks.
278 first_block->owner = &thread_cache();
279 SetThreadCacheBlock(first_block);
280 AddBlockInternal(first_block);
281 }
282 return space_allocated;
283 }
284
285 void Arena::CleanupList() {
286 Node* head =
287 reinterpret_cast<Node*>(google::protobuf::internal::NoBarrier_Load(&cleanu p_list_));
288 while (head != NULL) {
289 head->cleanup(head->elem);
290 head = head->next;
291 }
292 cleanup_list_ = 0;
293 }
294
295 Arena::Block* Arena::FindBlock(void* me) {
296 // TODO(sanjay): We might want to keep a separate list with one
297 // entry per thread.
298 Block* b = reinterpret_cast<Block*>(google::protobuf::internal::Acquire_Load(& blocks_));
299 while (b != NULL && b->owner != me) {
300 b = b->next;
301 }
302 return b;
303 }
304
305 } // namespace protobuf
306 } // namespace google
OLDNEW
« no previous file with comments | « third_party/protobuf/src/google/protobuf/arena.h ('k') | third_party/protobuf/src/google/protobuf/arena_nc.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698