xref: /aosp_15_r20/external/abseil-cpp/absl/container/internal/raw_hash_set.cc (revision 9356374a3709195abf420251b3e825997ff56c0f)
1 // Copyright 2018 The Abseil Authors.
2 //
3 // Licensed under the Apache License, Version 2.0 (the "License");
4 // you may not use this file except in compliance with the License.
5 // You may obtain a copy of the License at
6 //
7 //      https://www.apache.org/licenses/LICENSE-2.0
8 //
9 // Unless required by applicable law or agreed to in writing, software
10 // distributed under the License is distributed on an "AS IS" BASIS,
11 // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 // See the License for the specific language governing permissions and
13 // limitations under the License.
14 
15 #include "absl/container/internal/raw_hash_set.h"
16 
17 #include <atomic>
18 #include <cassert>
19 #include <cstddef>
20 #include <cstdint>
21 #include <cstring>
22 
23 #include "absl/base/attributes.h"
24 #include "absl/base/config.h"
25 #include "absl/base/dynamic_annotations.h"
26 #include "absl/base/internal/endian.h"
27 #include "absl/base/optimization.h"
28 #include "absl/container/internal/container_memory.h"
29 #include "absl/container/internal/hashtablez_sampler.h"
30 #include "absl/hash/hash.h"
31 
32 namespace absl {
33 ABSL_NAMESPACE_BEGIN
34 namespace container_internal {
35 
36 // Represents a control byte corresponding to a full slot with arbitrary hash.
ZeroCtrlT()37 constexpr ctrl_t ZeroCtrlT() { return static_cast<ctrl_t>(0); }
38 
39 // We have space for `growth_info` before a single block of control bytes. A
40 // single block of empty control bytes for tables without any slots allocated.
41 // This enables removing a branch in the hot path of find(). In order to ensure
42 // that the control bytes are aligned to 16, we have 16 bytes before the control
43 // bytes even though growth_info only needs 8.
44 alignas(16) ABSL_CONST_INIT ABSL_DLL const ctrl_t kEmptyGroup[32] = {
45     ZeroCtrlT(),       ZeroCtrlT(),    ZeroCtrlT(),    ZeroCtrlT(),
46     ZeroCtrlT(),       ZeroCtrlT(),    ZeroCtrlT(),    ZeroCtrlT(),
47     ZeroCtrlT(),       ZeroCtrlT(),    ZeroCtrlT(),    ZeroCtrlT(),
48     ZeroCtrlT(),       ZeroCtrlT(),    ZeroCtrlT(),    ZeroCtrlT(),
49     ctrl_t::kSentinel, ctrl_t::kEmpty, ctrl_t::kEmpty, ctrl_t::kEmpty,
50     ctrl_t::kEmpty,    ctrl_t::kEmpty, ctrl_t::kEmpty, ctrl_t::kEmpty,
51     ctrl_t::kEmpty,    ctrl_t::kEmpty, ctrl_t::kEmpty, ctrl_t::kEmpty,
52     ctrl_t::kEmpty,    ctrl_t::kEmpty, ctrl_t::kEmpty, ctrl_t::kEmpty};
53 
54 // We need one full byte followed by a sentinel byte for iterator::operator++ to
55 // work. We have a full group after kSentinel to be safe (in case operator++ is
56 // changed to read a full group).
57 ABSL_CONST_INIT ABSL_DLL const ctrl_t kSooControl[17] = {
58     ZeroCtrlT(),    ctrl_t::kSentinel, ZeroCtrlT(),    ctrl_t::kEmpty,
59     ctrl_t::kEmpty, ctrl_t::kEmpty,    ctrl_t::kEmpty, ctrl_t::kEmpty,
60     ctrl_t::kEmpty, ctrl_t::kEmpty,    ctrl_t::kEmpty, ctrl_t::kEmpty,
61     ctrl_t::kEmpty, ctrl_t::kEmpty,    ctrl_t::kEmpty, ctrl_t::kEmpty,
62     ctrl_t::kEmpty};
63 static_assert(NumControlBytes(SooCapacity()) <= 17,
64               "kSooControl capacity too small");
65 
66 #ifdef ABSL_INTERNAL_NEED_REDUNDANT_CONSTEXPR_DECL
67 constexpr size_t Group::kWidth;
68 #endif
69 
70 namespace {
71 
72 // Returns "random" seed.
RandomSeed()73 inline size_t RandomSeed() {
74 #ifdef ABSL_HAVE_THREAD_LOCAL
75   static thread_local size_t counter = 0;
76   // On Linux kernels >= 5.4 the MSAN runtime has a false-positive when
77   // accessing thread local storage data from loaded libraries
78   // (https://github.com/google/sanitizers/issues/1265), for this reason counter
79   // needs to be annotated as initialized.
80   ABSL_ANNOTATE_MEMORY_IS_INITIALIZED(&counter, sizeof(size_t));
81   size_t value = ++counter;
82 #else   // ABSL_HAVE_THREAD_LOCAL
83   static std::atomic<size_t> counter(0);
84   size_t value = counter.fetch_add(1, std::memory_order_relaxed);
85 #endif  // ABSL_HAVE_THREAD_LOCAL
86   return value ^ static_cast<size_t>(reinterpret_cast<uintptr_t>(&counter));
87 }
88 
ShouldRehashForBugDetection(const ctrl_t * ctrl,size_t capacity)89 bool ShouldRehashForBugDetection(const ctrl_t* ctrl, size_t capacity) {
90   // Note: we can't use the abseil-random library because abseil-random
91   // depends on swisstable. We want to return true with probability
92   // `min(1, RehashProbabilityConstant() / capacity())`. In order to do this,
93   // we probe based on a random hash and see if the offset is less than
94   // RehashProbabilityConstant().
95   return probe(ctrl, capacity, absl::HashOf(RandomSeed())).offset() <
96          RehashProbabilityConstant();
97 }
98 
99 }  // namespace
100 
EmptyGeneration()101 GenerationType* EmptyGeneration() {
102   if (SwisstableGenerationsEnabled()) {
103     constexpr size_t kNumEmptyGenerations = 1024;
104     static constexpr GenerationType kEmptyGenerations[kNumEmptyGenerations]{};
105     return const_cast<GenerationType*>(
106         &kEmptyGenerations[RandomSeed() % kNumEmptyGenerations]);
107   }
108   return nullptr;
109 }
110 
111 bool CommonFieldsGenerationInfoEnabled::
should_rehash_for_bug_detection_on_insert(const ctrl_t * ctrl,size_t capacity) const112     should_rehash_for_bug_detection_on_insert(const ctrl_t* ctrl,
113                                               size_t capacity) const {
114   if (reserved_growth_ == kReservedGrowthJustRanOut) return true;
115   if (reserved_growth_ > 0) return false;
116   return ShouldRehashForBugDetection(ctrl, capacity);
117 }
118 
should_rehash_for_bug_detection_on_move(const ctrl_t * ctrl,size_t capacity) const119 bool CommonFieldsGenerationInfoEnabled::should_rehash_for_bug_detection_on_move(
120     const ctrl_t* ctrl, size_t capacity) const {
121   return ShouldRehashForBugDetection(ctrl, capacity);
122 }
123 
ShouldInsertBackwardsForDebug(size_t capacity,size_t hash,const ctrl_t * ctrl)124 bool ShouldInsertBackwardsForDebug(size_t capacity, size_t hash,
125                                    const ctrl_t* ctrl) {
126   // To avoid problems with weak hashes and single bit tests, we use % 13.
127   // TODO(kfm,sbenza): revisit after we do unconditional mixing
128   return !is_small(capacity) && (H1(hash, ctrl) ^ RandomSeed()) % 13 > 6;
129 }
130 
PrepareInsertAfterSoo(size_t hash,size_t slot_size,CommonFields & common)131 size_t PrepareInsertAfterSoo(size_t hash, size_t slot_size,
132                              CommonFields& common) {
133   assert(common.capacity() == NextCapacity(SooCapacity()));
134   // After resize from capacity 1 to 3, we always have exactly the slot with
135   // index 1 occupied, so we need to insert either at index 0 or index 2.
136   assert(HashSetResizeHelper::SooSlotIndex() == 1);
137   PrepareInsertCommon(common);
138   const size_t offset = H1(hash, common.control()) & 2;
139   common.growth_info().OverwriteEmptyAsFull();
140   SetCtrlInSingleGroupTable(common, offset, H2(hash), slot_size);
141   common.infoz().RecordInsert(hash, /*distance_from_desired=*/0);
142   return offset;
143 }
144 
ConvertDeletedToEmptyAndFullToDeleted(ctrl_t * ctrl,size_t capacity)145 void ConvertDeletedToEmptyAndFullToDeleted(ctrl_t* ctrl, size_t capacity) {
146   assert(ctrl[capacity] == ctrl_t::kSentinel);
147   assert(IsValidCapacity(capacity));
148   for (ctrl_t* pos = ctrl; pos < ctrl + capacity; pos += Group::kWidth) {
149     Group{pos}.ConvertSpecialToEmptyAndFullToDeleted(pos);
150   }
151   // Copy the cloned ctrl bytes.
152   std::memcpy(ctrl + capacity + 1, ctrl, NumClonedBytes());
153   ctrl[capacity] = ctrl_t::kSentinel;
154 }
155 // Extern template instantiation for inline function.
156 template FindInfo find_first_non_full(const CommonFields&, size_t);
157 
find_first_non_full_outofline(const CommonFields & common,size_t hash)158 FindInfo find_first_non_full_outofline(const CommonFields& common,
159                                        size_t hash) {
160   return find_first_non_full(common, hash);
161 }
162 
163 namespace {
164 
165 // Returns the address of the slot just after slot assuming each slot has the
166 // specified size.
NextSlot(void * slot,size_t slot_size)167 static inline void* NextSlot(void* slot, size_t slot_size) {
168   return reinterpret_cast<void*>(reinterpret_cast<uintptr_t>(slot) + slot_size);
169 }
170 
171 // Returns the address of the slot just before slot assuming each slot has the
172 // specified size.
PrevSlot(void * slot,size_t slot_size)173 static inline void* PrevSlot(void* slot, size_t slot_size) {
174   return reinterpret_cast<void*>(reinterpret_cast<uintptr_t>(slot) - slot_size);
175 }
176 
177 // Finds guaranteed to exists empty slot from the given position.
178 // NOTE: this function is almost never triggered inside of the
179 // DropDeletesWithoutResize, so we keep it simple.
180 // The table is rather sparse, so empty slot will be found very quickly.
FindEmptySlot(size_t start,size_t end,const ctrl_t * ctrl)181 size_t FindEmptySlot(size_t start, size_t end, const ctrl_t* ctrl) {
182   for (size_t i = start; i < end; ++i) {
183     if (IsEmpty(ctrl[i])) {
184       return i;
185     }
186   }
187   assert(false && "no empty slot");
188   return ~size_t{};
189 }
190 
DropDeletesWithoutResize(CommonFields & common,const PolicyFunctions & policy)191 void DropDeletesWithoutResize(CommonFields& common,
192                               const PolicyFunctions& policy) {
193   void* set = &common;
194   void* slot_array = common.slot_array();
195   const size_t capacity = common.capacity();
196   assert(IsValidCapacity(capacity));
197   assert(!is_small(capacity));
198   // Algorithm:
199   // - mark all DELETED slots as EMPTY
200   // - mark all FULL slots as DELETED
201   // - for each slot marked as DELETED
202   //     hash = Hash(element)
203   //     target = find_first_non_full(hash)
204   //     if target is in the same group
205   //       mark slot as FULL
206   //     else if target is EMPTY
207   //       transfer element to target
208   //       mark slot as EMPTY
209   //       mark target as FULL
210   //     else if target is DELETED
211   //       swap current element with target element
212   //       mark target as FULL
213   //       repeat procedure for current slot with moved from element (target)
214   ctrl_t* ctrl = common.control();
215   ConvertDeletedToEmptyAndFullToDeleted(ctrl, capacity);
216   const void* hash_fn = policy.hash_fn(common);
217   auto hasher = policy.hash_slot;
218   auto transfer = policy.transfer;
219   const size_t slot_size = policy.slot_size;
220 
221   size_t total_probe_length = 0;
222   void* slot_ptr = SlotAddress(slot_array, 0, slot_size);
223 
224   // The index of an empty slot that can be used as temporary memory for
225   // the swap operation.
226   constexpr size_t kUnknownId = ~size_t{};
227   size_t tmp_space_id = kUnknownId;
228 
229   for (size_t i = 0; i != capacity;
230        ++i, slot_ptr = NextSlot(slot_ptr, slot_size)) {
231     assert(slot_ptr == SlotAddress(slot_array, i, slot_size));
232     if (IsEmpty(ctrl[i])) {
233       tmp_space_id = i;
234       continue;
235     }
236     if (!IsDeleted(ctrl[i])) continue;
237     const size_t hash = (*hasher)(hash_fn, slot_ptr);
238     const FindInfo target = find_first_non_full(common, hash);
239     const size_t new_i = target.offset;
240     total_probe_length += target.probe_length;
241 
242     // Verify if the old and new i fall within the same group wrt the hash.
243     // If they do, we don't need to move the object as it falls already in the
244     // best probe we can.
245     const size_t probe_offset = probe(common, hash).offset();
246     const auto probe_index = [probe_offset, capacity](size_t pos) {
247       return ((pos - probe_offset) & capacity) / Group::kWidth;
248     };
249 
250     // Element doesn't move.
251     if (ABSL_PREDICT_TRUE(probe_index(new_i) == probe_index(i))) {
252       SetCtrl(common, i, H2(hash), slot_size);
253       continue;
254     }
255 
256     void* new_slot_ptr = SlotAddress(slot_array, new_i, slot_size);
257     if (IsEmpty(ctrl[new_i])) {
258       // Transfer element to the empty spot.
259       // SetCtrl poisons/unpoisons the slots so we have to call it at the
260       // right time.
261       SetCtrl(common, new_i, H2(hash), slot_size);
262       (*transfer)(set, new_slot_ptr, slot_ptr);
263       SetCtrl(common, i, ctrl_t::kEmpty, slot_size);
264       // Initialize or change empty space id.
265       tmp_space_id = i;
266     } else {
267       assert(IsDeleted(ctrl[new_i]));
268       SetCtrl(common, new_i, H2(hash), slot_size);
269       // Until we are done rehashing, DELETED marks previously FULL slots.
270 
271       if (tmp_space_id == kUnknownId) {
272         tmp_space_id = FindEmptySlot(i + 1, capacity, ctrl);
273       }
274       void* tmp_space = SlotAddress(slot_array, tmp_space_id, slot_size);
275       SanitizerUnpoisonMemoryRegion(tmp_space, slot_size);
276 
277       // Swap i and new_i elements.
278       (*transfer)(set, tmp_space, new_slot_ptr);
279       (*transfer)(set, new_slot_ptr, slot_ptr);
280       (*transfer)(set, slot_ptr, tmp_space);
281 
282       SanitizerPoisonMemoryRegion(tmp_space, slot_size);
283 
284       // repeat the processing of the ith slot
285       --i;
286       slot_ptr = PrevSlot(slot_ptr, slot_size);
287     }
288   }
289   ResetGrowthLeft(common);
290   common.infoz().RecordRehash(total_probe_length);
291 }
292 
WasNeverFull(CommonFields & c,size_t index)293 static bool WasNeverFull(CommonFields& c, size_t index) {
294   if (is_single_group(c.capacity())) {
295     return true;
296   }
297   const size_t index_before = (index - Group::kWidth) & c.capacity();
298   const auto empty_after = Group(c.control() + index).MaskEmpty();
299   const auto empty_before = Group(c.control() + index_before).MaskEmpty();
300 
301   // We count how many consecutive non empties we have to the right and to the
302   // left of `it`. If the sum is >= kWidth then there is at least one probe
303   // window that might have seen a full group.
304   return empty_before && empty_after &&
305          static_cast<size_t>(empty_after.TrailingZeros()) +
306                  empty_before.LeadingZeros() <
307              Group::kWidth;
308 }
309 
310 }  // namespace
311 
EraseMetaOnly(CommonFields & c,size_t index,size_t slot_size)312 void EraseMetaOnly(CommonFields& c, size_t index, size_t slot_size) {
313   assert(IsFull(c.control()[index]) && "erasing a dangling iterator");
314   c.decrement_size();
315   c.infoz().RecordErase();
316 
317   if (WasNeverFull(c, index)) {
318     SetCtrl(c, index, ctrl_t::kEmpty, slot_size);
319     c.growth_info().OverwriteFullAsEmpty();
320     return;
321   }
322 
323   c.growth_info().OverwriteFullAsDeleted();
324   SetCtrl(c, index, ctrl_t::kDeleted, slot_size);
325 }
326 
ClearBackingArray(CommonFields & c,const PolicyFunctions & policy,bool reuse,bool soo_enabled)327 void ClearBackingArray(CommonFields& c, const PolicyFunctions& policy,
328                        bool reuse, bool soo_enabled) {
329   c.set_size(0);
330   if (reuse) {
331     assert(!soo_enabled || c.capacity() > SooCapacity());
332     ResetCtrl(c, policy.slot_size);
333     ResetGrowthLeft(c);
334     c.infoz().RecordStorageChanged(0, c.capacity());
335   } else {
336     // We need to record infoz before calling dealloc, which will unregister
337     // infoz.
338     c.infoz().RecordClearedReservation();
339     c.infoz().RecordStorageChanged(0, soo_enabled ? SooCapacity() : 0);
340     (*policy.dealloc)(c, policy);
341     c = soo_enabled ? CommonFields{soo_tag_t{}} : CommonFields{};
342   }
343 }
344 
GrowIntoSingleGroupShuffleControlBytes(ctrl_t * __restrict new_ctrl,size_t new_capacity) const345 void HashSetResizeHelper::GrowIntoSingleGroupShuffleControlBytes(
346     ctrl_t* __restrict new_ctrl, size_t new_capacity) const {
347   assert(is_single_group(new_capacity));
348   constexpr size_t kHalfWidth = Group::kWidth / 2;
349   constexpr size_t kQuarterWidth = Group::kWidth / 4;
350   assert(old_capacity_ < kHalfWidth);
351   static_assert(sizeof(uint64_t) >= kHalfWidth,
352                 "Group size is too large. The ctrl bytes for half a group must "
353                 "fit into a uint64_t for this implementation.");
354   static_assert(sizeof(uint64_t) <= Group::kWidth,
355                 "Group size is too small. The ctrl bytes for a group must "
356                 "cover a uint64_t for this implementation.");
357 
358   const size_t half_old_capacity = old_capacity_ / 2;
359 
360   // NOTE: operations are done with compile time known size = kHalfWidth.
361   // Compiler optimizes that into single ASM operation.
362 
363   // Load the bytes from half_old_capacity + 1. This contains the last half of
364   // old_ctrl bytes, followed by the sentinel byte, and then the first half of
365   // the cloned bytes. This effectively shuffles the control bytes.
366   uint64_t copied_bytes = 0;
367   copied_bytes =
368       absl::little_endian::Load64(old_ctrl() + half_old_capacity + 1);
369 
370   // We change the sentinel byte to kEmpty before storing to both the start of
371   // the new_ctrl, and past the end of the new_ctrl later for the new cloned
372   // bytes. Note that this is faster than setting the sentinel byte to kEmpty
373   // after the copy directly in new_ctrl because we are limited on store
374   // bandwidth.
375   constexpr uint64_t kEmptyXorSentinel =
376       static_cast<uint8_t>(ctrl_t::kEmpty) ^
377       static_cast<uint8_t>(ctrl_t::kSentinel);
378   const uint64_t mask_convert_old_sentinel_to_empty =
379       kEmptyXorSentinel << (half_old_capacity * 8);
380   copied_bytes ^= mask_convert_old_sentinel_to_empty;
381 
382   // Copy second half of bytes to the beginning. This correctly sets the bytes
383   // [0, old_capacity]. We potentially copy more bytes in order to have compile
384   // time known size. Mirrored bytes from the old_ctrl() will also be copied. In
385   // case of old_capacity_ == 3, we will copy 1st element twice.
386   // Examples:
387   // (old capacity = 1)
388   // old_ctrl = 0S0EEEEEEE...
389   // new_ctrl = E0EEEEEE??...
390   //
391   // (old capacity = 3)
392   // old_ctrl = 012S012EEEEE...
393   // new_ctrl = 12E012EE????...
394   //
395   // (old capacity = 7)
396   // old_ctrl = 0123456S0123456EE...
397   // new_ctrl = 456E0123?????????...
398   absl::little_endian::Store64(new_ctrl, copied_bytes);
399 
400   // Set the space [old_capacity + 1, new_capacity] to empty as these bytes will
401   // not be written again. This is safe because
402   // NumControlBytes = new_capacity + kWidth and new_capacity >=
403   // old_capacity+1.
404   // Examples:
405   // (old_capacity = 3, new_capacity = 15)
406   // new_ctrl  = 12E012EE?????????????...??
407   // *new_ctrl = 12E0EEEEEEEEEEEEEEEE?...??
408   // position        /          S
409   //
410   // (old_capacity = 7, new_capacity = 15)
411   // new_ctrl  = 456E0123?????????????????...??
412   // *new_ctrl = 456E0123EEEEEEEEEEEEEEEE?...??
413   // position            /      S
414   std::memset(new_ctrl + old_capacity_ + 1, static_cast<int8_t>(ctrl_t::kEmpty),
415               Group::kWidth);
416 
417   // Set the last kHalfWidth bytes to empty, to ensure the bytes all the way to
418   // the end are initialized.
419   // Examples:
420   // new_ctrl  = 12E0EEEEEEEEEEEEEEEE?...???????
421   // *new_ctrl = 12E0EEEEEEEEEEEEEEEE???EEEEEEEE
422   // position                   S       /
423   //
424   // new_ctrl  = 456E0123EEEEEEEEEEEEEEEE???????
425   // *new_ctrl = 456E0123EEEEEEEEEEEEEEEEEEEEEEE
426   // position                   S       /
427   std::memset(new_ctrl + NumControlBytes(new_capacity) - kHalfWidth,
428               static_cast<int8_t>(ctrl_t::kEmpty), kHalfWidth);
429 
430   // Copy the first bytes to the end (starting at new_capacity +1) to set the
431   // cloned bytes. Note that we use the already copied bytes from old_ctrl here
432   // rather than copying from new_ctrl to avoid a Read-after-Write hazard, since
433   // new_ctrl was just written to. The first old_capacity-1 bytes are set
434   // correctly. Then there may be up to old_capacity bytes that need to be
435   // overwritten, and any remaining bytes will be correctly set to empty. This
436   // sets [new_capacity + 1, new_capacity +1 + old_capacity] correctly.
437   // Examples:
438   // new_ctrl  = 12E0EEEEEEEEEEEEEEEE?...???????
439   // *new_ctrl = 12E0EEEEEEEEEEEE12E012EEEEEEEEE
440   // position                   S/
441   //
442   // new_ctrl  = 456E0123EEEEEEEE?...???EEEEEEEE
443   // *new_ctrl = 456E0123EEEEEEEE456E0123EEEEEEE
444   // position                   S/
445   absl::little_endian::Store64(new_ctrl + new_capacity + 1, copied_bytes);
446 
447   // Set The remaining bytes at the end past the cloned bytes to empty. The
448   // incorrectly set bytes are [new_capacity + old_capacity + 2,
449   // min(new_capacity + 1 + kHalfWidth, new_capacity + old_capacity + 2 +
450   // half_old_capacity)]. Taking the difference, we need to set min(kHalfWidth -
451   // (old_capacity + 1), half_old_capacity)]. Since old_capacity < kHalfWidth,
452   // half_old_capacity < kQuarterWidth, so we set kQuarterWidth beginning at
453   // new_capacity + old_capacity + 2 to kEmpty.
454   // Examples:
455   // new_ctrl  = 12E0EEEEEEEEEEEE12E012EEEEEEEEE
456   // *new_ctrl = 12E0EEEEEEEEEEEE12E0EEEEEEEEEEE
457   // position                   S    /
458   //
459   // new_ctrl  = 456E0123EEEEEEEE456E0123EEEEEEE
460   // *new_ctrl = 456E0123EEEEEEEE456E0123EEEEEEE (no change)
461   // position                   S        /
462   std::memset(new_ctrl + new_capacity + old_capacity_ + 2,
463               static_cast<int8_t>(ctrl_t::kEmpty), kQuarterWidth);
464 
465   // Finally, we set the new sentinel byte.
466   new_ctrl[new_capacity] = ctrl_t::kSentinel;
467 }
468 
InitControlBytesAfterSoo(ctrl_t * new_ctrl,ctrl_t h2,size_t new_capacity)469 void HashSetResizeHelper::InitControlBytesAfterSoo(ctrl_t* new_ctrl, ctrl_t h2,
470                                                    size_t new_capacity) {
471   assert(is_single_group(new_capacity));
472   std::memset(new_ctrl, static_cast<int8_t>(ctrl_t::kEmpty),
473               NumControlBytes(new_capacity));
474   assert(HashSetResizeHelper::SooSlotIndex() == 1);
475   // This allows us to avoid branching on had_soo_slot_.
476   assert(had_soo_slot_ || h2 == ctrl_t::kEmpty);
477   new_ctrl[1] = new_ctrl[new_capacity + 2] = h2;
478   new_ctrl[new_capacity] = ctrl_t::kSentinel;
479 }
480 
GrowIntoSingleGroupShuffleTransferableSlots(void * new_slots,size_t slot_size) const481 void HashSetResizeHelper::GrowIntoSingleGroupShuffleTransferableSlots(
482     void* new_slots, size_t slot_size) const {
483   assert(old_capacity_ > 0);
484   const size_t half_old_capacity = old_capacity_ / 2;
485 
486   SanitizerUnpoisonMemoryRegion(old_slots(), slot_size * old_capacity_);
487   std::memcpy(new_slots,
488               SlotAddress(old_slots(), half_old_capacity + 1, slot_size),
489               slot_size * half_old_capacity);
490   std::memcpy(SlotAddress(new_slots, half_old_capacity + 1, slot_size),
491               old_slots(), slot_size * (half_old_capacity + 1));
492 }
493 
GrowSizeIntoSingleGroupTransferable(CommonFields & c,size_t slot_size)494 void HashSetResizeHelper::GrowSizeIntoSingleGroupTransferable(
495     CommonFields& c, size_t slot_size) {
496   assert(old_capacity_ < Group::kWidth / 2);
497   assert(is_single_group(c.capacity()));
498   assert(IsGrowingIntoSingleGroupApplicable(old_capacity_, c.capacity()));
499 
500   GrowIntoSingleGroupShuffleControlBytes(c.control(), c.capacity());
501   GrowIntoSingleGroupShuffleTransferableSlots(c.slot_array(), slot_size);
502 
503   // We poison since GrowIntoSingleGroupShuffleTransferableSlots
504   // may leave empty slots unpoisoned.
505   PoisonSingleGroupEmptySlots(c, slot_size);
506 }
507 
TransferSlotAfterSoo(CommonFields & c,size_t slot_size)508 void HashSetResizeHelper::TransferSlotAfterSoo(CommonFields& c,
509                                                size_t slot_size) {
510   assert(was_soo_);
511   assert(had_soo_slot_);
512   assert(is_single_group(c.capacity()));
513   std::memcpy(SlotAddress(c.slot_array(), SooSlotIndex(), slot_size),
514               old_soo_data(), slot_size);
515   PoisonSingleGroupEmptySlots(c, slot_size);
516 }
517 
518 namespace {
519 
520 // Called whenever the table needs to vacate empty slots either by removing
521 // tombstones via rehash or growth.
522 ABSL_ATTRIBUTE_NOINLINE
FindInsertPositionWithGrowthOrRehash(CommonFields & common,size_t hash,const PolicyFunctions & policy)523 FindInfo FindInsertPositionWithGrowthOrRehash(CommonFields& common, size_t hash,
524                                               const PolicyFunctions& policy) {
525   const size_t cap = common.capacity();
526   if (cap > Group::kWidth &&
527       // Do these calculations in 64-bit to avoid overflow.
528       common.size() * uint64_t{32} <= cap * uint64_t{25}) {
529     // Squash DELETED without growing if there is enough capacity.
530     //
531     // Rehash in place if the current size is <= 25/32 of capacity.
532     // Rationale for such a high factor: 1) DropDeletesWithoutResize() is
533     // faster than resize, and 2) it takes quite a bit of work to add
534     // tombstones.  In the worst case, seems to take approximately 4
535     // insert/erase pairs to create a single tombstone and so if we are
536     // rehashing because of tombstones, we can afford to rehash-in-place as
537     // long as we are reclaiming at least 1/8 the capacity without doing more
538     // than 2X the work.  (Where "work" is defined to be size() for rehashing
539     // or rehashing in place, and 1 for an insert or erase.)  But rehashing in
540     // place is faster per operation than inserting or even doubling the size
541     // of the table, so we actually afford to reclaim even less space from a
542     // resize-in-place.  The decision is to rehash in place if we can reclaim
543     // at about 1/8th of the usable capacity (specifically 3/28 of the
544     // capacity) which means that the total cost of rehashing will be a small
545     // fraction of the total work.
546     //
547     // Here is output of an experiment using the BM_CacheInSteadyState
548     // benchmark running the old case (where we rehash-in-place only if we can
549     // reclaim at least 7/16*capacity) vs. this code (which rehashes in place
550     // if we can recover 3/32*capacity).
551     //
552     // Note that although in the worst-case number of rehashes jumped up from
553     // 15 to 190, but the number of operations per second is almost the same.
554     //
555     // Abridged output of running BM_CacheInSteadyState benchmark from
556     // raw_hash_set_benchmark.   N is the number of insert/erase operations.
557     //
558     //      | OLD (recover >= 7/16        | NEW (recover >= 3/32)
559     // size |    N/s LoadFactor NRehashes |    N/s LoadFactor NRehashes
560     //  448 | 145284       0.44        18 | 140118       0.44        19
561     //  493 | 152546       0.24        11 | 151417       0.48        28
562     //  538 | 151439       0.26        11 | 151152       0.53        38
563     //  583 | 151765       0.28        11 | 150572       0.57        50
564     //  628 | 150241       0.31        11 | 150853       0.61        66
565     //  672 | 149602       0.33        12 | 150110       0.66        90
566     //  717 | 149998       0.35        12 | 149531       0.70       129
567     //  762 | 149836       0.37        13 | 148559       0.74       190
568     //  807 | 149736       0.39        14 | 151107       0.39        14
569     //  852 | 150204       0.42        15 | 151019       0.42        15
570     DropDeletesWithoutResize(common, policy);
571   } else {
572     // Otherwise grow the container.
573     policy.resize(common, NextCapacity(cap), HashtablezInfoHandle{});
574   }
575   // This function is typically called with tables containing deleted slots.
576   // The table will be big and `FindFirstNonFullAfterResize` will always
577   // fallback to `find_first_non_full`. So using `find_first_non_full` directly.
578   return find_first_non_full(common, hash);
579 }
580 
581 }  // namespace
582 
GetHashRefForEmptyHasher(const CommonFields & common)583 const void* GetHashRefForEmptyHasher(const CommonFields& common) {
584   // Empty base optimization typically make the empty base class address to be
585   // the same as the first address of the derived class object.
586   // But we generally assume that for empty hasher we can return any valid
587   // pointer.
588   return &common;
589 }
590 
PrepareInsertNonSoo(CommonFields & common,size_t hash,FindInfo target,const PolicyFunctions & policy)591 size_t PrepareInsertNonSoo(CommonFields& common, size_t hash, FindInfo target,
592                            const PolicyFunctions& policy) {
593   // When there are no deleted slots in the table
594   // and growth_left is positive, we can insert at the first
595   // empty slot in the probe sequence (target).
596   const bool use_target_hint =
597       // Optimization is disabled when generations are enabled.
598       // We have to rehash even sparse tables randomly in such mode.
599       !SwisstableGenerationsEnabled() &&
600       common.growth_info().HasNoDeletedAndGrowthLeft();
601   if (ABSL_PREDICT_FALSE(!use_target_hint)) {
602     // Notes about optimized mode when generations are disabled:
603     // We do not enter this branch if table has no deleted slots
604     // and growth_left is positive.
605     // We enter this branch in the following cases listed in decreasing
606     // frequency:
607     // 1. Table without deleted slots (>95% cases) that needs to be resized.
608     // 2. Table with deleted slots that has space for the inserting element.
609     // 3. Table with deleted slots that needs to be rehashed or resized.
610     if (ABSL_PREDICT_TRUE(common.growth_info().HasNoGrowthLeftAndNoDeleted())) {
611       const size_t old_capacity = common.capacity();
612       policy.resize(common, NextCapacity(old_capacity), HashtablezInfoHandle{});
613       target = HashSetResizeHelper::FindFirstNonFullAfterResize(
614           common, old_capacity, hash);
615     } else {
616       // Note: the table may have no deleted slots here when generations
617       // are enabled.
618       const bool rehash_for_bug_detection =
619           common.should_rehash_for_bug_detection_on_insert();
620       if (rehash_for_bug_detection) {
621         // Move to a different heap allocation in order to detect bugs.
622         const size_t cap = common.capacity();
623         policy.resize(common,
624                       common.growth_left() > 0 ? cap : NextCapacity(cap),
625                       HashtablezInfoHandle{});
626       }
627       if (ABSL_PREDICT_TRUE(common.growth_left() > 0)) {
628         target = find_first_non_full(common, hash);
629       } else {
630         target = FindInsertPositionWithGrowthOrRehash(common, hash, policy);
631       }
632     }
633   }
634   PrepareInsertCommon(common);
635   common.growth_info().OverwriteControlAsFull(common.control()[target.offset]);
636   SetCtrl(common, target.offset, H2(hash), policy.slot_size);
637   common.infoz().RecordInsert(hash, target.probe_length);
638   return target.offset;
639 }
640 
641 }  // namespace container_internal
642 ABSL_NAMESPACE_END
643 }  // namespace absl
644