///////////////////////////////////////////////////////////////////////////// // // (C) Copyright Ion Gaztanaga 2006-2022 // (C) Copyright 2022 Joaquin M Lopez Munoz. // // Distributed under the Boost Software License, Version 1.0. // (See accompanying file LICENSE_1_0.txt or copy at // http://www.boost.org/LICENSE_1_0.txt) // // See http://www.boost.org/libs/intrusive for documentation. // ///////////////////////////////////////////////////////////////////////////// // fastmod_buckets option is implemented reusing parts of Joaquin M. Lopez // Munoz's "fxa_unordered" library (proof of concept of closed- and // open-addressing unordered associative containers), released under // Boost Software License: // // https://github.com/joaquintides/fxa_unordered/ // // On cases and systems that can't take advantage of Daniel Lemire's // "fastmod" (https://github.com/lemire/fastmod) approach, // precomputed divisions are used. // // As always, thanks Joaquin for your great work! #ifndef BOOST_INTRUSIVE_HASHTABLE_HPP #define BOOST_INTRUSIVE_HASHTABLE_HPP #include #include //General intrusive utilities #include #include #include #include #include #include #include #include #include #include #include #include #include #include //Implementation utilities #include #include #include #include #include #include //boost #include #include #include #include #include #include //std C++ #include //std::pair #include //std::size_t #include //std::uint64_t #if defined(BOOST_HAS_PRAGMA_ONCE) # pragma once #endif #ifdef _MSC_VER #include #endif namespace boost { namespace intrusive { #if !defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) /// @cond #if !defined(BOOST_NO_INT64_T)&&\ (defined(BOOST_HAS_INT128) || (defined(_MSC_VER) && defined(_WIN64))) #define BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT #endif //We only support LLP64(Win64) or LP64(most Unix) data models #ifdef _WIN64 //In 64 bit windows sizeof(size_t) == sizeof(unsigned long long) # define BOOST_INTRUSIVE_SIZE_C(NUMBER) NUMBER##ULL # define BOOST_INTRUSIVE_64_BIT_SIZE_T 1 #else //In 32 bit windows and 32/64 bit unixes sizeof(size_t) == sizeof(unsigned long) # define BOOST_INTRUSIVE_SIZE_C(NUMBER) NUMBER##UL # define BOOST_INTRUSIVE_64_BIT_SIZE_T (((((ULONG_MAX>>16)>>16)>>16)>>15) != 0) #endif template struct prime_list_holder { private: template // sizeof(SizeType) < sizeof(std::size_t) static BOOST_INTRUSIVE_FORCEINLINE SizeType truncate_size_type(std::size_t n, detail::true_) { return n < std::size_t(SizeType(-1)) ? static_cast(n) : SizeType(-1); } template // sizeof(SizeType) == sizeof(std::size_t) static BOOST_INTRUSIVE_FORCEINLINE SizeType truncate_size_type(std::size_t n, detail::false_) { return static_cast(n); } static const std::size_t prime_list[]; static const std::size_t prime_list_size; static const std::size_t *suggested_lower_bucket_count_ptr(std::size_t n) { const std::size_t *primes = &prime_list[0]; const std::size_t *primes_end = primes + prime_list_size; std::size_t const* bound = boost::movelib::lower_bound(primes, primes_end, n, value_less()); bound -= std::size_t(bound == primes_end); return bound; } static const std::size_t *suggested_upper_bucket_count_ptr(std::size_t n) { const std::size_t *primes = &prime_list[0]; const std::size_t *primes_end = primes + prime_list_size; std::size_t const* bound = boost::movelib::upper_bound(primes, primes_end, n, value_less()); bound -= std::size_t(bound == primes_end); return bound; } static std::size_t suggested_lower_bucket_count_impl(std::size_t n) { return *suggested_lower_bucket_count_ptr(n); } static std::size_t suggested_upper_bucket_count_impl(std::size_t n) { return *suggested_upper_bucket_count_ptr(n); } public: template static BOOST_INTRUSIVE_FORCEINLINE SizeType suggested_upper_bucket_count(SizeType n) { std::size_t const c = suggested_upper_bucket_count_impl(static_cast(n)); return truncate_size_type(c, detail::bool_<(sizeof(SizeType) < sizeof(std::size_t))>()); } template static BOOST_INTRUSIVE_FORCEINLINE SizeType suggested_lower_bucket_count(SizeType n) { std::size_t const c = suggested_lower_bucket_count_impl(static_cast(n)); return truncate_size_type(c, detail::bool_<(sizeof(SizeType) < sizeof(std::size_t))>()); } static BOOST_INTRUSIVE_FORCEINLINE std::size_t suggested_lower_bucket_count_idx(std::size_t n) { return static_cast(suggested_lower_bucket_count_ptr(n) - &prime_list[0]); } static BOOST_INTRUSIVE_FORCEINLINE std::size_t suggested_upper_bucket_count_idx(std::size_t n) { return static_cast(suggested_upper_bucket_count_ptr(n) - &prime_list[0]); } static BOOST_INTRUSIVE_FORCEINLINE std::size_t size_from_index(std::size_t n) { return prime_list[std::ptrdiff_t(n)]; } template BOOST_INTRUSIVE_FORCEINLINE static std::size_t modfunc(std::size_t hash) { return hash % prime_list[SizeIndex]; } static std::size_t(*const positions[])(std::size_t); #if defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) static const uint64_t inv_sizes32[]; static const std::size_t inv_sizes32_size; #endif BOOST_INTRUSIVE_FORCEINLINE static std::size_t lower_size_index(std::size_t n) { return prime_list_holder<>::suggested_lower_bucket_count_idx(n); } BOOST_INTRUSIVE_FORCEINLINE static std::size_t upper_size_index(std::size_t n) { return prime_list_holder<>::suggested_upper_bucket_count_idx(n); } BOOST_INTRUSIVE_FORCEINLINE static std::size_t size(std::size_t size_index) { return prime_list_holder<>::size_from_index(size_index); } #if defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) // https://github.com/lemire/fastmod BOOST_INTRUSIVE_FORCEINLINE static uint64_t mul128_u32(uint64_t lowbits, uint32_t d) { #if defined(_MSC_VER) return __umulh(lowbits, d); #else __extension__ typedef unsigned __int128 ext_uint128_t; return (ext_uint128_t(lowbits) * d) >> 64; #endif } BOOST_INTRUSIVE_FORCEINLINE static uint32_t fastmod_u32(uint32_t a, uint64_t M, uint32_t d) { uint64_t lowbits = M * a; return (uint32_t)(mul128_u32(lowbits, d)); } #endif // defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) BOOST_INTRUSIVE_FORCEINLINE static std::size_t position(std::size_t hash,std::size_t size_index) { #if defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) && BOOST_INTRUSIVE_64_BIT_SIZE_T const std::size_t sizes_under_32bit = sizeof(inv_sizes32)/sizeof(inv_sizes32[0]); if(BOOST_LIKELY(size_index < sizes_under_32bit)){ return fastmod_u32( uint32_t(hash)+uint32_t(hash>>32) , inv_sizes32[size_index] , uint32_t(prime_list[size_index]) ); } else{ return positions[size_index-sizes_under_32bit](hash); } #elif defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) return fastmod_u32(hash, inv_sizes32[size_index], uint32_t(sizes[size_index])); #else return positions[size_index](hash); #endif // defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) } }; template std::size_t(* const prime_list_holder::positions[])(std::size_t) = { #if !defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) modfunc< 0>,modfunc< 1>,modfunc< 2>,modfunc< 3>,modfunc< 4>, modfunc< 5>,modfunc< 6>,modfunc< 7>,modfunc< 8>,modfunc< 9>, modfunc<10>,modfunc<11>,modfunc<12>,modfunc<13>,modfunc<14>, modfunc<15>,modfunc<16>,modfunc<17>,modfunc<18>,modfunc<19>, modfunc<20>,modfunc<21>,modfunc<22>,modfunc<23>,modfunc<24>, modfunc<25>,modfunc<26>,modfunc<27>,modfunc<28>,modfunc<29>, modfunc<30>, #endif # if BOOST_INTRUSIVE_64_BIT_SIZE_T modfunc<31>,modfunc<32>,modfunc<33>,modfunc<34>, modfunc<35>,modfunc<36>,modfunc<37>,modfunc<38>,modfunc<39>, modfunc<40>,modfunc<41>,modfunc<42>,modfunc<43>,modfunc<44>, modfunc<45>,modfunc<46>,modfunc<47>,modfunc<48>,modfunc<49>, modfunc<50>,modfunc<51>,modfunc<52>,modfunc<53>,modfunc<54>, modfunc<55>,modfunc<56>,modfunc<57>,modfunc<58>,modfunc<59>, modfunc<60>,modfunc<61>,modfunc<62>,modfunc<63> # else modfunc<31> # endif }; template const std::size_t prime_list_holder::prime_list[] = { BOOST_INTRUSIVE_SIZE_C(3), BOOST_INTRUSIVE_SIZE_C(7), BOOST_INTRUSIVE_SIZE_C(11), BOOST_INTRUSIVE_SIZE_C(17), BOOST_INTRUSIVE_SIZE_C(29), BOOST_INTRUSIVE_SIZE_C(53), BOOST_INTRUSIVE_SIZE_C(97), BOOST_INTRUSIVE_SIZE_C(193), BOOST_INTRUSIVE_SIZE_C(389), BOOST_INTRUSIVE_SIZE_C(769), BOOST_INTRUSIVE_SIZE_C(1543), BOOST_INTRUSIVE_SIZE_C(3079), BOOST_INTRUSIVE_SIZE_C(6151), BOOST_INTRUSIVE_SIZE_C(12289), BOOST_INTRUSIVE_SIZE_C(24593), BOOST_INTRUSIVE_SIZE_C(49157), BOOST_INTRUSIVE_SIZE_C(98317), BOOST_INTRUSIVE_SIZE_C(196613), BOOST_INTRUSIVE_SIZE_C(393241), BOOST_INTRUSIVE_SIZE_C(786433), BOOST_INTRUSIVE_SIZE_C(1572869), BOOST_INTRUSIVE_SIZE_C(3145739), BOOST_INTRUSIVE_SIZE_C(6291469), BOOST_INTRUSIVE_SIZE_C(12582917), BOOST_INTRUSIVE_SIZE_C(25165843), BOOST_INTRUSIVE_SIZE_C(50331653), BOOST_INTRUSIVE_SIZE_C(100663319), BOOST_INTRUSIVE_SIZE_C(201326611), BOOST_INTRUSIVE_SIZE_C(402653189), BOOST_INTRUSIVE_SIZE_C(805306457), BOOST_INTRUSIVE_SIZE_C(1610612741), //0-30 indexes #if BOOST_INTRUSIVE_64_BIT_SIZE_T //Taken from Boost.MultiIndex code, thanks to Joaquin M. Lopez Munoz. BOOST_INTRUSIVE_SIZE_C(3221225473), //<- 32 bit values stop here (index 31) BOOST_INTRUSIVE_SIZE_C(6442450939), BOOST_INTRUSIVE_SIZE_C(12884901893), BOOST_INTRUSIVE_SIZE_C(25769803751), BOOST_INTRUSIVE_SIZE_C(51539607551), BOOST_INTRUSIVE_SIZE_C(103079215111), BOOST_INTRUSIVE_SIZE_C(206158430209), BOOST_INTRUSIVE_SIZE_C(412316860441), BOOST_INTRUSIVE_SIZE_C(824633720831), BOOST_INTRUSIVE_SIZE_C(1649267441651), BOOST_INTRUSIVE_SIZE_C(3298534883309), BOOST_INTRUSIVE_SIZE_C(6597069766657), BOOST_INTRUSIVE_SIZE_C(13194139533299), BOOST_INTRUSIVE_SIZE_C(26388279066623), BOOST_INTRUSIVE_SIZE_C(52776558133303), BOOST_INTRUSIVE_SIZE_C(105553116266489), BOOST_INTRUSIVE_SIZE_C(211106232532969), BOOST_INTRUSIVE_SIZE_C(422212465066001), BOOST_INTRUSIVE_SIZE_C(844424930131963), BOOST_INTRUSIVE_SIZE_C(1688849860263953), BOOST_INTRUSIVE_SIZE_C(3377699720527861), BOOST_INTRUSIVE_SIZE_C(6755399441055731), BOOST_INTRUSIVE_SIZE_C(13510798882111483), BOOST_INTRUSIVE_SIZE_C(27021597764222939), BOOST_INTRUSIVE_SIZE_C(54043195528445957), BOOST_INTRUSIVE_SIZE_C(108086391056891903), BOOST_INTRUSIVE_SIZE_C(216172782113783843), BOOST_INTRUSIVE_SIZE_C(432345564227567621), BOOST_INTRUSIVE_SIZE_C(864691128455135207), BOOST_INTRUSIVE_SIZE_C(1729382256910270481), BOOST_INTRUSIVE_SIZE_C(3458764513820540933), BOOST_INTRUSIVE_SIZE_C(6917529027641081903), BOOST_INTRUSIVE_SIZE_C(9223372036854775783) //(index 63) #else BOOST_INTRUSIVE_SIZE_C(2147483647) //<- 32 bit stops here (index 31) as ptrdiff_t is signed #endif }; template const std::size_t prime_list_holder::prime_list_size = sizeof(prime_list) / sizeof(std::size_t); #if defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) template const uint64_t prime_list_holder::inv_sizes32[] = { BOOST_INTRUSIVE_SIZE_C(6148914691236517206), //3 BOOST_INTRUSIVE_SIZE_C(2635249153387078803), //7 BOOST_INTRUSIVE_SIZE_C(1676976733973595602), //11 BOOST_INTRUSIVE_SIZE_C(1085102592571150096), //17 BOOST_INTRUSIVE_SIZE_C(636094623231363849), //29 BOOST_INTRUSIVE_SIZE_C(348051774975651918), //53 BOOST_INTRUSIVE_SIZE_C(190172619316593316), //97 BOOST_INTRUSIVE_SIZE_C(95578984837873325), //193 BOOST_INTRUSIVE_SIZE_C(47420935922132524), //389 BOOST_INTRUSIVE_SIZE_C(23987963684927896), //769 BOOST_INTRUSIVE_SIZE_C(11955116055547344), //1543 BOOST_INTRUSIVE_SIZE_C(5991147799191151), //3079 BOOST_INTRUSIVE_SIZE_C(2998982941588287), //6151 BOOST_INTRUSIVE_SIZE_C(1501077717772769), //12289 BOOST_INTRUSIVE_SIZE_C(750081082979285), //24593 BOOST_INTRUSIVE_SIZE_C(375261795343686), //49157 BOOST_INTRUSIVE_SIZE_C(187625172388393), //98317 BOOST_INTRUSIVE_SIZE_C(93822606204624), //196613 BOOST_INTRUSIVE_SIZE_C(46909513691883), //393241 BOOST_INTRUSIVE_SIZE_C(23456218233098), //786433 BOOST_INTRUSIVE_SIZE_C(11728086747027), //1572869 BOOST_INTRUSIVE_SIZE_C(5864041509391), //3145739 BOOST_INTRUSIVE_SIZE_C(2932024948977), //6291469 BOOST_INTRUSIVE_SIZE_C(1466014921160), //12582917 BOOST_INTRUSIVE_SIZE_C(733007198436), //25165843 BOOST_INTRUSIVE_SIZE_C(366503839517), //50331653 BOOST_INTRUSIVE_SIZE_C(183251896093), //100663319 BOOST_INTRUSIVE_SIZE_C(91625960335), //201326611 BOOST_INTRUSIVE_SIZE_C(45812983922), //402653189 BOOST_INTRUSIVE_SIZE_C(22906489714), //805306457 BOOST_INTRUSIVE_SIZE_C(11453246088), //1610612741 BOOST_INTRUSIVE_SIZE_C(5726623060) //3221225473 }; template const std::size_t prime_list_holder::inv_sizes32_size = sizeof(inv_sizes32) / sizeof(uint64_t); #endif // defined(BOOST_INTRUSIVE_FCA_FASTMOD_SUPPORT) struct prime_fmod_size : prime_list_holder<> { }; #undef BOOST_INTRUSIVE_SIZE_C #undef BOOST_INTRUSIVE_64_BIT_SIZE_T #endif //#if !defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) template InputIt priv_algo_find(InputIt first, InputIt last, const T& value) { for (; first != last; ++first) { if (*first == value) { return first; } } return last; } template typename boost::intrusive::iterator_traits::difference_type priv_algo_count(InputIt first, InputIt last, const T& value) { typename boost::intrusive::iterator_traits::difference_type ret = 0; for (; first != last; ++first) { if (*first == value) { ret++; } } return ret; } template bool priv_algo_is_permutation(ForwardIterator1 first1, ForwardIterator1 last1, ForwardIterator2 first2) { typedef typename boost::intrusive::iterator_traits::difference_type distance_type; //Efficiently compare identical prefixes: O(N) if sequences //have the same elements in the same order. for ( ; first1 != last1; ++first1, ++first2){ if (! (*first1 == *first2)) break; } if (first1 == last1){ return true; } //Establish last2 assuming equal ranges by iterating over the //rest of the list. ForwardIterator2 last2 = first2; boost::intrusive::iterator_advance(last2, boost::intrusive::iterator_distance(first1, last1)); for(ForwardIterator1 scan = first1; scan != last1; ++scan){ if (scan != (priv_algo_find)(first1, scan, *scan)){ continue; //We've seen this one before. } distance_type matches = (priv_algo_count)(first2, last2, *scan); if (0 == matches || (priv_algo_count)(scan, last1, *scan) != matches){ return false; } } return true; } struct hash_bool_flags { static const std::size_t unique_keys_pos = 1u; static const std::size_t constant_time_size_pos = 2u; static const std::size_t power_2_buckets_pos = 4u; static const std::size_t cache_begin_pos = 8u; static const std::size_t compare_hash_pos = 16u; static const std::size_t incremental_pos = 32u; static const std::size_t linear_buckets_pos = 64u; static const std::size_t fastmod_buckets_pos = 128u; }; template class exception_bucket_disposer { Bucket *cont_; Disposer &disp_; const SizeType &constructed_; exception_bucket_disposer(const exception_bucket_disposer&); exception_bucket_disposer &operator=(const exception_bucket_disposer&); public: exception_bucket_disposer (Bucket &cont, Disposer &disp, const SizeType &constructed) : cont_(&cont), disp_(disp), constructed_(constructed) {} BOOST_INTRUSIVE_FORCEINLINE void release() { cont_ = 0; } ~exception_bucket_disposer() { SizeType n = constructed_; if(cont_){ while(n--){ Algo::detach_and_dispose(cont_[n].get_node_ptr(), disp_); } } } }; template struct unordered_bucket_impl { typedef typename detail::get_node_traits ::type node_traits; typedef typename reduced_slist_node_traits ::type reduced_node_traits; typedef bucket_impl type; typedef typename pointer_traits ::template rebind_pointer::type pointer; }; template struct unordered_bucket_ptr_impl { typedef typename unordered_bucket_impl::pointer type; }; template struct bucket_traits_impl { private: BOOST_COPYABLE_AND_MOVABLE(bucket_traits_impl) public: /// @cond typedef BucketPtr bucket_ptr; typedef SizeType size_type; /// @endcond BOOST_INTRUSIVE_FORCEINLINE bucket_traits_impl(bucket_ptr buckets, size_type len) : buckets_(buckets), buckets_len_(len) {} BOOST_INTRUSIVE_FORCEINLINE bucket_traits_impl(const bucket_traits_impl& x) : buckets_(x.buckets_), buckets_len_(x.buckets_len_) {} BOOST_INTRUSIVE_FORCEINLINE bucket_traits_impl(BOOST_RV_REF(bucket_traits_impl) x) : buckets_(x.buckets_), buckets_len_(x.buckets_len_) { x.buckets_ = bucket_ptr(); x.buckets_len_ = 0u; } BOOST_INTRUSIVE_FORCEINLINE bucket_traits_impl& operator=(BOOST_RV_REF(bucket_traits_impl) x) { buckets_ = x.buckets_; buckets_len_ = x.buckets_len_; x.buckets_ = bucket_ptr(); x.buckets_len_ = 0u; return *this; } BOOST_INTRUSIVE_FORCEINLINE bucket_traits_impl& operator=(BOOST_COPY_ASSIGN_REF(bucket_traits_impl) x) { buckets_ = x.buckets_; buckets_len_ = x.buckets_len_; return *this; } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr bucket_begin() const { return buckets_; } BOOST_INTRUSIVE_FORCEINLINE size_type bucket_count() const BOOST_NOEXCEPT { return buckets_len_; } private: bucket_ptr buckets_; size_type buckets_len_; }; template struct store_hash_is_true { template struct two_or_three {detail::yes_type _[2u + (unsigned)Add];}; template static detail::yes_type test(...); template static two_or_three test (int); static const bool value = sizeof(test(0)) > sizeof(detail::yes_type)*2u; }; template struct optimize_multikey_is_true { template struct two_or_three { detail::yes_type _[2u + (unsigned)Add];}; template static detail::yes_type test(...); template static two_or_three test (int); static const bool value = sizeof(test(0)) > sizeof(detail::yes_type)*2u; }; template struct insert_commit_data_impl { std::size_t hash; std::size_t bucket_idx; BOOST_INTRUSIVE_FORCEINLINE std::size_t get_hash() const { return hash; } BOOST_INTRUSIVE_FORCEINLINE void set_hash(std::size_t h) { hash = h; } }; template<> struct insert_commit_data_impl { std::size_t bucket_idx; BOOST_INTRUSIVE_FORCEINLINE std::size_t get_hash() const { return 0U; } BOOST_INTRUSIVE_FORCEINLINE void set_hash(std::size_t) {} }; template BOOST_INTRUSIVE_FORCEINLINE typename pointer_traits::template rebind_pointer::type dcast_bucket_ptr(const SlistNodePtr &p) { typedef typename pointer_traits::template rebind_pointer::type node_ptr; return pointer_traits::pointer_to(static_cast(*p)); } template struct group_functions { // A group is reverse-linked // // A is "first in group" // C is "last in group" // __________________ // | _____ _____ | // | | | | | | <- Group links // ^ V ^ V ^ V // _ _ _ _ // A|_| B|_| C|_| D|_| // // ^ | ^ | ^ | ^ V <- Bucket links // _ _____| |_____| |______| |____| | // |B| | // ^________________________________| // typedef NodeTraits node_traits; typedef unordered_group_adapter group_traits; typedef typename node_traits::node_ptr node_ptr; typedef typename node_traits::node node; typedef typename reduced_slist_node_traits ::type reduced_node_traits; typedef typename reduced_node_traits::node_ptr slist_node_ptr; typedef typename reduced_node_traits::node slist_node; typedef circular_slist_algorithms group_algorithms; typedef circular_slist_algorithms node_algorithms; static slist_node_ptr get_bucket_before_begin (slist_node_ptr bucket_beg, slist_node_ptr bucket_last, slist_node_ptr sp, detail::true_) { //First find the last node of p's group. //This requires checking the first node of the next group or //the bucket node. node_ptr p = dcast_bucket_ptr(sp); node_ptr prev_node = p; node_ptr nxt(node_traits::get_next(p)); while(!(bucket_beg <= nxt && nxt <= bucket_last) && (group_traits::get_next(nxt) == prev_node)){ prev_node = nxt; nxt = node_traits::get_next(nxt); } //If we've reached the bucket node just return it. if(bucket_beg <= nxt && nxt <= bucket_last){ return nxt; } //Otherwise, iterate using group links until the bucket node node_ptr first_node_of_group = nxt; node_ptr last_node_group = group_traits::get_next(first_node_of_group); slist_node_ptr possible_end = node_traits::get_next(last_node_group); while(!(bucket_beg <= possible_end && possible_end <= bucket_last)){ first_node_of_group = dcast_bucket_ptr(possible_end); last_node_group = group_traits::get_next(first_node_of_group); possible_end = node_traits::get_next(last_node_group); } return possible_end; } static slist_node_ptr get_bucket_before_begin (slist_node_ptr bucket_beg, slist_node_ptr bucket_last, slist_node_ptr sp, detail::false_) { //The end node is embedded in the singly linked list: //iterate until we reach it. while (!(bucket_beg <= sp && sp <= bucket_last)){ sp = reduced_node_traits::get_next(sp); } return sp; } static node_ptr get_prev_to_first_in_group(slist_node_ptr bucket_node, node_ptr first_in_group) { node_ptr nb = dcast_bucket_ptr(bucket_node); node_ptr n; while((n = node_traits::get_next(nb)) != first_in_group){ nb = group_traits::get_next(n); //go to last in group } return nb; } static void erase_from_group(slist_node_ptr end_ptr, node_ptr to_erase_ptr, detail::true_) { node_ptr const nxt_ptr(node_traits::get_next(to_erase_ptr)); //Check if the next node is in the group (not end node) and reverse linked to //'to_erase_ptr'. Erase if that's the case. if(nxt_ptr != end_ptr && to_erase_ptr == group_traits::get_next(nxt_ptr)){ group_algorithms::unlink_after(nxt_ptr); } } BOOST_INTRUSIVE_FORCEINLINE static void erase_from_group(slist_node_ptr, node_ptr, detail::false_) {} BOOST_INTRUSIVE_FORCEINLINE static node_ptr get_last_in_group(node_ptr first_in_group, detail::true_) { return group_traits::get_next(first_in_group); } BOOST_INTRUSIVE_FORCEINLINE static node_ptr get_last_in_group(node_ptr n, detail::false_) { return n; } static node_ptr get_first_in_group(node_ptr n, detail::true_) { node_ptr ng; while(n == node_traits::get_next((ng = group_traits::get_next(n)))){ n = ng; } return n; } BOOST_INTRUSIVE_FORCEINLINE static node_ptr get_first_in_group(node_ptr n, detail::false_) { return n; } BOOST_INTRUSIVE_FORCEINLINE static bool is_first_in_group(node_ptr ptr) { return node_traits::get_next(group_traits::get_next(ptr)) != ptr; } BOOST_INTRUSIVE_FORCEINLINE static void insert_in_group(node_ptr first_in_group, node_ptr n, detail::true_) { group_algorithms::link_after(first_in_group, n); } BOOST_INTRUSIVE_FORCEINLINE static void insert_in_group(node_ptr, node_ptr, detail::false_) {} //Splits a group in two groups, and makes "new_first" the first node in the second group. //Returns the first element of the first group static node_ptr split_group(node_ptr const new_first) { node_ptr const old_first((get_first_in_group)(new_first, detail::true_())); //Check new_first was not the first in group if(old_first != new_first){ node_ptr const last = group_traits::get_next(old_first); group_traits::set_next(old_first, group_traits::get_next(new_first)); group_traits::set_next(new_first, last); } return old_first; } }; template class incremental_rehash_rollback { private: typedef BucketType bucket_type; typedef SplitTraits split_traits; incremental_rehash_rollback(); incremental_rehash_rollback & operator=(const incremental_rehash_rollback &); incremental_rehash_rollback (const incremental_rehash_rollback &); public: incremental_rehash_rollback (bucket_type &source_bucket, bucket_type &destiny_bucket, split_traits &split_tr) : source_bucket_(source_bucket), destiny_bucket_(destiny_bucket) , split_traits_(split_tr), released_(false) {} BOOST_INTRUSIVE_FORCEINLINE void release() { released_ = true; } ~incremental_rehash_rollback() { if(!released_){ //If an exception is thrown, just put all moved nodes back in the old bucket //and move back the split mark. SlistNodeAlgorithms::transfer_after(destiny_bucket_.get_node_ptr(), source_bucket_.get_node_ptr()); split_traits_.decrement(); } } private: bucket_type &source_bucket_; bucket_type &destiny_bucket_; split_traits &split_traits_; bool released_; }; template struct node_functions { BOOST_INTRUSIVE_FORCEINLINE static void store_hash(typename NodeTraits::node_ptr p, std::size_t h, detail::true_) { return NodeTraits::set_hash(p, h); } BOOST_INTRUSIVE_FORCEINLINE static void store_hash(typename NodeTraits::node_ptr, std::size_t, detail::false_) {} }; BOOST_INTRUSIVE_FORCEINLINE std::size_t hash_to_bucket(std::size_t hash_value, std::size_t bucket_cnt, detail::false_) { return hash_value % bucket_cnt; } BOOST_INTRUSIVE_FORCEINLINE std::size_t hash_to_bucket(std::size_t hash_value, std::size_t bucket_cnt, detail::true_) { return hash_value & (bucket_cnt - 1); } template //!fastmod_buckets BOOST_INTRUSIVE_FORCEINLINE std::size_t hash_to_bucket_split(std::size_t hash_value, std::size_t bucket_cnt, std::size_t split, detail::false_) { std::size_t bucket_number = hash_to_bucket(hash_value, bucket_cnt, detail::bool_()); BOOST_IF_CONSTEXPR(Incremental) bucket_number -= static_cast(bucket_number >= split)*(bucket_cnt/2); return bucket_number; } template //fastmod_buckets BOOST_INTRUSIVE_FORCEINLINE std::size_t hash_to_bucket_split(std::size_t hash_value, std::size_t bucket_cnt, std::size_t split, detail::true_) { return prime_fmod_size::position(hash_value, split); (void)bucket_cnt; } //!This metafunction will obtain the type of a bucket //!from the value_traits or hook option to be used with //!a hash container. template struct unordered_bucket : public unordered_bucket_impl < typename ValueTraitsOrHookOption:: template pack::proto_value_traits> {}; //!This metafunction will obtain the type of a bucket pointer //!from the value_traits or hook option to be used with //!a hash container. template struct unordered_bucket_ptr : public unordered_bucket_ptr_impl < typename ValueTraitsOrHookOption:: template pack::proto_value_traits> {}; //!This metafunction will obtain the type of the default bucket traits //!(when the user does not specify the bucket_traits<> option) from the //!value_traits or hook option to be used with //!a hash container. template struct unordered_default_bucket_traits { typedef typename ValueTraitsOrHookOption:: template pack::proto_value_traits supposed_value_traits; typedef bucket_traits_impl < typename unordered_bucket_ptr_impl ::type , std::size_t> type; }; struct default_bucket_traits; //hashtable default hook traits struct default_hashtable_hook_applier { template struct apply{ typedef typename T::default_hashtable_hook type; }; }; template<> struct is_default_hook_tag { static const bool value = true; }; struct hashtable_defaults { typedef default_hashtable_hook_applier proto_value_traits; typedef std::size_t size_type; typedef void key_of_value; typedef void equal; typedef void hash; typedef default_bucket_traits bucket_traits; static const bool constant_time_size = true; static const bool power_2_buckets = false; static const bool cache_begin = false; static const bool compare_hash = false; static const bool incremental = false; static const bool linear_buckets = false; static const bool fastmod_buckets = false; }; template struct downcast_node_to_value_t : public detail::node_to_value { typedef detail::node_to_value base_t; typedef typename base_t::result_type result_type; typedef ValueTraits value_traits; typedef typename unordered_bucket_impl ::type::node_traits::node node; typedef typename detail::add_const_if_c ::type &first_argument_type; typedef typename detail::add_const_if_c < typename ValueTraits::node_traits::node , IsConst>::type &intermediate_argument_type; typedef typename pointer_traits :: template rebind_pointer ::type const_value_traits_ptr; BOOST_INTRUSIVE_FORCEINLINE downcast_node_to_value_t(const_value_traits_ptr ptr) : base_t(ptr) {} BOOST_INTRUSIVE_FORCEINLINE result_type operator()(first_argument_type arg) const { return this->base_t::operator()(static_cast(arg)); } }; template struct node_cast_adaptor //Use public inheritance to avoid MSVC bugs with closures : public detail::ebo_functor_holder { typedef detail::ebo_functor_holder base_t; typedef typename pointer_traits::element_type slist_node; typedef typename pointer_traits::element_type node; template BOOST_INTRUSIVE_FORCEINLINE node_cast_adaptor(const ConvertibleToF &c2f, const RealValuTraits *traits) : base_t(base_t(c2f, traits)) {} BOOST_INTRUSIVE_FORCEINLINE typename base_t::node_ptr operator()(const slist_node &to_clone) { return base_t::operator()(static_cast(to_clone)); } BOOST_INTRUSIVE_FORCEINLINE void operator()(SlistNodePtr to_clone) { base_t::operator()(pointer_traits::pointer_to(static_cast(*to_clone))); } }; //bucket_plus_vtraits stores ValueTraits + BucketTraits //this data is needed by iterators to obtain the //value from the iterator and detect the bucket template struct bucket_plus_vtraits { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(bucket_plus_vtraits) struct data_type : public ValueTraits, BucketTraits { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(data_type) public: BOOST_INTRUSIVE_FORCEINLINE data_type(const ValueTraits& val_traits, const BucketTraits& b_traits) : ValueTraits(val_traits), BucketTraits(b_traits) {} BOOST_INTRUSIVE_FORCEINLINE data_type(BOOST_RV_REF(data_type) other) : ValueTraits (BOOST_MOVE_BASE(ValueTraits, other)) , BucketTraits(BOOST_MOVE_BASE(BucketTraits, other)) {} } m_data; public: typedef BucketTraits bucket_traits; typedef ValueTraits value_traits; static const bool safemode_or_autounlink = is_safe_autounlink::value; typedef typename unordered_bucket_impl ::type bucket_type; typedef typename unordered_bucket_ptr_impl ::type bucket_ptr; typedef typename value_traits::node_traits node_traits; typedef typename bucket_type::node_traits slist_node_traits; typedef unordered_group_adapter group_traits; typedef group_functions group_functions_t; typedef typename detail::if_c < LinearBuckets , linear_slist_algorithms , circular_slist_algorithms >::type slist_node_algorithms; typedef typename slist_node_traits::node_ptr slist_node_ptr; typedef trivial_value_traits slist_value_traits; typedef slist_iterator siterator; typedef slist_iterator const_siterator; typedef typename node_traits::node_ptr node_ptr; typedef typename node_traits::const_node_ptr const_node_ptr; typedef typename node_traits::node node; typedef typename value_traits::value_type value_type; typedef typename value_traits::pointer pointer; typedef typename value_traits::const_pointer const_pointer; typedef typename pointer_traits::reference reference; typedef typename pointer_traits ::reference const_reference; typedef circular_slist_algorithms group_algorithms; typedef typename pointer_traits :: template rebind_pointer ::type const_value_traits_ptr; typedef typename pointer_traits :: template rebind_pointer ::type const_bucket_value_traits_ptr; typedef detail::bool_ linear_buckets_t; typedef bucket_plus_vtraits& this_ref; static const std::size_t bucket_overhead = LinearBuckets ? 1u : 0u; BOOST_INTRUSIVE_FORCEINLINE bucket_plus_vtraits(const ValueTraits &val_traits, const bucket_traits &b_traits) : m_data(val_traits, b_traits) {} BOOST_INTRUSIVE_FORCEINLINE bucket_plus_vtraits(BOOST_RV_REF(bucket_plus_vtraits) other) : m_data(boost::move(((bucket_plus_vtraits&)other).m_data)) {} BOOST_INTRUSIVE_FORCEINLINE const_value_traits_ptr priv_value_traits_ptr() const { return pointer_traits::pointer_to(this->priv_value_traits()); } //bucket_value_traits // BOOST_INTRUSIVE_FORCEINLINE const bucket_plus_vtraits &get_bucket_value_traits() const { return *this; } BOOST_INTRUSIVE_FORCEINLINE bucket_plus_vtraits &get_bucket_value_traits() { return *this; } BOOST_INTRUSIVE_FORCEINLINE const_bucket_value_traits_ptr bucket_value_traits_ptr() const { return pointer_traits::pointer_to(this->get_bucket_value_traits()); } //value traits // BOOST_INTRUSIVE_FORCEINLINE const value_traits &priv_value_traits() const { return static_cast(this->m_data); } BOOST_INTRUSIVE_FORCEINLINE value_traits &priv_value_traits() { return static_cast(this->m_data); } //value traits // BOOST_INTRUSIVE_FORCEINLINE const bucket_traits &priv_bucket_traits() const { return static_cast(this->m_data); } BOOST_INTRUSIVE_FORCEINLINE bucket_traits& priv_bucket_traits() { return static_cast(this->m_data); } //bucket operations BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_bucket_pointer() const BOOST_NOEXCEPT { return this->priv_bucket_traits().bucket_begin(); } BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_usable_bucket_count() const BOOST_NOEXCEPT { BOOST_IF_CONSTEXPR(bucket_overhead){ const std::size_t n = this->priv_bucket_traits().bucket_count(); return n - std::size_t(n != 0)*bucket_overhead; } else{ return this->priv_bucket_traits().bucket_count(); } } BOOST_INTRUSIVE_FORCEINLINE bucket_type &priv_bucket(std::size_t n) const BOOST_NOEXCEPT { BOOST_INTRUSIVE_INVARIANT_ASSERT(n < this->priv_usable_bucket_count()); return this->priv_bucket_pointer()[std::ptrdiff_t(n)]; } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_bucket_ptr(std::size_t n) const BOOST_NOEXCEPT { return pointer_traits::pointer_to(this->priv_bucket(n)); } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_past_usable_bucket_ptr() const { return this->priv_bucket_pointer() + std::ptrdiff_t(priv_usable_bucket_count()); } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_invalid_bucket_ptr() const { BOOST_IF_CONSTEXPR(LinearBuckets) { return bucket_ptr(); } else{ return this->priv_past_usable_bucket_ptr(); } } BOOST_INTRUSIVE_FORCEINLINE void priv_set_sentinel_bucket() const { BOOST_IF_CONSTEXPR(LinearBuckets) { BOOST_INTRUSIVE_INVARIANT_ASSERT(this->priv_bucket_traits().bucket_count() > 1); bucket_type &b = this->priv_bucket_pointer()[std::ptrdiff_t(this->priv_usable_bucket_count())]; slist_node_algorithms::set_sentinel(b.get_node_ptr()); } } BOOST_INTRUSIVE_FORCEINLINE void priv_unset_sentinel_bucket() const { BOOST_IF_CONSTEXPR(LinearBuckets) { BOOST_INTRUSIVE_INVARIANT_ASSERT(this->priv_bucket_traits().bucket_count() > 1); bucket_type& b = this->priv_bucket_pointer()[std::ptrdiff_t(this->priv_usable_bucket_count())]; slist_node_algorithms::init_header(b.get_node_ptr()); } } BOOST_INTRUSIVE_FORCEINLINE siterator priv_end_sit() const { return priv_end_sit(linear_buckets_t()); } BOOST_INTRUSIVE_FORCEINLINE siterator priv_end_sit(detail::true_) const { return siterator(this->priv_bucket_pointer() + std::ptrdiff_t(this->priv_bucket_traits().bucket_count() - bucket_overhead)); } BOOST_INTRUSIVE_FORCEINLINE siterator priv_end_sit(detail::false_) const { return siterator(this->priv_bucket_pointer()->get_node_ptr()); } BOOST_INTRUSIVE_FORCEINLINE siterator priv_bucket_lbegin(std::size_t n) const { siterator s(this->priv_bucket_lbbegin(n)); return ++s; } BOOST_INTRUSIVE_FORCEINLINE siterator priv_bucket_lbbegin(std::size_t n) const { return this->sit_bbegin(this->priv_bucket(n)); } BOOST_INTRUSIVE_FORCEINLINE siterator priv_bucket_lend(std::size_t n) const { return this->sit_end(this->priv_bucket(n)); } BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_bucket_size(std::size_t n) const { return slist_node_algorithms::count(this->priv_bucket(n).get_node_ptr())-1u; } BOOST_INTRUSIVE_FORCEINLINE bool priv_bucket_empty(std::size_t n) const { return slist_node_algorithms::is_empty(this->priv_bucket(n).get_node_ptr()); } BOOST_INTRUSIVE_FORCEINLINE bool priv_bucket_empty(bucket_ptr p) const { return slist_node_algorithms::is_empty(p->get_node_ptr()); } static BOOST_INTRUSIVE_FORCEINLINE siterator priv_bucket_lbegin(bucket_type &b) { return siterator(slist_node_traits::get_next(b.get_node_ptr())); } static BOOST_INTRUSIVE_FORCEINLINE siterator priv_bucket_lbbegin(bucket_type& b) { return siterator(b.get_node_ptr()); } static BOOST_INTRUSIVE_FORCEINLINE siterator priv_bucket_lend(bucket_type& b) { return siterator(slist_node_algorithms::end_node(b.get_node_ptr())); } static BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_bucket_size(const bucket_type& b) { return slist_node_algorithms::count(b.get_node_ptr())-1u; } static BOOST_INTRUSIVE_FORCEINLINE bool priv_bucket_empty(const bucket_type& b) { return slist_node_algorithms::is_empty(b.get_node_ptr()); } template static std::size_t priv_erase_from_single_bucket (bucket_type &b, siterator sbefore_first, siterator slast, NodeDisposer node_disposer, detail::true_) //optimize multikey { std::size_t n = 0; siterator const sfirst(++siterator(sbefore_first)); if(sfirst != slast){ node_ptr const nf = dcast_bucket_ptr(sfirst.pointed_node()); node_ptr const nl = dcast_bucket_ptr(slast.pointed_node()); slist_node_ptr const ne = (priv_bucket_lend(b)).pointed_node(); if(group_functions_t::is_first_in_group(nf)) { // The first node is at the beginning of a group. if(nl != ne){ group_functions_t::split_group(nl); } } else { node_ptr const group1 = group_functions_t::split_group(nf); if(nl != ne) { node_ptr const group2 = group_functions_t::split_group(nl); if(nf == group2) { //Both first and last in the same group //so join group1 and group2 node_ptr const end1 = group_traits::get_next(group1); node_ptr const end2 = group_traits::get_next(group2); group_traits::set_next(group1, end2); group_traits::set_next(nl, end1); } } } n = slist_node_algorithms::unlink_after_and_dispose(sbefore_first.pointed_node(), slast.pointed_node(), node_disposer); } return n; } template static std::size_t priv_erase_from_single_bucket (bucket_type &, siterator sbefore_first, siterator slast, NodeDisposer node_disposer, detail::false_) //optimize multikey { return slist_node_algorithms::unlink_after_and_dispose(sbefore_first.pointed_node(), slast.pointed_node(), node_disposer); } template static void priv_erase_node(bucket_type &b, siterator i, NodeDisposer node_disposer, detail::true_) //optimize multikey { slist_node_ptr const ne(priv_bucket_lend(b).pointed_node()); slist_node_ptr const nbb(priv_bucket_lbbegin(b).pointed_node()); node_ptr n(dcast_bucket_ptr(i.pointed_node())); node_ptr pos = node_traits::get_next(group_traits::get_next(n)); node_ptr bn; node_ptr nn(node_traits::get_next(n)); if(pos != n) { //Node is the first of the group bn = group_functions_t::get_prev_to_first_in_group(nbb, n); //Unlink the rest of the group if it's not the last node of its group if(nn != ne && group_traits::get_next(nn) == n){ group_algorithms::unlink_after(nn); } } else if(nn != ne && group_traits::get_next(nn) == n){ //Node is not the end of the group bn = group_traits::get_next(n); group_algorithms::unlink_after(nn); } else{ //Node is the end of the group bn = group_traits::get_next(n); node_ptr const x(group_algorithms::get_previous_node(n)); group_algorithms::unlink_after(x); } slist_node_algorithms::unlink_after_and_dispose(bn, node_disposer); } template BOOST_INTRUSIVE_FORCEINLINE static void priv_erase_node(bucket_type &b, siterator i, NodeDisposer node_disposer, detail::false_) //!optimize multikey { slist_node_ptr bi = slist_node_algorithms::get_previous_node(b.get_node_ptr(), i.pointed_node()); slist_node_algorithms::unlink_after_and_dispose(bi, node_disposer); } template std::size_t priv_erase_node_range( siterator const &before_first_it, std::size_t const first_bucket , siterator const &last_it, std::size_t const last_bucket , NodeDisposer node_disposer, detail::bool_ optimize_multikey_tag) { std::size_t num_erased(0); siterator last_step_before_it; if(first_bucket != last_bucket){ bucket_type *b = &this->priv_bucket(0); num_erased += this->priv_erase_from_single_bucket (b[first_bucket], before_first_it, this->priv_bucket_lend(first_bucket), node_disposer, optimize_multikey_tag); for(std::size_t i = 0, n = (last_bucket - first_bucket - 1); i != n; ++i){ num_erased += this->priv_erase_whole_bucket(b[first_bucket+i+1], node_disposer); } last_step_before_it = this->priv_bucket_lbbegin(last_bucket); } else{ last_step_before_it = before_first_it; } num_erased += this->priv_erase_from_single_bucket (this->priv_bucket(last_bucket), last_step_before_it, last_it, node_disposer, optimize_multikey_tag); return num_erased; } static siterator priv_get_last(bucket_type &b, detail::true_) //optimize multikey { //First find the last node of p's group. //This requires checking the first node of the next group or //the bucket node. slist_node_ptr end_ptr(sit_end(b).pointed_node()); slist_node_ptr last_node_group(b.get_node_ptr()); slist_node_ptr possible_end(slist_node_traits::get_next(last_node_group)); while(end_ptr != possible_end){ last_node_group = group_traits::get_next(dcast_bucket_ptr(possible_end)); possible_end = slist_node_traits::get_next(last_node_group); } return siterator(last_node_group); } BOOST_INTRUSIVE_FORCEINLINE static siterator priv_get_last(bucket_type &b, detail::false_) //NOT optimize multikey { slist_node_ptr p = b.get_node_ptr(); return siterator(slist_node_algorithms::get_previous_node(p, slist_node_algorithms::end_node(p))); } template static BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_erase_whole_bucket(bucket_type &b, NodeDisposer node_disposer) { return slist_node_algorithms::detach_and_dispose(b.get_node_ptr(), node_disposer); } static siterator priv_get_previous(bucket_type &b, siterator i, detail::true_) //optimize multikey { node_ptr const elem(dcast_bucket_ptr(i.pointed_node())); node_ptr const prev_in_group(group_traits::get_next(elem)); bool const first_in_group = node_traits::get_next(prev_in_group) != elem; slist_node_ptr n = first_in_group ? group_functions_t::get_prev_to_first_in_group(b.get_node_ptr(), elem) : group_traits::get_next(elem) ; return siterator(n); } BOOST_INTRUSIVE_FORCEINLINE static siterator priv_get_previous(bucket_type &b, siterator i, detail::false_) //NOT optimize multikey { return siterator(slist_node_algorithms::get_previous_node(b.get_node_ptr(), i.pointed_node())); } template struct typeof_node_disposer { typedef node_cast_adaptor < detail::node_disposer< Disposer, value_traits, CommonSListAlgorithms> , slist_node_ptr, node_ptr > type; }; template BOOST_INTRUSIVE_FORCEINLINE typename typeof_node_disposer::type make_node_disposer(const Disposer &disposer) const { typedef typename typeof_node_disposer::type return_t; return return_t(disposer, &this->priv_value_traits()); } static BOOST_INTRUSIVE_FORCEINLINE bucket_ptr to_ptr(bucket_type &b) { return pointer_traits::pointer_to(b); } static BOOST_INTRUSIVE_FORCEINLINE siterator sit_bbegin(bucket_type& b) { return siterator(b.get_node_ptr()); } static BOOST_INTRUSIVE_FORCEINLINE siterator sit_begin(bucket_type& b) { return siterator(b.begin_ptr()); } static BOOST_INTRUSIVE_FORCEINLINE siterator sit_end(bucket_type& b) { return siterator(slist_node_algorithms::end_node(b.get_node_ptr())); } BOOST_INTRUSIVE_FORCEINLINE static std::size_t priv_stored_hash(siterator s, detail::true_) //store_hash { return node_traits::get_hash(dcast_bucket_ptr(s.pointed_node())); } BOOST_INTRUSIVE_FORCEINLINE static std::size_t priv_stored_hash(siterator, detail::false_) //NO store_hash { return std::size_t(-1); } BOOST_INTRUSIVE_FORCEINLINE node &priv_value_to_node(reference v) { return *this->priv_value_traits().to_node_ptr(v); } BOOST_INTRUSIVE_FORCEINLINE const node &priv_value_to_node(const_reference v) const { return *this->priv_value_traits().to_node_ptr(v); } BOOST_INTRUSIVE_FORCEINLINE node_ptr priv_value_to_node_ptr(reference v) { return this->priv_value_traits().to_node_ptr(v); } BOOST_INTRUSIVE_FORCEINLINE const_node_ptr priv_value_to_node_ptr(const_reference v) const { return this->priv_value_traits().to_node_ptr(v); } BOOST_INTRUSIVE_FORCEINLINE reference priv_value_from_siterator(siterator s) { return *this->priv_value_traits().to_value_ptr(dcast_bucket_ptr(s.pointed_node())); } BOOST_INTRUSIVE_FORCEINLINE const_reference priv_value_from_siterator(siterator s) const { return *this->priv_value_traits().to_value_ptr(dcast_bucket_ptr(s.pointed_node())); } static void priv_init_buckets(const bucket_ptr buckets_ptr, const std::size_t bucket_cnt) { bucket_ptr buckets_it = buckets_ptr; for (std::size_t bucket_i = 0; bucket_i != bucket_cnt; ++buckets_it, ++bucket_i) { slist_node_algorithms::init_header(buckets_it->get_node_ptr()); } } void priv_clear_buckets(const bucket_ptr buckets_ptr, const std::size_t bucket_cnt) { bucket_ptr buckets_it = buckets_ptr; for(std::size_t bucket_i = 0; bucket_i != bucket_cnt; ++buckets_it, ++bucket_i){ bucket_type &b = *buckets_it; BOOST_IF_CONSTEXPR(safemode_or_autounlink){ slist_node_algorithms::detach_and_dispose(b.get_node_ptr(), this->make_node_disposer(detail::null_disposer())); } else{ slist_node_algorithms::init_header(b.get_node_ptr()); } } } BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_stored_or_compute_hash(const value_type &v, detail::true_) const //For store_hash == true { return node_traits::get_hash(this->priv_value_traits().to_node_ptr(v)); } typedef hashtable_iterator iterator; typedef hashtable_iterator const_iterator; BOOST_INTRUSIVE_FORCEINLINE iterator end() BOOST_NOEXCEPT { return this->build_iterator(this->priv_end_sit(), bucket_ptr()); } BOOST_INTRUSIVE_FORCEINLINE const_iterator end() const BOOST_NOEXCEPT { return this->cend(); } BOOST_INTRUSIVE_FORCEINLINE const_iterator cend() const BOOST_NOEXCEPT { return this->build_const_iterator(this->priv_end_sit(), bucket_ptr()); } BOOST_INTRUSIVE_FORCEINLINE iterator build_iterator(siterator s, bucket_ptr p) { return this->build_iterator(s, p, linear_buckets_t()); } BOOST_INTRUSIVE_FORCEINLINE iterator build_iterator(siterator s, bucket_ptr p, detail::true_) //linear buckets { return iterator(s, p, this->priv_value_traits_ptr()); } BOOST_INTRUSIVE_FORCEINLINE iterator build_iterator(siterator s, bucket_ptr, detail::false_) //!linear buckets { return iterator(s, &this->get_bucket_value_traits()); } BOOST_INTRUSIVE_FORCEINLINE const_iterator build_const_iterator(siterator s, bucket_ptr p) const { return this->build_const_iterator(s, p, linear_buckets_t()); } BOOST_INTRUSIVE_FORCEINLINE const_iterator build_const_iterator(siterator s, bucket_ptr p, detail::true_) const //linear buckets { return const_iterator(s, p, this->priv_value_traits_ptr()); } BOOST_INTRUSIVE_FORCEINLINE const_iterator build_const_iterator(siterator s, bucket_ptr, detail::false_) const //!linear buckets { return const_iterator(s, &this->get_bucket_value_traits()); } }; template struct get_hash { typedef Hash type; }; template struct get_hash { typedef ::boost::hash type; }; template struct get_equal_to { typedef EqualTo type; }; template struct get_equal_to { typedef value_equal type; }; template struct get_hash_key_of_value { typedef KeyOfValue type; }; template struct get_hash_key_of_value { typedef ::boost::intrusive::detail::identity type; }; template struct hash_key_types_base { typedef typename get_hash_key_of_value < VoidOrKeyOfValue, T>::type key_of_value; typedef typename key_of_value::type key_type; }; template struct hash_key_hash : get_hash < VoidOrKeyHash , typename hash_key_types_base::key_type > {}; template struct hash_key_equal : get_equal_to < VoidOrKeyEqual , typename hash_key_types_base::key_type > {}; //bucket_hash_t //Stores bucket_plus_vtraits plust the hash function template struct bucket_hash_t //Use public inheritance to avoid MSVC bugs with closures : public detail::ebo_functor_holder ::value_traits::value_type , VoidOrKeyOfValue , VoidOrKeyHash >::type > , bucket_plus_vtraits //4 { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(bucket_hash_t) public: typedef typename bucket_plus_vtraits ::value_traits value_traits; typedef typename value_traits::value_type value_type; typedef typename value_traits::node_traits node_traits; typedef hash_key_hash < value_type, VoidOrKeyOfValue, VoidOrKeyHash> hash_key_hash_t; typedef typename hash_key_hash_t::type hasher; typedef typename hash_key_types_base::key_of_value key_of_value; typedef BucketTraits bucket_traits; typedef bucket_plus_vtraits bucket_plus_vtraits_t; typedef detail::ebo_functor_holder base_t; BOOST_INTRUSIVE_FORCEINLINE bucket_hash_t(const ValueTraits &val_traits, const bucket_traits &b_traits, const hasher & h) : base_t(h) , bucket_plus_vtraits_t(val_traits, b_traits) {} BOOST_INTRUSIVE_FORCEINLINE bucket_hash_t(BOOST_RV_REF(bucket_hash_t) other) : base_t(BOOST_MOVE_BASE(base_t, other)) , bucket_plus_vtraits_t(BOOST_MOVE_BASE(bucket_plus_vtraits_t, other)) {} BOOST_INTRUSIVE_FORCEINLINE const hasher &priv_hasher() const { return this->base_t::get(); } BOOST_INTRUSIVE_FORCEINLINE hasher &priv_hasher() { return this->base_t::get(); } using bucket_plus_vtraits_t::priv_stored_or_compute_hash; //For store_hash == true BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_stored_or_compute_hash(const value_type &v, detail::false_) const //For store_hash == false { return this->priv_hasher()(key_of_value()(v)); } }; template struct hashtable_equal_holder { typedef detail::ebo_functor_holder < typename hash_key_equal < typename bucket_plus_vtraits ::value_traits::value_type , VoidOrKeyOfValue , VoidOrKeyEqual >::type > type; }; //bucket_hash_equal_t //Stores bucket_hash_t and the equality function when the first //non-empty bucket shall not be cached. template struct bucket_hash_equal_t //Use public inheritance to avoid MSVC bugs with closures : public bucket_hash_t //3 , public hashtable_equal_holder::type //equal { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(bucket_hash_equal_t) public: typedef typename hashtable_equal_holder < ValueTraits, BucketTraits, VoidOrKeyOfValue , VoidOrKeyEqual, LinearBuckets>::type equal_holder_t; typedef bucket_hash_t< ValueTraits, VoidOrKeyOfValue , VoidOrKeyHash, BucketTraits , LinearBuckets> bucket_hash_type; typedef bucket_plus_vtraits bucket_plus_vtraits_t; typedef ValueTraits value_traits; typedef typename equal_holder_t::functor_type key_equal; typedef typename bucket_hash_type::hasher hasher; typedef BucketTraits bucket_traits; typedef typename bucket_plus_vtraits_t::siterator siterator; typedef typename bucket_plus_vtraits_t::const_siterator const_siterator; typedef typename bucket_plus_vtraits_t::bucket_type bucket_type; typedef typename bucket_plus_vtraits_t::slist_node_algorithms slist_node_algorithms; typedef typename unordered_bucket_ptr_impl ::type bucket_ptr; bucket_hash_equal_t(const ValueTraits &val_traits, const bucket_traits &b_traits, const hasher & h, const key_equal &e) : bucket_hash_type(val_traits, b_traits, h) , equal_holder_t(e) {} BOOST_INTRUSIVE_FORCEINLINE bucket_hash_equal_t(BOOST_RV_REF(bucket_hash_equal_t) other) : bucket_hash_type(BOOST_MOVE_BASE(bucket_hash_type, other)) , equal_holder_t(BOOST_MOVE_BASE(equal_holder_t, other)) {} BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_get_cache() { return this->priv_bucket_pointer(); } BOOST_INTRUSIVE_FORCEINLINE void priv_set_cache(bucket_ptr) {} BOOST_INTRUSIVE_FORCEINLINE void priv_set_cache_bucket_num(std::size_t) {} BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_get_cache_bucket_num() { return 0u; } BOOST_INTRUSIVE_FORCEINLINE void priv_init_cache() {} BOOST_INTRUSIVE_FORCEINLINE void priv_swap_cache(bucket_hash_equal_t &) {} siterator priv_begin(bucket_ptr &pbucketptr) const { std::size_t n = 0; std::size_t bucket_cnt = this->priv_usable_bucket_count(); for (n = 0; n < bucket_cnt; ++n){ bucket_type &b = this->priv_bucket(n); if(!slist_node_algorithms::is_empty(b.get_node_ptr())){ pbucketptr = this->to_ptr(b); return siterator(b.begin_ptr()); } } pbucketptr = this->priv_invalid_bucket_ptr(); return this->priv_end_sit(); } BOOST_INTRUSIVE_FORCEINLINE void priv_insertion_update_cache(std::size_t) {} BOOST_INTRUSIVE_FORCEINLINE void priv_erasure_update_cache_range(std::size_t, std::size_t) {} BOOST_INTRUSIVE_FORCEINLINE void priv_erasure_update_cache(bucket_ptr) {} BOOST_INTRUSIVE_FORCEINLINE void priv_erasure_update_cache() {} BOOST_INTRUSIVE_FORCEINLINE const key_equal &priv_equal() const { return this->equal_holder_t::get(); } BOOST_INTRUSIVE_FORCEINLINE key_equal &priv_equal() { return this->equal_holder_t::get(); } }; //bucket_hash_equal_t //Stores bucket_hash_t and the equality function when the first //non-empty bucket shall be cached. template //cache_begin == true version struct bucket_hash_equal_t //Use public inheritance to avoid MSVC bugs with closures : public bucket_hash_t //2 , public hashtable_equal_holder::type { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(bucket_hash_equal_t) public: typedef typename hashtable_equal_holder < ValueTraits, BucketTraits , VoidOrKeyOfValue, VoidOrKeyEqual, LinearBuckets>::type equal_holder_t; typedef bucket_plus_vtraits < ValueTraits, BucketTraits, LinearBuckets> bucket_plus_vtraits_t; typedef ValueTraits value_traits; typedef typename equal_holder_t::functor_type key_equal; typedef bucket_hash_t < ValueTraits, VoidOrKeyOfValue , VoidOrKeyHash, BucketTraits, LinearBuckets> bucket_hash_type; typedef typename bucket_hash_type::hasher hasher; typedef BucketTraits bucket_traits; typedef typename bucket_plus_vtraits_t::siterator siterator; typedef typename bucket_plus_vtraits_t::slist_node_algorithms slist_node_algorithms; bucket_hash_equal_t(const ValueTraits &val_traits, const bucket_traits &b_traits, const hasher & h, const key_equal &e) : bucket_hash_type(val_traits, b_traits, h) , equal_holder_t(e) {} BOOST_INTRUSIVE_FORCEINLINE bucket_hash_equal_t(BOOST_RV_REF(bucket_hash_equal_t) other) : bucket_hash_type(BOOST_MOVE_BASE(bucket_hash_type, other)) , equal_holder_t(BOOST_MOVE_BASE(equal_holder_t, other)) {} typedef typename unordered_bucket_ptr_impl ::type bucket_ptr; BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_get_cache() const { return cached_begin_; } BOOST_INTRUSIVE_FORCEINLINE void priv_set_cache(bucket_ptr p) { cached_begin_ = p; } BOOST_INTRUSIVE_FORCEINLINE void priv_set_cache_bucket_num(std::size_t insertion_bucket) { BOOST_INTRUSIVE_INVARIANT_ASSERT(insertion_bucket <= this->priv_usable_bucket_count()); this->cached_begin_ = this->priv_bucket_pointer() + std::ptrdiff_t(insertion_bucket); } BOOST_INTRUSIVE_FORCEINLINE std::size_t priv_get_cache_bucket_num() { return std::size_t(this->cached_begin_ - this->priv_bucket_pointer()); } BOOST_INTRUSIVE_FORCEINLINE void priv_init_cache() { this->cached_begin_ = this->priv_past_usable_bucket_ptr(); } BOOST_INTRUSIVE_FORCEINLINE void priv_swap_cache(bucket_hash_equal_t &other) { ::boost::adl_move_swap(this->cached_begin_, other.cached_begin_); } siterator priv_begin(bucket_ptr& pbucketptr) const { pbucketptr = this->cached_begin_; if(this->cached_begin_ == this->priv_past_usable_bucket_ptr()){ return this->priv_end_sit(); } else{ return siterator(cached_begin_->begin_ptr()); } } void priv_insertion_update_cache(std::size_t insertion_bucket) { BOOST_INTRUSIVE_INVARIANT_ASSERT(insertion_bucket < this->priv_usable_bucket_count()); bucket_ptr p = this->priv_bucket_pointer() + std::ptrdiff_t(insertion_bucket); if(p < this->cached_begin_){ this->cached_begin_ = p; } } BOOST_INTRUSIVE_FORCEINLINE const key_equal &priv_equal() const { return this->equal_holder_t::get(); } BOOST_INTRUSIVE_FORCEINLINE key_equal &priv_equal() { return this->equal_holder_t::get(); } void priv_erasure_update_cache_range(std::size_t first_bucket_num, std::size_t last_bucket_num) { //If the last bucket is the end, the cache must be updated //to the last position if all if(this->priv_get_cache_bucket_num() == first_bucket_num && this->priv_bucket_empty(first_bucket_num) ){ this->priv_set_cache(this->priv_bucket_pointer() + std::ptrdiff_t(last_bucket_num)); this->priv_erasure_update_cache(); } } void priv_erasure_update_cache(bucket_ptr first_bucket) { //If the last bucket is the end, the cache must be updated //to the last position if all if (this->priv_get_cache() == first_bucket && this->priv_bucket_empty(first_bucket)) { this->priv_erasure_update_cache(); } } void priv_erasure_update_cache() { const bucket_ptr cache_end = this->priv_past_usable_bucket_ptr(); while( cached_begin_ != cache_end) { if (!slist_node_algorithms::is_empty(cached_begin_->get_node_ptr())) { return; } ++cached_begin_; } } bucket_ptr cached_begin_; }; //This wrapper around size_traits is used //to maintain minimal container size with compilers like MSVC //that have problems with EBO and multiple empty base classes template struct hashtable_size_wrapper : public DeriveFrom { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(hashtable_size_wrapper) public: template hashtable_size_wrapper( BOOST_FWD_REF(Base) base, BOOST_FWD_REF(Arg0) arg0 , BOOST_FWD_REF(Arg1) arg1, BOOST_FWD_REF(Arg2) arg2) : DeriveFrom( ::boost::forward(base) , ::boost::forward(arg0) , ::boost::forward(arg1) , ::boost::forward(arg2)) {} typedef detail::size_holder < true, SizeType> size_traits;//size_traits BOOST_INTRUSIVE_FORCEINLINE hashtable_size_wrapper(BOOST_RV_REF(hashtable_size_wrapper) other) : DeriveFrom(BOOST_MOVE_BASE(DeriveFrom, other)) {} size_traits size_traits_; typedef const size_traits & size_traits_const_t; typedef size_traits & size_traits_t; BOOST_INTRUSIVE_FORCEINLINE size_traits_const_t priv_size_traits() const { return size_traits_; } BOOST_INTRUSIVE_FORCEINLINE size_traits_t priv_size_traits() { return size_traits_; } }; template struct hashtable_size_wrapper : public DeriveFrom { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(hashtable_size_wrapper) public: template hashtable_size_wrapper( BOOST_FWD_REF(Base) base, BOOST_FWD_REF(Arg0) arg0 , BOOST_FWD_REF(Arg1) arg1, BOOST_FWD_REF(Arg2) arg2) : DeriveFrom( ::boost::forward(base) , ::boost::forward(arg0) , ::boost::forward(arg1) , ::boost::forward(arg2)) {} BOOST_INTRUSIVE_FORCEINLINE hashtable_size_wrapper(BOOST_RV_REF(hashtable_size_wrapper) other) : DeriveFrom(BOOST_MOVE_BASE(DeriveFrom, other)) {} typedef detail::size_holder< false, SizeType> size_traits; typedef size_traits size_traits_const_t; typedef size_traits size_traits_t; BOOST_INTRUSIVE_FORCEINLINE size_traits priv_size_traits() const { return size_traits(); } }; template< class ValueTraits, class VoidOrKeyOfValue, class VoidOrKeyHash , class VoidOrKeyEqual, class BucketTraits, class SizeType , std::size_t BoolFlags> struct get_hashtable_size_wrapper_bucket { typedef hashtable_size_wrapper < bucket_hash_equal_t < ValueTraits, VoidOrKeyOfValue, VoidOrKeyHash, VoidOrKeyEqual , BucketTraits , 0 != (BoolFlags & hash_bool_flags::linear_buckets_pos) , 0 != (BoolFlags & hash_bool_flags::cache_begin_pos) > //2 , SizeType , (BoolFlags & hash_bool_flags::incremental_pos) != 0 || (BoolFlags & hash_bool_flags::fastmod_buckets_pos) != 0 > type; }; //hashdata_internal //Stores bucket_hash_equal_t and split_traits template struct hashdata_internal : public get_hashtable_size_wrapper_bucket ::type { private: BOOST_MOVABLE_BUT_NOT_COPYABLE(hashdata_internal) public: static const bool linear_buckets = 0 != (BoolFlags & hash_bool_flags::linear_buckets_pos); typedef typename get_hashtable_size_wrapper_bucket ::type internal_type; typedef typename internal_type::key_equal key_equal; typedef typename internal_type::hasher hasher; typedef bucket_plus_vtraits bucket_plus_vtraits_t; typedef SizeType size_type; typedef typename internal_type::size_traits split_traits; typedef typename bucket_plus_vtraits_t::bucket_ptr bucket_ptr; typedef typename bucket_plus_vtraits_t::const_value_traits_ptr const_value_traits_ptr; typedef typename bucket_plus_vtraits_t::siterator siterator; typedef typename bucket_plus_vtraits_t::bucket_traits bucket_traits; typedef typename bucket_plus_vtraits_t::value_traits value_traits; typedef typename bucket_plus_vtraits_t::bucket_type bucket_type; typedef typename value_traits::value_type value_type; typedef typename value_traits::pointer pointer; typedef typename value_traits::const_pointer const_pointer; typedef typename pointer_traits::reference reference; typedef typename pointer_traits ::reference const_reference; typedef typename value_traits::node_traits node_traits; typedef typename node_traits::node node; typedef typename node_traits::node_ptr node_ptr; typedef typename node_traits::const_node_ptr const_node_ptr; typedef typename bucket_plus_vtraits_t::slist_node_algorithms slist_node_algorithms; typedef typename bucket_plus_vtraits_t::slist_node_ptr slist_node_ptr; typedef hash_key_types_base < typename ValueTraits::value_type , VoidOrKeyOfValue > hash_types_base; typedef typename hash_types_base::key_of_value key_of_value; static const bool store_hash = store_hash_is_true::value; static const bool safemode_or_autounlink = is_safe_autounlink::value; static const bool stateful_value_traits = detail::is_stateful_value_traits::value; typedef detail::bool_ store_hash_t; typedef detail::transform_iterator < siterator , downcast_node_to_value_t > local_iterator; typedef detail::transform_iterator < siterator , downcast_node_to_value_t > const_local_iterator; typedef detail::bool_ linear_buckets_t; hashdata_internal( const ValueTraits &val_traits, const bucket_traits &b_traits , const hasher & h, const key_equal &e) : internal_type(val_traits, b_traits, h, e) {} BOOST_INTRUSIVE_FORCEINLINE hashdata_internal(BOOST_RV_REF(hashdata_internal) other) : internal_type(BOOST_MOVE_BASE(internal_type, other)) {} BOOST_INTRUSIVE_FORCEINLINE typename internal_type::size_traits_t priv_split_traits() { return this->priv_size_traits(); } BOOST_INTRUSIVE_FORCEINLINE typename internal_type::size_traits_const_t priv_split_traits() const { return this->priv_size_traits(); } ~hashdata_internal() { this->priv_clear_buckets(); } using internal_type::priv_clear_buckets; void priv_clear_buckets() { const std::size_t cache_num = this->priv_get_cache_bucket_num(); this->priv_clear_buckets(this->priv_get_cache(), this->priv_usable_bucket_count() - cache_num); } void priv_clear_buckets_and_cache() { this->priv_clear_buckets(); this->priv_init_cache(); } void priv_init_buckets_and_cache() { this->priv_init_buckets(this->priv_bucket_pointer(), this->priv_usable_bucket_count()); this->priv_init_cache(); } typedef typename bucket_plus_vtraits_t::iterator iterator; typedef typename bucket_plus_vtraits_t::const_iterator const_iterator; //public functions BOOST_INTRUSIVE_FORCEINLINE SizeType split_count() const BOOST_NOEXCEPT { return this->priv_split_traits().get_size(); } BOOST_INTRUSIVE_FORCEINLINE static SizeType initial_split_from_bucket_count(SizeType bc) BOOST_NOEXCEPT { BOOST_IF_CONSTEXPR(fastmod_buckets) { size_type split; split = static_cast(prime_fmod_size::lower_size_index(bc)); //The passed bucket size must be exactly the supported one BOOST_ASSERT(prime_fmod_size::size(split) == bc); return split; } else { BOOST_IF_CONSTEXPR(incremental) { BOOST_ASSERT(0 == (std::size_t(bc) & (std::size_t(bc) - 1u))); return size_type(bc >> 1u); } else{ return bc; } } } BOOST_INTRUSIVE_FORCEINLINE static SizeType rehash_split_from_bucket_count(SizeType bc) BOOST_NOEXCEPT { BOOST_IF_CONSTEXPR(fastmod_buckets) { return (initial_split_from_bucket_count)(bc); } else { BOOST_IF_CONSTEXPR(incremental) { BOOST_ASSERT(0 == (std::size_t(bc) & (std::size_t(bc) - 1u))); return bc; } else{ return bc; } } } BOOST_INTRUSIVE_FORCEINLINE iterator iterator_to(reference value) BOOST_NOEXCEPT_IF(!linear_buckets) { return iterator_to(value, linear_buckets_t()); } const_iterator iterator_to(const_reference value) const BOOST_NOEXCEPT_IF(!linear_buckets) { return iterator_to(value, linear_buckets_t()); } iterator iterator_to(reference value, detail::true_) //linear_buckets { const std::size_t h = this->priv_stored_or_compute_hash(value, store_hash_t()); siterator sit(this->priv_value_to_node_ptr(value)); return this->build_iterator(sit, this->priv_hash_to_bucket_ptr(h)); } const_iterator iterator_to(const_reference value, detail::true_) const //linear_buckets { const std::size_t h = this->priv_stored_or_compute_hash(value, store_hash_t()); siterator const sit = siterator ( pointer_traits::const_cast_from(this->priv_value_to_node_ptr(value)) ); return this->build_const_iterator(sit, this->priv_hash_to_bucket_ptr(h)); } static const bool incremental = 0 != (BoolFlags & hash_bool_flags::incremental_pos); static const bool power_2_buckets = incremental || (0 != (BoolFlags & hash_bool_flags::power_2_buckets_pos)); static const bool fastmod_buckets = 0 != (BoolFlags & hash_bool_flags::fastmod_buckets_pos); typedef detail::bool_ fastmod_buckets_t; BOOST_INTRUSIVE_FORCEINLINE bucket_type &priv_hash_to_bucket(std::size_t hash_value) const { return this->priv_bucket(this->priv_hash_to_nbucket(hash_value)); } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_hash_to_bucket_ptr(std::size_t hash_value) const { return this->priv_bucket_ptr(this->priv_hash_to_nbucket(hash_value)); } BOOST_INTRUSIVE_FORCEINLINE size_type priv_hash_to_nbucket(std::size_t hash_value) const { return (priv_hash_to_nbucket)(hash_value, fastmod_buckets_t()); } BOOST_INTRUSIVE_FORCEINLINE size_type priv_hash_to_nbucket(std::size_t hash_value, detail::true_) const //fastmod_buckets_t { return static_cast(hash_to_bucket_split (hash_value, this->priv_usable_bucket_count(), this->split_count(), detail::true_())); } BOOST_INTRUSIVE_FORCEINLINE size_type priv_hash_to_nbucket(std::size_t hash_value, detail::false_) const //!fastmod_buckets_t { return static_cast(hash_to_bucket_split (hash_value, this->priv_usable_bucket_count(), this->split_count(), detail::false_())); } BOOST_INTRUSIVE_FORCEINLINE iterator iterator_to(reference value, detail::false_) BOOST_NOEXCEPT { return iterator( siterator(this->priv_value_to_node_ptr(value)) , &this->get_bucket_value_traits()); } const_iterator iterator_to(const_reference value, detail::false_) const BOOST_NOEXCEPT { siterator const sit = siterator ( pointer_traits::const_cast_from(this->priv_value_to_node_ptr(value)) ); return const_iterator(sit, &this->get_bucket_value_traits()); } static local_iterator s_local_iterator_to(reference value) BOOST_NOEXCEPT { BOOST_STATIC_ASSERT((!stateful_value_traits)); siterator sit(value_traits::to_node_ptr(value)); return local_iterator(sit, const_value_traits_ptr()); } static const_local_iterator s_local_iterator_to(const_reference value) BOOST_NOEXCEPT { BOOST_STATIC_ASSERT((!stateful_value_traits)); siterator const sit = siterator ( pointer_traits::const_cast_from (value_traits::to_node_ptr(value)) ); return const_local_iterator(sit, const_value_traits_ptr()); } local_iterator local_iterator_to(reference value) BOOST_NOEXCEPT { siterator sit(this->priv_value_to_node_ptr(value)); return local_iterator(sit, this->priv_value_traits_ptr()); } const_local_iterator local_iterator_to(const_reference value) const BOOST_NOEXCEPT { siterator sit ( pointer_traits::const_cast_from(this->priv_value_to_node_ptr(value)) ); return const_local_iterator(sit, this->priv_value_traits_ptr()); } BOOST_INTRUSIVE_FORCEINLINE size_type bucket_count() const BOOST_NOEXCEPT { return size_type(this->priv_usable_bucket_count()); } BOOST_INTRUSIVE_FORCEINLINE size_type bucket_size(size_type n) const BOOST_NOEXCEPT { return (size_type)this->priv_bucket_size(n); } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr bucket_pointer() const BOOST_NOEXCEPT { return this->priv_bucket_pointer(); } BOOST_INTRUSIVE_FORCEINLINE local_iterator begin(size_type n) BOOST_NOEXCEPT { return local_iterator(this->priv_bucket_lbegin(n), this->priv_value_traits_ptr()); } BOOST_INTRUSIVE_FORCEINLINE const_local_iterator begin(size_type n) const BOOST_NOEXCEPT { return this->cbegin(n); } static BOOST_INTRUSIVE_FORCEINLINE size_type suggested_upper_bucket_count(size_type n) BOOST_NOEXCEPT { BOOST_IF_CONSTEXPR(fastmod_buckets){ std::size_t s = prime_fmod_size::upper_size_index(n); return static_cast(prime_fmod_size::size(s)); } else{ return prime_list_holder<0>::suggested_upper_bucket_count(n); } } static BOOST_INTRUSIVE_FORCEINLINE size_type suggested_lower_bucket_count(size_type n) BOOST_NOEXCEPT { BOOST_IF_CONSTEXPR(fastmod_buckets){ std::size_t s = prime_fmod_size::lower_size_index(n); return static_cast(prime_fmod_size::size(s)); } else{ return prime_list_holder<0>::suggested_lower_bucket_count(n); } } const_local_iterator cbegin(size_type n) const BOOST_NOEXCEPT { return const_local_iterator (this->priv_bucket_lbegin(n) , this->priv_value_traits_ptr()); } using internal_type::end; using internal_type::cend; local_iterator end(size_type n) BOOST_NOEXCEPT { return local_iterator(this->priv_bucket_lend(n), this->priv_value_traits_ptr()); } BOOST_INTRUSIVE_FORCEINLINE const_local_iterator end(size_type n) const BOOST_NOEXCEPT { return this->cend(n); } const_local_iterator cend(size_type n) const BOOST_NOEXCEPT { return const_local_iterator ( this->priv_bucket_lend(n) , this->priv_value_traits_ptr()); } //Public functions for hashtable_impl BOOST_INTRUSIVE_FORCEINLINE iterator begin() BOOST_NOEXCEPT { bucket_ptr p; siterator s = this->priv_begin(p); return this->build_iterator(s, p); } BOOST_INTRUSIVE_FORCEINLINE const_iterator begin() const BOOST_NOEXCEPT { return this->cbegin(); } BOOST_INTRUSIVE_FORCEINLINE const_iterator cbegin() const BOOST_NOEXCEPT { bucket_ptr p; siterator s = this->priv_begin(p); return this->build_const_iterator(s, p); } BOOST_INTRUSIVE_FORCEINLINE hasher hash_function() const { return this->priv_hasher(); } BOOST_INTRUSIVE_FORCEINLINE key_equal key_eq() const { return this->priv_equal(); } }; template< class ValueTraits, class VoidOrKeyOfValue, class VoidOrKeyHash , class VoidOrKeyEqual, class BucketTraits, class SizeType , std::size_t BoolFlags> struct get_hashtable_size_wrapper_internal { typedef hashtable_size_wrapper < hashdata_internal < ValueTraits , VoidOrKeyOfValue, VoidOrKeyHash, VoidOrKeyEqual , BucketTraits, SizeType , BoolFlags & ~(hash_bool_flags::constant_time_size_pos) //1 > , SizeType , (BoolFlags& hash_bool_flags::constant_time_size_pos) != 0 > type; }; /// @endcond //! The class template hashtable is an intrusive hash table container, that //! is used to construct intrusive unordered_set and unordered_multiset containers. The //! no-throw guarantee holds only, if the VoidOrKeyEqual object and Hasher don't throw. //! //! hashtable is a semi-intrusive container: each object to be stored in the //! container must contain a proper hook, but the container also needs //! additional auxiliary memory to work: hashtable needs a pointer to an array //! of type `bucket_type` to be passed in the constructor. This bucket array must //! have at least the same lifetime as the container. This makes the use of //! hashtable more complicated than purely intrusive containers. //! `bucket_type` is default-constructible, copyable and assignable //! //! The template parameter \c T is the type to be managed by the container. //! The user can specify additional options and if no options are provided //! default options are used. //! //! The container supports the following options: //! \c base_hook<>/member_hook<>/value_traits<>, //! \c constant_time_size<>, \c size_type<>, \c hash<> and \c equal<> //! \c bucket_traits<>, power_2_buckets<>, cache_begin<> and incremental<>. //! //! hashtable only provides forward iterators but it provides 4 iterator types: //! iterator and const_iterator to navigate through the whole container and //! local_iterator and const_local_iterator to navigate through the values //! stored in a single bucket. Local iterators are faster and smaller. //! //! It's not recommended to use non constant-time size hashtables because several //! key functions, like "empty()", become non-constant time functions. Non //! constant_time size hashtables are mainly provided to support auto-unlink hooks. //! //! hashtables, does not make automatic rehashings nor //! offers functions related to a load factor. Rehashing can be explicitly requested //! and the user must provide a new bucket array that will be used from that moment. //! //! Since no automatic rehashing is done, iterators are never invalidated when //! inserting or erasing elements. Iterators are only invalidated when rehashing. #if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) template #else template #endif class hashtable_impl : private get_hashtable_size_wrapper_internal ::type { static const bool linear_buckets_flag = (BoolFlags & hash_bool_flags::linear_buckets_pos) != 0; typedef typename get_hashtable_size_wrapper_internal ::type internal_type; typedef typename internal_type::size_traits size_traits; typedef hash_key_types_base < typename ValueTraits::value_type , VoidOrKeyOfValue > hash_types_base; public: typedef ValueTraits value_traits; /// @cond typedef BucketTraits bucket_traits; typedef bucket_plus_vtraits bucket_plus_vtraits_t; typedef typename bucket_plus_vtraits_t::const_value_traits_ptr const_value_traits_ptr; typedef detail::bool_ linear_buckets_t; typedef typename internal_type::siterator siterator; typedef typename internal_type::const_siterator const_siterator; using internal_type::begin; using internal_type::cbegin; using internal_type::end; using internal_type::cend; using internal_type::hash_function; using internal_type::key_eq; using internal_type::bucket_size; using internal_type::bucket_count; using internal_type::local_iterator_to; using internal_type::s_local_iterator_to; using internal_type::iterator_to; using internal_type::bucket_pointer; using internal_type::suggested_upper_bucket_count; using internal_type::suggested_lower_bucket_count; using internal_type::split_count; /// @endcond typedef typename value_traits::pointer pointer; typedef typename value_traits::const_pointer const_pointer; typedef typename value_traits::value_type value_type; typedef typename hash_types_base::key_type key_type; typedef typename hash_types_base::key_of_value key_of_value; typedef typename pointer_traits::reference reference; typedef typename pointer_traits::reference const_reference; typedef typename pointer_traits::difference_type difference_type; typedef SizeType size_type; typedef typename internal_type::key_equal key_equal; typedef typename internal_type::hasher hasher; typedef typename internal_type::bucket_type bucket_type; typedef typename internal_type::bucket_ptr bucket_ptr; typedef typename internal_type::iterator iterator; typedef typename internal_type::const_iterator const_iterator; typedef typename internal_type::local_iterator local_iterator; typedef typename internal_type::const_local_iterator const_local_iterator; typedef typename value_traits::node_traits node_traits; typedef typename node_traits::node node; typedef typename pointer_traits ::template rebind_pointer < node >::type node_ptr; typedef typename pointer_traits ::template rebind_pointer < const node >::type const_node_ptr; typedef typename pointer_traits ::reference node_reference; typedef typename pointer_traits ::reference const_node_reference; typedef typename internal_type::slist_node_algorithms slist_node_algorithms; static const bool stateful_value_traits = internal_type::stateful_value_traits; static const bool store_hash = internal_type::store_hash; static const bool unique_keys = 0 != (BoolFlags & hash_bool_flags::unique_keys_pos); static const bool constant_time_size = 0 != (BoolFlags & hash_bool_flags::constant_time_size_pos); static const bool cache_begin = 0 != (BoolFlags & hash_bool_flags::cache_begin_pos); static const bool compare_hash = 0 != (BoolFlags & hash_bool_flags::compare_hash_pos); static const bool incremental = 0 != (BoolFlags & hash_bool_flags::incremental_pos); static const bool power_2_buckets = incremental || (0 != (BoolFlags & hash_bool_flags::power_2_buckets_pos)); static const bool optimize_multikey = optimize_multikey_is_true::value && !unique_keys; static const bool linear_buckets = linear_buckets_flag; static const bool fastmod_buckets = 0 != (BoolFlags & hash_bool_flags::fastmod_buckets_pos); static const std::size_t bucket_overhead = internal_type::bucket_overhead; /// @cond static const bool is_multikey = !unique_keys; private: //Configuration error: compare_hash<> can't be specified without store_hash<> //See documentation for more explanations BOOST_STATIC_ASSERT((!compare_hash || store_hash)); //Configuration error: fasmod_buckets<> can't be specified with incremental<> or power_2_buckets<> //See documentation for more explanations BOOST_STATIC_ASSERT(!(fastmod_buckets && power_2_buckets)); typedef typename internal_type::slist_node_ptr slist_node_ptr; typedef typename pointer_traits ::template rebind_pointer < void >::type void_pointer; //We'll define group traits, but these won't be instantiated if //optimize_multikey is not true typedef unordered_group_adapter group_traits; typedef circular_slist_algorithms group_algorithms; typedef typename internal_type::store_hash_t store_hash_t; typedef detail::bool_ optimize_multikey_t; typedef detail::bool_ cache_begin_t; typedef detail::bool_ power_2_buckets_t; typedef detail::bool_ fastmod_buckets_t; typedef detail::bool_ compare_hash_t; typedef typename internal_type::split_traits split_traits; typedef group_functions group_functions_t; typedef node_functions node_functions_t; private: //noncopyable, movable BOOST_MOVABLE_BUT_NOT_COPYABLE(hashtable_impl) static const bool safemode_or_autounlink = internal_type::safemode_or_autounlink; //Constant-time size is incompatible with auto-unlink hooks! BOOST_STATIC_ASSERT(!(constant_time_size && ((int)value_traits::link_mode == (int)auto_unlink))); //Cache begin is incompatible with auto-unlink hooks! BOOST_STATIC_ASSERT(!(cache_begin && ((int)value_traits::link_mode == (int)auto_unlink))); /// @endcond public: typedef insert_commit_data_impl insert_commit_data; void default_init_actions() { this->priv_set_sentinel_bucket(); this->priv_init_buckets_and_cache(); this->priv_size_traits().set_size(size_type(0)); size_type bucket_sz = this->bucket_count(); BOOST_INTRUSIVE_INVARIANT_ASSERT(bucket_sz != 0); //Check power of two bucket array if the option is activated BOOST_INTRUSIVE_INVARIANT_ASSERT (!power_2_buckets || (0 == (bucket_sz & (bucket_sz - 1)))); this->priv_split_traits().set_size(this->initial_split_from_bucket_count(bucket_sz)); } public: //! Requires: buckets must not be being used by any other resource. //! //! Effects: Constructs an empty unordered_set, storing a reference //! to the bucket array and copies of the key_hasher and equal_func functors. //! //! Complexity: Constant. //! //! Throws: If value_traits::node_traits::node //! constructor throws (this does not happen with predefined Boost.Intrusive hooks) //! or the copy constructor or invocation of hash_func or equal_func throws. //! //! Notes: buckets array must be disposed only after //! *this is disposed. explicit hashtable_impl ( const bucket_traits &b_traits , const hasher & hash_func = hasher() , const key_equal &equal_func = key_equal() , const value_traits &v_traits = value_traits()) : internal_type(v_traits, b_traits, hash_func, equal_func) { this->default_init_actions(); } //! Requires: buckets must not be being used by any other resource //! and dereferencing iterator must yield an lvalue of type value_type. //! //! Effects: Constructs an empty container and inserts elements from //! [b, e). //! //! Complexity: If N is distance(b, e): Average case is O(N) //! (with a good hash function and with buckets_len >= N),worst case O(N^2). //! //! Throws: If value_traits::node_traits::node //! constructor throws (this does not happen with predefined Boost.Intrusive hooks) //! or the copy constructor or invocation of hasher or key_equal throws. //! //! Notes: buckets array must be disposed only after //! *this is disposed. template hashtable_impl ( bool unique, Iterator b, Iterator e , const bucket_traits &b_traits , const hasher & hash_func = hasher() , const key_equal &equal_func = key_equal() , const value_traits &v_traits = value_traits()) : internal_type(v_traits, b_traits, hash_func, equal_func) { this->default_init_actions(); //Now insert if(unique) this->insert_unique(b, e); else this->insert_equal(b, e); } //! Effects: Constructs a container moving resources from another container. //! Internal value traits, bucket traits, hasher and comparison are move constructed and //! nodes belonging to x are linked to *this. //! //! Complexity: Constant. //! //! Throws: If value_traits::node_traits::node's //! move constructor throws (this does not happen with predefined Boost.Intrusive hooks) //! or the move constructor of value traits, bucket traits, hasher or comparison throws. hashtable_impl(BOOST_RV_REF(hashtable_impl) x) : internal_type(BOOST_MOVE_BASE(internal_type, x)) { this->priv_swap_cache(x); x.priv_init_cache(); this->priv_size_traits().set_size(x.priv_size_traits().get_size()); x.priv_size_traits().set_size(size_type(0)); this->priv_split_traits().set_size(x.split_count()); x.priv_split_traits().set_size(size_type(0)); } //! Effects: Equivalent to swap. //! hashtable_impl& operator=(BOOST_RV_REF(hashtable_impl) x) { this->swap(x); return *this; } //! Effects: Detaches all elements from this. The objects in the unordered_set //! are not deleted (i.e. no destructors are called). //! //! Complexity: Linear to the number of elements in the unordered_set, if //! it's a safe-mode or auto-unlink value. Otherwise constant. //! //! Throws: Nothing. ~hashtable_impl() {} #if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) //! Effects: Returns an iterator pointing to the beginning of the unordered_set. //! //! Complexity: Amortized constant time. //! Worst case (empty unordered_set): O(this->bucket_count()) //! //! Throws: Nothing. iterator begin() BOOST_NOEXCEPT; //! Effects: Returns a const_iterator pointing to the beginning //! of the unordered_set. //! //! Complexity: Amortized constant time. //! Worst case (empty unordered_set): O(this->bucket_count()) //! //! Throws: Nothing. const_iterator begin() const BOOST_NOEXCEPT; //! Effects: Returns a const_iterator pointing to the beginning //! of the unordered_set. //! //! Complexity: Amortized constant time. //! Worst case (empty unordered_set): O(this->bucket_count()) //! //! Throws: Nothing. const_iterator cbegin() const BOOST_NOEXCEPT; //! Effects: Returns an iterator pointing to the end of the unordered_set. //! //! Complexity: Constant. //! //! Throws: Nothing. iterator end() BOOST_NOEXCEPT; //! Effects: Returns a const_iterator pointing to the end of the unordered_set. //! //! Complexity: Constant. //! //! Throws: Nothing. const_iterator end() const BOOST_NOEXCEPT; //! Effects: Returns a const_iterator pointing to the end of the unordered_set. //! //! Complexity: Constant. //! //! Throws: Nothing. const_iterator cend() const BOOST_NOEXCEPT; //! Effects: Returns the hasher object used by the unordered_set. //! //! Complexity: Constant. //! //! Throws: If hasher copy-constructor throws. hasher hash_function() const; //! Effects: Returns the key_equal object used by the unordered_set. //! //! Complexity: Constant. //! //! Throws: If key_equal copy-constructor throws. key_equal key_eq() const; #endif //! Effects: Returns true if the container is empty. //! //! Complexity: if constant-time size and cache_begin options are disabled, //! average constant time (worst case, with empty() == true: O(this->bucket_count()). //! Otherwise constant. //! //! Throws: Nothing. bool empty() const BOOST_NOEXCEPT { BOOST_IF_CONSTEXPR(constant_time_size){ return !this->size(); } else if(cache_begin){ return this->begin() == this->end(); } else{ size_type bucket_cnt = this->bucket_count(); const bucket_type *b = boost::movelib::to_raw_pointer(this->priv_bucket_pointer()); for (size_type n = 0; n < bucket_cnt; ++n, ++b){ if(!slist_node_algorithms::is_empty(b->get_node_ptr())){ return false; } } return true; } } //! Effects: Returns the number of elements stored in the unordered_set. //! //! Complexity: Linear to elements contained in *this if //! constant_time_size is false. Constant-time otherwise. //! //! Throws: Nothing. size_type size() const BOOST_NOEXCEPT { BOOST_IF_CONSTEXPR(constant_time_size) return this->priv_size_traits().get_size(); else{ std::size_t len = 0; std::size_t bucket_cnt = this->bucket_count(); const bucket_type *b = boost::movelib::to_raw_pointer(this->priv_bucket_pointer()); for (std::size_t n = 0; n < bucket_cnt; ++n, ++b){ len += slist_node_algorithms::count(b->get_node_ptr()) - 1u; } BOOST_INTRUSIVE_INVARIANT_ASSERT((len <= SizeType(-1))); return size_type(len); } } //! Requires: the hasher and the equality function unqualified swap //! call should not throw. //! //! Effects: Swaps the contents of two unordered_sets. //! Swaps also the contained bucket array and equality and hasher functors. //! //! Complexity: Constant. //! //! Throws: If the swap() call for the comparison or hash functors //! found using ADL throw. Basic guarantee. void swap(hashtable_impl& other) { //These can throw ::boost::adl_move_swap(this->priv_equal(), other.priv_equal()); ::boost::adl_move_swap(this->priv_hasher(), other.priv_hasher()); //These can't throw ::boost::adl_move_swap(this->priv_bucket_traits(), other.priv_bucket_traits()); ::boost::adl_move_swap(this->priv_value_traits(), other.priv_value_traits()); this->priv_swap_cache(other); this->priv_size_traits().swap(other.priv_size_traits()); this->priv_split_traits().swap(other.priv_split_traits()); } //! Requires: Disposer::operator()(pointer) shouldn't throw //! Cloner should yield to nodes that compare equal and produce the same //! hash than the original node. //! //! Effects: Erases all the elements from *this //! calling Disposer::operator()(pointer), clones all the //! elements from src calling Cloner::operator()(const_reference ) //! and inserts them on *this. The hash function and the equality //! predicate are copied from the source. //! //! If store_hash option is true, this method does not use the hash function. //! //! If any operation throws, all cloned elements are unlinked and disposed //! calling Disposer::operator()(pointer). //! //! Complexity: Linear to erased plus inserted elements. //! //! Throws: If cloner or hasher throw or hash or equality predicate copying //! throws. Basic guarantee. template BOOST_INTRUSIVE_FORCEINLINE void clone_from(const hashtable_impl &src, Cloner cloner, Disposer disposer) { this->priv_clone_from(src, cloner, disposer); } //! Requires: Disposer::operator()(pointer) shouldn't throw //! Cloner should yield to nodes that compare equal and produce the same //! hash than the original node. //! //! Effects: Erases all the elements from *this //! calling Disposer::operator()(pointer), clones all the //! elements from src calling Cloner::operator()(reference) //! and inserts them on *this. The hash function and the equality //! predicate are copied from the source. //! //! If store_hash option is true, this method does not use the hash function. //! //! If any operation throws, all cloned elements are unlinked and disposed //! calling Disposer::operator()(pointer). //! //! Complexity: Linear to erased plus inserted elements. //! //! Throws: If cloner or hasher throw or hash or equality predicate copying //! throws. Basic guarantee. template BOOST_INTRUSIVE_FORCEINLINE void clone_from(BOOST_RV_REF(hashtable_impl) src, Cloner cloner, Disposer disposer) { this->priv_clone_from(static_cast(src), cloner, disposer); } //! Requires: value must be an lvalue //! //! Effects: Inserts the value into the unordered_set. //! //! Returns: An iterator to the inserted value. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. Strong guarantee. //! //! Note: Does not affect the validity of iterators and references. //! No copy-constructors are called. iterator insert_equal(reference value) { size_type bucket_num; std::size_t hash_value; siterator prev; siterator const it = this->priv_find (key_of_value()(value), this->priv_hasher(), this->priv_equal(), bucket_num, hash_value, prev); bool const next_is_in_group = optimize_multikey && it != this->priv_end_sit(); return this->priv_insert_equal_after_find(value, bucket_num, hash_value, prev, next_is_in_group); } //! Requires: Dereferencing iterator must yield an lvalue //! of type value_type. //! //! Effects: Equivalent to this->insert_equal(t) for each element in [b, e). //! //! Complexity: Average case O(N), where N is distance(b, e). //! Worst case O(N*this->size()). //! //! Throws: If the internal hasher or the equality functor throws. Basic guarantee. //! //! Note: Does not affect the validity of iterators and references. //! No copy-constructors are called. template void insert_equal(Iterator b, Iterator e) { for (; b != e; ++b) this->insert_equal(*b); } //! Requires: value must be an lvalue //! //! Effects: Tries to inserts value into the unordered_set. //! //! Returns: If the value //! is not already present inserts it and returns a pair containing the //! iterator to the new value and true. If there is an equivalent value //! returns a pair containing an iterator to the already present value //! and false. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. Strong guarantee. //! //! Note: Does not affect the validity of iterators and references. //! No copy-constructors are called. std::pair insert_unique(reference value) { insert_commit_data commit_data; std::pair ret = this->insert_unique_check(key_of_value()(value), commit_data); if(ret.second){ ret.first = this->insert_unique_commit(value, commit_data); } return ret; } //! Requires: Dereferencing iterator must yield an lvalue //! of type value_type. //! //! Effects: Equivalent to this->insert_unique(t) for each element in [b, e). //! //! Complexity: Average case O(N), where N is distance(b, e). //! Worst case O(N*this->size()). //! //! Throws: If the internal hasher or the equality functor throws. Basic guarantee. //! //! Note: Does not affect the validity of iterators and references. //! No copy-constructors are called. template void insert_unique(Iterator b, Iterator e) { for (; b != e; ++b) this->insert_unique(*b); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! "equal_func" must be a equality function that induces //! the same equality as key_equal. The difference is that //! "equal_func" compares an arbitrary key with the contained values. //! //! Effects: Checks if a value can be inserted in the unordered_set, using //! a user provided key instead of the value itself. //! //! Returns: If there is an equivalent value //! returns a pair containing an iterator to the already present value //! and false. If the value can be inserted returns true in the returned //! pair boolean and fills "commit_data" that is meant to be used with //! the "insert_commit" function. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If hash_func or equal_func throw. Strong guarantee. //! //! Notes: This function is used to improve performance when constructing //! a value_type is expensive: if there is an equivalent value //! the constructed object must be discarded. Many times, the part of the //! node that is used to impose the hash or the equality is much cheaper to //! construct than the value_type and this function offers the possibility to //! use that the part to check if the insertion will be successful. //! //! If the check is successful, the user can construct the value_type and use //! "insert_commit" to insert the object in constant-time. //! //! "commit_data" remains valid for a subsequent "insert_commit" only if no more //! objects are inserted or erased from the unordered_set. //! //! After a successful rehashing insert_commit_data remains valid. template std::pair insert_unique_check ( const KeyType &key , KeyHasher hash_func , KeyEqual equal_func , insert_commit_data &commit_data) { const std::size_t h = hash_func(key); const std::size_t bn = this->priv_hash_to_nbucket(h); commit_data.bucket_idx = bn; commit_data.set_hash(h); bucket_ptr bp = this->priv_bucket_ptr(bn); siterator const s = this->priv_find_in_bucket(*bp, key, equal_func, h); return std::pair(this->build_iterator(s, bp), s == this->priv_end_sit()); } //! Effects: Checks if a value can be inserted in the unordered_set, using //! a user provided key instead of the value itself. //! //! Returns: If there is an equivalent value //! returns a pair containing an iterator to the already present value //! and false. If the value can be inserted returns true in the returned //! pair boolean and fills "commit_data" that is meant to be used with //! the "insert_commit" function. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If hasher or key_compare throw. Strong guarantee. //! //! Notes: This function is used to improve performance when constructing //! a value_type is expensive: if there is an equivalent value //! the constructed object must be discarded. Many times, the part of the //! node that is used to impose the hash or the equality is much cheaper to //! construct than the value_type and this function offers the possibility to //! use that the part to check if the insertion will be successful. //! //! If the check is successful, the user can construct the value_type and use //! "insert_commit" to insert the object in constant-time. //! //! "commit_data" remains valid for a subsequent "insert_commit" only if no more //! objects are inserted or erased from the unordered_set. //! //! After a successful rehashing insert_commit_data remains valid. BOOST_INTRUSIVE_FORCEINLINE std::pair insert_unique_check ( const key_type &key, insert_commit_data &commit_data) { return this->insert_unique_check(key, this->priv_hasher(), this->priv_equal(), commit_data); } //! Requires: value must be an lvalue of type value_type. commit_data //! must have been obtained from a previous call to "insert_check". //! No objects should have been inserted or erased from the unordered_set between //! the "insert_check" that filled "commit_data" and the call to "insert_commit". //! //! Effects: Inserts the value in the unordered_set using the information obtained //! from the "commit_data" that a previous "insert_check" filled. //! //! Returns: An iterator to the newly inserted object. //! //! Complexity: Constant time. //! //! Throws: Nothing. //! //! Notes: This function has only sense if a "insert_check" has been //! previously executed to fill "commit_data". No value should be inserted or //! erased between the "insert_check" and "insert_commit" calls. //! //! After a successful rehashing insert_commit_data remains valid. iterator insert_unique_commit(reference value, const insert_commit_data &commit_data) BOOST_NOEXCEPT { this->priv_size_traits().increment(); node_ptr const n = this->priv_value_to_node_ptr(value); BOOST_INTRUSIVE_SAFE_HOOK_DEFAULT_ASSERT(!safemode_or_autounlink || slist_node_algorithms::unique(n)); node_functions_t::store_hash(n, commit_data.get_hash(), store_hash_t()); this->priv_insertion_update_cache(static_cast(commit_data.bucket_idx)); group_functions_t::insert_in_group(n, n, optimize_multikey_t()); bucket_type& b = this->priv_bucket(commit_data.bucket_idx); slist_node_algorithms::link_after(b.get_node_ptr(), n); return this->build_iterator(siterator(n), this->to_ptr(b)); } //! Effects: Erases the element pointed to by i. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: Nothing. //! //! Note: Invalidates the iterators (but not the references) //! to the erased element. No destructors are called. BOOST_INTRUSIVE_FORCEINLINE void erase(const_iterator i) BOOST_NOEXCEPT { this->erase_and_dispose(i, detail::null_disposer()); } //! Effects: Erases the range pointed to by b end e. //! //! Complexity: Average case O(distance(b, e)), //! worst case O(this->size()). //! //! Throws: Nothing. //! //! Note: Invalidates the iterators (but not the references) //! to the erased elements. No destructors are called. BOOST_INTRUSIVE_FORCEINLINE void erase(const_iterator b, const_iterator e) BOOST_NOEXCEPT { this->erase_and_dispose(b, e, detail::null_disposer()); } //! Effects: Erases all the elements with the given value. //! //! Returns: The number of erased elements. //! //! Complexity: Average case O(this->count(value)). //! Worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. //! Basic guarantee. //! //! Note: Invalidates the iterators (but not the references) //! to the erased elements. No destructors are called. BOOST_INTRUSIVE_FORCEINLINE size_type erase(const key_type &key) { return this->erase(key, this->priv_hasher(), this->priv_equal()); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! "equal_func" must be a equality function that induces //! the same equality as key_equal. The difference is that //! "equal_func" compares an arbitrary key with the contained values. //! //! Effects: Erases all the elements that have the same hash and //! compare equal with the given key. //! //! Returns: The number of erased elements. //! //! Complexity: Average case O(this->count(value)). //! Worst case O(this->size()). //! //! Throws: If hash_func or equal_func throw. Basic guarantee. //! //! Note: Invalidates the iterators (but not the references) //! to the erased elements. No destructors are called. template BOOST_INTRUSIVE_FORCEINLINE size_type erase(const KeyType& key, KeyHasher hash_func, KeyEqual equal_func) { return this->erase_and_dispose(key, hash_func, equal_func, detail::null_disposer()); } //! Requires: Disposer::operator()(pointer) shouldn't throw. //! //! Effects: Erases the element pointed to by i. //! Disposer::operator()(pointer) is called for the removed element. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: Nothing. //! //! Note: Invalidates the iterators //! to the erased elements. template BOOST_INTRUSIVE_DOC1ST(void , typename detail::disable_if_convertible::type) erase_and_dispose(const_iterator i, Disposer disposer) BOOST_NOEXCEPT { //Get the bucket number and local iterator for both iterators const bucket_ptr bp = this->priv_get_bucket_ptr(i); this->priv_erase_node(*bp, i.slist_it(), this->make_node_disposer(disposer), optimize_multikey_t()); this->priv_size_traits().decrement(); this->priv_erasure_update_cache(bp); } //! Requires: Disposer::operator()(pointer) shouldn't throw. //! //! Effects: Erases the range pointed to by b end e. //! Disposer::operator()(pointer) is called for the removed elements. //! //! Complexity: Average case O(distance(b, e)), //! worst case O(this->size()). //! //! Throws: Nothing. //! //! Note: Invalidates the iterators //! to the erased elements. template void erase_and_dispose(const_iterator b, const_iterator e, Disposer disposer) BOOST_NOEXCEPT { if(b != e){ //Get the bucket number and local iterator for both iterators size_type first_bucket_num = this->priv_get_bucket_num(b); siterator before_first_local_it = this->priv_get_previous(this->priv_bucket(first_bucket_num), b.slist_it(), optimize_multikey_t()); size_type last_bucket_num; siterator last_local_it; //For the end iterator, we will assign the end iterator //of the last bucket if(e == this->end()){ last_bucket_num = size_type(this->bucket_count() - 1u); last_local_it = this->sit_end(this->priv_bucket(last_bucket_num)); } else{ last_local_it = e.slist_it(); last_bucket_num = this->priv_get_bucket_num(e); } size_type const num_erased = (size_type)this->priv_erase_node_range ( before_first_local_it, first_bucket_num, last_local_it, last_bucket_num , this->make_node_disposer(disposer), optimize_multikey_t()); this->priv_size_traits().set_size(size_type(this->priv_size_traits().get_size()-num_erased)); this->priv_erasure_update_cache_range(first_bucket_num, last_bucket_num); } } //! Requires: Disposer::operator()(pointer) shouldn't throw. //! //! Effects: Erases all the elements with the given value. //! Disposer::operator()(pointer) is called for the removed elements. //! //! Returns: The number of erased elements. //! //! Complexity: Average case O(this->count(value)). //! Worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. //! Basic guarantee. //! //! Note: Invalidates the iterators (but not the references) //! to the erased elements. No destructors are called. template BOOST_INTRUSIVE_FORCEINLINE size_type erase_and_dispose(const key_type &key, Disposer disposer) { return this->erase_and_dispose(key, this->priv_hasher(), this->priv_equal(), disposer); } //! Requires: Disposer::operator()(pointer) shouldn't throw. //! //! Effects: Erases all the elements with the given key. //! according to the comparison functor "equal_func". //! Disposer::operator()(pointer) is called for the removed elements. //! //! Returns: The number of erased elements. //! //! Complexity: Average case O(this->count(value)). //! Worst case O(this->size()). //! //! Throws: If hash_func or equal_func throw. Basic guarantee. //! //! Note: Invalidates the iterators //! to the erased elements. template size_type erase_and_dispose(const KeyType& key, KeyHasher hash_func ,KeyEqual equal_func, Disposer disposer) { size_type bucket_num; std::size_t h; siterator prev; siterator it = this->priv_find(key, hash_func, equal_func, bucket_num, h, prev); bool const success = it != this->priv_end_sit(); std::size_t cnt(0); if(success){ if(optimize_multikey){ siterator past_last_in_group = it; (priv_go_to_last_in_group)(past_last_in_group, optimize_multikey_t()); ++past_last_in_group; cnt = this->priv_erase_from_single_bucket ( this->priv_bucket(bucket_num), prev , past_last_in_group , this->make_node_disposer(disposer), optimize_multikey_t()); } else{ siterator const end_sit = this->priv_bucket_lend(bucket_num); do{ ++cnt; ++it; }while(it != end_sit && this->priv_is_value_equal_to_key (this->priv_value_from_siterator(it), h, key, equal_func, compare_hash_t())); slist_node_algorithms::unlink_after_and_dispose(prev.pointed_node(), it.pointed_node(), this->make_node_disposer(disposer)); } this->priv_size_traits().set_size(size_type(this->priv_size_traits().get_size()-cnt)); this->priv_erasure_update_cache(); } return static_cast(cnt); } //! Effects: Erases all of the elements. //! //! Complexity: Linear to the number of elements on the container. //! if it's a safe-mode or auto-unlink value_type. Constant time otherwise. //! //! Throws: Nothing. //! //! Note: Invalidates the iterators (but not the references) //! to the erased elements. No destructors are called. void clear() BOOST_NOEXCEPT { this->priv_clear_buckets_and_cache(); this->priv_size_traits().set_size(size_type(0)); } //! Requires: Disposer::operator()(pointer) shouldn't throw. //! //! Effects: Erases all of the elements. //! //! Complexity: Linear to the number of elements on the container. //! Disposer::operator()(pointer) is called for the removed elements. //! //! Throws: Nothing. //! //! Note: Invalidates the iterators (but not the references) //! to the erased elements. No destructors are called. template void clear_and_dispose(Disposer disposer) BOOST_NOEXCEPT { if(!constant_time_size || !this->empty()){ size_type num_buckets = this->bucket_count(); bucket_ptr b = this->priv_bucket_pointer(); typename internal_type::template typeof_node_disposer::type d(disposer, &this->priv_value_traits()); for(; num_buckets; ++b){ --num_buckets; slist_node_algorithms::detach_and_dispose(b->get_node_ptr(), d); } this->priv_size_traits().set_size(size_type(0)); } this->priv_init_cache(); } //! Effects: Returns the number of contained elements with the given value //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. BOOST_INTRUSIVE_FORCEINLINE size_type count(const key_type &key) const { return this->count(key, this->priv_hasher(), this->priv_equal()); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! "equal_func" must be a equality function that induces //! the same equality as key_equal. The difference is that //! "equal_func" compares an arbitrary key with the contained values. //! //! Effects: Returns the number of contained elements with the given key //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If hash_func or equal throw. template size_type count(const KeyType &key, KeyHasher hash_func, KeyEqual equal_func) const { size_type cnt; size_type n_bucket; this->priv_local_equal_range(key, hash_func, equal_func, n_bucket, cnt); return cnt; } //! Effects: Finds an iterator to the first element is equal to //! "value" or end() if that element does not exist. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. BOOST_INTRUSIVE_FORCEINLINE iterator find(const key_type &key) { return this->find(key, this->priv_hasher(), this->priv_equal()); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! "equal_func" must be a equality function that induces //! the same equality as key_equal. The difference is that //! "equal_func" compares an arbitrary key with the contained values. //! //! Effects: Finds an iterator to the first element whose key is //! "key" according to the given hash and equality functor or end() if //! that element does not exist. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If hash_func or equal_func throw. //! //! Note: This function is used when constructing a value_type //! is expensive and the value_type can be compared with a cheaper //! key type. Usually this key is part of the value_type. template iterator find(const KeyType &key, KeyHasher hash_func, KeyEqual equal_func) { std::size_t h = hash_func(key); bucket_ptr bp = this->priv_hash_to_bucket_ptr(h); siterator s = this->priv_find_in_bucket(*bp, key, equal_func, h); return this->build_iterator(s, bp); } //! Effects: Finds a const_iterator to the first element whose key is //! "key" or end() if that element does not exist. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. BOOST_INTRUSIVE_FORCEINLINE const_iterator find(const key_type &key) const { return this->find(key, this->priv_hasher(), this->priv_equal()); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! "equal_func" must be a equality function that induces //! the same equality as key_equal. The difference is that //! "equal_func" compares an arbitrary key with the contained values. //! //! Effects: Finds an iterator to the first element whose key is //! "key" according to the given hasher and equality functor or end() if //! that element does not exist. //! //! Complexity: Average case O(1), worst case O(this->size()). //! //! Throws: If hash_func or equal_func throw. //! //! Note: This function is used when constructing a value_type //! is expensive and the value_type can be compared with a cheaper //! key type. Usually this key is part of the value_type. template const_iterator find (const KeyType &key, KeyHasher hash_func, KeyEqual equal_func) const { std::size_t h = hash_func(key); bucket_ptr bp = this->priv_hash_to_bucket_ptr(h); siterator s = this->priv_find_in_bucket(*bp, key, equal_func, h); return this->build_const_iterator(s, bp); } //! Effects: Returns a range containing all elements with values equivalent //! to value. Returns std::make_pair(this->end(), this->end()) if no such //! elements exist. //! //! Complexity: Average case O(this->count(value)). Worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. BOOST_INTRUSIVE_FORCEINLINE std::pair equal_range(const key_type &key) { return this->equal_range(key, this->priv_hasher(), this->priv_equal()); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! "equal_func" must be a equality function that induces //! the same equality as key_equal. The difference is that //! "equal_func" compares an arbitrary key with the contained values. //! //! Effects: Returns a range containing all elements with equivalent //! keys. Returns std::make_pair(this->end(), this->end()) if no such //! elements exist. //! //! Complexity: Average case O(this->count(key, hash_func, equal_func)). //! Worst case O(this->size()). //! //! Throws: If hash_func or the equal_func throw. //! //! Note: This function is used when constructing a value_type //! is expensive and the value_type can be compared with a cheaper //! key type. Usually this key is part of the value_type. template std::pair equal_range (const KeyType &key, KeyHasher hash_func, KeyEqual equal_func) { priv_equal_range_result ret = this->priv_equal_range(key, hash_func, equal_func); return std::pair ( this->build_iterator(ret.first, ret.bucket_first) , this->build_iterator(ret.second, ret.bucket_second)); } //! Effects: Returns a range containing all elements with values equivalent //! to value. Returns std::make_pair(this->end(), this->end()) if no such //! elements exist. //! //! Complexity: Average case O(this->count(value)). Worst case O(this->size()). //! //! Throws: If the internal hasher or the equality functor throws. BOOST_INTRUSIVE_FORCEINLINE std::pair equal_range(const key_type &key) const { return this->equal_range(key, this->priv_hasher(), this->priv_equal()); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! "equal_func" must be a equality function that induces //! the same equality as key_equal. The difference is that //! "equal_func" compares an arbitrary key with the contained values. //! //! Effects: Returns a range containing all elements with equivalent //! keys. Returns std::make_pair(this->end(), this->end()) if no such //! elements exist. //! //! Complexity: Average case O(this->count(key, hash_func, equal_func)). //! Worst case O(this->size()). //! //! Throws: If the hasher or equal_func throw. //! //! Note: This function is used when constructing a value_type //! is expensive and the value_type can be compared with a cheaper //! key type. Usually this key is part of the value_type. template std::pair equal_range (const KeyType &key, KeyHasher hash_func, KeyEqual equal_func) const { priv_equal_range_result ret = this->priv_equal_range(key, hash_func, equal_func); return std::pair ( this->build_const_iterator(ret.first, ret.bucket_first) , this->build_const_iterator(ret.second, ret.bucket_second)); } #if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) //! Requires: value must be an lvalue and shall be in a unordered_set of //! appropriate type. Otherwise the behavior is undefined. //! //! Effects: Returns: a valid iterator belonging to the unordered_set //! that points to the value //! //! Complexity: Constant. //! //! Throws: If the internal hash function throws. iterator iterator_to(reference value) BOOST_NOEXCEPT; //! Requires: value must be an lvalue and shall be in a unordered_set of //! appropriate type. Otherwise the behavior is undefined. //! //! Effects: Returns: a valid const_iterator belonging to the //! unordered_set that points to the value //! //! Complexity: Constant. //! //! Throws: If the internal hash function throws. const_iterator iterator_to(const_reference value) const BOOST_NOEXCEPT; //! Requires: value must be an lvalue and shall be in a unordered_set of //! appropriate type. Otherwise the behavior is undefined. //! //! Effects: Returns: a valid local_iterator belonging to the unordered_set //! that points to the value //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: This static function is available only if the value traits //! is stateless. static local_iterator s_local_iterator_to(reference value) BOOST_NOEXCEPT; //! Requires: value must be an lvalue and shall be in a unordered_set of //! appropriate type. Otherwise the behavior is undefined. //! //! Effects: Returns: a valid const_local_iterator belonging to //! the unordered_set that points to the value //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: This static function is available only if the value traits //! is stateless. static const_local_iterator s_local_iterator_to(const_reference value) BOOST_NOEXCEPT; //! Requires: value must be an lvalue and shall be in a unordered_set of //! appropriate type. Otherwise the behavior is undefined. //! //! Effects: Returns: a valid local_iterator belonging to the unordered_set //! that points to the value //! //! Complexity: Constant. //! //! Throws: Nothing. local_iterator local_iterator_to(reference value) BOOST_NOEXCEPT; //! Requires: value must be an lvalue and shall be in a unordered_set of //! appropriate type. Otherwise the behavior is undefined. //! //! Effects: Returns: a valid const_local_iterator belonging to //! the unordered_set that points to the value //! //! Complexity: Constant. //! //! Throws: Nothing. const_local_iterator local_iterator_to(const_reference value) const BOOST_NOEXCEPT; //! Effects: Returns the number of buckets passed in the constructor //! or the last rehash function. //! //! Complexity: Constant. //! //! Throws: Nothing. size_type bucket_count() const BOOST_NOEXCEPT; //! Requires: n is in the range [0, this->bucket_count()). //! //! Effects: Returns the number of elements in the nth bucket. //! //! Complexity: Constant. //! //! Throws: Nothing. size_type bucket_size(size_type n) const BOOST_NOEXCEPT; #endif //#if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) //! Effects: Returns the index of the bucket in which elements //! with keys equivalent to k would be found, if any such element existed. //! //! Complexity: Constant. //! //! Throws: If the hash functor throws. //! //! Note: the return value is in the range [0, this->bucket_count()). BOOST_INTRUSIVE_FORCEINLINE size_type bucket(const key_type& k) const { return this->bucket(k, this->priv_hasher()); } //! Requires: "hash_func" must be a hash function that induces //! the same hash values as the stored hasher. The difference is that //! "hash_func" hashes the given key instead of the value_type. //! //! Effects: Returns the index of the bucket in which elements //! with keys equivalent to k would be found, if any such element existed. //! //! Complexity: Constant. //! //! Throws: If hash_func throws. //! //! Note: the return value is in the range [0, this->bucket_count()). template BOOST_INTRUSIVE_FORCEINLINE size_type bucket(const KeyType& k, KeyHasher hash_func) const { return this->priv_hash_to_nbucket(hash_func(k)); } #if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) //! Effects: Returns the bucket array pointer passed in the constructor //! or the last rehash function. //! //! Complexity: Constant. //! //! Throws: Nothing. bucket_ptr bucket_pointer() const BOOST_NOEXCEPT; //! Requires: n is in the range [0, this->bucket_count()). //! //! Effects: Returns a local_iterator pointing to the beginning //! of the sequence stored in the bucket n. //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: [this->begin(n), this->end(n)) is a valid range //! containing all of the elements in the nth bucket. local_iterator begin(size_type n) BOOST_NOEXCEPT; //! Requires: n is in the range [0, this->bucket_count()). //! //! Effects: Returns a const_local_iterator pointing to the beginning //! of the sequence stored in the bucket n. //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: [this->begin(n), this->end(n)) is a valid range //! containing all of the elements in the nth bucket. const_local_iterator begin(size_type n) const BOOST_NOEXCEPT; //! Requires: n is in the range [0, this->bucket_count()). //! //! Effects: Returns a const_local_iterator pointing to the beginning //! of the sequence stored in the bucket n. //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: [this->begin(n), this->end(n)) is a valid range //! containing all of the elements in the nth bucket. const_local_iterator cbegin(size_type n) const BOOST_NOEXCEPT; //! Requires: n is in the range [0, this->bucket_count()). //! //! Effects: Returns a local_iterator pointing to the end //! of the sequence stored in the bucket n. //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: [this->begin(n), this->end(n)) is a valid range //! containing all of the elements in the nth bucket. local_iterator end(size_type n) BOOST_NOEXCEPT; //! Requires: n is in the range [0, this->bucket_count()). //! //! Effects: Returns a const_local_iterator pointing to the end //! of the sequence stored in the bucket n. //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: [this->begin(n), this->end(n)) is a valid range //! containing all of the elements in the nth bucket. const_local_iterator end(size_type n) const BOOST_NOEXCEPT; //! Requires: n is in the range [0, this->bucket_count()). //! //! Effects: Returns a const_local_iterator pointing to the end //! of the sequence stored in the bucket n. //! //! Complexity: Constant. //! //! Throws: Nothing. //! //! Note: [this->begin(n), this->end(n)) is a valid range //! containing all of the elements in the nth bucket. const_local_iterator cend(size_type n) const BOOST_NOEXCEPT; #endif //#if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) //! Requires: new_bucket_traits can hold a pointer to a new bucket array //! or the same as the old bucket array with a different length. new_size is the length of the //! the array pointed by new_buckets. If new_bucket_traits.bucket_begin() == this->bucket_pointer() //! new_bucket_traits.bucket_count() can be bigger or smaller than this->bucket_count(). //! 'new_bucket_traits' copy constructor should not throw. //! //! Effects: //! If `new_bucket_traits.bucket_begin() == this->bucket_pointer()` is false, //! unlinks values from the old bucket and inserts then in the new one according //! to the hash value of values. //! //! If `new_bucket_traits.bucket_begin() == this->bucket_pointer()` is true, //! the implementations avoids moving values as much as possible. //! //! Bucket traits hold by *this is assigned from new_bucket_traits. //! If the container is configured as incremental<>, the split bucket is set //! to the new bucket_count(). //! //! If store_hash option is true, this method does not use the hash function. //! If false, the implementation tries to minimize calls to the hash function //! (e.g. once for equivalent values if optimize_multikey is true). //! //! If rehash is successful updates the internal bucket_traits with new_bucket_traits. //! //! Complexity: Average case linear in this->size(), worst case quadratic. //! //! Throws: If the hasher functor throws. Basic guarantee. BOOST_INTRUSIVE_FORCEINLINE void rehash(const bucket_traits &new_bucket_traits) { this->priv_rehash_impl(new_bucket_traits, false); } //! Note: This function is used when keys from inserted elements are changed //! (e.g. a language change when key is a string) but uniqueness and hash properties are //! preserved so a fast full rehash recovers invariants for *this without extracting and //! reinserting all elements again. //! //! Requires: Calls produced to the hash function should not alter the value uniqueness //! properties of already inserted elements. If hasher(key1) == hasher(key2) was true when //! elements were inserted, it shall be true during calls produced in the execution of this function. //! //! key_equal is not called inside this function so it is assumed that key_equal(value1, value2) //! should produce the same results as before for inserted elements. //! //! Effects: Reprocesses all values hold by *this, recalculating their hash values //! and redistributing them though the buckets. //! //! If store_hash option is true, this method uses the hash function and updates the stored hash value. //! //! Complexity: Average case linear in this->size(), worst case quadratic. //! //! Throws: If the hasher functor throws. Basic guarantee. BOOST_INTRUSIVE_FORCEINLINE void full_rehash() { this->priv_rehash_impl(this->priv_bucket_traits(), true); } //! Requires: //! //! Effects: //! //! Complexity: //! //! Throws: //! //! Note: this method is only available if incremental option is activated. bool incremental_rehash(bool grow = true) { //This function is only available for containers with incremental hashing BOOST_STATIC_ASSERT(( incremental && power_2_buckets )); const std::size_t split_idx = this->split_count(); const std::size_t bucket_cnt = this->bucket_count(); bool ret = false; if(grow){ //Test if the split variable can be changed if((ret = split_idx < bucket_cnt)){ const std::size_t bucket_to_rehash = split_idx - bucket_cnt/2u; bucket_type &old_bucket = this->priv_bucket(bucket_to_rehash); this->priv_split_traits().increment(); //Anti-exception stuff: if an exception is thrown while //moving elements from old_bucket to the target bucket, all moved //elements are moved back to the original one. incremental_rehash_rollback rollback ( this->priv_bucket(split_idx), old_bucket, this->priv_split_traits()); siterator before_i(old_bucket.get_node_ptr()); siterator i(before_i); ++i; siterator end_sit = linear_buckets ? siterator() : before_i; for( ; i != end_sit; i = before_i, ++i){ const value_type &v = this->priv_value_from_siterator(i); const std::size_t hash_value = this->priv_stored_or_compute_hash(v, store_hash_t()); const std::size_t new_n = this->priv_hash_to_nbucket(hash_value); siterator last = i; (priv_go_to_last_in_group)(last, optimize_multikey_t()); if(new_n == bucket_to_rehash){ before_i = last; } else{ bucket_type &new_b = this->priv_bucket(new_n); slist_node_algorithms::transfer_after(new_b.get_node_ptr(), before_i.pointed_node(), last.pointed_node()); } } rollback.release(); this->priv_erasure_update_cache(); } } else if((ret = split_idx > bucket_cnt/2u)){ //!grow const std::size_t target_bucket_num = split_idx - 1u - bucket_cnt/2u; bucket_type &target_bucket = this->priv_bucket(target_bucket_num); bucket_type &source_bucket = this->priv_bucket(split_idx-1u); slist_node_algorithms::transfer_after(target_bucket.get_node_ptr(), source_bucket.get_node_ptr()); this->priv_split_traits().decrement(); this->priv_insertion_update_cache(target_bucket_num); } return ret; } //! Effects: If new_bucket_traits.bucket_count() is not //! this->bucket_count()/2 or this->bucket_count()*2, or //! this->split_bucket() != new_bucket_traits.bucket_count() returns false //! and does nothing. //! //! Otherwise, copy assigns new_bucket_traits to the internal bucket_traits //! and transfers all the objects from old buckets to the new ones. //! //! Complexity: Linear to size(). //! //! Throws: Nothing //! //! Note: this method is only available if incremental option is activated. bool incremental_rehash(const bucket_traits &new_bucket_traits) BOOST_NOEXCEPT { //This function is only available for containers with incremental hashing BOOST_STATIC_ASSERT(( incremental && power_2_buckets )); const bucket_ptr new_buckets = new_bucket_traits.bucket_begin(); const size_type new_bucket_count_stdszt = static_cast(new_bucket_traits.bucket_count() - bucket_overhead); BOOST_INTRUSIVE_INVARIANT_ASSERT(sizeof(size_type) >= sizeof(std::size_t) || new_bucket_count_stdszt <= size_type(-1)); size_type new_bucket_count = static_cast(new_bucket_count_stdszt); const size_type old_bucket_count = static_cast(this->priv_usable_bucket_count()); const size_type split_idx = this->split_count(); //Test new bucket size is consistent with internal bucket size and split count if(new_bucket_count/2 == old_bucket_count){ if(!(split_idx >= old_bucket_count)) return false; } else if(new_bucket_count == old_bucket_count/2){ if(!(split_idx <= new_bucket_count)) return false; } else{ return false; } const size_type ini_n = (size_type)this->priv_get_cache_bucket_num(); const bucket_ptr old_buckets = this->priv_bucket_pointer(); this->priv_unset_sentinel_bucket(); this->priv_initialize_new_buckets(old_buckets, old_bucket_count, new_buckets, new_bucket_count); if (&new_bucket_traits != &this->priv_bucket_traits()) this->priv_bucket_traits() = new_bucket_traits; if(old_buckets != new_buckets){ for(size_type n = ini_n; n < split_idx; ++n){ slist_node_ptr new_bucket_nodeptr = new_bucket_traits.bucket_begin()[difference_type(n)].get_node_ptr(); slist_node_ptr old_bucket_node_ptr = old_buckets[difference_type(n)].get_node_ptr(); slist_node_algorithms::transfer_after(new_bucket_nodeptr, old_bucket_node_ptr); } //Reset cache to safe position this->priv_set_cache_bucket_num(ini_n); } this->priv_set_sentinel_bucket(); return true; } #if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) //! Requires: incremental<> option must be set //! //! Effects: returns the current split count //! //! Complexity: Constant //! //! Throws: Nothing size_type split_count() const BOOST_NOEXCEPT; //! Effects: Returns the nearest new bucket count optimized for //! the container that is bigger or equal than n. This suggestion can be //! used to create bucket arrays with a size that will usually improve //! container's performance. If such value does not exist, the //! higher possible value is returned. //! //! Complexity: Amortized constant time. //! //! Throws: Nothing. static size_type suggested_upper_bucket_count(size_type n) BOOST_NOEXCEPT; //! Effects: Returns the nearest new bucket count optimized for //! the container that is smaller or equal than n. This suggestion can be //! used to create bucket arrays with a size that will usually improve //! container's performance. If such value does not exist, the //! lowest possible value is returned. //! //! Complexity: Amortized constant time. //! //! Throws: Nothing. static size_type suggested_lower_bucket_count(size_type n) BOOST_NOEXCEPT; #endif //#if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) friend bool operator==(const hashtable_impl &x, const hashtable_impl &y) { //Taken from N3068 if(constant_time_size && x.size() != y.size()){ return false; } if (boost::intrusive::iterator_udistance(x.begin(), x.end()) != x.size()) return false; for (const_iterator ix = x.cbegin(), ex = x.cend(); ix != ex; ++ix){ std::pair eqx(x.equal_range(key_of_value()(*ix))), eqy(y.equal_range(key_of_value()(*ix))); if (boost::intrusive::iterator_distance(eqx.first, eqx.second) != boost::intrusive::iterator_distance(eqy.first, eqy.second) || !(priv_algo_is_permutation)(eqx.first, eqx.second, eqy.first) ){ return false; } ix = eqx.second; } return true; } friend bool operator!=(const hashtable_impl &x, const hashtable_impl &y) { return !(x == y); } friend bool operator<(const hashtable_impl &x, const hashtable_impl &y) { return ::boost::intrusive::algo_lexicographical_compare(x.begin(), x.end(), y.begin(), y.end()); } friend bool operator>(const hashtable_impl &x, const hashtable_impl &y) { return y < x; } friend bool operator<=(const hashtable_impl &x, const hashtable_impl &y) { return !(y < x); } friend bool operator>=(const hashtable_impl &x, const hashtable_impl &y) { return !(x < y); } /// @cond BOOST_INTRUSIVE_FORCEINLINE void check() const {} private: static void priv_initialize_new_buckets ( bucket_ptr old_buckets, size_type old_bucket_count , bucket_ptr new_buckets, size_type new_bucket_count) { //Initialize new buckets const bool same_buffer = old_buckets == new_buckets; if (same_buffer && new_bucket_count <= old_bucket_count) { //Nothing to do here } else { bucket_ptr p; size_type c; if (same_buffer) { p = old_buckets + std::ptrdiff_t(old_bucket_count); c = size_type(new_bucket_count - old_bucket_count); } else { p = new_buckets; c = new_bucket_count; } internal_type::priv_init_buckets(p, c); } } void priv_rehash_impl(const bucket_traits &new_bucket_traits, bool do_full_rehash) { const std::size_t nbc = new_bucket_traits.bucket_count() - bucket_overhead; BOOST_INTRUSIVE_INVARIANT_ASSERT(sizeof(SizeType) >= sizeof(std::size_t) || nbc <= SizeType(-1)); const bucket_ptr new_buckets = new_bucket_traits.bucket_begin(); const size_type new_bucket_count = static_cast(nbc); const bucket_ptr old_buckets = this->priv_bucket_pointer(); const size_type old_bucket_count = this->bucket_count(); //Check power of two bucket array if the option is activated BOOST_INTRUSIVE_INVARIANT_ASSERT (!power_2_buckets || (0 == (new_bucket_count & (new_bucket_count-1u)))); const bool same_buffer = old_buckets == new_buckets; //If the new bucket length is a common factor //of the old one we can avoid hash calculations. const bool fast_shrink = (!do_full_rehash) && (!incremental) && (old_bucket_count >= new_bucket_count) && (power_2_buckets || (old_bucket_count % new_bucket_count) == 0); //If we are shrinking the same bucket array and it's //is a fast shrink, just rehash the last nodes size_type new_first_bucket_num = new_bucket_count; size_type old_bucket_cache = (size_type)this->priv_get_cache_bucket_num(); if(same_buffer && fast_shrink && (old_bucket_cache < new_bucket_count)){ new_first_bucket_num = old_bucket_cache; old_bucket_cache = new_bucket_count; } if (!do_full_rehash) this->priv_initialize_new_buckets(old_buckets, old_bucket_count, new_buckets, new_bucket_count); //Anti-exception stuff: they destroy the elements if something goes wrong. //If the source and destination buckets are the same, the second rollback function //is harmless, because all elements have been already unlinked and destroyed typedef typename internal_type::template typeof_node_disposer::type NodeDisposer; typedef exception_bucket_disposer ArrayDisposer; NodeDisposer nd(this->make_node_disposer(detail::null_disposer())); ArrayDisposer rollback1(new_buckets[0], nd, new_bucket_count); ArrayDisposer rollback2(old_buckets[0], nd, old_bucket_count); //Put size in a safe value for rollback exception size_type const size_backup = this->priv_size_traits().get_size(); this->priv_size_traits().set_size(0); //Put cache to safe position this->priv_init_cache(); this->priv_unset_sentinel_bucket(); const size_type split = this->rehash_split_from_bucket_count(new_bucket_count); //Iterate through nodes for(size_type n = old_bucket_cache; n < old_bucket_count; ++n){ bucket_type &old_bucket = old_buckets[difference_type(n)]; if(!fast_shrink){ siterator before_i(old_bucket.get_node_ptr()); siterator i(before_i); ++i; siterator end_sit(this->sit_end(old_bucket)); for( // ; i != end_sit ; i = before_i, ++i){ //First obtain hash value (and store it if do_full_rehash) std::size_t hash_value; if(do_full_rehash){ value_type &v = this->priv_value_from_siterator(i); hash_value = this->priv_hasher()(key_of_value()(v)); node_functions_t::store_hash(this->priv_value_to_node_ptr(v), hash_value, store_hash_t()); } else{ const value_type &v = this->priv_value_from_siterator(i); hash_value = this->priv_stored_or_compute_hash(v, store_hash_t()); } //Now calculate the new bucket position const size_type new_n = (size_type)hash_to_bucket_split (hash_value, new_bucket_count, split, fastmod_buckets_t()); //Update first used bucket cache if(cache_begin && new_n < new_first_bucket_num) new_first_bucket_num = new_n; //If the target bucket is new, transfer the whole group siterator last = i; (priv_go_to_last_in_group)(i, optimize_multikey_t()); if(same_buffer && new_n == n){ before_i = last; } else{ bucket_type &new_b = new_buckets[difference_type(new_n)]; slist_node_algorithms::transfer_after(new_b.get_node_ptr(), before_i.pointed_node(), last.pointed_node()); } } } else{ const size_type new_n = (size_type)hash_to_bucket_split (n, new_bucket_count, split, fastmod_buckets_t()); if(cache_begin && new_n < new_first_bucket_num) new_first_bucket_num = new_n; bucket_type &new_b = new_buckets[difference_type(new_n)]; siterator last = this->priv_get_last(old_bucket, optimize_multikey_t()); slist_node_algorithms::transfer_after(new_b.get_node_ptr(), old_bucket.get_node_ptr(), last.pointed_node()); } } this->priv_size_traits().set_size(size_backup); this->priv_split_traits().set_size(split); if(&new_bucket_traits != &this->priv_bucket_traits()) this->priv_bucket_traits() = new_bucket_traits; this->priv_set_sentinel_bucket(); this->priv_set_cache_bucket_num(new_first_bucket_num); rollback1.release(); rollback2.release(); } template void priv_clone_from(MaybeConstHashtableImpl &src, Cloner cloner, Disposer disposer) { this->clear_and_dispose(disposer); if(!constant_time_size || !src.empty()){ const size_type src_bucket_count = src.bucket_count(); const size_type dst_bucket_count = this->bucket_count(); //Check power of two bucket array if the option is activated BOOST_INTRUSIVE_INVARIANT_ASSERT (!power_2_buckets || (0 == (src_bucket_count & (src_bucket_count-1)))); BOOST_INTRUSIVE_INVARIANT_ASSERT (!power_2_buckets || (0 == (dst_bucket_count & (dst_bucket_count-1)))); //If src bucket count is bigger or equal, structural copy is possible const bool structural_copy = (!incremental) && (src_bucket_count >= dst_bucket_count) && (power_2_buckets || (src_bucket_count % dst_bucket_count) == 0); if(structural_copy){ this->priv_structural_clone_from(src, cloner, disposer); } else{ //Unlike previous cloning algorithm, this can throw //if cloner, hasher or comparison functor throw typedef typename detail::if_c< detail::is_const::value , typename MaybeConstHashtableImpl::const_iterator , typename MaybeConstHashtableImpl::iterator >::type clone_iterator; clone_iterator b(src.begin()), e(src.end()); detail::exception_disposer rollback(*this, disposer); for(; b != e; ++b){ //No need to check for duplicates and insert it in the first position //as this is an unordered container. So use minimal insertion code std::size_t const hash_to_store = this->priv_stored_or_compute_hash(*b, store_hash_t()); size_type const bucket_number = this->priv_hash_to_nbucket(hash_to_store); typedef typename detail::if_c ::value, const_reference, reference>::type reference_type; reference_type r = *b; this->priv_clone_front_in_bucket(bucket_number, r, hash_to_store, cloner); } rollback.release(); } } } template void priv_clone_front_in_bucket( size_type const bucket_number , typename detail::identity::type src_ref , std::size_t const hash_to_store, Cloner cloner) { //No need to check for duplicates and insert it in the first position //as this is an unordered container. So use minimal insertion code bucket_type &cur_bucket = this->priv_bucket(bucket_number); siterator const prev(cur_bucket.get_node_ptr()); //Just check if the cloned node is equal to the first inserted value in the new bucket //as equal src values were contiguous and they should be already inserted in the //destination bucket. bool const next_is_in_group = optimize_multikey && !this->priv_bucket_empty(bucket_number) && this->priv_equal()( key_of_value()(src_ref) , key_of_value()(this->priv_value_from_siterator(++siterator(prev)))); this->priv_insert_equal_after_find(*cloner(src_ref), bucket_number, hash_to_store, prev, next_is_in_group); } template void priv_structural_clone_from(MaybeConstHashtableImpl &src, Cloner cloner, Disposer disposer) { //First clone the first ones const size_type src_bucket_count = src.bucket_count(); const size_type dst_bucket_count = this->bucket_count(); size_type constructed = 0; typedef typename internal_type::template typeof_node_disposer::type NodeDisposer; NodeDisposer node_disp(disposer, &this->priv_value_traits()); exception_bucket_disposer rollback(this->priv_bucket(0), node_disp, constructed); //Now insert the remaining ones using the modulo trick for( //"constructed" already initialized ; constructed < src_bucket_count ; ++constructed){ const size_type new_n = (size_type)hash_to_bucket_split (constructed, dst_bucket_count, this->split_count(), fastmod_buckets_t()); bucket_type &src_b = src.priv_bucket(constructed); for( siterator b(this->priv_bucket_lbegin(src_b)), e(this->priv_bucket_lend(src_b)); b != e; ++b){ typedef typename detail::if_c ::value, const_reference, reference>::type reference_type; reference_type r = this->priv_value_from_siterator(b); this->priv_clone_front_in_bucket (new_n, r, this->priv_stored_hash(b, store_hash_t()), cloner); } } this->priv_hasher() = src.priv_hasher(); this->priv_equal() = src.priv_equal(); rollback.release(); this->priv_size_traits().set_size(src.priv_size_traits().get_size()); this->priv_split_traits().set_size(dst_bucket_count); this->priv_set_cache_bucket_num(0u); this->priv_erasure_update_cache(); } iterator priv_insert_equal_after_find(reference value, size_type bucket_num, std::size_t hash_value, siterator prev, bool const next_is_in_group) { //Now store hash if needed node_ptr n = this->priv_value_to_node_ptr(value); node_functions_t::store_hash(n, hash_value, store_hash_t()); //Checks for some modes BOOST_INTRUSIVE_SAFE_HOOK_DEFAULT_ASSERT(!safemode_or_autounlink || slist_node_algorithms::unique(n)); //Shortcut to optimize_multikey cases group_functions_t::insert_in_group ( next_is_in_group ? dcast_bucket_ptr((++siterator(prev)).pointed_node()) : n , n, optimize_multikey_t()); //Update cache and increment size if needed this->priv_insertion_update_cache(bucket_num); this->priv_size_traits().increment(); slist_node_algorithms::link_after(prev.pointed_node(), n); return this->build_iterator(siterator(n), this->priv_bucket_ptr(bucket_num)); } template siterator priv_find //In case it is not found previt is priv_end_sit() ( const KeyType &key, KeyHasher hash_func , KeyEqual equal_func, size_type &bucket_number, std::size_t &h, siterator &previt) const { h = hash_func(key); bucket_number = this->priv_hash_to_nbucket(h); bucket_type& b = this->priv_bucket(bucket_number); siterator prev = this->sit_bbegin(b); siterator it = prev; siterator const endit = this->sit_end(b); while (++it != endit) { if (this->priv_is_value_equal_to_key (this->priv_value_from_siterator(it), h, key, equal_func, compare_hash_t())) { previt = prev; return it; } (priv_go_to_last_in_group)(it, optimize_multikey_t()); prev = it; } previt = b.get_node_ptr(); return this->priv_end_sit(); } template siterator priv_find_in_bucket //In case it is not found previt is priv_end_sit() (bucket_type &b, const KeyType& key, KeyEqual equal_func, const std::size_t h) const { siterator it(this->sit_begin(b)); siterator const endit(this->sit_end(b)); for (; it != endit; (priv_go_to_last_in_group)(it, optimize_multikey_t()), ++it) { if (BOOST_LIKELY(this->priv_is_value_equal_to_key (this->priv_value_from_siterator(it), h, key, equal_func, compare_hash_t()))) { return it; } } return this->priv_end_sit(); } template BOOST_INTRUSIVE_FORCEINLINE bool priv_is_value_equal_to_key (const value_type &v, const std::size_t h, const KeyType &key, KeyEqual equal_func, detail::true_) const //compare_hash { return this->priv_stored_or_compute_hash(v, store_hash_t()) == h && equal_func(key, key_of_value()(v)); } template BOOST_INTRUSIVE_FORCEINLINE bool priv_is_value_equal_to_key (const value_type& v, const std::size_t , const KeyType& key, KeyEqual equal_func, detail::false_) const //compare_hash { return equal_func(key, key_of_value()(v)); } //return previous iterator to the next equal range group in case BOOST_INTRUSIVE_FORCEINLINE static void priv_go_to_last_in_group (siterator &it_first_in_group, detail::true_) BOOST_NOEXCEPT //optimize_multikey { it_first_in_group = (group_functions_t::get_last_in_group (dcast_bucket_ptr(it_first_in_group.pointed_node()), optimize_multikey_t())); } //return previous iterator to the next equal range group in case BOOST_INTRUSIVE_FORCEINLINE static void priv_go_to_last_in_group //!optimize_multikey (siterator /*&it_first_in_group*/, detail::false_) BOOST_NOEXCEPT { } template std::pair priv_local_equal_range ( const KeyType &key , KeyHasher hash_func , KeyEqual equal_func , size_type &found_bucket , size_type &cnt) const { std::size_t internal_cnt = 0; //Let's see if the element is present siterator prev; size_type n_bucket; std::size_t h; std::pair to_return ( this->priv_find(key, hash_func, equal_func, n_bucket, h, prev) , this->priv_end_sit()); if(to_return.first != to_return.second){ found_bucket = n_bucket; //If it's present, find the first that it's not equal in //the same bucket siterator it = to_return.first; siterator const bend = this->priv_bucket_lend(n_bucket); BOOST_IF_CONSTEXPR(optimize_multikey){ siterator past_last_in_group_it = it; (priv_go_to_last_in_group)(past_last_in_group_it, optimize_multikey_t()); ++past_last_in_group_it; internal_cnt += boost::intrusive::iterator_udistance(++it, past_last_in_group_it) + 1u; if (past_last_in_group_it != bend) to_return.second = past_last_in_group_it; } else{ do { ++internal_cnt; //At least one is found ++it; } while(it != bend && this->priv_is_value_equal_to_key (this->priv_value_from_siterator(it), h, key, equal_func, compare_hash_t())); if (it != bend) to_return.second = it; } } cnt = size_type(internal_cnt); return to_return; } struct priv_equal_range_result { siterator first; siterator second; bucket_ptr bucket_first; bucket_ptr bucket_second; }; template priv_equal_range_result priv_equal_range ( const KeyType &key , KeyHasher hash_func , KeyEqual equal_func) const { size_type n_bucket; size_type cnt; //Let's see if the element is present const std::pair to_return (this->priv_local_equal_range(key, hash_func, equal_func, n_bucket, cnt)); priv_equal_range_result r; r.first = to_return.first; r.second = to_return.second; //If not, find the next element as ".second" if ".second" local iterator //is not pointing to an element. if(to_return.first == to_return.second) { r.bucket_first = r.bucket_second = this->priv_invalid_bucket_ptr(); } else if (to_return.second != this->priv_end_sit()) { r.bucket_first = this->priv_bucket_ptr(n_bucket); } else{ r.bucket_first = this->priv_bucket_ptr(n_bucket); const size_type max_bucket = this->bucket_count(); do{ ++n_bucket; } while (n_bucket != max_bucket && this->priv_bucket_empty(n_bucket)); if (n_bucket == max_bucket){ r.bucket_second = this->priv_invalid_bucket_ptr(); } else{ r.bucket_second = this->priv_bucket_ptr(n_bucket); r.second = siterator(r.bucket_second->begin_ptr()); } } return r; } BOOST_INTRUSIVE_FORCEINLINE size_type priv_get_bucket_num(const_iterator it) BOOST_NOEXCEPT { return this->priv_get_bucket_num(it, linear_buckets_t()); } BOOST_INTRUSIVE_FORCEINLINE size_type priv_get_bucket_num(const_iterator it, detail::true_) BOOST_NOEXCEPT //linear { return size_type(it.get_bucket_ptr() - this->priv_bucket_pointer()); } BOOST_INTRUSIVE_FORCEINLINE size_type priv_get_bucket_num(const_iterator it, detail::false_) BOOST_NOEXCEPT //!linear { return this->priv_get_bucket_num_hash_dispatch(it.slist_it(), store_hash_t()); } BOOST_INTRUSIVE_FORCEINLINE size_type priv_get_bucket_num_hash_dispatch(siterator it, detail::true_) BOOST_NOEXCEPT //store_hash { return (size_type)this->priv_hash_to_nbucket(this->priv_stored_hash(it, store_hash_t())); } size_type priv_get_bucket_num_hash_dispatch(siterator it, detail::false_) BOOST_NOEXCEPT //NO store_hash { const bucket_type &f = this->priv_bucket(0u); slist_node_ptr bb = group_functions_t::get_bucket_before_begin ( this->priv_bucket_lbbegin(0u).pointed_node() , this->priv_bucket_lbbegin(this->priv_usable_bucket_count() - 1u).pointed_node() , it.pointed_node() , optimize_multikey_t()); //Now get the bucket_impl from the iterator const bucket_type &b = static_cast(*bb); //Now just calculate the index b has in the bucket array return static_cast(&b - &f); } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_get_bucket_ptr(const_iterator it) BOOST_NOEXCEPT { return this->priv_get_bucket_ptr(it, linear_buckets_t()); } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_get_bucket_ptr(const_iterator it, detail::true_) BOOST_NOEXCEPT //linear { return it.get_bucket_ptr(); } BOOST_INTRUSIVE_FORCEINLINE bucket_ptr priv_get_bucket_ptr(const_iterator it, detail::false_) BOOST_NOEXCEPT //!linear { return this->priv_bucket_ptr(this->priv_get_bucket_num_hash_dispatch(it.slist_it(), store_hash_t())); } /// @endcond }; /// @cond template < class T , class PackedOptions > struct make_bucket_traits { //Real value traits must be calculated from options typedef typename detail::get_value_traits ::type value_traits; typedef typename PackedOptions::bucket_traits specified_bucket_traits; //Real bucket traits must be calculated from options and calculated value_traits typedef bucket_traits_impl < typename unordered_bucket_ptr_impl ::type , std::size_t> bucket_traits_t; typedef typename detail::if_c< detail::is_same < specified_bucket_traits , default_bucket_traits >::value , bucket_traits_t , specified_bucket_traits >::type type; }; /// @endcond //! Helper metafunction to define a \c hashtable that yields to the same type when the //! same options (either explicitly or implicitly) are used. #if defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) || defined(BOOST_INTRUSIVE_VARIADIC_TEMPLATES) template #else template #endif struct make_hashtable { /// @cond typedef typename pack_options < hashtable_defaults, #if !defined(BOOST_INTRUSIVE_VARIADIC_TEMPLATES) O1, O2, O3, O4, O5, O6, O7, O8, O9, O10, O11 #else Options... #endif >::type packed_options; typedef typename detail::get_value_traits ::type value_traits; typedef typename make_bucket_traits ::type bucket_traits; typedef hashtable_impl < value_traits , typename packed_options::key_of_value , typename packed_options::hash , typename packed_options::equal , bucket_traits , typename packed_options::size_type , (std::size_t(false)*hash_bool_flags::unique_keys_pos) |(std::size_t(packed_options::constant_time_size)*hash_bool_flags::constant_time_size_pos) |(std::size_t(packed_options::power_2_buckets)*hash_bool_flags::power_2_buckets_pos) |(std::size_t(packed_options::cache_begin)*hash_bool_flags::cache_begin_pos) |(std::size_t(packed_options::compare_hash)*hash_bool_flags::compare_hash_pos) |(std::size_t(packed_options::incremental)*hash_bool_flags::incremental_pos) |(std::size_t(packed_options::linear_buckets)*hash_bool_flags::linear_buckets_pos) |(std::size_t(packed_options::fastmod_buckets)*hash_bool_flags::fastmod_buckets_pos) > implementation_defined; /// @endcond typedef implementation_defined type; }; #if !defined(BOOST_INTRUSIVE_DOXYGEN_INVOKED) #if defined(BOOST_INTRUSIVE_VARIADIC_TEMPLATES) template #else template #endif class hashtable : public make_hashtable::type { typedef typename make_hashtable::type Base; BOOST_MOVABLE_BUT_NOT_COPYABLE(hashtable) public: typedef typename Base::value_traits value_traits; typedef typename Base::iterator iterator; typedef typename Base::const_iterator const_iterator; typedef typename Base::bucket_ptr bucket_ptr; typedef typename Base::size_type size_type; typedef typename Base::hasher hasher; typedef typename Base::bucket_traits bucket_traits; typedef typename Base::key_equal key_equal; //Assert if passed value traits are compatible with the type BOOST_STATIC_ASSERT((detail::is_same::value)); BOOST_INTRUSIVE_FORCEINLINE explicit hashtable ( const bucket_traits &b_traits , const hasher & hash_func = hasher() , const key_equal &equal_func = key_equal() , const value_traits &v_traits = value_traits()) : Base(b_traits, hash_func, equal_func, v_traits) {} BOOST_INTRUSIVE_FORCEINLINE hashtable(BOOST_RV_REF(hashtable) x) : Base(BOOST_MOVE_BASE(Base, x)) {} BOOST_INTRUSIVE_FORCEINLINE hashtable& operator=(BOOST_RV_REF(hashtable) x) { return static_cast(this->Base::operator=(BOOST_MOVE_BASE(Base, x))); } template BOOST_INTRUSIVE_FORCEINLINE void clone_from(const hashtable &src, Cloner cloner, Disposer disposer) { Base::clone_from(src, cloner, disposer); } template BOOST_INTRUSIVE_FORCEINLINE void clone_from(BOOST_RV_REF(hashtable) src, Cloner cloner, Disposer disposer) { Base::clone_from(BOOST_MOVE_BASE(Base, src), cloner, disposer); } }; #endif } //namespace intrusive } //namespace boost #include #endif //BOOST_INTRUSIVE_HASHTABLE_HPP