318 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
			
		
		
	
	
			318 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
| https://bugs.gentoo.org/887037
 | |
| 
 | |
| Workaround https://github.com/boostorg/container/commit/99091420ae553b27345e04279fd19fe24fb684c1
 | |
| in Boost 1.81.
 | |
| 
 | |
| Upstream s2 (as in real upstream, not MongoDB) has deviated substantially
 | |
| from the version vendored.
 | |
| --- a/src/third_party/s2/base/stl_decl_msvc.h
 | |
| +++ b/src/third_party/s2/base/stl_decl_msvc.h
 | |
| @@ -118,8 +118,8 @@ namespace msvchash {
 | |
|    class hash_multimap;
 | |
|  }  // end namespace  msvchash
 | |
|  
 | |
| -using msvchash::hash_set;
 | |
| -using msvchash::hash_map;
 | |
| +using msvchash::hash_set = my_hash_set;
 | |
| +using msvchash::hash_map = my_hash_map;
 | |
|  using msvchash::hash;
 | |
|  using msvchash::hash_multimap;
 | |
|  using msvchash::hash_multiset;
 | |
| --- a/src/third_party/s2/base/stl_decl_osx.h
 | |
| +++ b/src/third_party/s2/base/stl_decl_osx.h
 | |
| @@ -68,8 +68,8 @@ using std::string;
 | |
|  
 | |
|  using namespace std;
 | |
|  using __gnu_cxx::hash;
 | |
| -using __gnu_cxx::hash_set;
 | |
| -using __gnu_cxx::hash_map;
 | |
| +using __gnu_cxx::hash_set = my_hash_set;
 | |
| +using __gnu_cxx::hash_map = my_hash_map;
 | |
|  using __gnu_cxx::select1st;
 | |
|  
 | |
|  /* On Linux (and gdrive on OSX), this comes from places like
 | |
| --- a/src/third_party/s2/hash.h
 | |
| +++ b/src/third_party/s2/hash.h
 | |
| @@ -2,10 +2,10 @@
 | |
|  #define THIRD_PARTY_S2_HASH_H_
 | |
|  
 | |
|  #include <unordered_map>
 | |
| -#define hash_map std::unordered_map
 | |
| +#define my_hash_map std::unordered_map
 | |
|  
 | |
|  #include <unordered_set>
 | |
| -#define hash_set std::unordered_set
 | |
| +#define my_hash_set std::unordered_set
 | |
|  
 | |
|  #define HASH_NAMESPACE_START namespace std {
 | |
|  #define HASH_NAMESPACE_END }
 | |
| --- a/src/third_party/s2/s2_test.cc
 | |
| +++ b/src/third_party/s2/s2_test.cc
 | |
| @@ -10,7 +10,7 @@ using std::reverse;
 | |
|  
 | |
|  #include <hash_set>
 | |
|  #include <hash_map>
 | |
| -using __gnu_cxx::hash_set;
 | |
| +using __gnu_cxx::hash_set = my_hash_map;
 | |
|  
 | |
|  #include "s2.h"
 | |
|  #include "base/logging.h"
 | |
| @@ -709,8 +709,8 @@ TEST(S2, Frames) {
 | |
|  #if 0
 | |
|  TEST(S2, S2PointHashSpreads) {
 | |
|    int kTestPoints = 1 << 16;
 | |
| -  hash_set<size_t> set;
 | |
| -  hash_set<S2Point> points;
 | |
| +  my_hash_set<size_t> set;
 | |
| +  my_hash_set<S2Point> points;
 | |
|    hash<S2Point> hasher;
 | |
|    S2Point base = S2Point(1, 1, 1);
 | |
|    for (int i = 0; i < kTestPoints; ++i) {
 | |
| @@ -733,7 +733,7 @@ TEST(S2, S2PointHashCollapsesZero) {
 | |
|    double minus_zero = -zero;
 | |
|    EXPECT_NE(*reinterpret_cast<uint64 const*>(&zero),
 | |
|              *reinterpret_cast<uint64 const*>(&minus_zero));
 | |
| -  hash_map<S2Point, int> map;
 | |
| +  my_hash_map<S2Point, int> map;
 | |
|    S2Point zero_pt(zero, zero, zero);
 | |
|    S2Point minus_zero_pt(minus_zero, minus_zero, minus_zero);
 | |
|  
 | |
| --- a/src/third_party/s2/s2cellid_test.cc
 | |
| +++ b/src/third_party/s2/s2cellid_test.cc
 | |
| @@ -10,7 +10,7 @@ using std::reverse;
 | |
|  
 | |
|  #include <cstdio>
 | |
|  #include <hash_map>
 | |
| -using __gnu_cxx::hash_map;
 | |
| +using __gnu_cxx::hash_map = my_hash_map;
 | |
|  
 | |
|  #include <sstream>
 | |
|  #include <vector>
 | |
| @@ -170,7 +170,7 @@ TEST(S2CellId, Tokens) {
 | |
|  static const int kMaxExpandLevel = 3;
 | |
|  
 | |
|  static void ExpandCell(S2CellId const& parent, vector<S2CellId>* cells,
 | |
| -                       hash_map<S2CellId, S2CellId>* parent_map) {
 | |
| +                       my_hash_map<S2CellId, S2CellId>* parent_map) {
 | |
|    cells->push_back(parent);
 | |
|    if (parent.level() == kMaxExpandLevel) return;
 | |
|    int i, j, orientation;
 | |
| @@ -194,7 +194,7 @@ static void ExpandCell(S2CellId const& parent, vector<S2CellId>* cells,
 | |
|  
 | |
|  TEST(S2CellId, Containment) {
 | |
|    // Test contains() and intersects().
 | |
| -  hash_map<S2CellId, S2CellId> parent_map;
 | |
| +  my_hash_map<S2CellId, S2CellId> parent_map;
 | |
|    vector<S2CellId> cells;
 | |
|    for (int face = 0; face < 6; ++face) {
 | |
|      ExpandCell(S2CellId::FromFacePosLevel(face, 0, 0), &cells, &parent_map);
 | |
| --- a/src/third_party/s2/s2loop.cc
 | |
| +++ b/src/third_party/s2/s2loop.cc
 | |
| @@ -120,7 +120,7 @@ bool S2Loop::IsValid(string* err) const {
 | |
|      }
 | |
|    }
 | |
|    // Loops are not allowed to have any duplicate vertices.
 | |
| -  hash_map<S2Point, int> vmap;
 | |
| +  my_hash_map<S2Point, int> vmap;
 | |
|    for (int i = 0; i < num_vertices(); ++i) {
 | |
|      if (!vmap.insert(make_pair(vertex(i), i)).second) {
 | |
|        VLOG(2) << "Duplicate vertices: " << vmap[vertex(i)] << " and " << i;
 | |
| --- a/src/third_party/s2/s2polygon.cc
 | |
| +++ b/src/third_party/s2/s2polygon.cc
 | |
| @@ -117,7 +117,7 @@ HASH_NAMESPACE_END
 | |
|  bool S2Polygon::IsValid(const vector<S2Loop*>& loops, string* err) {
 | |
|    // If a loop contains an edge AB, then no other loop may contain AB or BA.
 | |
|    if (loops.size() > 1) {
 | |
| -    hash_map<S2PointPair, pair<int, int> > edges;
 | |
| +    my_hash_map<S2PointPair, pair<int, int> > edges;
 | |
|      for (size_t i = 0; i < loops.size(); ++i) {
 | |
|        S2Loop* lp = loops[i];
 | |
|        for (int j = 0; j < lp->num_vertices(); ++j) {
 | |
| --- a/src/third_party/s2/s2polygonbuilder.cc
 | |
| +++ b/src/third_party/s2/s2polygonbuilder.cc
 | |
| @@ -175,7 +175,7 @@ S2Loop* S2PolygonBuilder::AssembleLoop(S2Point const& v0, S2Point const& v1,
 | |
|    // This ensures that only CCW loops are constructed when possible.
 | |
|  
 | |
|    vector<S2Point> path;          // The path so far.
 | |
| -  hash_map<S2Point, int> index;  // Maps a vertex to its index in "path".
 | |
| +  my_hash_map<S2Point, int> index;  // Maps a vertex to its index in "path".
 | |
|    path.push_back(v0);
 | |
|    path.push_back(v1);
 | |
|    index[v1] = 1;
 | |
| @@ -361,7 +361,7 @@ void S2PolygonBuilder::BuildMergeMap(PointIndex* index, MergeMap* merge_map) {
 | |
|  
 | |
|    // First, we build the set of all the distinct vertices in the input.
 | |
|    // We need to include the source and destination of every edge.
 | |
| -  hash_set<S2Point> vertices;
 | |
| +  my_hash_set<S2Point> vertices;
 | |
|    for (EdgeSet::const_iterator i = edges_->begin(); i != edges_->end(); ++i) {
 | |
|      vertices.insert(i->first);
 | |
|      VertexSet const& vset = i->second;
 | |
| @@ -370,7 +370,7 @@ void S2PolygonBuilder::BuildMergeMap(PointIndex* index, MergeMap* merge_map) {
 | |
|    }
 | |
|  
 | |
|    // Build a spatial index containing all the distinct vertices.
 | |
| -  for (hash_set<S2Point>::const_iterator i = vertices.begin();
 | |
| +  for (my_hash_set<S2Point>::const_iterator i = vertices.begin();
 | |
|         i != vertices.end(); ++i) {
 | |
|      index->Insert(*i);
 | |
|    }
 | |
| @@ -378,7 +378,7 @@ void S2PolygonBuilder::BuildMergeMap(PointIndex* index, MergeMap* merge_map) {
 | |
|    // Next, we loop through all the vertices and attempt to grow a maximial
 | |
|    // mergeable group starting from each vertex.
 | |
|    vector<S2Point> frontier, mergeable;
 | |
| -  for (hash_set<S2Point>::const_iterator vstart = vertices.begin();
 | |
| +  for (my_hash_set<S2Point>::const_iterator vstart = vertices.begin();
 | |
|         vstart != vertices.end(); ++vstart) {
 | |
|      // Skip any vertices that have already been merged with another vertex.
 | |
|      if (merge_map->find(*vstart) != merge_map->end()) continue;
 | |
| --- a/src/third_party/s2/s2polygonbuilder.h
 | |
| +++ b/src/third_party/s2/s2polygonbuilder.h
 | |
| @@ -262,7 +262,7 @@ class S2PolygonBuilder {
 | |
|    // current position to a new position, and also returns a spatial index
 | |
|    // containing all of the vertices that do not need to be moved.
 | |
|    class PointIndex;
 | |
| -  typedef hash_map<S2Point, S2Point> MergeMap;
 | |
| +  typedef my_hash_map<S2Point, S2Point> MergeMap;
 | |
|    void BuildMergeMap(PointIndex* index, MergeMap* merge_map);
 | |
|  
 | |
|    // Moves a set of vertices from old to new positions.
 | |
| @@ -282,7 +282,7 @@ class S2PolygonBuilder {
 | |
|    // once.  We could have also used a multiset<pair<S2Point, S2Point> >,
 | |
|    // but this representation is a bit more convenient.
 | |
|    typedef multiset<S2Point> VertexSet;
 | |
| -  typedef hash_map<S2Point, VertexSet> EdgeSet;
 | |
| +  typedef my_hash_map<S2Point, VertexSet> EdgeSet;
 | |
|    scoped_ptr<EdgeSet> edges_;
 | |
|  
 | |
|    // Unique collection of the starting (first) vertex of all edges,
 | |
| --- a/src/third_party/s2/s2regioncoverer.cc
 | |
| +++ b/src/third_party/s2/s2regioncoverer.cc
 | |
| @@ -321,7 +321,7 @@ void S2RegionCoverer::GetInteriorCellUnion(S2Region const& region,
 | |
|  
 | |
|  void S2RegionCoverer::FloodFill(
 | |
|      S2Region const& region, S2CellId const& start, vector<S2CellId>* output) {
 | |
| -  hash_set<S2CellId> all;
 | |
| +  my_hash_set<S2CellId> all;
 | |
|    vector<S2CellId> frontier;
 | |
|    output->clear();
 | |
|    all.insert(start);
 | |
| --- a/src/third_party/s2/s2regioncoverer_test.cc
 | |
| +++ b/src/third_party/s2/s2regioncoverer_test.cc
 | |
| @@ -11,7 +11,7 @@ using std::swap;
 | |
|  using std::reverse;
 | |
|  
 | |
|  #include <hash_map>
 | |
| -using __gnu_cxx::hash_map;
 | |
| +using __gnu_cxx::hash_map = my_hash_map;
 | |
|  
 | |
|  #include <queue>
 | |
|  using std::priority_queue;
 | |
| @@ -65,7 +65,7 @@ static void CheckCovering(S2RegionCoverer const& coverer,
 | |
|                            vector<S2CellId> const& covering,
 | |
|                            bool interior) {
 | |
|    // Keep track of how many cells have the same coverer.min_level() ancestor.
 | |
| -  hash_map<S2CellId, int> min_level_cells;
 | |
| +  my_hash_map<S2CellId, int> min_level_cells;
 | |
|    for (int i = 0; i < covering.size(); ++i) {
 | |
|      int level = covering[i].level();
 | |
|      EXPECT_GE(level, coverer.min_level());
 | |
| @@ -76,7 +76,7 @@ static void CheckCovering(S2RegionCoverer const& coverer,
 | |
|    if (covering.size() > coverer.max_cells()) {
 | |
|      // If the covering has more than the requested number of cells, then check
 | |
|      // that the cell count cannot be reduced by using the parent of some cell.
 | |
| -    for (hash_map<S2CellId, int>::const_iterator i = min_level_cells.begin();
 | |
| +    for (my_hash_map<S2CellId, int>::const_iterator i = min_level_cells.begin();
 | |
|           i != min_level_cells.end(); ++i) {
 | |
|        EXPECT_EQ(i->second, 1);
 | |
|      }
 | |
| --- a/src/third_party/s2/strings/split.cc
 | |
| +++ b/src/third_party/s2/strings/split.cc
 | |
| @@ -156,7 +156,7 @@ struct simple_insert_iterator {
 | |
|  // SplitStringToIterator{Using|AllowEmpty}().
 | |
|  template <typename T>
 | |
|  struct simple_hash_map_iterator {
 | |
| -  typedef hash_map<T, T> hashmap;
 | |
| +  typedef my_hash_map<T, T> hashmap;
 | |
|    hashmap* t;
 | |
|    bool even;
 | |
|    typename hashmap::iterator curr;
 | |
| @@ -246,8 +246,8 @@ void SplitStringAllowEmpty(const string& full, const char* delim,
 | |
|  }
 | |
|  
 | |
|  void SplitStringToHashsetAllowEmpty(const string& full, const char* delim,
 | |
| -                                    hash_set<string>* result) {
 | |
| -  simple_insert_iterator<hash_set<string> > it(result);
 | |
| +                                    my_hash_set<string>* result) {
 | |
| +  simple_insert_iterator<my_hash_set<string> > it(result);
 | |
|    SplitStringToIteratorAllowEmpty(full, delim, 0, it);
 | |
|  }
 | |
|  
 | |
| @@ -258,7 +258,7 @@ void SplitStringToSetAllowEmpty(const string& full, const char* delim,
 | |
|  }
 | |
|  
 | |
|  void SplitStringToHashmapAllowEmpty(const string& full, const char* delim,
 | |
| -                                    hash_map<string, string>* result) {
 | |
| +                                    my_hash_map<string, string>* result) {
 | |
|    simple_hash_map_iterator<string> it(result);
 | |
|    SplitStringToIteratorAllowEmpty(full, delim, 0, it);
 | |
|  }
 | |
| @@ -352,8 +352,8 @@ void SplitStringUsing(const string& full,
 | |
|  }
 | |
|  
 | |
|  void SplitStringToHashsetUsing(const string& full, const char* delim,
 | |
| -                               hash_set<string>* result) {
 | |
| -  simple_insert_iterator<hash_set<string> > it(result);
 | |
| +                               my_hash_set<string>* result) {
 | |
| +  simple_insert_iterator<my_hash_set<string> > it(result);
 | |
|    SplitStringToIteratorUsing(full, delim, it);
 | |
|  }
 | |
|  
 | |
| @@ -364,7 +364,7 @@ void SplitStringToSetUsing(const string& full, const char* delim,
 | |
|  }
 | |
|  
 | |
|  void SplitStringToHashmapUsing(const string& full, const char* delim,
 | |
| -                               hash_map<string, string>* result) {
 | |
| +                               my_hash_map<string, string>* result) {
 | |
|    simple_hash_map_iterator<string> it(result);
 | |
|    SplitStringToIteratorUsing(full, delim, it);
 | |
|  }
 | |
| --- a/src/third_party/s2/strings/split.h
 | |
| +++ b/src/third_party/s2/strings/split.h
 | |
| @@ -41,7 +41,7 @@ using namespace std;
 | |
|  void SplitStringAllowEmpty(const string& full, const char* delim,
 | |
|                             vector<string>* res);
 | |
|  void SplitStringToHashsetAllowEmpty(const string& full, const char* delim,
 | |
| -                                    hash_set<string>* res);
 | |
| +                                    my_hash_set<string>* res);
 | |
|  void SplitStringToSetAllowEmpty(const string& full, const char* delim,
 | |
|                                  set<string>* res);
 | |
|  // The even-positioned (0-based) components become the keys for the
 | |
| @@ -50,7 +50,7 @@ void SplitStringToSetAllowEmpty(const string& full, const char* delim,
 | |
|  // if the key was already present in the hash table, or will be the
 | |
|  // empty string if the key is a newly inserted key.
 | |
|  void SplitStringToHashmapAllowEmpty(const string& full, const char* delim,
 | |
| -                                    hash_map<string, string>* result);
 | |
| +                                    my_hash_map<string, string>* result);
 | |
|  
 | |
|  // ----------------------------------------------------------------------
 | |
|  // SplitStringUsing()
 | |
| @@ -66,7 +66,7 @@ void SplitStringToHashmapAllowEmpty(const string& full, const char* delim,
 | |
|  void SplitStringUsing(const string& full, const char* delim,
 | |
|                        vector<string>* res);
 | |
|  void SplitStringToHashsetUsing(const string& full, const char* delim,
 | |
| -                               hash_set<string>* res);
 | |
| +                               my_hash_set<string>* res);
 | |
|  void SplitStringToSetUsing(const string& full, const char* delim,
 | |
|                             set<string>* res);
 | |
|  // The even-positioned (0-based) components become the keys for the
 | |
| @@ -75,7 +75,7 @@ void SplitStringToSetUsing(const string& full, const char* delim,
 | |
|  // if the key was already present in the hash table, or will be the
 | |
|  // empty string if the key is a newly inserted key.
 | |
|  void SplitStringToHashmapUsing(const string& full, const char* delim,
 | |
| -                               hash_map<string, string>* result);
 | |
| +                               my_hash_map<string, string>* result);
 | |
|  
 | |
|  // ----------------------------------------------------------------------
 | |
|  // SplitOneIntToken()
 |