#ifndef Magnum_MeshTools_RemoveDuplicates_h #define Magnum_MeshTools_RemoveDuplicates_h /* This file is part of Magnum. Copyright © 2010, 2011, 2012, 2013, 2014, 2015, 2016, 2017 Vladimír Vondruš Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ /** @file * @brief Function @ref Magnum::MeshTools::removeDuplicates() */ #include #include #include #include #include #include "Magnum/Magnum.h" #include "Magnum/Math/Functions.h" namespace Magnum { namespace MeshTools { namespace Implementation { template class VectorHash { public: std::size_t operator()(const Math::Vector& data) const { return *reinterpret_cast(Utility::MurmurHash2()(reinterpret_cast(&data), sizeof(data)).byteArray()); } }; } /** @brief Remove duplicate floating-point vector data from given array @param[in,out] data Input data array @param[in] epsilon Epsilon value, vertices nearer than this distance will be melt together @return Index array and unique data Removes duplicate data from the array by collapsing them into buckets of size @p epsilon. First vector in given bucket is used, other ones are thrown away, no interpolation is done. Note that this function is meant to be used for floating-point data (or generally with non-zero @p epsilon), for discrete data the usual sorting method is much more efficient. If you want to remove duplicate data from already indexed array, first remove duplicates as if the array wasn't indexed at all and then use @ref duplicate() to combine the two index arrays: @code{.cpp} std::vector indices; std::vector positions; indices = MeshTools::duplicate(indices, MeshTools::removeDuplicates(positions)); @endcode Removing duplicates in multiple indcidental arrays is also possible -- first remove duplicates in each array separately and then use @ref combineIndexedArrays() to combine the resulting index arrays to single index array and reorder the data accordingly: @code{.cpp} std::vector positions; std::vector texCoords; std::vector positionIndices = MeshTools::removeDuplicates(positions); std::vector texCoordIndices = MeshTools::removeDuplicates(texCoords); std::vector indices = MeshTools::combineIndexedArrays( std::make_pair(std::cref(positionIndices), std::ref(positions)), std::make_pair(std::cref(texCoordIndices), std::ref(texCoords)) ); @endcode */ template std::vector removeDuplicates(std::vector& data, typename Vector::Type epsilon = Math::TypeTraits::epsilon()) { /* Get bounds */ Vector min = data[0], max = data[0]; for(const auto& v: data) { min = Math::min(v, min); max = Math::max(v, max); } /* Make epsilon so large that std::size_t can index all vectors inside the bounds. */ epsilon = Math::max(epsilon, typename Vector::Type((max-min).max()/~std::size_t{})); /* Resulting index array */ std::vector resultIndices(data.size()); std::iota(resultIndices.begin(), resultIndices.end(), 0); /* Table containing original vector index for each discretized vector. Reserving more buckets than necessary (i.e. as if each vector was unique). */ std::unordered_map, UnsignedInt, Implementation::VectorHash> table(data.size()); /* Index array for each pass, new data array */ std::vector indices; indices.reserve(data.size()); /* First go with original coordinates, then move them by epsilon/2 in each direction. */ Vector moved; for(std::size_t moving = 0; moving <= Vector::Size; ++moving) { /* Go through all vectors */ for(std::size_t i = 0; i != data.size(); ++i) { /* Try to insert new vertex to the table */ const Math::Vector v((data[i] + moved - min)/epsilon); const auto result = table.emplace(v, table.size()); /* Add the (either new or already existing) index to index array */ indices.push_back(result.first->second); /* If this is new combination, copy the data to new (earlier) possition in the array */ if(result.second && i != table.size()-1) data[table.size()-1] = data[i]; } /* Shrink the data array */ CORRADE_INTERNAL_ASSERT(data.size() >= table.size()); data.resize(table.size()); /* Remap the resulting index array */ for(auto& i: resultIndices) i = indices[i]; /* Finished */ if(moving == Vector::Size) continue; /* Move vertex coordinates by epsilon/2 in next direction */ moved = Vector(); moved[moving] = epsilon/2; /* Clear the structures for next pass */ table.clear(); indices.clear(); } return resultIndices; } }} #endif