1
0
Fork 0
mirror of https://github.com/OpenMW/openmw.git synced 2025-06-05 02:41:32 +00:00

Remove redundant job distribution between threads

Instead don't take jobs from queue until job for the same tile is processing.
This commit is contained in:
elsid 2021-08-07 11:25:01 +02:00
parent 77e58abf0d
commit 431501e23a
No known key found for this signature in database
GPG key ID: B845CB9FEE18AB40
2 changed files with 31 additions and 74 deletions

View file

@ -102,7 +102,6 @@ namespace DetourNavigator
{ {
mShouldStop = true; mShouldStop = true;
std::unique_lock<std::mutex> lock(mMutex); std::unique_lock<std::mutex> lock(mMutex);
mThreadsQueues.clear();
mWaiting.clear(); mWaiting.clear();
mHasJob.notify_all(); mHasJob.notify_all();
lock.unlock(); lock.unlock();
@ -340,64 +339,38 @@ namespace DetourNavigator
{ {
std::unique_lock<std::mutex> lock(mMutex); std::unique_lock<std::mutex> lock(mMutex);
const auto threadId = std::this_thread::get_id(); bool shouldStop = false;
auto& threadQueue = mThreadsQueues[threadId]; const auto hasJob = [&]
while (true)
{ {
bool shouldStop = false; shouldStop = mShouldStop;
return shouldStop
|| (!mWaiting.empty() && mWaiting.front()->mProcessTime <= std::chrono::steady_clock::now());
};
const auto hasJob = [&] { if (!mHasJob.wait_for(lock, std::chrono::milliseconds(10), hasJob))
shouldStop = mShouldStop; {
return shouldStop if (mJobs.empty())
|| (!mWaiting.empty() && mWaiting.front()->mProcessTime <= std::chrono::steady_clock::now()) mDone.notify_all();
|| !threadQueue.empty(); return mJobs.end();
};
if (!mHasJob.wait_for(lock, std::chrono::milliseconds(10), hasJob))
{
if (mJobs.empty())
mDone.notify_all();
return mJobs.end();
}
if (shouldStop)
return mJobs.end();
Log(Debug::Debug) << "Got " << mJobs.size() << " navigator jobs and "
<< threadQueue.size() << " thread jobs by thread=" << std::this_thread::get_id();
const JobIt job = threadQueue.empty()
? getJob(mWaiting, true)
: getJob(threadQueue, false);
if (job == mJobs.end())
continue;
const auto owner = lockTile(job->mAgentHalfExtents, job->mChangedTile);
if (owner == threadId)
{
mPushed.erase(getAgentAndTile(*job));
return job;
}
postThreadJob(job, mThreadsQueues[owner]);
} }
}
JobIt AsyncNavMeshUpdater::getJob(std::deque<JobIt>& jobs, bool changeLastUpdate) if (shouldStop)
{
const auto now = std::chrono::steady_clock::now();
JobIt job = jobs.front();
if (job->mProcessTime > now)
return mJobs.end(); return mJobs.end();
jobs.pop_front(); const JobIt job = mWaiting.front();
if (changeLastUpdate && job->mChangeType == ChangeType::update) mWaiting.pop_front();
mLastUpdates[getAgentAndTile(*job)] = now;
if (!lockTile(job->mAgentHalfExtents, job->mChangedTile))
{
++job->mTryNumber;
insertPrioritizedJob(job, mWaiting);
return mJobs.end();
}
if (job->mChangeType == ChangeType::update)
mLastUpdates[getAgentAndTile(*job)] = std::chrono::steady_clock::now();
mPushed.erase(getAgentAndTile(*job));
return job; return job;
} }
@ -435,7 +408,7 @@ namespace DetourNavigator
if (mPushed.emplace(job->mAgentHalfExtents, job->mChangedTile).second) if (mPushed.emplace(job->mAgentHalfExtents, job->mChangedTile).second)
{ {
++job->mTryNumber; ++job->mTryNumber;
mWaiting.push_back(job); insertPrioritizedJob(job, mWaiting);
mHasJob.notify_all(); mHasJob.notify_all();
return; return;
} }
@ -443,26 +416,11 @@ namespace DetourNavigator
mJobs.erase(job); mJobs.erase(job);
} }
void AsyncNavMeshUpdater::postThreadJob(JobIt job, std::deque<JobIt>& queue) bool AsyncNavMeshUpdater::lockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile)
{
queue.push_back(job);
mHasJob.notify_all();
}
std::thread::id AsyncNavMeshUpdater::lockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile)
{ {
if (mSettings.get().mAsyncNavMeshUpdaterThreads <= 1) if (mSettings.get().mAsyncNavMeshUpdaterThreads <= 1)
return std::this_thread::get_id(); return true;
return mProcessingTiles.lock()->emplace(agentHalfExtents, changedTile).second;
auto locked = mProcessingTiles.lock();
const auto tile = locked->find(std::make_tuple(agentHalfExtents, changedTile));
if (tile == locked->end())
{
const auto threadId = std::this_thread::get_id();
locked->emplace(std::tie(agentHalfExtents, changedTile), threadId);
return threadId;
}
return tile->second;
} }
void AsyncNavMeshUpdater::unlockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile) void AsyncNavMeshUpdater::unlockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile)

View file

@ -99,10 +99,9 @@ namespace DetourNavigator
std::set<std::tuple<osg::Vec3f, TilePosition>> mPushed; std::set<std::tuple<osg::Vec3f, TilePosition>> mPushed;
Misc::ScopeGuarded<TilePosition> mPlayerTile; Misc::ScopeGuarded<TilePosition> mPlayerTile;
NavMeshTilesCache mNavMeshTilesCache; NavMeshTilesCache mNavMeshTilesCache;
Misc::ScopeGuarded<std::map<std::tuple<osg::Vec3f, TilePosition>, std::thread::id>> mProcessingTiles; Misc::ScopeGuarded<std::set<std::tuple<osg::Vec3f, TilePosition>>> mProcessingTiles;
std::map<std::tuple<osg::Vec3f, TilePosition>, std::chrono::steady_clock::time_point> mLastUpdates; std::map<std::tuple<osg::Vec3f, TilePosition>, std::chrono::steady_clock::time_point> mLastUpdates;
std::set<std::tuple<osg::Vec3f, TilePosition>> mPresentTiles; std::set<std::tuple<osg::Vec3f, TilePosition>> mPresentTiles;
std::map<std::thread::id, std::deque<JobIt>> mThreadsQueues;
std::vector<std::thread> mThreads; std::vector<std::thread> mThreads;
void process() noexcept; void process() noexcept;
@ -119,7 +118,7 @@ namespace DetourNavigator
void repost(JobIt job); void repost(JobIt job);
std::thread::id lockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile); bool lockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile);
void unlockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile); void unlockTile(const osg::Vec3f& agentHalfExtents, const TilePosition& changedTile);