diff options
| author | git perforce import user <a@b> | 2016-10-25 12:29:14 -0600 |
|---|---|---|
| committer | Sheikh Dawood Abdul Ajees <Sheikh Dawood Abdul Ajees> | 2016-10-25 18:56:37 -0500 |
| commit | 3dfe2108cfab31ba3ee5527e217d0d8e99a51162 (patch) | |
| tree | fa6485c169e50d7415a651bf838f5bcd0fd3bfbd /PhysX_3.4/Source/Common/src/CmPreallocatingPool.h | |
| download | physx-3.4-3dfe2108cfab31ba3ee5527e217d0d8e99a51162.tar.xz physx-3.4-3dfe2108cfab31ba3ee5527e217d0d8e99a51162.zip | |
Initial commit:
PhysX 3.4.0 Update @ 21294896
APEX 1.4.0 Update @ 21275617
[CL 21300167]
Diffstat (limited to 'PhysX_3.4/Source/Common/src/CmPreallocatingPool.h')
| -rw-r--r-- | PhysX_3.4/Source/Common/src/CmPreallocatingPool.h | 435 |
1 files changed, 435 insertions, 0 deletions
diff --git a/PhysX_3.4/Source/Common/src/CmPreallocatingPool.h b/PhysX_3.4/Source/Common/src/CmPreallocatingPool.h new file mode 100644 index 00000000..da8bba92 --- /dev/null +++ b/PhysX_3.4/Source/Common/src/CmPreallocatingPool.h @@ -0,0 +1,435 @@ +// This code contains NVIDIA Confidential Information and is disclosed to you +// under a form of NVIDIA software license agreement provided separately to you. +// +// Notice +// NVIDIA Corporation and its licensors retain all intellectual property and +// proprietary rights in and to this software and related documentation and +// any modifications thereto. Any use, reproduction, disclosure, or +// distribution of this software and related documentation without an express +// license agreement from NVIDIA Corporation is strictly prohibited. +// +// ALL NVIDIA DESIGN SPECIFICATIONS, CODE ARE PROVIDED "AS IS.". NVIDIA MAKES +// NO WARRANTIES, EXPRESSED, IMPLIED, STATUTORY, OR OTHERWISE WITH RESPECT TO +// THE MATERIALS, AND EXPRESSLY DISCLAIMS ALL IMPLIED WARRANTIES OF NONINFRINGEMENT, +// MERCHANTABILITY, AND FITNESS FOR A PARTICULAR PURPOSE. +// +// Information and code furnished is believed to be accurate and reliable. +// However, NVIDIA Corporation assumes no responsibility for the consequences of use of such +// information or for any infringement of patents or other rights of third parties that may +// result from its use. No license is granted by implication or otherwise under any patent +// or patent rights of NVIDIA Corporation. Details are subject to change without notice. +// This code supersedes and replaces all information previously supplied. +// NVIDIA Corporation products are not authorized for use as critical +// components in life support devices or systems without express written approval of +// NVIDIA Corporation. +// +// Copyright (c) 2008-2016 NVIDIA Corporation. All rights reserved. +// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved. +// Copyright (c) 2001-2004 NovodeX AG. All rights reserved. + + +#ifndef PX_PHYSICS_COMMON_PREALLOCATINGPOOL +#define PX_PHYSICS_COMMON_PREALLOCATINGPOOL + +#include "foundation/Px.h" +#include "PsUserAllocated.h" +#include "CmPhysXCommon.h" +#include "PsSort.h" +#include "PsArray.h" + +/* +Pool used to allocate variable sized tasks. It's intended to be cleared after a short period (time step). +*/ + +namespace physx +{ +namespace Cm +{ + +class PreallocatingRegion +{ +public: + PX_FORCE_INLINE PreallocatingRegion() : mMemory(NULL), mFirstFree(NULL), mNbElements(0) {} + + void init(PxU32 maxElements, PxU32 elementSize, const char* typeName) + { + mFirstFree = NULL; + mNbElements = 0; + PX_ASSERT(typeName); + PX_UNUSED(typeName); + mMemory = reinterpret_cast<PxU8*>(PX_ALLOC(sizeof(PxU8)*elementSize*maxElements, typeName?typeName:"SceneSim Pool")); // ### addActor alloc + PX_ASSERT(elementSize*maxElements>=sizeof(void*)); + } + + void reset() + { + PX_FREE_AND_RESET(mMemory); + } + + PX_FORCE_INLINE PxU8* allocateMemory(PxU32 maxElements, PxU32 elementSize) + { + if(mFirstFree) + { + PxU8* recycled = reinterpret_cast<PxU8*>(mFirstFree); + + void** recycled32 = reinterpret_cast<void**>(recycled); + mFirstFree = *recycled32; + + return recycled; + } + else + { + if(mNbElements==maxElements) + return NULL; // Out of memory + + const PxU32 freeIndex = mNbElements++; + return mMemory + freeIndex * elementSize; + } + } + + void deallocateMemory(PxU32 maxElements, PxU32 elementSize, PxU8* element) + { + PX_ASSERT(element); + PX_ASSERT(element>=mMemory && element<mMemory + maxElements * elementSize); + PX_UNUSED(elementSize); + PX_UNUSED(maxElements); + + void** recycled32 = reinterpret_cast<void**>(element); + *recycled32 = mFirstFree; + + mFirstFree = element; + } + + PX_FORCE_INLINE bool operator < (const PreallocatingRegion& p) const + { + return mMemory < p.mMemory; + } + + PX_FORCE_INLINE bool operator > (const PreallocatingRegion& p) const + { + return mMemory > p.mMemory; + } + + PxU8* mMemory; + void* mFirstFree; + PxU32 mNbElements; +}; + + + +class PreallocatingRegionManager +{ + public: + PreallocatingRegionManager(PxU32 maxElements, PxU32 elementSize, const char* typeName) + : mMaxElements (maxElements) + , mElementSize (elementSize) + , mActivePoolIndex (0) + , mPools(PX_DEBUG_EXP("MyPoolManagerPools")) + , mNeedsSorting (true) + , mTypeName (typeName) + { + PreallocatingRegion tmp; + tmp.init(maxElements, elementSize, mTypeName); + mPools.pushBack(tmp); + } + + ~PreallocatingRegionManager() + { + const PxU32 nbPools = mPools.size(); + for(PxU32 i=0;i<nbPools;i++) + mPools[i].reset(); + } + + void preAllocate(PxU32 n) + { + if(!n) + return; + + const PxU32 nbPools = mPools.size(); + const PxU32 maxElements = mMaxElements; + const PxU32 elementSize = mElementSize; + PxU32 availableSpace = nbPools * maxElements; + + while(n>availableSpace) + { + PreallocatingRegion tmp; + tmp.init(maxElements, elementSize, mTypeName); + mPools.pushBack(tmp); + + availableSpace += maxElements; + } + } + + PX_FORCE_INLINE PxU8* allocateMemory() + { + PX_ASSERT(mActivePoolIndex<mPools.size()); + PxU8* memory = mPools[mActivePoolIndex].allocateMemory(mMaxElements, mElementSize); + return memory ? memory : searchForMemory(); + } + + void deallocateMemory(PxU8* element) + { + if(!element) + return; + + if(mNeedsSorting) + Ps::sort(mPools.begin(), mPools.size()); + + const PxU32 maxElements = mMaxElements; + const PxU32 elementSize = mElementSize; + const PxU32 slabSize = maxElements * elementSize; + const PxU32 nbPools = mPools.size(); + + // O(log n) search + int first = 0; + int last = int(nbPools-1); + + while(first<=last) + { + const int mid = (first+last)>>1; + + PreallocatingRegion& candidate = mPools[PxU32(mid)]; + if(contains(candidate.mMemory, slabSize, element)) + { + candidate.deallocateMemory(maxElements, elementSize, element); + + // when we sorted earlier we trashed the active index, but at least this region has a free element + if(mNeedsSorting) + mActivePoolIndex = PxU32(mid); + + mNeedsSorting = false; + return; + } + + if(candidate.mMemory<element) + first = mid+1; + else + last = mid-1; + } + + PX_ASSERT(0); + } + + +private: + + PreallocatingRegionManager& operator=(const PreallocatingRegionManager&); + PxU8* searchForMemory() + { + const PxU32 nbPools = mPools.size(); + const PxU32 activePoolIndex = mActivePoolIndex; + const PxU32 maxElements = mMaxElements; + const PxU32 elementSize = mElementSize; + + + for(PxU32 i=0;i<nbPools;i++) + { + if(i==activePoolIndex) + continue; + + PxU8* memory = mPools[i].allocateMemory(maxElements, elementSize); + if(memory) + { + mActivePoolIndex = i; + return memory; + } + } + + mActivePoolIndex = nbPools; + mNeedsSorting = true; + + PreallocatingRegion tmp; + tmp.init(maxElements, elementSize, mTypeName); + + PreallocatingRegion& newPool = mPools.pushBack(tmp); // ### addActor alloc (StaticSim, ShapeSim, SceneQueryShapeData) + return newPool.allocateMemory(maxElements, elementSize); + } + + + + PX_FORCE_INLINE bool contains(PxU8* memory, const PxU32 slabSize, PxU8* element) + { + return element>=memory && element<memory+slabSize; + } + + + + const PxU32 mMaxElements; + const PxU32 mElementSize; + PxU32 mActivePoolIndex; + + Ps::Array<PreallocatingRegion> mPools; + bool mNeedsSorting; + const char* mTypeName; +}; + +template<class T> +class PreallocatingPool : public Ps::UserAllocated +{ + PreallocatingPool<T>& operator=(const PreallocatingPool<T>&); + +public: + PreallocatingPool(PxU32 maxElements, const char* typeName) : mPool(maxElements, sizeof(T), typeName) + { + } + + ~PreallocatingPool() + { + } + + PX_FORCE_INLINE void preAllocate(PxU32 n) + { + mPool.preAllocate(n); + } + + PX_INLINE T* allocate() + { + return reinterpret_cast<T*>(mPool.allocateMemory()); + } + + PX_FORCE_INLINE T* allocateAndPrefetch() + { + T* t = reinterpret_cast<T*>(mPool.allocateMemory()); + Ps::prefetch(t, sizeof(T)); + return t; + } + + PX_INLINE T* construct() + { + T* t = reinterpret_cast<T*>(mPool.allocateMemory()); + return t ? new (t) T() : 0; + } + + template<class A1> + PX_INLINE T* construct(A1& a) + { + T* t = reinterpret_cast<T*>(mPool.allocateMemory()); + return t ? new (t) T(a) : 0; + } + + template<class A1, class A2> + PX_INLINE T* construct(A1& a, A2& b) + { + T* t = reinterpret_cast<T*>(mPool.allocateMemory()); + return t ? new (t) T(a,b) : 0; + } + + template<class A1, class A2, class A3> + PX_INLINE T* construct(A1& a, A2& b, A3& c) + { + T* t = reinterpret_cast<T*>(mPool.allocateMemory()); + return t ? new (t) T(a,b,c) : 0; + } + + template<class A1, class A2, class A3, class A4> + PX_INLINE T* construct(A1& a, A2& b, A3& c, A4& d) + { + T* t = reinterpret_cast<T*>(mPool.allocateMemory()); + return t ? new (t) T(a,b,c,d) : 0; + } + + template<class A1, class A2, class A3, class A4, class A5> + PX_INLINE T* construct(A1& a, A2& b, A3& c, A4& d, A5& e) + { + T* t = reinterpret_cast<T*>(mPool.allocateMemory()); + return t ? new (t) T(a,b,c,d,e) : 0; + } + + //// + + PX_INLINE T* construct(T* t) + { + PX_ASSERT(t); + return new (t) T(); + } + + template<class A1> + PX_INLINE T* construct(T* t, A1& a) + { + PX_ASSERT(t); + return new (t) T(a); + } + + template<class A1, class A2> + PX_INLINE T* construct(T* t, A1& a, A2& b) + { + PX_ASSERT(t); + return new (t) T(a,b); + } + + template<class A1, class A2, class A3> + PX_INLINE T* construct(T* t, A1& a, A2& b, A3& c) + { + PX_ASSERT(t); + return new (t) T(a,b,c); + } + + template<class A1, class A2, class A3, class A4> + PX_INLINE T* construct(T* t, A1& a, A2& b, A3& c, A4& d) + { + PX_ASSERT(t); + return new (t) T(a,b,c,d); + } + + template<class A1, class A2, class A3, class A4, class A5> + PX_INLINE T* construct(T* t, A1& a, A2& b, A3& c, A4& d, A5& e) + { + PX_ASSERT(t); + return new (t) T(a,b,c,d,e); + } + + PX_INLINE void destroy(T* const p) + { + if(p) + { + p->~T(); + mPool.deallocateMemory(reinterpret_cast<PxU8*>(p)); + } + } + + PX_INLINE void releasePreallocated(T* const p) + { + if(p) + mPool.deallocateMemory(reinterpret_cast<PxU8*>(p)); + } +protected: + PreallocatingRegionManager mPool; +}; + +template<class T> +class BufferedPreallocatingPool : public PreallocatingPool<T> +{ + Ps::Array<T*> mDeletedElems; + PX_NOCOPY(BufferedPreallocatingPool<T>) +public: + BufferedPreallocatingPool(PxU32 maxElements, const char* typeName) : PreallocatingPool<T>(maxElements, typeName) + { + } + + PX_INLINE void destroy(T* const p) + { + if (p) + { + p->~T(); + mDeletedElems.pushBack(p); + } + } + + void processPendingDeletedElems() + { + for (PxU32 i = 0; i < mDeletedElems.size(); ++i) + this->mPool.deallocateMemory(reinterpret_cast<PxU8*>(mDeletedElems[i])); + mDeletedElems.clear(); + } + + +}; + + + + +} // namespace Cm + +} + +#endif |