C++程序  |  338行  |  8.69 KB

// Copyright 2016 The SwiftShader Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//    http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#include "PoolAlloc.h"

#ifndef _MSC_VER
#include <stdint.h>
#endif
#include <stdio.h>
#include <stdlib.h>

#include "InitializeGlobals.h"
#include "osinclude.h"

OS_TLSIndex PoolIndex = OS_INVALID_TLS_INDEX;

bool InitializePoolIndex()
{
	assert(PoolIndex == OS_INVALID_TLS_INDEX);

	PoolIndex = OS_AllocTLSIndex();
	return PoolIndex != OS_INVALID_TLS_INDEX;
}

void FreePoolIndex()
{
	assert(PoolIndex != OS_INVALID_TLS_INDEX);

	OS_FreeTLSIndex(PoolIndex);
	PoolIndex = OS_INVALID_TLS_INDEX;
}

TPoolAllocator* GetGlobalPoolAllocator()
{
	assert(PoolIndex != OS_INVALID_TLS_INDEX);
	return static_cast<TPoolAllocator*>(OS_GetTLSValue(PoolIndex));
}

void SetGlobalPoolAllocator(TPoolAllocator* poolAllocator)
{
	assert(PoolIndex != OS_INVALID_TLS_INDEX);
	OS_SetTLSValue(PoolIndex, poolAllocator);
}

//
// Implement the functionality of the TPoolAllocator class, which
// is documented in PoolAlloc.h.
//
TPoolAllocator::TPoolAllocator(int growthIncrement, int allocationAlignment) :
	alignment(allocationAlignment)
#if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	, pageSize(growthIncrement),
	freeList(0),
	inUseList(0),
	numCalls(0),
	totalBytes(0)
#endif
{
	//
	// Adjust alignment to be at least pointer aligned and
	// power of 2.
	//
	size_t minAlign = sizeof(void*);
	alignment &= ~(minAlign - 1);
	if (alignment < minAlign)
		alignment = minAlign;
	size_t a = 1;
	while (a < alignment)
		a <<= 1;
	alignment = a;
	alignmentMask = a - 1;

#if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	//
	// Don't allow page sizes we know are smaller than all common
	// OS page sizes.
	//
	if (pageSize < 4*1024)
		pageSize = 4*1024;

	//
	// A large currentPageOffset indicates a new page needs to
	// be obtained to allocate memory.
	//
	currentPageOffset = pageSize;

	//
	// Align header skip
	//
	headerSkip = minAlign;
	if (headerSkip < sizeof(tHeader)) {
		headerSkip = (sizeof(tHeader) + alignmentMask) & ~alignmentMask;
	}
#else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	mStack.push_back({});
#endif
}

TPoolAllocator::~TPoolAllocator()
{
#if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	while (inUseList) {
		tHeader* next = inUseList->nextPage;
		inUseList->~tHeader();
		delete [] reinterpret_cast<char*>(inUseList);
		inUseList = next;
	}

	// We should not check the guard blocks
	// here, because we did it already when the block was
	// placed into the free list.
	//
	while (freeList) {
		tHeader* next = freeList->nextPage;
		delete [] reinterpret_cast<char*>(freeList);
		freeList = next;
	}
#else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	for (auto& allocs : mStack) {
		for (auto alloc : allocs) {
			free(alloc);
		}
	}
	mStack.clear();
#endif
}

// Support MSVC++ 6.0
const unsigned char TAllocation::guardBlockBeginVal = 0xfb;
const unsigned char TAllocation::guardBlockEndVal   = 0xfe;
const unsigned char TAllocation::userDataFill       = 0xcd;

#ifdef GUARD_BLOCKS
	const size_t TAllocation::guardBlockSize = 16;
#else
	const size_t TAllocation::guardBlockSize = 0;
#endif

//
// Check a single guard block for damage
//
void TAllocation::checkGuardBlock(unsigned char* blockMem, unsigned char val, const char* locText) const
{
#ifdef GUARD_BLOCKS
	for (size_t x = 0; x < guardBlockSize; x++) {
		if (blockMem[x] != val) {
			char assertMsg[80];

			// We don't print the assert message.  It's here just to be helpful.
			#if defined(_MSC_VER)
				_snprintf(assertMsg, sizeof(assertMsg), "PoolAlloc: Damage %s %Iu byte allocation at 0x%p\n",
						  locText, size, data());
			#else
				snprintf(assertMsg, sizeof(assertMsg), "PoolAlloc: Damage %s %zu byte allocation at 0x%p\n",
						 locText, size, data());
			#endif
			assert(0 && "PoolAlloc: Damage in guard block");
		}
	}
#endif
}


void TPoolAllocator::push()
{
#if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	tAllocState state = { currentPageOffset, inUseList };

	mStack.push_back(state);

	//
	// Indicate there is no current page to allocate from.
	//
	currentPageOffset = pageSize;
#else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	mStack.push_back({});
#endif
}

//
// Do a mass-deallocation of all the individual allocations
// that have occurred since the last push(), or since the
// last pop(), or since the object's creation.
//
// The deallocated pages are saved for future allocations.
//
void TPoolAllocator::pop()
{
	if (mStack.size() < 1)
		return;

#if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	tHeader* page = mStack.back().page;
	currentPageOffset = mStack.back().offset;

	while (inUseList != page) {
		// invoke destructor to free allocation list
		inUseList->~tHeader();

		tHeader* nextInUse = inUseList->nextPage;
		if (inUseList->pageCount > 1)
			delete [] reinterpret_cast<char*>(inUseList);
		else {
			inUseList->nextPage = freeList;
			freeList = inUseList;
		}
		inUseList = nextInUse;
	}

	mStack.pop_back();
#else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	for (auto alloc : mStack.back()) {
		free(alloc);
	}
	mStack.pop_back();
#endif
}

//
// Do a mass-deallocation of all the individual allocations
// that have occurred.
//
void TPoolAllocator::popAll()
{
	while (mStack.size() > 0)
		pop();
}

void* TPoolAllocator::allocate(size_t numBytes)
{
#if !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	//
	// Just keep some interesting statistics.
	//
	++numCalls;
	totalBytes += numBytes;

	// If we are using guard blocks, all allocations are bracketed by
	// them: [guardblock][allocation][guardblock].  numBytes is how
	// much memory the caller asked for.  allocationSize is the total
	// size including guard blocks.  In release build,
	// guardBlockSize=0 and this all gets optimized away.
	size_t allocationSize = TAllocation::allocationSize(numBytes);
	// Detect integer overflow.
	if (allocationSize < numBytes)
		return 0;

	//
	// Do the allocation, most likely case first, for efficiency.
	// This step could be moved to be inline sometime.
	//
	if (allocationSize <= pageSize - currentPageOffset) {
		//
		// Safe to allocate from currentPageOffset.
		//
		unsigned char* memory = reinterpret_cast<unsigned char *>(inUseList) + currentPageOffset;
		currentPageOffset += allocationSize;
		currentPageOffset = (currentPageOffset + alignmentMask) & ~alignmentMask;

		return initializeAllocation(inUseList, memory, numBytes);
	}

	if (allocationSize > pageSize - headerSkip) {
		//
		// Do a multi-page allocation.  Don't mix these with the others.
		// The OS is efficient and allocating and free-ing multiple pages.
		//
		size_t numBytesToAlloc = allocationSize + headerSkip;
		// Detect integer overflow.
		if (numBytesToAlloc < allocationSize)
			return 0;

		tHeader* memory = reinterpret_cast<tHeader*>(::new char[numBytesToAlloc]);
		if (memory == 0)
			return 0;

		// Use placement-new to initialize header
		new(memory) tHeader(inUseList, (numBytesToAlloc + pageSize - 1) / pageSize);
		inUseList = memory;

		currentPageOffset = pageSize;  // make next allocation come from a new page

		// No guard blocks for multi-page allocations (yet)
		return reinterpret_cast<void*>(reinterpret_cast<uintptr_t>(memory) + headerSkip);
	}

	//
	// Need a simple page to allocate from.
	//
	tHeader* memory;
	if (freeList) {
		memory = freeList;
		freeList = freeList->nextPage;
	} else {
		memory = reinterpret_cast<tHeader*>(::new char[pageSize]);
		if (memory == 0)
			return 0;
	}

	// Use placement-new to initialize header
	new(memory) tHeader(inUseList, 1);
	inUseList = memory;

	unsigned char* ret = reinterpret_cast<unsigned char *>(inUseList) + headerSkip;
	currentPageOffset = (headerSkip + allocationSize + alignmentMask) & ~alignmentMask;

	return initializeAllocation(inUseList, ret, numBytes);
#else  // !defined(SWIFTSHADER_TRANSLATOR_DISABLE_POOL_ALLOC)
	void *alloc = malloc(numBytes + alignmentMask);
	mStack.back().push_back(alloc);

	intptr_t intAlloc = reinterpret_cast<intptr_t>(alloc);
	intAlloc = (intAlloc + alignmentMask) & ~alignmentMask;
	return reinterpret_cast<void *>(intAlloc);
#endif
}


//
// Check all allocations in a list for damage by calling check on each.
//
void TAllocation::checkAllocList() const
{
	for (const TAllocation* alloc = this; alloc != 0; alloc = alloc->prevAlloc)
		alloc->check();
}