177 lines
6.0 KiB
C++
177 lines
6.0 KiB
C++
// Copyright 2017 The Dawn Authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "dawn_native/CommandAllocator.h"
|
|
|
|
#include "common/Assert.h"
|
|
#include "common/Math.h"
|
|
|
|
#include <algorithm>
|
|
#include <climits>
|
|
#include <cstdlib>
|
|
|
|
namespace dawn_native {
|
|
|
|
// TODO(cwallez@chromium.org): figure out a way to have more type safety for the iterator
|
|
|
|
CommandIterator::CommandIterator() {
|
|
Reset();
|
|
}
|
|
|
|
CommandIterator::~CommandIterator() {
|
|
ASSERT(IsEmpty());
|
|
}
|
|
|
|
CommandIterator::CommandIterator(CommandIterator&& other) {
|
|
if (!other.IsEmpty()) {
|
|
mBlocks = std::move(other.mBlocks);
|
|
other.Reset();
|
|
}
|
|
Reset();
|
|
}
|
|
|
|
CommandIterator& CommandIterator::operator=(CommandIterator&& other) {
|
|
ASSERT(IsEmpty());
|
|
mBlocks = std::move(other.mBlocks);
|
|
other.Reset();
|
|
Reset();
|
|
return *this;
|
|
}
|
|
|
|
CommandIterator::CommandIterator(CommandAllocator&& allocator)
|
|
: mBlocks(allocator.AcquireBlocks()) {
|
|
Reset();
|
|
}
|
|
|
|
CommandIterator& CommandIterator::operator=(CommandAllocator&& allocator) {
|
|
ASSERT(IsEmpty());
|
|
mBlocks = allocator.AcquireBlocks();
|
|
Reset();
|
|
return *this;
|
|
}
|
|
|
|
bool CommandIterator::NextCommandIdInNewBlock(uint32_t* commandId) {
|
|
mCurrentBlock++;
|
|
if (mCurrentBlock >= mBlocks.size()) {
|
|
Reset();
|
|
*commandId = detail::kEndOfBlock;
|
|
return false;
|
|
}
|
|
mCurrentPtr = AlignPtr(mBlocks[mCurrentBlock].block, alignof(uint32_t));
|
|
return NextCommandId(commandId);
|
|
}
|
|
|
|
void CommandIterator::Reset() {
|
|
mCurrentBlock = 0;
|
|
|
|
if (mBlocks.empty()) {
|
|
// This will case the first NextCommandId call to try to move to the next block and stop
|
|
// the iteration immediately, without special casing the initialization.
|
|
mCurrentPtr = reinterpret_cast<uint8_t*>(&mEndOfBlock);
|
|
mBlocks.emplace_back();
|
|
mBlocks[0].size = sizeof(mEndOfBlock);
|
|
mBlocks[0].block = mCurrentPtr;
|
|
} else {
|
|
mCurrentPtr = AlignPtr(mBlocks[0].block, alignof(uint32_t));
|
|
}
|
|
}
|
|
|
|
void CommandIterator::MakeEmptyAsDataWasDestroyed() {
|
|
if (IsEmpty()) {
|
|
return;
|
|
}
|
|
|
|
for (auto& block : mBlocks) {
|
|
free(block.block);
|
|
}
|
|
mBlocks.clear();
|
|
Reset();
|
|
ASSERT(IsEmpty());
|
|
}
|
|
|
|
bool CommandIterator::IsEmpty() const {
|
|
return mBlocks[0].block == reinterpret_cast<const uint8_t*>(&mEndOfBlock);
|
|
}
|
|
|
|
// Potential TODO(cwallez@chromium.org):
|
|
// - Host the size and pointer to next block in the block itself to avoid having an allocation
|
|
// in the vector
|
|
// - Assume T's alignof is, say 64bits, static assert it, and make commandAlignment a constant
|
|
// in Allocate
|
|
// - Be able to optimize allocation to one block, for command buffers expected to live long to
|
|
// avoid cache misses
|
|
// - Better block allocation, maybe have Dawn API to say command buffer is going to have size
|
|
// close to another
|
|
|
|
CommandAllocator::CommandAllocator()
|
|
: mCurrentPtr(reinterpret_cast<uint8_t*>(&mDummyEnum[0])),
|
|
mEndPtr(reinterpret_cast<uint8_t*>(&mDummyEnum[1])) {
|
|
}
|
|
|
|
CommandAllocator::~CommandAllocator() {
|
|
ASSERT(mBlocks.empty());
|
|
}
|
|
|
|
CommandBlocks&& CommandAllocator::AcquireBlocks() {
|
|
ASSERT(mCurrentPtr != nullptr && mEndPtr != nullptr);
|
|
ASSERT(IsPtrAligned(mCurrentPtr, alignof(uint32_t)));
|
|
ASSERT(mCurrentPtr + sizeof(uint32_t) <= mEndPtr);
|
|
*reinterpret_cast<uint32_t*>(mCurrentPtr) = detail::kEndOfBlock;
|
|
|
|
mCurrentPtr = nullptr;
|
|
mEndPtr = nullptr;
|
|
return std::move(mBlocks);
|
|
}
|
|
|
|
uint8_t* CommandAllocator::AllocateInNewBlock(uint32_t commandId,
|
|
size_t commandSize,
|
|
size_t commandAlignment) {
|
|
// When there is not enough space, we signal the kEndOfBlock, so that the iterator knows
|
|
// to move to the next one. kEndOfBlock on the last block means the end of the commands.
|
|
uint32_t* idAlloc = reinterpret_cast<uint32_t*>(mCurrentPtr);
|
|
*idAlloc = detail::kEndOfBlock;
|
|
|
|
// We'll request a block that can contain at least the command ID, the command and an
|
|
// additional ID to contain the kEndOfBlock tag.
|
|
size_t requestedBlockSize = commandSize + kWorstCaseAdditionalSize;
|
|
|
|
// The computation of the request could overflow.
|
|
if (DAWN_UNLIKELY(requestedBlockSize <= commandSize)) {
|
|
return nullptr;
|
|
}
|
|
|
|
if (DAWN_UNLIKELY(!GetNewBlock(requestedBlockSize))) {
|
|
return nullptr;
|
|
}
|
|
return Allocate(commandId, commandSize, commandAlignment);
|
|
}
|
|
|
|
bool CommandAllocator::GetNewBlock(size_t minimumSize) {
|
|
// Allocate blocks doubling sizes each time, to a maximum of 16k (or at least minimumSize).
|
|
mLastAllocationSize =
|
|
std::max(minimumSize, std::min(mLastAllocationSize * 2, size_t(16384)));
|
|
|
|
uint8_t* block = static_cast<uint8_t*>(malloc(mLastAllocationSize));
|
|
if (DAWN_UNLIKELY(block == nullptr)) {
|
|
return false;
|
|
}
|
|
|
|
mBlocks.push_back({mLastAllocationSize, block});
|
|
mCurrentPtr = AlignPtr(block, alignof(uint32_t));
|
|
mEndPtr = block + mLastAllocationSize;
|
|
return true;
|
|
}
|
|
|
|
} // namespace dawn_native
|