mirror of
https://github.com/baldurk/renderdoc.git
synced 2026-05-05 01:20:42 +00:00
1218 lines
32 KiB
C++
1218 lines
32 KiB
C++
/******************************************************************************
|
|
* The MIT License (MIT)
|
|
*
|
|
* Copyright (c) 2019-2021 Baldur Karlsson
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
******************************************************************************/
|
|
|
|
// used to avoid instantiating templates too early in the header
|
|
#define SERIALISER_IMPL
|
|
|
|
#include "serialiser.h"
|
|
#include "core/core.h"
|
|
#include "strings/string_utils.h"
|
|
|
|
#if ENABLED(RDOC_DEVEL)
|
|
|
|
int64_t Chunk::m_LiveChunks = 0;
|
|
int64_t Chunk::m_TotalMem = 0;
|
|
|
|
#endif
|
|
|
|
void DumpObject(FileIO::LogFileHandle *log, const rdcstr &indent, SDObject *obj)
|
|
{
|
|
if(obj->NumChildren() > 0)
|
|
{
|
|
rdcstr msg =
|
|
StringFormat::Fmt("%s%s%s %s:\n", indent.c_str(), obj->type.name.c_str(),
|
|
obj->type.basetype == SDBasic::Array ? "[]" : "", obj->name.c_str());
|
|
FileIO::logfile_append(log, msg.c_str(), msg.length());
|
|
for(size_t i = 0; i < obj->NumChildren(); i++)
|
|
DumpObject(log, indent + " ", obj->GetChild(i));
|
|
}
|
|
else
|
|
{
|
|
rdcstr val;
|
|
switch(obj->type.basetype)
|
|
{
|
|
case SDBasic::Chunk: val = "{Chunk}"; break;
|
|
case SDBasic::Struct: val = "{Struct}"; break;
|
|
case SDBasic::Array:
|
|
// this must be an empty array, or it would have children above
|
|
val = "{}";
|
|
break;
|
|
case SDBasic::Buffer: val = "[buffer]"; break;
|
|
case SDBasic::Null: val = "NULL"; break;
|
|
case SDBasic::String: val = obj->data.str; break;
|
|
case SDBasic::Enum: val = obj->data.str; break;
|
|
case SDBasic::UnsignedInteger: val = ToStr(obj->data.basic.u); break;
|
|
case SDBasic::SignedInteger: val = ToStr(obj->data.basic.i); break;
|
|
case SDBasic::Float: val = ToStr(obj->data.basic.d); break;
|
|
case SDBasic::Boolean: val = ToStr(obj->data.basic.b); break;
|
|
case SDBasic::Character: val = ToStr(obj->data.basic.c); break;
|
|
case SDBasic::Resource: val = ToStr(obj->data.basic.id); break;
|
|
}
|
|
rdcstr msg = StringFormat::Fmt("%s%s %s = %s\n", indent.c_str(), obj->type.name.c_str(),
|
|
obj->name.c_str(), val.c_str());
|
|
FileIO::logfile_append(log, msg.c_str(), msg.length());
|
|
}
|
|
}
|
|
|
|
void DumpChunk(bool reading, FileIO::LogFileHandle *log, SDChunk *chunk)
|
|
{
|
|
rdcstr msg = StringFormat::Fmt("%s %s @ %llu:\n", reading ? "Read" : "Wrote", chunk->name.c_str(),
|
|
chunk->metadata.timestampMicro);
|
|
FileIO::logfile_append(log, msg.c_str(), msg.length());
|
|
DumpObject(log, " ", chunk);
|
|
}
|
|
|
|
/////////////////////////////////////////////////////////////
|
|
// Read Serialiser functions
|
|
|
|
template <>
|
|
Serialiser<SerialiserMode::Reading>::Serialiser(StreamReader *reader, Ownership own,
|
|
SDObject *rootStructuredObj)
|
|
{
|
|
m_Read = reader;
|
|
m_Write = NULL;
|
|
|
|
m_Ownership = own;
|
|
|
|
if(rootStructuredObj)
|
|
m_StructureStack.push_back(rootStructuredObj);
|
|
}
|
|
|
|
template <>
|
|
Serialiser<SerialiserMode::Reading>::~Serialiser()
|
|
{
|
|
if(m_Ownership == Ownership::Stream && m_Read)
|
|
delete m_Read;
|
|
}
|
|
|
|
template <>
|
|
uint32_t Serialiser<SerialiserMode::Reading>::BeginChunk(uint32_t, uint64_t)
|
|
{
|
|
uint32_t chunkID = 0;
|
|
|
|
m_ChunkMetadata = SDChunkMetaData();
|
|
|
|
{
|
|
uint32_t c = 0;
|
|
bool success = m_Read->Read(c);
|
|
|
|
// Chunk index 0 is not allowed in normal situations, and allows us to indicate some control
|
|
// bytes. Currently this is unused
|
|
RDCASSERT(c != 0 || !success);
|
|
|
|
chunkID = c & ChunkIndexMask;
|
|
|
|
/////////////////
|
|
|
|
m_ChunkMetadata.chunkID = chunkID;
|
|
|
|
if(c & ChunkCallstack)
|
|
{
|
|
uint32_t numFrames = 0;
|
|
m_Read->Read(numFrames);
|
|
|
|
// try to sanity check the number of frames
|
|
if(numFrames < 4096)
|
|
{
|
|
m_ChunkMetadata.flags |= SDChunkFlags::HasCallstack;
|
|
|
|
m_ChunkMetadata.callstack.resize((size_t)numFrames);
|
|
m_Read->Read(m_ChunkMetadata.callstack.data(), m_ChunkMetadata.callstack.byteSize());
|
|
}
|
|
else
|
|
{
|
|
RDCERR("Read invalid number of callstack frames: %u", numFrames);
|
|
// still read the size that we should, even though we expect this to be broken after here
|
|
m_Read->Read(NULL, numFrames * sizeof(uint64_t));
|
|
}
|
|
}
|
|
|
|
if(c & ChunkThreadID)
|
|
m_Read->Read(m_ChunkMetadata.threadID);
|
|
|
|
if(c & ChunkDuration)
|
|
{
|
|
m_Read->Read(m_ChunkMetadata.durationMicro);
|
|
if(m_TimerFrequency != 1.0)
|
|
m_ChunkMetadata.durationMicro =
|
|
int64_t(double(m_ChunkMetadata.durationMicro) / m_TimerFrequency);
|
|
}
|
|
|
|
if(c & ChunkTimestamp)
|
|
{
|
|
m_Read->Read(m_ChunkMetadata.timestampMicro);
|
|
if(m_TimerFrequency != 1.0 || m_TimerBase != 0)
|
|
m_ChunkMetadata.timestampMicro =
|
|
int64_t(double(m_ChunkMetadata.timestampMicro - m_TimerBase) / m_TimerFrequency);
|
|
}
|
|
|
|
if(c & Chunk64BitSize)
|
|
{
|
|
m_Read->Read(m_ChunkMetadata.length);
|
|
}
|
|
else
|
|
{
|
|
uint32_t chunkSize = 0;
|
|
m_Read->Read(chunkSize);
|
|
m_ChunkMetadata.length = chunkSize;
|
|
}
|
|
|
|
m_LastChunkOffset = m_Read->GetOffset();
|
|
}
|
|
|
|
if(ExportStructure())
|
|
{
|
|
rdcstr name = m_ChunkLookup ? m_ChunkLookup(chunkID) : "";
|
|
|
|
if(name.empty())
|
|
name = "<Unknown Chunk>";
|
|
|
|
SDChunk *chunk = new SDChunk(name);
|
|
chunk->metadata = m_ChunkMetadata;
|
|
|
|
m_StructuredFile->chunks.push_back(chunk);
|
|
m_StructureStack.push_back(chunk);
|
|
|
|
m_InternalElement = 0;
|
|
}
|
|
|
|
return chunkID;
|
|
}
|
|
|
|
template <>
|
|
void Serialiser<SerialiserMode::Reading>::SkipCurrentChunk()
|
|
{
|
|
if(ExportStructure())
|
|
{
|
|
RDCASSERTMSG("Skipping chunk after we've begun serialising!", m_StructureStack.size() == 1,
|
|
m_StructureStack.size());
|
|
|
|
SDObject ¤t = *m_StructureStack.back();
|
|
|
|
SDObject &obj = *current.AddAndOwnChild(new SDObject("Opaque chunk"_lit, "Byte Buffer"_lit));
|
|
|
|
obj.type.basetype = SDBasic::Buffer;
|
|
obj.type.byteSize = m_ChunkMetadata.length;
|
|
|
|
if(m_StructureStack.size() == 1)
|
|
{
|
|
SDChunk *chunk = (SDChunk *)m_StructureStack.back();
|
|
chunk->metadata.flags |= SDChunkFlags::OpaqueChunk;
|
|
}
|
|
}
|
|
|
|
{
|
|
uint64_t readBytes = m_Read->GetOffset() - m_LastChunkOffset;
|
|
|
|
if(readBytes > m_ChunkMetadata.length)
|
|
{
|
|
RDCERR("Can't skip current chunk outside of {BeginChunk, EndChunk}");
|
|
return;
|
|
}
|
|
|
|
if(readBytes > 0)
|
|
{
|
|
RDCWARN("Partially consumed bytes at SkipCurrentChunk - blob data will be truncated");
|
|
}
|
|
|
|
uint64_t chunkBytes = m_ChunkMetadata.length - readBytes;
|
|
|
|
if(ExportStructure() && m_ExportBuffers)
|
|
{
|
|
SDObject ¤t = *m_StructureStack.back();
|
|
|
|
SDObject &obj = *current.GetChild(current.NumChildren() - 1);
|
|
|
|
obj.data.basic.u = m_StructuredFile->buffers.size();
|
|
|
|
bytebuf *alloc = new bytebuf;
|
|
alloc->resize((size_t)chunkBytes);
|
|
m_Read->Read(alloc->data(), (size_t)chunkBytes);
|
|
|
|
m_StructuredFile->buffers.push_back(alloc);
|
|
}
|
|
else
|
|
{
|
|
m_Read->SkipBytes(chunkBytes);
|
|
}
|
|
}
|
|
}
|
|
|
|
template <>
|
|
void Serialiser<SerialiserMode::Reading>::EndChunk()
|
|
{
|
|
if(ExportStructure())
|
|
{
|
|
RDCASSERTMSG("Object Stack is imbalanced!", m_StructureStack.size() <= 1,
|
|
m_StructureStack.size());
|
|
|
|
if(!m_StructureStack.empty())
|
|
{
|
|
m_StructureStack.back()->type.byteSize = m_ChunkMetadata.length;
|
|
m_StructureStack.pop_back();
|
|
}
|
|
|
|
if(m_DebugDumpLog && !m_StructuredFile->chunks.empty())
|
|
{
|
|
DumpChunk(true, m_DebugDumpLog, m_StructuredFile->chunks.back());
|
|
}
|
|
}
|
|
|
|
// only skip remaining bytes if we have a valid length - if we have a length of 0 we wrote this
|
|
// chunk in 'streaming mode' (see SetStreamingMode and the Writing EndChunk() impl) so there's
|
|
// nothing to skip.
|
|
if(m_ChunkMetadata.length > 0 && !m_Read->IsErrored())
|
|
{
|
|
// this will be a no-op if the last chunk length was accurate. If it was a
|
|
// conservative estimate of the length then we'll skip some padding bytes
|
|
uint64_t readBytes = m_Read->GetOffset() - m_LastChunkOffset;
|
|
|
|
if(m_ChunkMetadata.length < readBytes)
|
|
{
|
|
RDCERR(
|
|
"!!! "
|
|
"READ %llu BYTES, OVERRUNNING CHUNK LENGTH %u. "
|
|
"CAPTURE IS CORRUPTED, OR REPLAY MISMATCHED CAPTURED CHUNK. "
|
|
"!!!",
|
|
readBytes, m_ChunkMetadata.length);
|
|
}
|
|
else
|
|
{
|
|
m_Read->SkipBytes(size_t(m_ChunkMetadata.length - readBytes));
|
|
}
|
|
}
|
|
|
|
// align to the natural chunk alignment
|
|
m_Read->AlignTo<ChunkAlignment>();
|
|
}
|
|
|
|
/////////////////////////////////////////////////////////////
|
|
// Write Serialiser functions
|
|
|
|
template <>
|
|
Serialiser<SerialiserMode::Writing>::Serialiser(StreamWriter *writer, Ownership own)
|
|
{
|
|
m_Write = writer;
|
|
m_Read = NULL;
|
|
|
|
m_Ownership = own;
|
|
}
|
|
|
|
template <>
|
|
Serialiser<SerialiserMode::Writing>::~Serialiser()
|
|
{
|
|
if(m_Ownership == Ownership::Stream && m_Write)
|
|
{
|
|
m_Write->Finish();
|
|
delete m_Write;
|
|
}
|
|
}
|
|
|
|
template <>
|
|
void Serialiser<SerialiserMode::Writing>::SetChunkMetadataRecording(uint32_t flags)
|
|
{
|
|
// cannot change this mid-chunk
|
|
RDCASSERT(m_Write->GetOffset() == 0);
|
|
|
|
m_ChunkFlags = flags;
|
|
}
|
|
|
|
template <>
|
|
uint32_t Serialiser<SerialiserMode::Writing>::BeginChunk(uint32_t chunkID, uint64_t byteLength)
|
|
{
|
|
// cannot start a chunk inside a chunk
|
|
RDCASSERTMSG("Beginning a chunk inside another chunk", m_ChunkMetadata.chunkID == 0,
|
|
m_ChunkMetadata.chunkID);
|
|
|
|
{
|
|
// chunk index needs to be valid
|
|
RDCASSERT(chunkID > 0);
|
|
|
|
{
|
|
uint32_t c = chunkID & ChunkIndexMask;
|
|
RDCASSERT(chunkID <= ChunkIndexMask);
|
|
|
|
c |= m_ChunkFlags;
|
|
if(byteLength > 0xffffffff)
|
|
c |= Chunk64BitSize;
|
|
|
|
m_ChunkMetadata.chunkID = chunkID;
|
|
|
|
/////////////////
|
|
|
|
m_Write->Write(c);
|
|
|
|
if(c & ChunkCallstack)
|
|
{
|
|
if(m_ChunkMetadata.callstack.empty())
|
|
{
|
|
bool collect = RenderDoc::Inst().GetCaptureOptions().captureCallstacks;
|
|
|
|
if(RenderDoc::Inst().GetCaptureOptions().captureCallstacksOnlyActions)
|
|
collect = collect && m_ActionChunk;
|
|
|
|
if(collect)
|
|
{
|
|
Callstack::Stackwalk *stack = Callstack::Collect();
|
|
if(stack && stack->NumLevels() > 0)
|
|
{
|
|
m_ChunkMetadata.callstack.assign(stack->GetAddrs(), stack->NumLevels());
|
|
}
|
|
|
|
SAFE_DELETE(stack);
|
|
}
|
|
}
|
|
|
|
m_ChunkMetadata.flags |= SDChunkFlags::HasCallstack;
|
|
|
|
uint32_t numFrames = (uint32_t)m_ChunkMetadata.callstack.size();
|
|
m_Write->Write(numFrames);
|
|
|
|
m_Write->Write(m_ChunkMetadata.callstack.data(), m_ChunkMetadata.callstack.byteSize());
|
|
}
|
|
|
|
if(c & ChunkThreadID)
|
|
{
|
|
if(m_ChunkMetadata.threadID == 0)
|
|
m_ChunkMetadata.threadID = Threading::GetCurrentID();
|
|
|
|
m_Write->Write(m_ChunkMetadata.threadID);
|
|
}
|
|
|
|
if(c & ChunkDuration)
|
|
{
|
|
if(m_ChunkMetadata.durationMicro < 0)
|
|
m_ChunkMetadata.durationMicro = 0;
|
|
m_Write->Write(m_ChunkMetadata.durationMicro);
|
|
}
|
|
|
|
if(c & ChunkTimestamp)
|
|
{
|
|
if(m_ChunkMetadata.timestampMicro == 0)
|
|
m_ChunkMetadata.timestampMicro = Timing::GetTick();
|
|
|
|
m_Write->Write(m_ChunkMetadata.timestampMicro);
|
|
}
|
|
|
|
if(byteLength > 0 || m_DataStreaming)
|
|
{
|
|
// write length, assuming it is an upper bound
|
|
m_ChunkFixup = 0;
|
|
RDCASSERT(byteLength < 0x100000000 || (c & Chunk64BitSize) != 0);
|
|
if(c & Chunk64BitSize)
|
|
{
|
|
m_Write->Write(byteLength);
|
|
}
|
|
else
|
|
{
|
|
m_Write->Write(uint32_t(byteLength & 0xffffffff));
|
|
}
|
|
m_LastChunkOffset = m_Write->GetOffset();
|
|
m_ChunkMetadata.length = byteLength;
|
|
}
|
|
else
|
|
{
|
|
// length will be fixed up in EndChunk
|
|
// assume that this case will not produce chunks with size larger than can fit in 32 bit
|
|
// value
|
|
uint32_t chunkSize = 0xbeebfeed;
|
|
m_ChunkFixup = m_Write->GetOffset();
|
|
m_Write->Write(chunkSize);
|
|
}
|
|
}
|
|
}
|
|
|
|
if(ExportStructure())
|
|
{
|
|
rdcstr name = m_ChunkLookup ? m_ChunkLookup(chunkID) : "";
|
|
|
|
if(name.empty())
|
|
name = "<Unknown Chunk>";
|
|
|
|
SDChunk *chunk = new SDChunk(name);
|
|
chunk->metadata = m_ChunkMetadata;
|
|
|
|
m_StructuredFile->chunks.push_back(chunk);
|
|
m_StructureStack.push_back(chunk);
|
|
|
|
m_InternalElement = 0;
|
|
}
|
|
|
|
return chunkID;
|
|
}
|
|
|
|
template <>
|
|
void Serialiser<SerialiserMode::Writing>::EndChunk()
|
|
{
|
|
m_ActionChunk = false;
|
|
|
|
if(m_DataStreaming)
|
|
{
|
|
// nothing to fixup, length is unused
|
|
}
|
|
else if(m_ChunkFixup != 0)
|
|
{
|
|
// fix up the chunk header
|
|
uint64_t chunkOffset = m_ChunkFixup;
|
|
m_ChunkFixup = 0;
|
|
|
|
uint64_t curOffset = m_Write->GetOffset();
|
|
|
|
RDCASSERT(curOffset > chunkOffset);
|
|
|
|
uint64_t chunkLength = (curOffset - chunkOffset) - sizeof(uint32_t);
|
|
if(chunkLength > 0xffffffff)
|
|
{
|
|
RDCERR("!!! CHUNK LENGTH %llu EXCEEDED 32 BIT VALUE. CAPTURE WILL BE CORRUPTED. !!!",
|
|
chunkLength);
|
|
}
|
|
|
|
m_Write->WriteAt(chunkOffset, uint32_t(chunkLength & 0xffffffff));
|
|
|
|
m_ChunkMetadata.length = chunkLength;
|
|
}
|
|
else
|
|
{
|
|
uint64_t writtenLength = (m_Write->GetOffset() - m_LastChunkOffset);
|
|
|
|
if(writtenLength < m_ChunkMetadata.length)
|
|
{
|
|
uint64_t numPadBytes = m_ChunkMetadata.length - writtenLength;
|
|
|
|
// need to write some padding bytes so that the length is accurate
|
|
for(uint64_t i = 0; i < numPadBytes; i++)
|
|
{
|
|
byte padByte = 0xbb;
|
|
m_Write->Write(padByte);
|
|
}
|
|
|
|
// only log if there's more than 128 bytes of padding
|
|
if(m_ChunkMetadata.length - writtenLength > 128)
|
|
{
|
|
RDCDEBUG("Chunk estimated at %llu bytes, actual length %llu. Added %llu bytes padding.",
|
|
m_ChunkMetadata.length, writtenLength, numPadBytes);
|
|
}
|
|
}
|
|
else if(writtenLength > m_ChunkMetadata.length)
|
|
{
|
|
RDCERR(
|
|
"!!! "
|
|
"ESTIMATED UPPER BOUND CHUNK LENGTH %llu EXCEEDED: %llu. "
|
|
"CAPTURE WILL BE CORRUPTED. "
|
|
"!!!",
|
|
m_ChunkMetadata.length, writtenLength);
|
|
}
|
|
else
|
|
{
|
|
// RDCDEBUG("Chunk was exactly the estimate of %llu bytes.", m_ChunkMetadata.length);
|
|
}
|
|
}
|
|
|
|
if(ExportStructure())
|
|
{
|
|
RDCASSERTMSG("Object Stack is imbalanced!", m_StructureStack.size() <= 1,
|
|
m_StructureStack.size());
|
|
|
|
if(!m_StructureStack.empty())
|
|
{
|
|
m_StructureStack.back()->type.byteSize = m_ChunkMetadata.length;
|
|
m_StructureStack.pop_back();
|
|
}
|
|
|
|
if(m_DebugDumpLog && !m_StructuredFile->chunks.empty())
|
|
{
|
|
DumpChunk(false, m_DebugDumpLog, m_StructuredFile->chunks.back());
|
|
}
|
|
}
|
|
|
|
// align to the natural chunk alignment
|
|
m_Write->AlignTo<ChunkAlignment>();
|
|
|
|
m_ChunkMetadata = SDChunkMetaData();
|
|
|
|
m_Write->Flush();
|
|
}
|
|
|
|
template <>
|
|
void Serialiser<SerialiserMode::Writing>::WriteStructuredFile(const SDFile &file,
|
|
RENDERDOC_ProgressCallback progress)
|
|
{
|
|
Serialiser<SerialiserMode::Writing> scratchWriter(
|
|
new StreamWriter(StreamWriter::DefaultScratchSize), Ownership::Stream);
|
|
|
|
// slightly cheeky to cast away the const, but we don't modify it in a writing serialiser
|
|
scratchWriter.m_StructuredFile = m_StructuredFile = (SDFile *)&file;
|
|
|
|
for(size_t i = 0; i < file.chunks.size(); i++)
|
|
{
|
|
const SDChunk &chunk = *file.chunks[i];
|
|
|
|
m_ChunkMetadata = chunk.metadata;
|
|
|
|
m_ChunkFlags = 0;
|
|
|
|
if(m_ChunkMetadata.flags & SDChunkFlags::HasCallstack)
|
|
m_ChunkFlags |= ChunkCallstack;
|
|
|
|
if(m_ChunkMetadata.threadID != 0)
|
|
m_ChunkFlags |= ChunkThreadID;
|
|
|
|
if(m_ChunkMetadata.durationMicro >= 0)
|
|
m_ChunkFlags |= ChunkDuration;
|
|
|
|
if(m_ChunkMetadata.timestampMicro != 0)
|
|
m_ChunkFlags |= ChunkTimestamp;
|
|
|
|
Serialiser<SerialiserMode::Writing> *ser = this;
|
|
|
|
if(m_ChunkMetadata.length == 0)
|
|
{
|
|
ser = &scratchWriter;
|
|
scratchWriter.m_ChunkMetadata = m_ChunkMetadata;
|
|
scratchWriter.m_ChunkFlags = m_ChunkFlags;
|
|
}
|
|
|
|
m_ChunkMetadata.chunkID = 0;
|
|
|
|
ser->BeginChunk(chunk.metadata.chunkID, m_ChunkMetadata.length);
|
|
|
|
if(chunk.metadata.flags & SDChunkFlags::OpaqueChunk)
|
|
{
|
|
RDCASSERT(chunk.NumChildren() == 1);
|
|
|
|
size_t bufID = (size_t)chunk.GetChild(0)->data.basic.u;
|
|
byte *ptr = m_StructuredFile->buffers[bufID]->data();
|
|
size_t len = m_StructuredFile->buffers[bufID]->size();
|
|
|
|
ser->GetWriter()->Write(ptr, len);
|
|
}
|
|
else
|
|
{
|
|
for(size_t o = 0; o < chunk.NumChildren(); o++)
|
|
{
|
|
// note, we don't need names because we aren't exporting structured data
|
|
ser->Serialise(""_lit, (SDObject *)chunk.GetChild(o));
|
|
}
|
|
}
|
|
|
|
ser->EndChunk();
|
|
|
|
if(m_ChunkMetadata.length == 0)
|
|
{
|
|
m_Write->Write(scratchWriter.GetWriter()->GetData(), scratchWriter.GetWriter()->GetOffset());
|
|
scratchWriter.GetWriter()->Rewind();
|
|
}
|
|
|
|
if(progress)
|
|
progress(float(i) / float(file.chunks.size()));
|
|
}
|
|
|
|
if(progress)
|
|
progress(1.0f);
|
|
|
|
m_StructuredFile = &m_StructData;
|
|
scratchWriter.m_StructuredFile = &scratchWriter.m_StructData;
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const SDTypeFlags &el)
|
|
{
|
|
BEGIN_BITFIELD_STRINGISE(SDTypeFlags);
|
|
{
|
|
STRINGISE_BITFIELD_CLASS_VALUE(NoFlags);
|
|
|
|
STRINGISE_BITFIELD_CLASS_BIT(HasCustomString);
|
|
STRINGISE_BITFIELD_CLASS_BIT(Hidden);
|
|
STRINGISE_BITFIELD_CLASS_BIT(Nullable);
|
|
STRINGISE_BITFIELD_CLASS_BIT(NullString);
|
|
STRINGISE_BITFIELD_CLASS_BIT(FixedArray);
|
|
STRINGISE_BITFIELD_CLASS_BIT(Union);
|
|
}
|
|
END_BITFIELD_STRINGISE();
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const SDChunkFlags &el)
|
|
{
|
|
BEGIN_BITFIELD_STRINGISE(SDChunkFlags);
|
|
{
|
|
STRINGISE_BITFIELD_CLASS_VALUE(NoFlags);
|
|
|
|
STRINGISE_BITFIELD_CLASS_BIT(OpaqueChunk);
|
|
STRINGISE_BITFIELD_CLASS_BIT(HasCallstack);
|
|
}
|
|
END_BITFIELD_STRINGISE();
|
|
}
|
|
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDType &el)
|
|
{
|
|
SERIALISE_MEMBER(name);
|
|
SERIALISE_MEMBER(basetype);
|
|
SERIALISE_MEMBER(flags);
|
|
SERIALISE_MEMBER(byteSize);
|
|
}
|
|
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDChunkMetaData &el)
|
|
{
|
|
SERIALISE_MEMBER(chunkID);
|
|
SERIALISE_MEMBER(flags);
|
|
SERIALISE_MEMBER(length);
|
|
SERIALISE_MEMBER(threadID);
|
|
SERIALISE_MEMBER(durationMicro);
|
|
SERIALISE_MEMBER(timestampMicro);
|
|
SERIALISE_MEMBER(callstack);
|
|
}
|
|
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDObjectPODData &el)
|
|
{
|
|
SERIALISE_MEMBER(u);
|
|
}
|
|
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDObjectData &el)
|
|
{
|
|
SERIALISE_MEMBER(basic);
|
|
SERIALISE_MEMBER(str);
|
|
|
|
// this is deliberately not serialised here, it's serialised in the parent SDObject. See below
|
|
// SERIALISE_MEMBER(children);
|
|
}
|
|
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDObject &el, StructuredObjectList &children)
|
|
{
|
|
// serialising the data above doesn't serialise the children, so we can do it here using a
|
|
// potential lazy generator. This is so that we don't incur the full cost of populating lazy
|
|
// children all at once (which could be slow). This is a bit of a hack as this can take many
|
|
// seconds and cause a timeout during transfer, and it would be uglier to try and keep the
|
|
// connection alive while serialising chunks.
|
|
uint64_t childCount = children.size();
|
|
SERIALISE_ELEMENT(childCount).Hidden();
|
|
|
|
if(ser.IsReading())
|
|
children.resize((size_t)childCount);
|
|
|
|
for(size_t c = 0; c < el.NumChildren(); c++)
|
|
{
|
|
// we also assume that the caller serialising these objects will handle lifetime management.
|
|
if(ser.IsReading())
|
|
children[c] = new SDObject(""_lit, ""_lit);
|
|
else
|
|
el.PopulateChild(c);
|
|
|
|
ser.Serialise("$el"_lit, *children[c]);
|
|
|
|
if(ser.IsReading())
|
|
children[c]->m_Parent = ⪙
|
|
}
|
|
}
|
|
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDObject &el)
|
|
{
|
|
SERIALISE_MEMBER(name);
|
|
SERIALISE_MEMBER(type);
|
|
SERIALISE_MEMBER(data);
|
|
|
|
DoSerialise(ser, el, el.data.children);
|
|
}
|
|
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDChunk &el)
|
|
{
|
|
SERIALISE_MEMBER(name);
|
|
SERIALISE_MEMBER(type);
|
|
SERIALISE_MEMBER(metadata);
|
|
SERIALISE_MEMBER(data);
|
|
|
|
DoSerialise(ser, el, el.data.children);
|
|
}
|
|
|
|
INSTANTIATE_SERIALISE_TYPE(SDChunk);
|
|
|
|
// serialise the pointer version - special case for writing a structured file, so can assume writing
|
|
template <class SerialiserType>
|
|
void DoSerialise(SerialiserType &ser, SDObject *el)
|
|
{
|
|
// clang barfs if we try to do ser.IsWriting() here for some reason, claiming it isn't static and
|
|
// const enough for static_assert. As a workaround we call IsWriting as a static member of the
|
|
// type itself, and that works.
|
|
RDCCOMPILE_ASSERT(SerialiserType::IsWriting(),
|
|
"SDObject pointer only supported for writing serialisation");
|
|
if(el->type.flags & SDTypeFlags::Nullable)
|
|
{
|
|
bool present = el->type.basetype != SDBasic::Null;
|
|
ser.Serialise(""_lit, present);
|
|
}
|
|
|
|
const SDFile &file = ser.GetStructuredFile();
|
|
|
|
switch(el->type.basetype)
|
|
{
|
|
case SDBasic::Chunk: RDCERR("Unexpected chunk inside object!"); break;
|
|
case SDBasic::Struct:
|
|
for(size_t o = 0; o < el->NumChildren(); o++)
|
|
ser.Serialise(""_lit, el->GetChild(o));
|
|
break;
|
|
case SDBasic::Array:
|
|
{
|
|
uint64_t arraySize = el->NumChildren();
|
|
ser.Serialise(""_lit, arraySize);
|
|
// ensure all children are ready
|
|
for(size_t o = 0; o < el->NumChildren(); o++)
|
|
ser.Serialise(""_lit, el->GetChild(o));
|
|
break;
|
|
}
|
|
case SDBasic::Null:
|
|
// nothing to do, we serialised present flag above
|
|
RDCASSERT(el->type.flags & SDTypeFlags::Nullable);
|
|
break;
|
|
case SDBasic::Buffer:
|
|
{
|
|
size_t bufID = (size_t)el->data.basic.u;
|
|
byte *buf = file.buffers[bufID]->data();
|
|
uint64_t size = file.buffers[bufID]->size();
|
|
ser.Serialise(""_lit, buf, size);
|
|
break;
|
|
}
|
|
case SDBasic::String:
|
|
{
|
|
if(el->type.flags & SDTypeFlags::NullString)
|
|
{
|
|
const char *nullstring = NULL;
|
|
ser.Serialise(""_lit, nullstring);
|
|
}
|
|
else
|
|
{
|
|
ser.Serialise(""_lit, el->data.str);
|
|
}
|
|
break;
|
|
}
|
|
case SDBasic::Enum:
|
|
{
|
|
uint32_t e = (uint32_t)el->data.basic.u;
|
|
ser.Serialise(""_lit, e);
|
|
break;
|
|
}
|
|
case SDBasic::Boolean: ser.Serialise(""_lit, el->data.basic.b); break;
|
|
case SDBasic::Character: ser.Serialise(""_lit, el->data.basic.c); break;
|
|
case SDBasic::Resource: ser.Serialise(""_lit, el->data.basic.id); break;
|
|
case SDBasic::UnsignedInteger:
|
|
if(el->type.byteSize == 1)
|
|
{
|
|
uint8_t u = uint8_t(el->data.basic.u);
|
|
ser.Serialise(""_lit, u);
|
|
}
|
|
else if(el->type.byteSize == 2)
|
|
{
|
|
uint16_t u = uint16_t(el->data.basic.u);
|
|
ser.Serialise(""_lit, u);
|
|
}
|
|
else if(el->type.byteSize == 4)
|
|
{
|
|
uint32_t u = uint32_t(el->data.basic.u);
|
|
ser.Serialise(""_lit, u);
|
|
}
|
|
else if(el->type.byteSize == 8)
|
|
{
|
|
ser.Serialise(""_lit, el->data.basic.u);
|
|
}
|
|
else
|
|
{
|
|
RDCERR("Unexpeted integer size %u", el->type.byteSize);
|
|
}
|
|
break;
|
|
case SDBasic::SignedInteger:
|
|
if(el->type.byteSize == 1)
|
|
{
|
|
int8_t i = int8_t(el->data.basic.i);
|
|
ser.Serialise(""_lit, i);
|
|
}
|
|
else if(el->type.byteSize == 2)
|
|
{
|
|
int16_t i = int16_t(el->data.basic.i);
|
|
ser.Serialise(""_lit, i);
|
|
}
|
|
else if(el->type.byteSize == 4)
|
|
{
|
|
int32_t i = int32_t(el->data.basic.i);
|
|
ser.Serialise(""_lit, i);
|
|
}
|
|
else if(el->type.byteSize == 8)
|
|
{
|
|
ser.Serialise(""_lit, el->data.basic.i);
|
|
}
|
|
else
|
|
{
|
|
RDCERR("Unexpeted integer size %u", el->type.byteSize);
|
|
}
|
|
break;
|
|
case SDBasic::Float:
|
|
if(el->type.byteSize == 4)
|
|
{
|
|
float f = float(el->data.basic.d);
|
|
ser.Serialise(""_lit, f);
|
|
}
|
|
else if(el->type.byteSize == 8)
|
|
{
|
|
ser.Serialise(""_lit, el->data.basic.d);
|
|
}
|
|
else
|
|
{
|
|
RDCERR("Unexpeted float size %u", el->type.byteSize);
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
/////////////////////////////////////////////////////////////
|
|
// Basic types
|
|
|
|
template <>
|
|
rdcstr DoStringise(const rdcstr &el)
|
|
{
|
|
return el;
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const rdcinflexiblestr &el)
|
|
{
|
|
return el;
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const rdcliteral &el)
|
|
{
|
|
return el;
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(void *const &el)
|
|
{
|
|
return StringFormat::Fmt("%#p", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const int64_t &el)
|
|
{
|
|
return StringFormat::Fmt("%lld", el);
|
|
}
|
|
|
|
#if ENABLED(RDOC_SIZET_SEP_TYPE)
|
|
template <>
|
|
rdcstr DoStringise(const size_t &el)
|
|
{
|
|
return StringFormat::Fmt("%llu", (uint64_t)el);
|
|
}
|
|
#endif
|
|
|
|
template <>
|
|
rdcstr DoStringise(const uint64_t &el)
|
|
{
|
|
return StringFormat::Fmt("%llu", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const uint32_t &el)
|
|
{
|
|
return StringFormat::Fmt("%u", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const char &el)
|
|
{
|
|
return StringFormat::Fmt("'%c'", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const wchar_t &el)
|
|
{
|
|
return StringFormat::Fmt("'%lc'", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const byte &el)
|
|
{
|
|
return StringFormat::Fmt("%hhu", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const uint16_t &el)
|
|
{
|
|
return StringFormat::Fmt("%hu", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const int32_t &el)
|
|
{
|
|
return StringFormat::Fmt("%d", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const int16_t &el)
|
|
{
|
|
return StringFormat::Fmt("%hd", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const float &el)
|
|
{
|
|
return StringFormat::Fmt("%0.4f", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const double &el)
|
|
{
|
|
return StringFormat::Fmt("%0.4lf", el);
|
|
}
|
|
|
|
template <>
|
|
rdcstr DoStringise(const bool &el)
|
|
{
|
|
if(el)
|
|
return "True";
|
|
|
|
return "False";
|
|
}
|
|
|
|
Chunk *Chunk::Create(Serialiser<SerialiserMode::Writing> &ser, uint16_t chunkType,
|
|
ChunkAllocator *allocator)
|
|
{
|
|
RDCCOMPILE_ASSERT(sizeof(Chunk) <= 16, "Chunk should be no more than 16 bytes");
|
|
|
|
RDCASSERT(ser.GetWriter()->GetOffset() < 0xffffffff);
|
|
uint32_t length = (uint32_t)ser.GetWriter()->GetOffset();
|
|
|
|
byte *data = NULL;
|
|
if(allocator)
|
|
{
|
|
// try to allocate from the allocator
|
|
data = allocator->AllocAlignedBuffer(length);
|
|
|
|
// if we couldn't satisfy the allocation then pretend we never had an allocator in the first
|
|
// place. We'll externally allocate the chunk and the data.
|
|
if(!data)
|
|
allocator = NULL;
|
|
}
|
|
|
|
// if we don't have an allocator or we gave up on it above, allocate the data externally
|
|
if(!allocator)
|
|
data = AllocAlignedBuffer(length);
|
|
|
|
memcpy(data, ser.GetWriter()->GetData(), (size_t)length);
|
|
|
|
ser.GetWriter()->Rewind();
|
|
|
|
Chunk *ret = NULL;
|
|
|
|
// if allocator wasn't NULL'd above, use it to allocate the chunk as well. We always either
|
|
// allocate both chunk and data from the allocator (so we don't have anything to do on
|
|
// destruction) or neither. Otherwise if we allocated the chunk from the allocator and the data
|
|
// externally, our data pointer might be corrupted or the bool indicating that it's external.
|
|
// Consider the case where we allocate some chunks from an allocator - one of which allocated
|
|
// external data - then the allocator is reset. Now the chunk could be overwritten by subsequent
|
|
// recording before it is deleted. We don't want the allocator to have to explicitly delete all
|
|
// chunks that were allocated from it and external data allocations should be rare (only really
|
|
// massive chunks bigger than a page) so we can afford to externally allocate the chunk too.
|
|
if(allocator)
|
|
ret = new(allocator->AllocChunk()) Chunk(true);
|
|
else
|
|
ret = new Chunk(false);
|
|
|
|
ret->m_Length = length;
|
|
ret->m_ChunkType = chunkType;
|
|
ret->m_Data = data;
|
|
|
|
if(allocator == NULL)
|
|
{
|
|
#if ENABLED(RDOC_DEVEL)
|
|
Atomic::Inc64(&m_LiveChunks);
|
|
Atomic::ExchAdd64(&m_TotalMem, int64_t(length));
|
|
#endif
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
ChunkPagePool::~ChunkPagePool()
|
|
{
|
|
// all allocated pages are in precisely one list, so just free the contents of both lists
|
|
for(ChunkPage &p : freePages)
|
|
{
|
|
FreeAlignedBuffer(p.chunkBase);
|
|
FreeAlignedBuffer(p.bufferBase);
|
|
}
|
|
|
|
for(ChunkPage &p : allocatedPages)
|
|
{
|
|
FreeAlignedBuffer(p.chunkBase);
|
|
FreeAlignedBuffer(p.bufferBase);
|
|
}
|
|
}
|
|
|
|
ChunkPage ChunkPagePool::AllocPage()
|
|
{
|
|
if(!freePages.empty())
|
|
{
|
|
// if there's a free page, move it to the allocated list and return it
|
|
ChunkPage free = freePages.back();
|
|
freePages.pop_back();
|
|
allocatedPages.push_back(free);
|
|
}
|
|
else
|
|
{
|
|
// otherwise allocate a new one straight into the allocated list
|
|
byte *buffers = ::AllocAlignedBuffer(BufferPageSize);
|
|
byte *chunks = ::AllocAlignedBuffer(ChunkPageSize);
|
|
allocatedPages.push_back({m_ID++, buffers, buffers, chunks, chunks});
|
|
}
|
|
|
|
return allocatedPages.back();
|
|
}
|
|
|
|
void ChunkPagePool::Trim()
|
|
{
|
|
// truly release any currently free pages back to the system
|
|
for(ChunkPage &p : freePages)
|
|
{
|
|
FreeAlignedBuffer(p.chunkBase);
|
|
FreeAlignedBuffer(p.bufferBase);
|
|
}
|
|
|
|
freePages.clear();
|
|
}
|
|
|
|
void ChunkPagePool::Reset()
|
|
{
|
|
// forcibly move all allocated pages into the free list
|
|
freePages.append(allocatedPages);
|
|
allocatedPages.clear();
|
|
|
|
for(ChunkPage &p : freePages)
|
|
{
|
|
// reset head pointers
|
|
p.bufferHead = p.bufferBase;
|
|
p.chunkHead = p.chunkBase;
|
|
|
|
// assign a new ID so these pages can't get reset again by any allocator currently holding them
|
|
p.ID = m_ID++;
|
|
}
|
|
}
|
|
|
|
void ChunkPagePool::ResetPageSet(const rdcarray<ChunkPage> &pages)
|
|
{
|
|
// iterate over each page being freed
|
|
for(const ChunkPage &p : pages)
|
|
{
|
|
// try to find it in the allocated page list. This compares by ID, so if the page was already
|
|
// freed with a pool reset we won't find it at all because it will have a new ID - that's fine.
|
|
int32_t idx = allocatedPages.indexOf(p);
|
|
if(idx >= 0)
|
|
{
|
|
ChunkPage &alloc = allocatedPages[idx];
|
|
// give a new ID to be safe
|
|
alloc.ID = m_ID++;
|
|
// reset head pointers
|
|
alloc.bufferHead = alloc.bufferBase;
|
|
alloc.chunkHead = alloc.chunkBase;
|
|
// move to free list
|
|
freePages.push_back(alloc);
|
|
allocatedPages.erase(idx);
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
ChunkAllocator::~ChunkAllocator()
|
|
{
|
|
// move any pages we have back to the pool on destruction
|
|
Reset();
|
|
}
|
|
|
|
void ChunkAllocator::swap(ChunkAllocator &alloc)
|
|
{
|
|
if(&m_Pool != &alloc.m_Pool)
|
|
{
|
|
RDCERR(
|
|
"Allocator swap with allocator from another pool! Losing all pages to leak instead of "
|
|
"crashing");
|
|
pages.clear();
|
|
alloc.pages.clear();
|
|
return;
|
|
}
|
|
|
|
pages.swap(alloc.pages);
|
|
}
|
|
|
|
byte *ChunkAllocator::AllocAlignedBuffer(uint64_t size)
|
|
{
|
|
// always allocate 64-bytes at a time even if the size is smaller
|
|
return AllocateFromPages(false, AlignUp((size_t)size, (size_t)64));
|
|
}
|
|
|
|
byte *ChunkAllocator::AllocChunk()
|
|
{
|
|
RDCCOMPILE_ASSERT(sizeof(ChunkAllocator) <= 128, "foo");
|
|
return AllocateFromPages(true, sizeof(Chunk));
|
|
}
|
|
|
|
void ChunkAllocator::Reset()
|
|
{
|
|
m_Pool.ResetPageSet(pages);
|
|
pages.clear();
|
|
}
|
|
|
|
byte *ChunkAllocator::AllocateFromPages(bool chunkAlloc, size_t size)
|
|
{
|
|
// if the size can't be satisfied in a page, return NULL and we'll force a full allocation which
|
|
// will be freed on its own
|
|
if(size > m_Pool.GetBufferPageSize())
|
|
return NULL;
|
|
|
|
// if we don't have a current page, or it can't satisfy the allocation, get a new page from the
|
|
// pool
|
|
if(pages.empty() || GetRemainingBytes(chunkAlloc, pages.back()) < size)
|
|
pages.push_back(m_Pool.AllocPage());
|
|
|
|
ChunkPage &p = pages.back();
|
|
|
|
byte *ret = NULL;
|
|
|
|
if(chunkAlloc)
|
|
{
|
|
ret = p.chunkHead;
|
|
p.chunkHead += size;
|
|
}
|
|
else
|
|
{
|
|
ret = p.bufferHead;
|
|
p.bufferHead += size;
|
|
}
|
|
|
|
return ret;
|
|
}
|