/*
* This file is part of the TrinityCore Project. See AUTHORS file for Copyright information
*
* This program is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License as published by the
* Free Software Foundation; either version 2 of the License, or (at your
* option) any later version.
*
* This program is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
* more details.
*
* You should have received a copy of the GNU General Public License along
* with this program. If not, see .
*/
#include "DB2FileLoader.h"
#include "ByteConverter.h"
#include "DB2Meta.h"
#include "Errors.h"
#include "Log.h"
#include
#include
#include
#include
#include
class Temporary_10_0_7_metadata_bug_workaround : public std::exception
{
};
enum class DB2ColumnCompression : uint32
{
None,
Immediate,
CommonData,
Pallet,
PalletArray,
SignedImmediate
};
#pragma pack(push, 1)
struct DB2FieldEntry
{
int16 UnusedBits;
uint16 Offset;
};
struct DB2CatalogEntry
{
uint32 FileOffset;
uint16 RecordSize;
};
struct DB2ColumnMeta
{
uint16 BitOffset;
uint16 BitSize;
uint32 AdditionalDataSize;
DB2ColumnCompression CompressionType;
union
{
struct
{
uint32 BitOffset;
uint32 BitWidth;
bool Signed;
} immediate;
struct
{
uint32 Value;
} commonData;
struct
{
uint32 BitOffset;
uint32 BitWidth;
uint32 ArraySize;
} pallet;
} CompressionData;
};
struct DB2CommonValue
{
uint32 RecordId;
uint32 Value;
};
struct DB2PalletValue
{
uint32 Value;
};
struct DB2IndexDataInfo
{
uint32 NumEntries;
uint32 MinId;
uint32 MaxId;
};
struct DB2IndexEntry
{
uint32 ParentId;
uint32 RecordIndex;
};
#pragma pack(pop)
struct DB2IndexData
{
std::vector Entries;
};
uint32 DB2FileLoadInfo::GetStringFieldCount(bool localizedOnly) const
{
uint32 stringFields = 0;
for (std::size_t i = 0; i < FieldCount; ++i)
if (Fields[i].Type == FT_STRING || (Fields[i].Type == FT_STRING_NOT_LOCALIZED && !localizedOnly))
++stringFields;
return stringFields;
}
std::pair DB2FileLoadInfo::GetFieldIndexByName(char const* fieldName) const
{
std::size_t ourIndex = Meta->HasIndexFieldInData() ? 0 : 1;
for (uint32 i = 0; i < Meta->FieldCount; ++i)
{
for (uint8 arr = 0; arr < Meta->Fields[i].ArraySize; ++arr)
{
if (!strcmp(Fields[ourIndex].Name, fieldName))
return std::make_pair(int32(i), int32(arr));
++ourIndex;
}
}
return std::make_pair(-1, -1);
}
int32 DB2FileLoadInfo::GetFieldIndexByMetaIndex(uint32 metaIndex) const
{
ASSERT(metaIndex < Meta->FieldCount);
int32 ourIndex = Meta->HasIndexFieldInData() ? 0 : 1;
for (uint32 i = 0; i < metaIndex; ++i)
ourIndex += Meta->Fields[i].ArraySize;
return ourIndex;
}
DB2FileSource::DB2FileSource() = default;
DB2FileSource::~DB2FileSource() = default;
class DB2FileLoaderImpl
{
public:
DB2FileLoaderImpl() = default;
DB2FileLoaderImpl(DB2FileLoaderImpl const& other) = delete;
DB2FileLoaderImpl(DB2FileLoaderImpl&& other) noexcept = delete;
DB2FileLoaderImpl& operator=(DB2FileLoaderImpl const& other) = delete;
DB2FileLoaderImpl& operator=(DB2FileLoaderImpl&& other) noexcept = delete;
virtual ~DB2FileLoaderImpl() = default;
virtual void LoadColumnData(std::unique_ptr sections, std::unique_ptr fields, std::unique_ptr columnMeta,
std::unique_ptr[]> palletValues, std::unique_ptr[]> palletArrayValues,
std::unique_ptr[]> commonValues) = 0;
virtual void SkipEncryptedSection(uint32 section) = 0;
virtual bool LoadTableData(DB2FileSource* source, uint32 section) = 0;
virtual bool LoadCatalogData(DB2FileSource* source, uint32 section) = 0;
virtual void SetAdditionalData(std::vector idTable, std::vector copyTable, std::vector> parentIndexes) = 0;
virtual char* AutoProduceData(uint32& indexTableSize, char**& indexTable) = 0;
virtual char* AutoProduceStrings(char** indexTable, uint32 indexTableSize, uint32 locale) = 0;
virtual void AutoProduceRecordCopies(uint32 records, char** indexTable, char* dataTable) = 0;
virtual DB2Record GetRecord(uint32 recordNumber) const = 0;
virtual DB2RecordCopy GetRecordCopy(uint32 copyNumber) const = 0;
virtual uint32 GetRecordCount() const = 0;
virtual uint32 GetRecordCopyCount() const = 0;
virtual uint32 GetMinId() const = 0;
virtual uint32 GetMaxId() const = 0;
virtual DB2FileLoadInfo const* GetLoadInfo() const = 0;
virtual DB2SectionHeader& GetSection(uint32 section) const = 0;
virtual bool IsSignedField(uint32 field) const = 0;
virtual char const* GetExpectedSignMismatchReason(uint32 field) const = 0;
private:
friend class DB2Record;
virtual unsigned char const* GetRawRecordData(uint32 recordNumber, uint32 const* section) const = 0;
virtual uint32 RecordGetId(uint8 const* record, uint32 recordIndex) const = 0;
virtual uint8 RecordGetUInt8(uint8 const* record, uint32 field, uint32 arrayIndex) const = 0;
virtual uint16 RecordGetUInt16(uint8 const* record, uint32 field, uint32 arrayIndex) const = 0;
virtual uint32 RecordGetUInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const = 0;
virtual int32 RecordGetInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const = 0;
virtual uint64 RecordGetUInt64(uint8 const* record, uint32 field, uint32 arrayIndex) const = 0;
virtual float RecordGetFloat(uint8 const* record, uint32 field, uint32 arrayIndex) const = 0;
virtual char const* RecordGetString(uint8 const* record, uint32 field, uint32 arrayIndex) const = 0;
virtual std::size_t* RecordCreateDetachedFieldOffsets(std::size_t* oldOffsets) const = 0;
virtual std::size_t* RecordCopyDetachedFieldOffsets(std::size_t* oldOffsets) const = 0;
virtual void RecordDestroyFieldOffsets(std::size_t*& fieldOffsets) const = 0;
};
class DB2FileLoaderRegularImpl final : public DB2FileLoaderImpl
{
public:
DB2FileLoaderRegularImpl(char const* fileName, DB2FileLoadInfo const* loadInfo, DB2Header const* header);
DB2FileLoaderRegularImpl(DB2FileLoaderRegularImpl const& other) = delete;
DB2FileLoaderRegularImpl(DB2FileLoaderRegularImpl&& other) noexcept = delete;
DB2FileLoaderRegularImpl& operator=(DB2FileLoaderRegularImpl const& other) = delete;
DB2FileLoaderRegularImpl& operator=(DB2FileLoaderRegularImpl&& other) noexcept = delete;
~DB2FileLoaderRegularImpl();
void LoadColumnData(std::unique_ptr sections, std::unique_ptr fields, std::unique_ptr columnMeta,
std::unique_ptr[]> palletValues, std::unique_ptr[]> palletArrayValues,
std::unique_ptr[]> commonValues) override;
void SkipEncryptedSection(uint32 /*section*/) override { }
bool LoadTableData(DB2FileSource* source, uint32 section) override;
bool LoadCatalogData(DB2FileSource* /*source*/, uint32 /*section*/) override { return true; }
void SetAdditionalData(std::vector idTable, std::vector copyTable, std::vector> parentIndexes) override;
char* AutoProduceData(uint32& indexTableSize, char**& indexTable) override;
char* AutoProduceStrings(char** indexTable, uint32 indexTableSize, uint32 locale) override;
void AutoProduceRecordCopies(uint32 records, char** indexTable, char* dataTable) override;
DB2Record GetRecord(uint32 recordNumber) const override;
DB2RecordCopy GetRecordCopy(uint32 copyNumber) const override;
uint32 GetRecordCount() const override;
uint32 GetRecordCopyCount() const override;
uint32 GetMinId() const override;
uint32 GetMaxId() const override;
DB2FileLoadInfo const* GetLoadInfo() const override;
DB2SectionHeader& GetSection(uint32 section) const override;
bool IsSignedField(uint32 field) const override;
char const* GetExpectedSignMismatchReason(uint32 field) const override;
private:
void FillParentLookup(char* dataTable, char** indexTable);
uint32 GetRecordSection(uint32 recordNumber) const;
unsigned char const* GetRawRecordData(uint32 recordNumber, uint32 const* section) const override;
uint32 RecordGetId(uint8 const* record, uint32 recordIndex) const override;
uint8 RecordGetUInt8(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
uint16 RecordGetUInt16(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
uint32 RecordGetUInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
int32 RecordGetInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
uint64 RecordGetUInt64(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
float RecordGetFloat(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
char const* RecordGetString(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
template
T RecordGetVarInt(uint8 const* record, uint32 field, uint32 arrayIndex) const;
uint64 RecordGetPackedValue(uint8 const* packedRecordData, uint32 bitWidth, uint32 bitOffset) const;
uint16 GetFieldOffset(uint32 field) const;
std::size_t* RecordCreateDetachedFieldOffsets(std::size_t* oldOffsets) const override;
std::size_t* RecordCopyDetachedFieldOffsets(std::size_t* oldOffsets) const override;
void RecordDestroyFieldOffsets(std::size_t*& fieldOffsets) const override;
char const* _fileName;
DB2FileLoadInfo const* _loadInfo;
DB2Header const* _header;
std::unique_ptr _data;
uint8* _stringTable;
std::unique_ptr _sections;
std::unique_ptr _columnMeta;
std::unique_ptr[]> _palletValues;
std::unique_ptr[]> _palletArrayValues;
std::unique_ptr[]> _commonValues;
std::vector _idTable;
std::vector _copyTable;
std::vector> _parentIndexes;
};
class DB2FileLoaderSparseImpl final : public DB2FileLoaderImpl
{
public:
DB2FileLoaderSparseImpl(char const* fileName, DB2FileLoadInfo const* loadInfo, DB2Header const* header, DB2FileSource* source);
DB2FileLoaderSparseImpl(DB2FileLoaderSparseImpl const& other) = delete;
DB2FileLoaderSparseImpl(DB2FileLoaderSparseImpl&& other) noexcept = delete;
DB2FileLoaderSparseImpl& operator=(DB2FileLoaderSparseImpl const& other) = delete;
DB2FileLoaderSparseImpl& operator=(DB2FileLoaderSparseImpl&& other) noexcept = delete;
~DB2FileLoaderSparseImpl();
void LoadColumnData(std::unique_ptr sections, std::unique_ptr fields, std::unique_ptr columnMeta,
std::unique_ptr[]> palletValues, std::unique_ptr[]> palletArrayValues,
std::unique_ptr[]> commonValues) override;
void SkipEncryptedSection(uint32 section) override;
bool LoadTableData(DB2FileSource* /*source*/, uint32 /*section*/) override { return true; }
bool LoadCatalogData(DB2FileSource* source, uint32 section) override;
void SetAdditionalData(std::vector idTable, std::vector copyTable, std::vector> parentIndexes) override;
char* AutoProduceData(uint32& indexTableSize, char**& indexTable) override;
char* AutoProduceStrings(char** indexTable, uint32 indexTableSize, uint32 locale) override;
void AutoProduceRecordCopies(uint32 records, char** indexTable, char* dataTable) override;
DB2Record GetRecord(uint32 recordNumber) const override;
DB2RecordCopy GetRecordCopy(uint32 copyNumber) const override;
uint32 GetRecordCount() const override;
uint32 GetRecordCopyCount() const override;
uint32 GetMinId() const override;
uint32 GetMaxId() const override;
DB2FileLoadInfo const* GetLoadInfo() const override;
DB2SectionHeader& GetSection(uint32 section) const override;
bool IsSignedField(uint32 field) const override;
char const* GetExpectedSignMismatchReason(uint32 field) const override;
private:
void FillParentLookup(char* dataTable);
uint32 GetRecordSection(uint32 recordNumber) const;
unsigned char const* GetRawRecordData(uint32 recordNumber, uint32 const* section) const override;
uint32 RecordGetId(uint8 const* record, uint32 recordIndex) const override;
uint8 RecordGetUInt8(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
uint16 RecordGetUInt16(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
uint32 RecordGetUInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
int32 RecordGetInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
uint64 RecordGetUInt64(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
float RecordGetFloat(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
char const* RecordGetString(uint8 const* record, uint32 field, uint32 arrayIndex) const override;
uint32 RecordGetVarInt(uint8 const* record, uint32 field, uint32 arrayIndex, bool isSigned) const;
uint16 GetFieldOffset(uint32 field, uint32 arrayIndex) const;
uint16 GetFieldSize(uint32 field) const;
std::size_t* RecordCreateDetachedFieldOffsets(std::size_t* oldOffsets) const override;
std::size_t* RecordCopyDetachedFieldOffsets(std::size_t* oldOffsets) const override;
void RecordDestroyFieldOffsets(std::size_t*& fieldOffsets) const override;
void CalculateAndStoreFieldOffsets(uint8 const* rawRecord) const;
char const* _fileName;
DB2FileLoadInfo const* _loadInfo;
DB2Header const* _header;
DB2FileSource* const _source;
std::size_t _totalRecordSize;
uint16 _maxRecordSize;
std::unique_ptr _recordBuffer;
std::unique_ptr _sections;
std::unique_ptr _fields;
std::unique_ptr _fieldAndArrayOffsets;
std::vector _catalogIds;
std::vector _catalog;
std::vector _copyTable;
std::vector> _parentIndexes;
};
DB2FileLoaderRegularImpl::DB2FileLoaderRegularImpl(char const* fileName, DB2FileLoadInfo const* loadInfo, DB2Header const* header) :
_fileName(fileName),
_loadInfo(loadInfo),
_header(header),
_stringTable(nullptr)
{
}
void DB2FileLoaderRegularImpl::LoadColumnData(std::unique_ptr sections, std::unique_ptr /*fields*/, std::unique_ptr columnMeta,
std::unique_ptr[]> palletValues, std::unique_ptr[]> palletArrayValues,
std::unique_ptr[]> commonValues)
{
_sections = std::move(sections);
_columnMeta = std::move(columnMeta);
_palletValues = std::move(palletValues);
_palletArrayValues = std::move(palletArrayValues);
_commonValues = std::move(commonValues);
}
bool DB2FileLoaderRegularImpl::LoadTableData(DB2FileSource* source, uint32 section)
{
if (!_data)
{
_data = std::make_unique(_header->RecordSize * _header->RecordCount + _header->StringTableSize + 8);
_stringTable = &_data[_header->RecordSize * _header->RecordCount];
}
uint32 sectionDataStart = 0;
uint32 sectionStringTableStart = 0;
for (uint32 i = 0; i < section; ++i)
{
sectionDataStart += _header->RecordSize * _sections[i].RecordCount;
sectionStringTableStart += _sections[i].StringTableSize;
}
if (_sections[section].RecordCount && !source->Read(&_data[sectionDataStart], _header->RecordSize * _sections[section].RecordCount))
return false;
if (_sections[section].StringTableSize && !source->Read(&_stringTable[sectionStringTableStart], _sections[section].StringTableSize))
return false;
return true;
}
void DB2FileLoaderRegularImpl::SetAdditionalData(std::vector idTable, std::vector copyTable, std::vector> parentIndexes)
{
_idTable = std::move(idTable);
_copyTable = std::move(copyTable);
_parentIndexes = std::move(parentIndexes);
}
DB2FileLoaderRegularImpl::~DB2FileLoaderRegularImpl() = default;
static char const* const EmptyDb2String = "";
char* DB2FileLoaderRegularImpl::AutoProduceData(uint32& indexTableSize, char**& indexTable)
{
//get struct size and index pos
uint32 recordsize = _loadInfo->Meta->GetRecordSize();
uint32 maxi = GetMaxId() + 1;
using index_entry_t = char*;
indexTableSize = maxi;
indexTable = new index_entry_t[maxi];
memset(indexTable, 0, maxi * sizeof(index_entry_t));
char* dataTable = new char[(_header->RecordCount + _copyTable.size()) * recordsize];
uint32 offset = 0;
uint32 recordIndex = 0;
for (uint32 section = 0; section < _header->SectionCount; ++section)
{
DB2SectionHeader const& sectionHeader = GetSection(section);
if (sectionHeader.TactId)
{
offset += recordsize * sectionHeader.RecordCount;
recordIndex += sectionHeader.RecordCount;
continue;
}
for (uint32 sr = 0; sr < sectionHeader.RecordCount; ++sr, ++recordIndex)
{
unsigned char const* rawRecord = GetRawRecordData(recordIndex, §ion);
if (!rawRecord)
continue;
uint32 indexVal = RecordGetId(rawRecord, recordIndex);
indexTable[indexVal] = &dataTable[offset];
uint32 fieldIndex = 0;
if (!_loadInfo->Meta->HasIndexFieldInData())
{
*reinterpret_cast(&dataTable[offset]) = indexVal;
offset += 4;
++fieldIndex;
}
try
{
for (uint32 x = 0; x < _header->FieldCount; ++x)
{
for (uint32 z = 0; z < _loadInfo->Meta->Fields[x].ArraySize; ++z)
{
switch (_loadInfo->Fields[fieldIndex].Type)
{
case FT_FLOAT:
*reinterpret_cast(&dataTable[offset]) = RecordGetFloat(rawRecord, x, z);
offset += 4;
break;
case FT_INT:
*reinterpret_cast(&dataTable[offset]) = RecordGetVarInt(rawRecord, x, z);
offset += 4;
break;
case FT_BYTE:
*reinterpret_cast(&dataTable[offset]) = RecordGetUInt8(rawRecord, x, z);
offset += 1;
break;
case FT_SHORT:
*reinterpret_cast(&dataTable[offset]) = RecordGetUInt16(rawRecord, x, z);
offset += 2;
break;
case FT_LONG:
*reinterpret_cast(&dataTable[offset]) = RecordGetUInt64(rawRecord, x, z);
offset += 8;
break;
case FT_STRING:
for (char const*& localeStr : reinterpret_cast(&dataTable[offset])->Str)
localeStr = EmptyDb2String;
offset += sizeof(LocalizedString);
break;
case FT_STRING_NOT_LOCALIZED:
*reinterpret_cast(&dataTable[offset]) = EmptyDb2String;
offset += sizeof(char*);
break;
default:
ABORT_MSG("Unknown format character '%c' found in %s meta for field %s",
_loadInfo->Fields[fieldIndex].Type, _fileName, _loadInfo->Fields[fieldIndex].Name);
break;
}
++fieldIndex;
}
}
for (uint32 x = _header->FieldCount; x < _loadInfo->Meta->FieldCount; ++x)
{
for (uint32 z = 0; z < _loadInfo->Meta->Fields[x].ArraySize; ++z)
{
switch (_loadInfo->Fields[fieldIndex].Type)
{
case FT_INT:
*reinterpret_cast(&dataTable[offset]) = 0;
offset += 4;
break;
case FT_BYTE:
*reinterpret_cast(&dataTable[offset]) = 0;
offset += 1;
break;
case FT_SHORT:
*reinterpret_cast(&dataTable[offset]) = 0;
offset += 2;
break;
default:
ABORT_MSG("Unknown format character '%c' found in %s meta for parent field %s",
_loadInfo->Fields[fieldIndex].Type, _fileName, _loadInfo->Fields[fieldIndex].Name);
break;
}
++fieldIndex;
}
}
}
catch (Temporary_10_0_7_metadata_bug_workaround const&)
{
// pretend this record doesnt exist, has overflown column
indexTable[indexVal] = nullptr;
}
}
}
if (!_parentIndexes.empty())
FillParentLookup(dataTable, indexTable);
return dataTable;
}
char* DB2FileLoaderRegularImpl::AutoProduceStrings(char** indexTable, uint32 indexTableSize, uint32 locale)
{
if (!(_header->Locale & (1 << locale)))
{
char const* sep = "";
std::ostringstream str;
for (uint32 i = 0; i < TOTAL_LOCALES; ++i)
{
if (_header->Locale & (1 << i))
{
str << sep << localeNames[i];
sep = ", ";
}
}
TC_LOG_ERROR("", "Attempted to load {} which has locales {} as {}. Check if you placed your localized db2 files in correct directory.", _fileName, str.str(), localeNames[locale]);
return nullptr;
}
if (!_loadInfo->GetStringFieldCount(false))
return nullptr;
char* stringPool = new char[_header->StringTableSize];
memcpy(stringPool, _stringTable, _header->StringTableSize);
uint32 y = 0;
for (uint32 section = 0; section < _header->SectionCount; ++section)
{
DB2SectionHeader const& sectionHeader = GetSection(section);
if (sectionHeader.TactId)
{
y += sectionHeader.RecordCount;
continue;
}
for (uint32 sr = 0; sr < sectionHeader.RecordCount; ++sr, ++y)
{
unsigned char const* rawRecord = GetRawRecordData(y, §ion);
if (!rawRecord)
continue;
uint32 indexVal = RecordGetId(rawRecord, y);
if (indexVal >= indexTableSize)
continue;
char* recordData = indexTable[indexVal];
if (!recordData)
continue;
uint32 offset = 0;
uint32 fieldIndex = 0;
if (!_loadInfo->Meta->HasIndexFieldInData())
{
offset += 4;
++fieldIndex;
}
for (uint32 x = 0; x < _loadInfo->Meta->FieldCount; ++x)
{
for (uint32 z = 0; z < _loadInfo->Meta->Fields[x].ArraySize; ++z)
{
switch (_loadInfo->Fields[fieldIndex].Type)
{
case FT_FLOAT:
case FT_INT:
offset += 4;
break;
case FT_BYTE:
offset += 1;
break;
case FT_SHORT:
offset += 2;
break;
case FT_LONG:
offset += 8;
break;
case FT_STRING:
if (char const* string = RecordGetString(rawRecord, x, z))
reinterpret_cast(&recordData[offset])->Str[locale] = stringPool + (string - reinterpret_cast(_stringTable));
offset += sizeof(LocalizedString);
break;
case FT_STRING_NOT_LOCALIZED:
if (char const* string = RecordGetString(rawRecord, x, z))
*reinterpret_cast(&recordData[offset]) = stringPool + (string - reinterpret_cast(_stringTable));
offset += sizeof(char*);
break;
default:
ABORT_MSG("Unknown format character '%c' found in %s meta for field %s",
_loadInfo->Fields[fieldIndex].Type, _fileName, _loadInfo->Fields[fieldIndex].Name);
break;
}
++fieldIndex;
}
}
}
}
return stringPool;
}
void DB2FileLoaderRegularImpl::AutoProduceRecordCopies(uint32 records, char** indexTable, char* dataTable)
{
uint32 recordCopies = GetRecordCopyCount();
if (!recordCopies)
return;
uint32 recordsize = _loadInfo->Meta->GetRecordSize();
uint32 offset = _header->RecordCount * recordsize;
uint32 idFieldOffset = _loadInfo->Meta->HasIndexFieldInData() ? GetFieldOffset(_loadInfo->Meta->GetIndexField()) : 0;
for (uint32 c = 0; c < recordCopies; ++c)
{
DB2RecordCopy copy = GetRecordCopy(c);
if (copy.SourceRowId && copy.SourceRowId < records && copy.NewRowId < records && indexTable[copy.SourceRowId])
{
indexTable[copy.NewRowId] = &dataTable[offset];
memcpy(indexTable[copy.NewRowId], indexTable[copy.SourceRowId], recordsize);
*reinterpret_cast(&dataTable[offset + idFieldOffset]) = copy.NewRowId;
offset += recordsize;
}
}
}
void DB2FileLoaderRegularImpl::FillParentLookup(char* dataTable, char** indexTable)
{
int32 parentIdOffset = _loadInfo->Meta->GetParentIndexFieldOffset();
uint32 recordSize = _loadInfo->Meta->GetRecordSize();
uint32 recordIndexOffset = 0;
for (uint32 i = 0; i < _header->SectionCount; ++i)
{
DB2SectionHeader const& section = GetSection(i);
if (!section.TactId)
{
for (std::size_t j = 0; j < _parentIndexes[i][0].Entries.size(); ++j)
{
uint32 parentId = _parentIndexes[i][0].Entries[j].ParentId;
uint32 recordIndex = _parentIndexes[i][0].Entries[j].RecordIndex + recordIndexOffset;
char* recordData = &dataTable[recordIndex * recordSize];
// temporary workaround
uint32 id = GetRecord(recordIndex).GetId();
if (!indexTable[id])
continue;
switch (_loadInfo->Meta->Fields[_loadInfo->Meta->ParentIndexField].Type)
{
case FT_SHORT:
{
if (_loadInfo->Meta->ParentIndexField >= int32(_loadInfo->Meta->FileFieldCount))
{
// extra field at the end
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
else
{
// in data block, must fit
ASSERT(parentId <= std::numeric_limits::max(), "ParentId value %u does not fit into uint16 field (%s in %s)",
parentId, _loadInfo->Fields[_loadInfo->GetFieldIndexByMetaIndex(_loadInfo->Meta->ParentIndexField)].Name, _fileName);
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
break;
}
case FT_BYTE:
{
if (_loadInfo->Meta->ParentIndexField >= int32(_loadInfo->Meta->FileFieldCount))
{
// extra field at the end
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
else
{
// in data block, must fit
ASSERT(parentId <= std::numeric_limits::max(), "ParentId value %u does not fit into uint8 field (%s in %s)",
parentId, _loadInfo->Fields[_loadInfo->GetFieldIndexByMetaIndex(_loadInfo->Meta->ParentIndexField)].Name, _fileName);
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
break;
}
case FT_INT:
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
break;
default:
ABORT_MSG("Unhandled parent id type '%c' found in %s", _loadInfo->Meta->Fields[_loadInfo->Meta->ParentIndexField].Type, _fileName);
break;
}
}
}
recordIndexOffset += section.RecordCount;
}
}
DB2Record DB2FileLoaderRegularImpl::GetRecord(uint32 recordNumber) const
{
return DB2Record(*this, recordNumber, nullptr);
}
DB2RecordCopy DB2FileLoaderRegularImpl::GetRecordCopy(uint32 copyNumber) const
{
if (copyNumber >= GetRecordCopyCount())
return DB2RecordCopy{};
return _copyTable[copyNumber];
}
uint32 DB2FileLoaderRegularImpl::GetRecordCount() const
{
return _header->RecordCount;
}
uint32 DB2FileLoaderRegularImpl::GetRecordCopyCount() const
{
return _copyTable.size();
}
uint32 DB2FileLoaderRegularImpl::GetRecordSection(uint32 recordNumber) const
{
uint32 section = 0;
for (; section < _header->SectionCount; ++section)
{
DB2SectionHeader const& sectionHeader = GetSection(section);
if (recordNumber < sectionHeader.RecordCount)
break;
recordNumber -= sectionHeader.RecordCount;
}
return section;
}
unsigned char const* DB2FileLoaderRegularImpl::GetRawRecordData(uint32 recordNumber, uint32 const* section) const
{
if (recordNumber >= _header->RecordCount)
return nullptr;
if (GetSection(section ? *section : GetRecordSection(recordNumber)).TactId)
return nullptr;
return &_data[recordNumber * _header->RecordSize];
}
uint32 DB2FileLoaderRegularImpl::RecordGetId(uint8 const* record, uint32 recordIndex) const
{
if (_loadInfo->Meta->HasIndexFieldInData())
return RecordGetVarInt(record, _loadInfo->Meta->GetIndexField(), 0);
return _idTable[recordIndex];
}
uint8 DB2FileLoaderRegularImpl::RecordGetUInt8(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return RecordGetVarInt(record, field, arrayIndex);
}
uint16 DB2FileLoaderRegularImpl::RecordGetUInt16(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return RecordGetVarInt(record, field, arrayIndex);
}
uint32 DB2FileLoaderRegularImpl::RecordGetUInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return RecordGetVarInt(record, field, arrayIndex);
}
int32 DB2FileLoaderRegularImpl::RecordGetInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return RecordGetVarInt(record, field, arrayIndex);
}
uint64 DB2FileLoaderRegularImpl::RecordGetUInt64(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return RecordGetVarInt(record, field, arrayIndex);
}
float DB2FileLoaderRegularImpl::RecordGetFloat(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return RecordGetVarInt(record, field, arrayIndex);
}
char const* DB2FileLoaderRegularImpl::RecordGetString(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
uint32 fieldOffset = GetFieldOffset(field) + sizeof(uint32) * arrayIndex;
uint32 stringOffset = RecordGetVarInt(record, field, arrayIndex);
ASSERT(stringOffset < _header->RecordSize * _header->RecordCount + _header->StringTableSize);
return stringOffset ? reinterpret_cast(record + fieldOffset + stringOffset) : nullptr;
}
template
T DB2FileLoaderRegularImpl::RecordGetVarInt(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
ASSERT(field < _header->FieldCount);
DB2ColumnCompression compressionType = _columnMeta ? _columnMeta[field].CompressionType : DB2ColumnCompression::None;
switch (compressionType)
{
case DB2ColumnCompression::None:
{
T val = *reinterpret_cast(record + GetFieldOffset(field) + sizeof(T) * arrayIndex);
EndianConvert(val);
return val;
}
case DB2ColumnCompression::Immediate:
{
ASSERT(arrayIndex == 0);
uint64 immediateValue = RecordGetPackedValue(record + GetFieldOffset(field),
_columnMeta[field].CompressionData.immediate.BitWidth, _columnMeta[field].CompressionData.immediate.BitOffset);
EndianConvert(immediateValue);
T value;
memcpy(&value, &immediateValue, std::min(sizeof(T), sizeof(immediateValue)));
if constexpr (std::is_integral_v)
if (_columnMeta[field].CompressionData.immediate.BitWidth > sizeof(T) * 8 && int64(value) != int64(immediateValue))
throw Temporary_10_0_7_metadata_bug_workaround();
return value;
}
case DB2ColumnCompression::CommonData:
{
uint32 id = RecordGetId(record, (record - _data.get()) / _header->RecordSize);
T value;
auto valueItr = _commonValues[field].find(id);
if (valueItr != _commonValues[field].end())
memcpy(&value, &valueItr->second, std::min(sizeof(T), sizeof(uint32)));
else
memcpy(&value, &_columnMeta[field].CompressionData.commonData.Value, std::min(sizeof(T), sizeof(uint32)));
return value;
}
case DB2ColumnCompression::Pallet:
{
ASSERT(arrayIndex == 0);
uint64 palletIndex = RecordGetPackedValue(record + GetFieldOffset(field),
_columnMeta[field].CompressionData.pallet.BitWidth, _columnMeta[field].CompressionData.pallet.BitOffset);
EndianConvert(palletIndex);
uint32 palletValue = _palletValues[field][palletIndex].Value;
EndianConvert(palletValue);
T value;
memcpy(&value, &palletValue, std::min(sizeof(T), sizeof(palletValue)));
return value;
}
case DB2ColumnCompression::PalletArray:
{
uint64 palletIndex = RecordGetPackedValue(record + GetFieldOffset(field),
_columnMeta[field].CompressionData.pallet.BitWidth, _columnMeta[field].CompressionData.pallet.BitOffset);
EndianConvert(palletIndex);
uint32 palletValue = _palletArrayValues[field][palletIndex * _columnMeta[field].CompressionData.pallet.ArraySize + arrayIndex].Value;
EndianConvert(palletValue);
T value;
memcpy(&value, &palletValue, std::min(sizeof(T), sizeof(palletValue)));
return value;
}
case DB2ColumnCompression::SignedImmediate:
{
ASSERT(arrayIndex == 0);
uint64 immediateValue = RecordGetPackedValue(record + GetFieldOffset(field),
_columnMeta[field].CompressionData.immediate.BitWidth, _columnMeta[field].CompressionData.immediate.BitOffset);
EndianConvert(immediateValue);
uint64 mask = UI64LIT(1) << (_columnMeta[field].CompressionData.immediate.BitWidth - 1);
immediateValue = (immediateValue ^ mask) - mask;
T value;
memcpy(&value, &immediateValue, std::min(sizeof(T), sizeof(immediateValue)));
return value;
}
default:
ABORT_MSG("Unhandled compression type %u in %s", uint32(_columnMeta[field].CompressionType), _fileName);
break;
}
return 0;
}
uint64 DB2FileLoaderRegularImpl::RecordGetPackedValue(uint8 const* packedRecordData, uint32 bitWidth, uint32 bitOffset) const
{
uint32 bitsToRead = bitOffset & 7;
return *reinterpret_cast(packedRecordData) << (64 - bitsToRead - bitWidth) >> (64 - bitWidth);
}
uint16 DB2FileLoaderRegularImpl::GetFieldOffset(uint32 field) const
{
ASSERT(field < _header->FieldCount);
DB2ColumnCompression compressionType = _columnMeta ? _columnMeta[field].CompressionType : DB2ColumnCompression::None;
switch (compressionType)
{
case DB2ColumnCompression::None:
return _columnMeta[field].BitOffset / 8;
case DB2ColumnCompression::Immediate:
case DB2ColumnCompression::SignedImmediate:
return _columnMeta[field].CompressionData.immediate.BitOffset / 8 + _header->PackedDataOffset;
case DB2ColumnCompression::CommonData:
return 0xFFFF;
case DB2ColumnCompression::Pallet:
case DB2ColumnCompression::PalletArray:
return _columnMeta[field].CompressionData.pallet.BitOffset / 8 + _header->PackedDataOffset;
default:
ABORT_MSG("Unhandled compression type %u in %s", uint32(_columnMeta[field].CompressionType), _fileName);
break;
}
return 0;
}
std::size_t* DB2FileLoaderRegularImpl::RecordCreateDetachedFieldOffsets(std::size_t* /*oldOffsets*/) const
{
return nullptr;
}
std::size_t* DB2FileLoaderRegularImpl::RecordCopyDetachedFieldOffsets(std::size_t* /*oldOffsets*/) const
{
return nullptr;
}
void DB2FileLoaderRegularImpl::RecordDestroyFieldOffsets(std::size_t*& /*fieldOffsets*/) const
{
}
uint32 DB2FileLoaderRegularImpl::GetMinId() const
{
uint32 minId = std::numeric_limits::max();
for (uint32 row = 0; row < _header->RecordCount; ++row)
{
unsigned char const* rawRecord = GetRawRecordData(row, nullptr);
if (!rawRecord)
continue;
uint32 id = RecordGetId(rawRecord, row);
if (id < minId)
minId = id;
}
for (uint32 copy = 0; copy < GetRecordCopyCount(); ++copy)
{
uint32 id = GetRecordCopy(copy).NewRowId;
if (id < minId)
minId = id;
}
if (minId == std::numeric_limits::max())
minId = 0;
ASSERT(minId >= _header->MinId);
return minId;
}
uint32 DB2FileLoaderRegularImpl::GetMaxId() const
{
uint32 maxId = 0;
for (uint32 row = 0; row < _header->RecordCount; ++row)
{
unsigned char const* rawRecord = GetRawRecordData(row, nullptr);
if (!rawRecord)
continue;
uint32 id = RecordGetId(rawRecord, row);
if (id > maxId)
maxId = id;
}
for (uint32 copy = 0; copy < GetRecordCopyCount(); ++copy)
{
uint32 id = GetRecordCopy(copy).NewRowId;
if (id > maxId)
maxId = id;
}
ASSERT(maxId <= _header->MaxId);
return maxId;
}
DB2FileLoadInfo const* DB2FileLoaderRegularImpl::GetLoadInfo() const
{
return _loadInfo;
}
DB2SectionHeader& DB2FileLoaderRegularImpl::GetSection(uint32 section) const
{
return _sections[section];
}
bool DB2FileLoaderRegularImpl::IsSignedField(uint32 field) const
{
if (field >= _header->TotalFieldCount)
{
ASSERT(field == _header->TotalFieldCount);
ASSERT(int32(field) == _loadInfo->Meta->ParentIndexField);
return _loadInfo->Meta->IsSignedField(field);
}
DB2ColumnCompression compressionType = _columnMeta ? _columnMeta[field].CompressionType : DB2ColumnCompression::None;
switch (compressionType)
{
case DB2ColumnCompression::None:
case DB2ColumnCompression::CommonData:
case DB2ColumnCompression::Pallet:
case DB2ColumnCompression::PalletArray:
return _loadInfo->Meta->IsSignedField(field);
case DB2ColumnCompression::SignedImmediate:
return field != uint32(_loadInfo->Meta->IndexField);
case DB2ColumnCompression::Immediate:
return false;
default:
ABORT_MSG("Unhandled compression type %u in %s", uint32(_columnMeta[field].CompressionType), _fileName);
break;
}
return false;
}
char const* DB2FileLoaderRegularImpl::GetExpectedSignMismatchReason(uint32 field) const
{
if (field >= _header->TotalFieldCount)
{
ASSERT(field == _header->TotalFieldCount);
ASSERT(int32(field) == _loadInfo->Meta->ParentIndexField);
return " (ParentIndexField must always be unsigned)";
}
DB2ColumnCompression compressionType = _columnMeta ? _columnMeta[field].CompressionType : DB2ColumnCompression::None;
switch (compressionType)
{
case DB2ColumnCompression::None:
case DB2ColumnCompression::CommonData:
case DB2ColumnCompression::Pallet:
case DB2ColumnCompression::PalletArray:
if (int32(field) == _loadInfo->Meta->IndexField)
return " (IndexField must always be unsigned)";
if (int32(field) == _loadInfo->Meta->ParentIndexField)
return " (ParentIndexField must always be unsigned)";
return "";
case DB2ColumnCompression::SignedImmediate:
return " (CompressionType is SignedImmediate)";
case DB2ColumnCompression::Immediate:
return " (CompressionType is Immediate)";
default:
ABORT_MSG("Unhandled compression type %u in %s", uint32(_columnMeta[field].CompressionType), _fileName);
break;
}
return "";
}
DB2FileLoaderSparseImpl::DB2FileLoaderSparseImpl(char const* fileName, DB2FileLoadInfo const* loadInfo, DB2Header const* header, DB2FileSource* source) :
_fileName(fileName),
_loadInfo(loadInfo),
_header(header),
_source(source),
_totalRecordSize(0),
_maxRecordSize(0),
_fieldAndArrayOffsets(loadInfo ? (std::make_unique(loadInfo->Meta->FieldCount + loadInfo->FieldCount - (!loadInfo->Meta->HasIndexFieldInData() ? 1 : 0))) : nullptr)
{
}
DB2FileLoaderSparseImpl::~DB2FileLoaderSparseImpl() = default;
void DB2FileLoaderSparseImpl::LoadColumnData(std::unique_ptr sections, std::unique_ptr fields, std::unique_ptr /*columnMeta*/,
std::unique_ptr[]> /*palletValues*/, std::unique_ptr[]> /*palletArrayValues*/,
std::unique_ptr[]> /*commonValues*/)
{
_sections = std::move(sections);
_fields = std::move(fields);
}
void DB2FileLoaderSparseImpl::SkipEncryptedSection(uint32 section)
{
_catalogIds.resize(_catalogIds.size() + _sections[section].CatalogDataCount);
_catalog.resize(_catalog.size() + _sections[section].CatalogDataCount);
}
bool DB2FileLoaderSparseImpl::LoadCatalogData(DB2FileSource* source, uint32 section)
{
source->SetPosition(_sections[section].CatalogDataOffset);
std::size_t oldSize = _catalog.size();
_catalogIds.resize(oldSize + _sections[section].CatalogDataCount);
if (!source->Read(&_catalogIds[oldSize], sizeof(uint32) * _sections[section].CatalogDataCount))
return false;
if (_sections[section].CopyTableCount)
{
std::size_t oldCopyTableSize = _copyTable.size();
_copyTable.resize(oldCopyTableSize + _sections[section].CopyTableCount);
if (!source->Read(&_copyTable[oldCopyTableSize], sizeof(DB2RecordCopy) * _sections[section].CopyTableCount))
return false;
}
_catalog.resize(oldSize + _sections[section].CatalogDataCount);
if (!source->Read(&_catalog[oldSize], sizeof(DB2CatalogEntry) * _sections[section].CatalogDataCount))
return false;
for (uint32 i = 0; i < _sections[section].CatalogDataCount; ++i)
{
_totalRecordSize += _catalog[oldSize + i].RecordSize;
_maxRecordSize = std::max(_maxRecordSize, _catalog[oldSize + i].RecordSize);
}
return true;
}
void DB2FileLoaderSparseImpl::SetAdditionalData(std::vector /*idTable*/, std::vector /*copyTable*/, std::vector> parentIndexes)
{
_parentIndexes = std::move(parentIndexes);
_recordBuffer = std::make_unique(_maxRecordSize);
}
char* DB2FileLoaderSparseImpl::AutoProduceData(uint32& indexTableSize, char**& indexTable)
{
if (_loadInfo->Meta->FieldCount != _header->FieldCount)
throw DB2FileLoadException(Trinity::StringFormat("Found unsupported parent index in sparse db2 {}", _fileName));
//get struct size and index pos
uint32 recordsize = _loadInfo->Meta->GetRecordSize();
uint32 records = _catalog.size();
using index_entry_t = char*;
indexTableSize = _header->MaxId + 1;
indexTable = new index_entry_t[indexTableSize];
memset(indexTable, 0, indexTableSize * sizeof(index_entry_t));
char* dataTable = new char[(records + _copyTable.size()) * recordsize];
memset(dataTable, 0, (records + _copyTable.size()) * recordsize);
uint32 offset = 0;
uint32 y = 0;
for (uint32 section = 0; section < _header->SectionCount; ++section)
{
DB2SectionHeader const& sectionHeader = GetSection(section);
if (sectionHeader.TactId)
{
offset += recordsize * sectionHeader.RecordCount;
y += sectionHeader.RecordCount;
continue;
}
for (uint32 sr = 0; sr < sectionHeader.CatalogDataCount; ++sr, ++y)
{
unsigned char const* rawRecord = GetRawRecordData(y, §ion);
if (!rawRecord)
continue;
uint32 indexVal = _catalogIds[y];
indexTable[indexVal] = &dataTable[offset];
uint32 fieldIndex = 0;
if (!_loadInfo->Meta->HasIndexFieldInData())
{
*reinterpret_cast(&dataTable[offset]) = indexVal;
offset += 4;
++fieldIndex;
}
for (uint32 x = 0; x < _header->FieldCount; ++x)
{
for (uint32 z = 0; z < _loadInfo->Meta->Fields[x].ArraySize; ++z)
{
switch (_loadInfo->Fields[fieldIndex].Type)
{
case FT_FLOAT:
*reinterpret_cast(&dataTable[offset]) = RecordGetFloat(rawRecord, x, z);
offset += 4;
break;
case FT_INT:
*reinterpret_cast(&dataTable[offset]) = RecordGetVarInt(rawRecord, x, z, _loadInfo->Fields[fieldIndex].IsSigned);
offset += 4;
break;
case FT_BYTE:
*reinterpret_cast(&dataTable[offset]) = RecordGetUInt8(rawRecord, x, z);
offset += 1;
break;
case FT_SHORT:
*reinterpret_cast(&dataTable[offset]) = RecordGetUInt16(rawRecord, x, z);
offset += 2;
break;
case FT_LONG:
*reinterpret_cast(&dataTable[offset]) = RecordGetUInt64(rawRecord, x, z);
offset += 8;
break;
case FT_STRING:
for (char const*& localeStr : reinterpret_cast(&dataTable[offset])->Str)
localeStr = EmptyDb2String;
offset += sizeof(LocalizedString);
break;
case FT_STRING_NOT_LOCALIZED:
*reinterpret_cast(&dataTable[offset]) = EmptyDb2String;
offset += sizeof(char*);
break;
default:
ABORT_MSG("Unknown format character '%c' found in %s meta for field %s",
_loadInfo->Fields[fieldIndex].Type, _fileName, _loadInfo->Fields[fieldIndex].Name);
break;
}
++fieldIndex;
}
}
for (uint32 x = _header->FieldCount; x < _loadInfo->Meta->FieldCount; ++x)
{
for (uint32 z = 0; z < _loadInfo->Meta->Fields[x].ArraySize; ++z)
{
switch (_loadInfo->Fields[fieldIndex].Type)
{
case FT_INT:
*reinterpret_cast(&dataTable[offset]) = 0;
offset += 4;
break;
case FT_BYTE:
*reinterpret_cast(&dataTable[offset]) = 0;
offset += 1;
break;
case FT_SHORT:
*reinterpret_cast(&dataTable[offset]) = 0;
offset += 2;
break;
default:
ABORT_MSG("Unknown format character '%c' found in %s meta for parent field %s",
_loadInfo->Fields[fieldIndex].Type, _fileName, _loadInfo->Fields[fieldIndex].Name);
break;
}
++fieldIndex;
}
}
}
}
return dataTable;
}
char* DB2FileLoaderSparseImpl::AutoProduceStrings(char** indexTable, uint32 indexTableSize, uint32 locale)
{
if (_loadInfo->Meta->FieldCount != _header->FieldCount)
throw DB2FileLoadException(Trinity::StringFormat("Found unsupported parent index in sparse db2 {}", _fileName));
if (!(_header->Locale & (1 << locale)))
{
char const* sep = "";
std::ostringstream str;
for (uint32 i = 0; i < TOTAL_LOCALES; ++i)
{
if (_header->Locale & (1 << i))
{
str << sep << localeNames[i];
sep = ", ";
}
}
TC_LOG_ERROR("", "Attempted to load {} which has locales {} as {}. Check if you placed your localized db2 files in correct directory.", _fileName, str.str(), localeNames[locale]);
return nullptr;
}
uint32 records = _catalog.size();
uint32 recordsize = _loadInfo->Meta->GetRecordSize();
std::size_t stringFields = _loadInfo->GetStringFieldCount(false);
std::size_t localizedStringFields = _loadInfo->GetStringFieldCount(true);
if (!stringFields)
return nullptr;
std::size_t stringsInRecordSize = (stringFields - localizedStringFields) * sizeof(char*);
std::size_t localizedStringsInRecordSize = localizedStringFields * sizeof(LocalizedString);
// string table size is "total size of all records" - RecordCount * "size of record without strings"
std::size_t stringTableSize = _totalRecordSize - (records * ((recordsize - (!_loadInfo->Meta->HasIndexFieldInData() ? 4 : 0)) - stringsInRecordSize - localizedStringsInRecordSize));
char* stringTable = new char[stringTableSize];
memset(stringTable, 0, stringTableSize);
char* stringPtr = stringTable;
uint32 y = 0;
for (uint32 section = 0; section < _header->SectionCount; ++section)
{
DB2SectionHeader const& sectionHeader = GetSection(section);
if (sectionHeader.TactId)
{
y += sectionHeader.RecordCount;
continue;
}
for (uint32 sr = 0; sr < sectionHeader.CatalogDataCount; ++sr, ++y)
{
unsigned char const* rawRecord = GetRawRecordData(y, §ion);
if (!rawRecord)
continue;
uint32 indexVal = _catalogIds[y];
if (indexVal >= indexTableSize)
continue;
char* recordData = indexTable[indexVal];
uint32 offset = 0;
uint32 fieldIndex = 0;
if (!_loadInfo->Meta->HasIndexFieldInData())
{
offset += 4;
++fieldIndex;
}
for (uint32 x = 0; x < _header->FieldCount; ++x)
{
for (uint32 z = 0; z < _loadInfo->Meta->Fields[x].ArraySize; ++z)
{
switch (_loadInfo->Fields[fieldIndex].Type)
{
case FT_FLOAT:
offset += 4;
break;
case FT_INT:
offset += 4;
break;
case FT_BYTE:
offset += 1;
break;
case FT_SHORT:
offset += 2;
break;
case FT_LONG:
offset += 8;
break;
case FT_STRING:
{
LocalizedString* db2str = reinterpret_cast(&recordData[offset]);
db2str->Str[locale] = stringPtr;
strcpy(stringPtr, RecordGetString(rawRecord, x, z));
stringPtr += strlen(stringPtr) + 1;
offset += sizeof(LocalizedString);
break;
}
case FT_STRING_NOT_LOCALIZED:
{
char const** db2str = reinterpret_cast(&recordData[offset]);
*db2str = stringPtr;
strcpy(stringPtr, RecordGetString(rawRecord, x, z));
stringPtr += strlen(stringPtr) + 1;
offset += sizeof(char*);
break;
}
default:
ABORT_MSG("Unknown format character '%c' found in %s meta for field %s",
_loadInfo->Fields[fieldIndex].Type, _fileName, _loadInfo->Fields[fieldIndex].Name);
break;
}
++fieldIndex;
}
}
}
}
return stringTable;
}
void DB2FileLoaderSparseImpl::AutoProduceRecordCopies(uint32 records, char** indexTable, char* dataTable)
{
uint32 recordCopies = GetRecordCopyCount();
if (!recordCopies)
return;
uint32 recordsize = _loadInfo->Meta->GetRecordSize();
uint32 offset = _header->RecordCount * recordsize;
uint32 idFieldOffset = _loadInfo->Meta->HasIndexFieldInData() ? _loadInfo->Meta->GetIndexFieldOffset() : 0;
for (uint32 c = 0; c < recordCopies; ++c)
{
DB2RecordCopy copy = GetRecordCopy(c);
if (copy.SourceRowId && copy.SourceRowId < records && copy.NewRowId < records && indexTable[copy.SourceRowId])
{
indexTable[copy.NewRowId] = &dataTable[offset];
memcpy(indexTable[copy.NewRowId], indexTable[copy.SourceRowId], recordsize);
*reinterpret_cast(&dataTable[offset + idFieldOffset]) = copy.NewRowId;
offset += recordsize;
}
}
}
DB2Record DB2FileLoaderSparseImpl::GetRecord(uint32 recordNumber) const
{
return DB2Record(*this, recordNumber, _fieldAndArrayOffsets.get());
}
DB2RecordCopy DB2FileLoaderSparseImpl::GetRecordCopy(uint32 copyNumber) const
{
if (copyNumber >= GetRecordCopyCount())
return DB2RecordCopy{};
return _copyTable[copyNumber];
}
uint32 DB2FileLoaderSparseImpl::GetRecordCount() const
{
return _catalog.size();
}
uint32 DB2FileLoaderSparseImpl::GetRecordCopyCount() const
{
return _copyTable.size();
}
void DB2FileLoaderSparseImpl::FillParentLookup(char* dataTable)
{
int32 parentIdOffset = _loadInfo->Meta->GetParentIndexFieldOffset();
uint32 recordSize = _loadInfo->Meta->GetRecordSize();
uint32 recordIndexOffset = 0;
for (uint32 i = 0; i < _header->SectionCount; ++i)
{
DB2SectionHeader const& section = GetSection(i);
if (!section.TactId)
{
for (std::size_t j = 0; j < _parentIndexes[i][0].Entries.size(); ++j)
{
uint32 parentId = _parentIndexes[i][0].Entries[j].ParentId;
char* recordData = &dataTable[(_parentIndexes[i][0].Entries[j].RecordIndex + recordIndexOffset) * recordSize];
switch (_loadInfo->Meta->Fields[_loadInfo->Meta->ParentIndexField].Type)
{
case FT_SHORT:
{
if (_loadInfo->Meta->ParentIndexField >= int32(_loadInfo->Meta->FileFieldCount))
{
// extra field at the end
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
else
{
// in data block, must fit
ASSERT(parentId <= 0xFFFF, "ParentId value %u does not fit into uint16 field (%s in %s)",
parentId, _loadInfo->Fields[_loadInfo->GetFieldIndexByMetaIndex(_loadInfo->Meta->ParentIndexField)].Name, _fileName);
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
break;
}
case FT_BYTE:
{
if (_loadInfo->Meta->ParentIndexField >= int32(_loadInfo->Meta->FileFieldCount))
{
// extra field at the end
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
else
{
// in data block, must fit
ASSERT(parentId <= 0xFF, "ParentId value %u does not fit into uint8 field (%s in %s)",
parentId, _loadInfo->Fields[_loadInfo->GetFieldIndexByMetaIndex(_loadInfo->Meta->ParentIndexField)].Name, _fileName);
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
}
break;
}
case FT_INT:
*reinterpret_cast(&recordData[parentIdOffset]) = parentId;
break;
default:
ABORT_MSG("Unhandled parent id type '%c' found in %s", _loadInfo->Meta->Fields[_loadInfo->Meta->ParentIndexField].Type, _fileName);
break;
}
}
}
recordIndexOffset += section.RecordCount;
}
}
uint32 DB2FileLoaderSparseImpl::GetRecordSection(uint32 recordNumber) const
{
uint32 section = 0;
for (; section < _header->SectionCount; ++section)
{
DB2SectionHeader const& sectionHeader = GetSection(section);
if (recordNumber < sectionHeader.CatalogDataCount)
break;
recordNumber -= sectionHeader.CatalogDataCount;
}
return section;
}
unsigned char const* DB2FileLoaderSparseImpl::GetRawRecordData(uint32 recordNumber, uint32 const* section) const
{
if (recordNumber >= _catalog.size())
return nullptr;
if (GetSection(section ? *section : GetRecordSection(recordNumber)).TactId)
return nullptr;
_source->SetPosition(_catalog[recordNumber].FileOffset);
uint8* rawRecord = _recordBuffer.get();
if (!_source->Read(rawRecord, _catalog[recordNumber].RecordSize))
return nullptr;
CalculateAndStoreFieldOffsets(rawRecord);
return rawRecord;
}
uint32 DB2FileLoaderSparseImpl::RecordGetId(uint8 const* record, uint32 recordIndex) const
{
if (_loadInfo->Meta->HasIndexFieldInData())
return RecordGetVarInt(record, _loadInfo->Meta->GetIndexField(), 0, false);
return _catalogIds[recordIndex];
}
uint8 DB2FileLoaderSparseImpl::RecordGetUInt8(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
ASSERT(field < _header->FieldCount);
return *reinterpret_cast(record + GetFieldOffset(field, arrayIndex));
}
uint16 DB2FileLoaderSparseImpl::RecordGetUInt16(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
ASSERT(field < _header->FieldCount);
uint16 val = *reinterpret_cast(record + GetFieldOffset(field, arrayIndex));
EndianConvert(val);
return val;
}
uint32 DB2FileLoaderSparseImpl::RecordGetUInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return RecordGetVarInt(record, field, arrayIndex, false);
}
int32 DB2FileLoaderSparseImpl::RecordGetInt32(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
return int32(RecordGetVarInt(record, field, arrayIndex, true));
}
uint64 DB2FileLoaderSparseImpl::RecordGetUInt64(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
ASSERT(field < _header->FieldCount);
uint64 val = *reinterpret_cast(record + GetFieldOffset(field, arrayIndex));
EndianConvert(val);
return val;
}
float DB2FileLoaderSparseImpl::RecordGetFloat(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
ASSERT(field < _header->FieldCount);
float val = *reinterpret_cast(record + GetFieldOffset(field, arrayIndex));
EndianConvert(val);
return val;
}
char const* DB2FileLoaderSparseImpl::RecordGetString(uint8 const* record, uint32 field, uint32 arrayIndex) const
{
ASSERT(field < _header->FieldCount);
return reinterpret_cast(record + GetFieldOffset(field, arrayIndex));
}
uint32 DB2FileLoaderSparseImpl::RecordGetVarInt(uint8 const* record, uint32 field, uint32 arrayIndex, bool isSigned) const
{
ASSERT(field < _header->FieldCount);
uint32 val = 0;
memcpy(&val, record + GetFieldOffset(field, arrayIndex), GetFieldSize(field));
EndianConvert(val);
if (isSigned)
return int32(val) << _fields[field].UnusedBits >> _fields[field].UnusedBits;
return val << _fields[field].UnusedBits >> _fields[field].UnusedBits;
}
uint16 DB2FileLoaderSparseImpl::GetFieldOffset(uint32 field, uint32 arrayIndex) const
{
return uint16(_fieldAndArrayOffsets[_fieldAndArrayOffsets[field] + arrayIndex]);
}
uint16 DB2FileLoaderSparseImpl::GetFieldSize(uint32 field) const
{
ASSERT(field < _header->FieldCount);
return 4 - _fields[field].UnusedBits / 8;
}
std::size_t* DB2FileLoaderSparseImpl::RecordCreateDetachedFieldOffsets(std::size_t* oldOffsets) const
{
if (oldOffsets != _fieldAndArrayOffsets.get())
return oldOffsets;
std::size_t size = _loadInfo->Meta->FieldCount + _loadInfo->FieldCount - (!_loadInfo->Meta->HasIndexFieldInData() ? 1 : 0);
std::size_t* newOffsets = new std::size_t[size];
memcpy(newOffsets, _fieldAndArrayOffsets.get(), size * sizeof(std::size_t));
return newOffsets;
}
std::size_t* DB2FileLoaderSparseImpl::RecordCopyDetachedFieldOffsets(std::size_t* oldOffsets) const
{
if (oldOffsets == _fieldAndArrayOffsets.get())
return oldOffsets;
std::size_t size = _loadInfo->Meta->FieldCount + _loadInfo->FieldCount - (!_loadInfo->Meta->HasIndexFieldInData() ? 1 : 0);
std::size_t* newOffsets = new std::size_t[size];
memcpy(newOffsets, _fieldAndArrayOffsets.get(), size * sizeof(std::size_t));
return newOffsets;
}
void DB2FileLoaderSparseImpl::RecordDestroyFieldOffsets(std::size_t*& fieldOffsets) const
{
if (fieldOffsets == _fieldAndArrayOffsets.get())
return;
delete[] fieldOffsets;
fieldOffsets = nullptr;
}
void DB2FileLoaderSparseImpl::CalculateAndStoreFieldOffsets(uint8 const* rawRecord) const
{
std::size_t offset = 0;
uint32 combinedField = _loadInfo->Meta->FieldCount;
for (uint32 field = 0; field < _loadInfo->Meta->FieldCount; ++field)
{
_fieldAndArrayOffsets[field] = combinedField;
for (uint32 arr = 0; arr < _loadInfo->Meta->Fields[field].ArraySize; ++arr)
{
_fieldAndArrayOffsets[combinedField] = offset;
switch (_loadInfo->Meta->Fields[field].Type)
{
case FT_BYTE:
case FT_SHORT:
case FT_INT:
case FT_LONG:
offset += GetFieldSize(field);
break;
case FT_FLOAT:
offset += sizeof(float);
break;
case FT_STRING:
case FT_STRING_NOT_LOCALIZED:
offset += strlen(reinterpret_cast(rawRecord) + offset) + 1;
break;
default:
ABORT_MSG("Unknown format character '%c' found in %s meta", _loadInfo->Meta->Fields[field].Type, _fileName);
break;
}
++combinedField;
}
}
}
uint32 DB2FileLoaderSparseImpl::GetMinId() const
{
return _header->MinId;
}
uint32 DB2FileLoaderSparseImpl::GetMaxId() const
{
return _header->MaxId;
}
DB2FileLoadInfo const* DB2FileLoaderSparseImpl::GetLoadInfo() const
{
return _loadInfo;
}
DB2SectionHeader& DB2FileLoaderSparseImpl::GetSection(uint32 section) const
{
return _sections[section];
}
bool DB2FileLoaderSparseImpl::IsSignedField(uint32 field) const
{
ASSERT(field < _header->FieldCount);
return _loadInfo->Meta->IsSignedField(field);
}
char const* DB2FileLoaderSparseImpl::GetExpectedSignMismatchReason(uint32 field) const
{
ASSERT(field < _header->FieldCount);
if (int32(field) == _loadInfo->Meta->IndexField)
return " (IndexField must always be unsigned)";
if (int32(field) == _loadInfo->Meta->ParentIndexField)
return " (ParentIndexField must always be unsigned)";
return "";
}
DB2Record::DB2Record(DB2FileLoaderImpl const& db2, uint32 recordIndex, std::size_t* fieldOffsets)
: _db2(db2), _recordIndex(recordIndex), _recordData(db2.GetRawRecordData(recordIndex, nullptr)), _fieldOffsets(fieldOffsets)
{
}
DB2Record::DB2Record(DB2Record const& other)
: _db2(other._db2), _recordIndex(other._recordIndex), _recordData(other._recordData), _fieldOffsets(_db2.RecordCopyDetachedFieldOffsets(other._fieldOffsets))
{
}
DB2Record::DB2Record(DB2Record&& other) noexcept
: _db2(other._db2), _recordIndex(other._recordIndex), _recordData(other._recordData), _fieldOffsets(std::exchange(other._fieldOffsets, nullptr))
{
}
DB2Record::~DB2Record()
{
_db2.RecordDestroyFieldOffsets(_fieldOffsets);
}
DB2Record::operator bool() const
{
return _recordData != nullptr;
}
uint32 DB2Record::GetId() const
{
return _db2.RecordGetId(_recordData, _recordIndex);
}
uint8 DB2Record::GetUInt8(uint32 field, uint32 arrayIndex) const
{
return _db2.RecordGetUInt8(_recordData, field, arrayIndex);
}
uint8 DB2Record::GetUInt8(char const* fieldName) const
{
std::pair fieldIndex = _db2.GetLoadInfo()->GetFieldIndexByName(fieldName);
ASSERT(fieldIndex.first != -1, "Field with name %s does not exist!", fieldName);
return _db2.RecordGetUInt8(_recordData, uint32(fieldIndex.first), uint32(fieldIndex.second));
}
uint16 DB2Record::GetUInt16(uint32 field, uint32 arrayIndex) const
{
return _db2.RecordGetUInt16(_recordData, field, arrayIndex);
}
uint16 DB2Record::GetUInt16(char const* fieldName) const
{
std::pair fieldIndex = _db2.GetLoadInfo()->GetFieldIndexByName(fieldName);
ASSERT(fieldIndex.first != -1, "Field with name %s does not exist!", fieldName);
return _db2.RecordGetUInt16(_recordData, uint32(fieldIndex.first), uint32(fieldIndex.second));
}
uint32 DB2Record::GetUInt32(uint32 field, uint32 arrayIndex) const
{
return _db2.RecordGetUInt32(_recordData, field, arrayIndex);
}
uint32 DB2Record::GetUInt32(char const* fieldName) const
{
std::pair fieldIndex = _db2.GetLoadInfo()->GetFieldIndexByName(fieldName);
ASSERT(fieldIndex.first != -1, "Field with name %s does not exist!", fieldName);
return _db2.RecordGetUInt32(_recordData, uint32(fieldIndex.first), uint32(fieldIndex.second));
}
int32 DB2Record::GetInt32(uint32 field, uint32 arrayIndex) const
{
return _db2.RecordGetInt32(_recordData, field, arrayIndex);
}
int32 DB2Record::GetInt32(char const* fieldName) const
{
std::pair fieldIndex = _db2.GetLoadInfo()->GetFieldIndexByName(fieldName);
ASSERT(fieldIndex.first != -1, "Field with name %s does not exist!", fieldName);
return _db2.RecordGetInt32(_recordData, uint32(fieldIndex.first), uint32(fieldIndex.second));
}
uint64 DB2Record::GetUInt64(uint32 field, uint32 arrayIndex) const
{
return _db2.RecordGetUInt64(_recordData, field, arrayIndex);
}
uint64 DB2Record::GetUInt64(char const* fieldName) const
{
std::pair fieldIndex = _db2.GetLoadInfo()->GetFieldIndexByName(fieldName);
ASSERT(fieldIndex.first != -1, "Field with name %s does not exist!", fieldName);
return _db2.RecordGetUInt64(_recordData, uint32(fieldIndex.first), uint32(fieldIndex.second));
}
float DB2Record::GetFloat(uint32 field, uint32 arrayIndex) const
{
return _db2.RecordGetFloat(_recordData, field, arrayIndex);
}
float DB2Record::GetFloat(char const* fieldName) const
{
std::pair fieldIndex = _db2.GetLoadInfo()->GetFieldIndexByName(fieldName);
ASSERT(fieldIndex.first != -1, "Field with name %s does not exist!", fieldName);
return _db2.RecordGetFloat(_recordData, uint32(fieldIndex.first), uint32(fieldIndex.second));
}
char const* DB2Record::GetString(uint32 field, uint32 arrayIndex) const
{
return _db2.RecordGetString(_recordData, field, arrayIndex);
}
char const* DB2Record::GetString(char const* fieldName) const
{
std::pair fieldIndex = _db2.GetLoadInfo()->GetFieldIndexByName(fieldName);
ASSERT(fieldIndex.first != -1, "Field with name %s does not exist!", fieldName);
return _db2.RecordGetString(_recordData, uint32(fieldIndex.first), uint32(fieldIndex.second));
}
void DB2Record::MakePersistent()
{
_fieldOffsets = _db2.RecordCreateDetachedFieldOffsets(_fieldOffsets);
}
DB2FileLoader::DB2FileLoader() : _impl(nullptr), _header()
{
}
DB2FileLoader::~DB2FileLoader()
{
delete _impl;
}
void DB2FileLoader::LoadHeaders(DB2FileSource* source, DB2FileLoadInfo const* loadInfo)
{
if (!source->IsOpen())
throw std::system_error(std::make_error_code(std::errc::no_such_file_or_directory));
if (!source->Read(&_header, sizeof(DB2Header)))
throw DB2FileLoadException("Failed to read header");
EndianConvert(_header.Signature);
EndianConvert(_header.RecordCount);
EndianConvert(_header.FieldCount);
EndianConvert(_header.RecordSize);
EndianConvert(_header.StringTableSize);
EndianConvert(_header.TableHash);
EndianConvert(_header.LayoutHash);
EndianConvert(_header.MinId);
EndianConvert(_header.MaxId);
EndianConvert(_header.Locale);
EndianConvert(_header.Flags);
EndianConvert(_header.IndexField);
EndianConvert(_header.TotalFieldCount);
EndianConvert(_header.PackedDataOffset);
EndianConvert(_header.ParentLookupCount);
EndianConvert(_header.ColumnMetaSize);
EndianConvert(_header.CommonDataSize);
EndianConvert(_header.PalletDataSize);
EndianConvert(_header.SectionCount);
if (_header.Signature != 0x33434457) //'WDC3'
throw DB2FileLoadException(Trinity::StringFormat("Incorrect file signature in {}, expected 'WDC3', got %c%c%c%c", source->GetFileName(),
char(_header.Signature & 0xFF), char((_header.Signature >> 8) & 0xFF), char((_header.Signature >> 16) & 0xFF), char((_header.Signature >> 24) & 0xFF)));
if (loadInfo && _header.LayoutHash != loadInfo->Meta->LayoutHash)
throw DB2FileLoadException(Trinity::StringFormat("Incorrect layout hash in {}, expected 0x{:08X}, got 0x{:08X} (possibly wrong client version)",
source->GetFileName(), loadInfo->Meta->LayoutHash, _header.LayoutHash));
if (_header.ParentLookupCount > 1)
throw DB2FileLoadException(Trinity::StringFormat("Too many parent lookups in {}, only one is allowed, got {}",
source->GetFileName(), _header.ParentLookupCount));
if (loadInfo && (_header.TotalFieldCount + (loadInfo->Meta->ParentIndexField >= int32(_header.TotalFieldCount) ? 1 : 0) != loadInfo->Meta->FieldCount))
throw DB2FileLoadException(Trinity::StringFormat("Incorrect number of fields in {}, expected {}, got {}",
source->GetFileName(), loadInfo->Meta->FieldCount, _header.TotalFieldCount + (loadInfo->Meta->ParentIndexField >= int32(_header.TotalFieldCount) ? 1 : 0)));
if (loadInfo && (_header.ParentLookupCount && loadInfo->Meta->ParentIndexField == -1))
throw DB2FileLoadException(Trinity::StringFormat("Unexpected parent lookup found in {}", source->GetFileName()));
std::unique_ptr sections = std::make_unique(_header.SectionCount);
if (_header.SectionCount && !source->Read(sections.get(), sizeof(DB2SectionHeader) * _header.SectionCount))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read section headers from {}", source->GetFileName()));
uint32 totalCopyTableSize = 0;
uint32 totalParentLookupDataSize = 0;
for (uint32 i = 0; i < _header.SectionCount; ++i)
{
totalCopyTableSize += sections[i].CopyTableCount * sizeof(DB2RecordCopy);
totalParentLookupDataSize += sections[i].ParentLookupDataSize;
}
if (loadInfo && !(_header.Flags & 0x1))
{
int64 expectedFileSize =
sizeof(DB2Header) +
sizeof(DB2SectionHeader) * _header.SectionCount +
sizeof(DB2FieldEntry) * _header.FieldCount +
int64(_header.RecordSize) * _header.RecordCount +
_header.StringTableSize +
(loadInfo->Meta->IndexField == -1 ? sizeof(uint32) * _header.RecordCount : 0) +
totalCopyTableSize +
_header.ColumnMetaSize +
_header.PalletDataSize +
_header.CommonDataSize +
totalParentLookupDataSize;
if (source->GetFileSize() != expectedFileSize)
throw DB2FileLoadException(Trinity::StringFormat("{} failed size consistency check, expected {}, got {}", source->GetFileName(), expectedFileSize, source->GetFileSize()));
}
std::unique_ptr fieldData = std::make_unique(_header.FieldCount);
if (!source->Read(fieldData.get(), sizeof(DB2FieldEntry) * _header.FieldCount))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read field information from {}", source->GetFileName()));
std::unique_ptr columnMeta;
std::unique_ptr[]> palletValues;
std::unique_ptr[]> palletArrayValues;
std::unique_ptr[]> commonValues;
if (_header.ColumnMetaSize)
{
columnMeta = std::make_unique(_header.TotalFieldCount);
if (!source->Read(columnMeta.get(), _header.ColumnMetaSize))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read field metadata from {}", source->GetFileName()));
if (loadInfo && loadInfo->Meta->HasIndexFieldInData())
{
if (columnMeta[loadInfo->Meta->IndexField].CompressionType != DB2ColumnCompression::None
&& columnMeta[loadInfo->Meta->IndexField].CompressionType != DB2ColumnCompression::Immediate
&& columnMeta[loadInfo->Meta->IndexField].CompressionType != DB2ColumnCompression::SignedImmediate)
throw DB2FileLoadException(Trinity::StringFormat("Invalid compression type for index field in {}, expected one of None (0), Immediate (1), SignedImmediate (5), got {}",
source->GetFileName(), uint32(columnMeta[loadInfo->Meta->IndexField].CompressionType)));
}
palletValues = std::make_unique[]>(_header.TotalFieldCount);
for (uint32 i = 0; i < _header.TotalFieldCount; ++i)
{
if (columnMeta[i].CompressionType != DB2ColumnCompression::Pallet)
continue;
palletValues[i] = std::make_unique(columnMeta[i].AdditionalDataSize / sizeof(DB2PalletValue));
if (!source->Read(palletValues[i].get(), columnMeta[i].AdditionalDataSize))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read field pallet values from {} for field {}", source->GetFileName(), i));
}
palletArrayValues = std::make_unique[]>(_header.TotalFieldCount);
for (uint32 i = 0; i < _header.TotalFieldCount; ++i)
{
if (columnMeta[i].CompressionType != DB2ColumnCompression::PalletArray)
continue;
palletArrayValues[i] = std::make_unique(columnMeta[i].AdditionalDataSize / sizeof(DB2PalletValue));
if (!source->Read(palletArrayValues[i].get(), columnMeta[i].AdditionalDataSize))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read field pallet array values from {} for field {}", source->GetFileName(), i));
}
std::unique_ptr[]> commonData = std::make_unique[]>(_header.TotalFieldCount);
commonValues = std::make_unique[]>(_header.TotalFieldCount);
for (uint32 i = 0; i < _header.TotalFieldCount; ++i)
{
if (columnMeta[i].CompressionType != DB2ColumnCompression::CommonData)
continue;
if (!columnMeta[i].AdditionalDataSize)
continue;
commonData[i] = std::make_unique(columnMeta[i].AdditionalDataSize / sizeof(DB2CommonValue));
if (!source->Read(commonData[i].get(), columnMeta[i].AdditionalDataSize))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read field common values from {} for field {}", source->GetFileName(), i));
uint32 numExtraValuesForField = columnMeta[i].AdditionalDataSize / sizeof(DB2CommonValue);
for (uint32 record = 0; record < numExtraValuesForField; ++record)
{
uint32 recordId = commonData[i][record].RecordId;
uint32 value = commonData[i][record].Value;
EndianConvert(value);
commonValues[i][recordId] = value;
}
}
}
if (!(_header.Flags & 0x1))
_impl = new DB2FileLoaderRegularImpl(source->GetFileName(), loadInfo, &_header);
else
_impl = new DB2FileLoaderSparseImpl(source->GetFileName(), loadInfo, &_header, source);
_impl->LoadColumnData(std::move(sections), std::move(fieldData), std::move(columnMeta), std::move(palletValues), std::move(palletArrayValues), std::move(commonValues));
}
void DB2FileLoader::Load(DB2FileSource* source, DB2FileLoadInfo const* loadInfo)
{
LoadHeaders(source, loadInfo);
std::vector idTable;
std::vector copyTable;
std::vector> parentIndexes;
if (loadInfo && !loadInfo->Meta->HasIndexFieldInData() && _header.RecordCount)
idTable.reserve(_header.RecordCount);
if (_header.ParentLookupCount)
{
parentIndexes.resize(_header.SectionCount);
for (std::vector& parentIndexesForSection : parentIndexes)
parentIndexesForSection.resize(_header.ParentLookupCount);
}
for (uint32 i = 0; i < _header.SectionCount; ++i)
{
DB2SectionHeader& section = _impl->GetSection(i);
if (section.TactId)
{
switch (source->HandleEncryptedSection(section))
{
case DB2EncryptedSectionHandling::Skip:
_impl->SkipEncryptedSection(i);
idTable.resize(idTable.size() + section.IdTableSize / sizeof(uint32));
continue;
case DB2EncryptedSectionHandling::Process:
section.TactId = 0;
break;
default:
break;
}
}
if (!source->SetPosition(section.FileOffset))
throw DB2FileLoadException(Trinity::StringFormat("Unable to change {} read position for section {}", source->GetFileName(), i));
if (!_impl->LoadTableData(source, i))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read section table data from {} for section {}", source->GetFileName(), i));
if (!_impl->LoadCatalogData(source, i))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read section catalog data from {} for section {}", source->GetFileName(), i));
if (loadInfo)
{
if (loadInfo->Meta->HasIndexFieldInData())
{
if (section.IdTableSize != 0)
throw DB2FileLoadException(Trinity::StringFormat("Unexpected id table found in {} for section {}", source->GetFileName(), i));
}
else if (section.IdTableSize != 4 * section.RecordCount)
throw DB2FileLoadException(Trinity::StringFormat("Unexpected id table size in {} for section {}, expected {}, got {}",
source->GetFileName(), i, 4 * section.RecordCount, section.IdTableSize));
}
if (section.IdTableSize)
{
std::size_t idTableSize = idTable.size();
idTable.resize(idTableSize + section.IdTableSize / sizeof(uint32));
if (!source->Read(&idTable[idTableSize], section.IdTableSize))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read non-inline record ids from {} for section {}", source->GetFileName(), i));
// This is a hack fix for broken db2 files that have invalid id tables
for (std::size_t i = idTableSize; i < idTable.size(); ++i)
if (idTable[i] <= _header.MinId)
idTable[i] = _header.MinId + i;
}
if (!(_header.Flags & 0x1) && section.CopyTableCount)
{
std::size_t copyTableSize = copyTable.size();
copyTable.resize(copyTableSize + section.CopyTableCount);
if (!source->Read(©Table[copyTableSize], section.CopyTableCount * sizeof(DB2RecordCopy)))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read record copies from {} for section {}", source->GetFileName(), i));
}
if (_header.ParentLookupCount)
{
std::vector& parentIndexesForSection = parentIndexes[i];
for (uint32 j = 0; j < _header.ParentLookupCount; ++j)
{
DB2IndexDataInfo indexInfo;
if (!source->Read(&indexInfo, sizeof(DB2IndexDataInfo)))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read parent lookup info from {} for section {}", source->GetFileName(), i));
if (!indexInfo.NumEntries)
continue;
parentIndexesForSection[j].Entries.resize(indexInfo.NumEntries);
if (!source->Read(parentIndexesForSection[j].Entries.data(), sizeof(DB2IndexEntry) * indexInfo.NumEntries))
throw DB2FileLoadException(Trinity::StringFormat("Unable to read parent lookup content from {} for section {}", source->GetFileName(), i));
}
}
}
_impl->SetAdditionalData(std::move(idTable), std::move(copyTable), std::move(parentIndexes));
if (loadInfo)
{
uint32 fieldIndex = 0;
if (!loadInfo->Meta->HasIndexFieldInData())
{
ASSERT(!loadInfo->Fields[0].IsSigned, "ID must be unsigned in %s", source->GetFileName());
++fieldIndex;
}
for (uint32 f = 0; f < loadInfo->Meta->FieldCount; ++f)
{
ASSERT(loadInfo->Fields[fieldIndex].IsSigned == _impl->IsSignedField(f),
"Field %s in %s must be %s%s", loadInfo->Fields[fieldIndex].Name, source->GetFileName(), _impl->IsSignedField(f) ? "signed" : "unsigned",
_impl->GetExpectedSignMismatchReason(f));
fieldIndex += loadInfo->Meta->Fields[f].ArraySize;
}
}
}
char* DB2FileLoader::AutoProduceData(uint32& indexTableSize, char**& indexTable)
{
return _impl->AutoProduceData(indexTableSize, indexTable);
}
char* DB2FileLoader::AutoProduceStrings(char** indexTable, uint32 indexTableSize, LocaleConstant locale)
{
return _impl->AutoProduceStrings(indexTable, indexTableSize, locale);
}
void DB2FileLoader::AutoProduceRecordCopies(uint32 records, char** indexTable, char* dataTable)
{
_impl->AutoProduceRecordCopies(records, indexTable, dataTable);
}
uint32 DB2FileLoader::GetRecordCount() const
{
return _impl->GetRecordCount();
}
uint32 DB2FileLoader::GetRecordCopyCount() const
{
return _impl->GetRecordCopyCount();
}
uint32 DB2FileLoader::GetMinId() const
{
return _impl->GetMinId();
}
uint32 DB2FileLoader::GetMaxId() const
{
return _impl->GetMaxId();
}
DB2SectionHeader const& DB2FileLoader::GetSectionHeader(uint32 section) const
{
return _impl->GetSection(section);
}
DB2Record DB2FileLoader::GetRecord(uint32 recordNumber) const
{
return _impl->GetRecord(recordNumber);
}
DB2RecordCopy DB2FileLoader::GetRecordCopy(uint32 copyNumber) const
{
return _impl->GetRecordCopy(copyNumber);
}