/* * Copyright (c) 2023 Huawei Device Co., Ltd. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "becommon.h" #include "rt.h" #include "cg_option.h" #include "mir_builder.h" #include "mpl_logging.h" #include #include namespace maplebe { using namespace maple; BECommon::BECommon(MIRModule &mod) : mirModule(mod), typeSizeTable(GlobalTables::GetTypeTable().GetTypeTable().size(), 0, mirModule.GetMPAllocator().Adapter()), typeAlignTable(GlobalTables::GetTypeTable().GetTypeTable().size(), static_cast(mirModule.IsCModule()), mirModule.GetMPAllocator().Adapter()), typeHasFlexibleArray(GlobalTables::GetTypeTable().GetTypeTable().size(), 0, mirModule.GetMPAllocator().Adapter()), structFieldCountTable(GlobalTables::GetTypeTable().GetTypeTable().size(), 0, mirModule.GetMPAllocator().Adapter()), jClassLayoutTable(mirModule.GetMPAllocator().Adapter()), funcReturnType(mirModule.GetMPAllocator().Adapter()) { for (uint32 i = 1; i < GlobalTables::GetTypeTable().GetTypeTable().size(); ++i) { MIRType *ty = GlobalTables::GetTypeTable().GetTypeTable()[i]; ComputeTypeSizesAligns(*ty); } } /* * try to find an available padding slot, and allocate the given field in it. * return the offset of the allocated memory. 0 if not available * Note: this will update lists in paddingSlots * Note: padding slots is a list of un-occupied (small size) slots * available to allocate new fields. so far, just for 1, 2, 4 bytes * types (map to array index 0, 1, 2) */ static uint32 TryAllocInPaddingSlots(std::list paddingSlots[], uint32 fieldSize, uint32 fieldAlign, size_t paddingSlotsLength) { CHECK_FATAL(paddingSlotsLength > 0, "expect paddingSlotsLength > 0"); if (fieldSize > 4) { // padding slots are just for size 1/2/4 bytes return 0; } uint32 fieldOffset = 0; /* here is a greedy search */ for (size_t freeSlot = static_cast(fieldSize >> 1); freeSlot < paddingSlotsLength; ++freeSlot) { if (!paddingSlots[freeSlot].empty()) { uint32 paddingOffset = paddingSlots[freeSlot].front(); if (IsAlignedTo(paddingOffset, fieldAlign)) { /* reuse one padding slot */ paddingSlots[freeSlot].pop_front(); fieldOffset = paddingOffset; /* check whether there're still space left in this slot */ uint32 leftSize = (1u << freeSlot) - fieldSize; if (leftSize != 0) { uint32 leftOffset = paddingOffset + fieldSize; if (leftSize & 0x1) { /* check whether the last bit is 1 */ paddingSlots[0].push_front(leftOffset); leftOffset += 1; } if (leftSize & 0x2) { /* check whether the penultimate bit is 1 */ paddingSlots[1].push_front(leftOffset); } } break; } } } return fieldOffset; } static void AddPaddingSlot(std::list paddingSlots[], uint32 offset, uint32 size, size_t paddingSlotsLength) { CHECK_FATAL(paddingSlotsLength > 0, "expect paddingSlotsLength > 0"); /* * decompose the padding into 1/2/4 bytes slots. * to satisfy alignment constraints. */ for (size_t i = 0; i < paddingSlotsLength; ++i) { if (size & (1u << i)) { paddingSlots[i].push_front(offset); offset += (1u << i); } } } void BECommon::AddNewTypeAfterBecommon(uint32 oldTypeTableSize, uint32 newTypeTableSize) { for (auto i = oldTypeTableSize; i < newTypeTableSize; ++i) { MIRType *ty = GlobalTables::GetTypeTable().GetTypeFromTyIdx(i); CHECK_NULL_FATAL(ty); typeSizeTable.emplace_back(0); typeAlignTable.emplace_back(static_cast(mirModule.IsCModule())); typeHasFlexibleArray.emplace_back(0); structFieldCountTable.emplace_back(0); ComputeTypeSizesAligns(*ty); } } void BECommon::ComputeStructTypeSizesAligns(MIRType &ty, const TyIdx &tyIdx) { auto &structType = static_cast(ty); const FieldVector &fields = structType.GetFields(); uint64 allocedSize = 0; uint64 allocedSizeInBits = 0; SetStructFieldCount(structType.GetTypeIndex(), fields.size()); if (fields.size() == 0) { if (structType.IsCPlusPlus()) { SetTypeSize(tyIdx.GetIdx(), 1); /* empty struct in C++ has size 1 */ SetTypeAlign(tyIdx.GetIdx(), 1); } else { SetTypeSize(tyIdx.GetIdx(), 0); SetTypeAlign(tyIdx.GetIdx(), k8ByteSize); } return; } auto structAttr = structType.GetTypeAttrs(); auto structPack = static_cast(structAttr.GetPack()); for (uint32 j = 0; j < fields.size(); ++j) { TyIdx fieldTyIdx = fields[j].second.first; auto fieldAttr = fields[j].second.second; MIRType *fieldType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(fieldTyIdx); uint32 fieldTypeSize = GetTypeSize(fieldTyIdx); if (fieldTypeSize == 0) { ComputeTypeSizesAligns(*fieldType); fieldTypeSize = GetTypeSize(fieldTyIdx); } uint64 fieldSizeBits = fieldTypeSize * kBitsPerByte; auto attrAlign = static_cast(fieldAttr.GetAlign()); auto originAlign = std::max(attrAlign, GetTypeAlign(fieldTyIdx)); uint8 fieldAlign = fieldAttr.IsPacked() ? 1 : std::min(originAlign, structPack); uint64 fieldAlignBits = fieldAlign * kBitsPerByte; CHECK_FATAL(fieldAlign != 0, "expect fieldAlign not equal 0"); MIRStructType *subStructType = fieldType->EmbeddedStructType(); if (subStructType != nullptr) { AppendStructFieldCount(structType.GetTypeIndex(), GetStructFieldCount(subStructType->GetTypeIndex())); } if (structType.GetKind() != kTypeUnion) { if (fieldType->GetKind() == kTypeBitField) { uint32 fieldSize = static_cast(fieldType)->GetFieldSize(); /* is this field is crossing the align boundary of its base type? */ if ((!structAttr.IsPacked() && ((allocedSizeInBits / fieldSizeBits) != ((allocedSizeInBits + fieldSize - 1u) / fieldSizeBits))) || fieldSize == 0) { allocedSizeInBits = RoundUp(allocedSizeInBits, fieldSizeBits); } /* allocate the bitfield */ allocedSizeInBits += fieldSize; allocedSize = std::max(allocedSize, RoundUp(allocedSizeInBits, fieldAlignBits) / kBitsPerByte); } else { bool leftoverbits = false; if (allocedSizeInBits == allocedSize * kBitsPerByte) { allocedSize = RoundUp(allocedSize, fieldAlign); } else { /* still some leftover bits on allocated words, we calculate things based on bits then. */ if (allocedSizeInBits / fieldAlignBits != (allocedSizeInBits + fieldSizeBits - 1) / fieldAlignBits) { /* the field is crossing the align boundary of its base type */ allocedSizeInBits = RoundUp(allocedSizeInBits, fieldAlignBits); } leftoverbits = true; } if (leftoverbits) { allocedSizeInBits += fieldSizeBits; allocedSize = std::max(allocedSize, RoundUp(allocedSizeInBits, fieldAlignBits) / kBitsPerByte); } else { /* pad alloced_size according to the field alignment */ allocedSize = RoundUp(allocedSize, fieldAlign); allocedSize += fieldTypeSize; allocedSizeInBits = allocedSize * kBitsPerByte; } } } else { /* for unions, bitfields are treated as non-bitfields */ allocedSize = std::max(allocedSize, static_cast(fieldTypeSize)); } SetTypeAlign(tyIdx, std::max(GetTypeAlign(tyIdx), fieldAlign)); /* C99 * Last struct element of a struct with more than one member * is a flexible array if it is an array of size 0. */ if ((j != 0) && ((j + 1) == fields.size()) && (fieldType->GetKind() == kTypeArray) && (GetTypeSize(fieldTyIdx.GetIdx()) == 0)) { SetHasFlexibleArray(tyIdx.GetIdx(), true); } } SetTypeSize(tyIdx, RoundUp(allocedSize, GetTypeAlign(tyIdx.GetIdx()))); } void BECommon::ComputeClassTypeSizesAligns(MIRType &ty, const TyIdx &tyIdx, uint8 align) { uint64 allocedSize = 0; const FieldVector &fields = static_cast(ty).GetFields(); auto &classType = static_cast(ty); TyIdx prntTyIdx = classType.GetParentTyIdx(); /* process parent class */ if (prntTyIdx != 0u) { MIRClassType *parentType = static_cast(GlobalTables::GetTypeTable().GetTypeFromTyIdx(prntTyIdx)); uint32 prntSize = GetTypeSize(prntTyIdx); if (prntSize == 0) { ComputeTypeSizesAligns(*parentType); prntSize = GetTypeSize(prntTyIdx); } uint8 prntAlign = GetTypeAlign(prntTyIdx); AppendStructFieldCount(tyIdx, GetStructFieldCount(prntTyIdx) + 1); /* pad alloced_size according to the field alignment */ allocedSize = RoundUp(allocedSize, prntAlign); JClassLayout *layout = mirModule.GetMemPool()->New(mirModule.GetMPAllocator().Adapter()); /* add parent's record to the front */ layout->emplace_back(JClassFieldInfo(false, false, false, allocedSize)); /* copy parent's layout plan into my plan */ if (HasJClassLayout(*parentType)) { /* parent may have incomplete type definition. */ const JClassLayout &parentLayout = GetJClassLayout(*parentType); layout->insert(layout->end(), parentLayout.begin(), parentLayout.end()); allocedSize += prntSize; SetTypeAlign(tyIdx, std::max(GetTypeAlign(tyIdx), prntAlign)); } else { LogInfo::MapleLogger() << "Warning:try to layout class with incomplete type:" << parentType->GetName() << "\n"; } jClassLayoutTable[&classType] = layout; } else { /* This is the root class, say, The Object */ jClassLayoutTable[&classType] = mirModule.GetMemPool()->New(mirModule.GetMPAllocator().Adapter()); } /* * a list of un-occupied (small size) slots available for insertion * so far, just for 1, 2, 4 bytes types (map to array index 0, 1, 2) */ std::list paddingSlots[3]; // padding slots are just 3 types for size 1/2/4 bytes /* process fields */ AppendStructFieldCount(tyIdx, fields.size()); if (fields.size() == 0 && mirModule.IsCModule()) { SetTypeAlign(tyIdx.GetIdx(), 1); SetTypeSize(tyIdx.GetIdx(), 1); return; } for (uint32 j = 0; j < fields.size(); ++j) { TyIdx fieldTyIdx = fields[j].second.first; MIRType *fieldType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(fieldTyIdx); FieldAttrs fieldAttr = fields[j].second.second; uint32 fieldSize = GetTypeSize(fieldTyIdx); if (fieldSize == 0) { ComputeTypeSizesAligns(*fieldType); fieldSize = GetTypeSize(fieldTyIdx); } uint8 fieldAlign = GetTypeAlign(fieldTyIdx); if ((fieldType->GetKind() == kTypePointer) && (fieldType->GetPrimType() == PTY_a64)) { /* handle class reference field */ fieldSize = static_cast(RTSupport::GetRTSupportInstance().GetFieldSize()); fieldAlign = RTSupport::GetRTSupportInstance().GetFieldAlign(); } /* try to alloc the field in one of previously created padding slots */ uint32 currentFieldOffset = TryAllocInPaddingSlots(paddingSlots, fieldSize, fieldAlign, sizeof(paddingSlots) / sizeof(paddingSlots[0])); /* cannot reuse one padding slot. layout to current end */ if (currentFieldOffset == 0) { /* pad alloced_size according to the field alignment */ currentFieldOffset = RoundUp(allocedSize, fieldAlign); if (currentFieldOffset != allocedSize) { /* rounded up, create one padding-slot */ uint32 paddingSize = currentFieldOffset - allocedSize; AddPaddingSlot(paddingSlots, allocedSize, paddingSize, sizeof(paddingSlots) / sizeof(paddingSlots[0])); allocedSize = currentFieldOffset; } /* need new memory for this field */ allocedSize += fieldSize; } AddElementToJClassLayout(classType, JClassFieldInfo(fieldType->GetKind() == kTypePointer, fieldAttr.GetAttr(FLDATTR_rcunowned), fieldAttr.GetAttr(FLDATTR_rcweak), currentFieldOffset)); SetTypeAlign(tyIdx, std::max(GetTypeAlign(tyIdx), fieldAlign)); } SetTypeSize(tyIdx, RoundUp(allocedSize, align)); } void BECommon::ComputeArrayTypeSizesAligns(MIRType &ty, const TyIdx &tyIdx) { MIRArrayType &arrayType = static_cast(ty); MIRType *elemType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(arrayType.GetElemTyIdx()); uint32 elemSize = GetTypeSize(elemType->GetTypeIndex()); if (elemSize == 0) { ComputeTypeSizesAligns(*elemType); elemSize = GetTypeSize(elemType->GetTypeIndex()); } if (!mirModule.IsCModule()) { CHECK_FATAL(elemSize != 0, "elemSize should not equal 0"); CHECK_FATAL(elemType->GetTypeIndex() != 0u, "elemType's idx should not equal 0"); } uint32 arrayAlign = arrayType.GetTypeAttrs().GetAlign(); elemSize = std::max(elemSize, static_cast(GetTypeAlign(elemType->GetTypeIndex()))); elemSize = std::max(elemSize, arrayAlign); /* compute total number of elements from the multipel dimensions */ uint64 numElems = 1; for (int d = 0; d < arrayType.GetDim(); ++d) { numElems *= arrayType.GetSizeArrayItem(d); } auto typeSize = elemSize * numElems; SetTypeSize(tyIdx, typeSize); if (typeSize == 0) { SetTypeAlign(tyIdx, static_cast(arrayAlign)); } else { auto maxAlign = std::max(static_cast(GetTypeAlign(elemType->GetTypeIndex())), arrayAlign); SetTypeAlign(tyIdx, static_cast(maxAlign)); } } void BECommon::ComputeFArrayOrJArrayTypeSizesAligns(MIRType &ty, const TyIdx &tyIdx) { MIRFarrayType &arrayType = static_cast(ty); MIRType *elemType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(arrayType.GetElemTyIdx()); uint32 elemSize = GetTypeSize(elemType->GetTypeIndex()); if (elemSize == 0) { ComputeTypeSizesAligns(*elemType); elemSize = GetTypeSize(elemType->GetTypeIndex()); } CHECK_FATAL(elemSize != 0, "elemSize should not equal 0"); CHECK_FATAL(GetTypeAlign(elemType->GetTypeIndex()) != 0u, "GetTypeAlign return 0 is not expected"); elemSize = std::max(elemSize, static_cast(GetTypeAlign(elemType->GetTypeIndex()))); SetTypeSize(tyIdx, 0); SetTypeAlign(tyIdx, GetTypeAlign(elemType->GetTypeIndex())); } void BECommon::ComputeTypeSizesAligns(MIRType &ty, uint8 align) { TyIdx tyIdx = ty.GetTypeIndex(); if ((structFieldCountTable.size() > tyIdx) && (GetStructFieldCount(tyIdx) != 0)) { return; /* processed before */ } if ((ty.GetPrimType() == PTY_ptr) || (ty.GetPrimType() == PTY_ref)) { ty.SetPrimType(GetLoweredPtrType()); } switch (ty.GetKind()) { case kTypeScalar: case kTypePointer: case kTypeBitField: case kTypeFunction: SetTypeSize(tyIdx, GetPrimTypeSize(ty.GetPrimType())); SetTypeAlign(tyIdx, GetTypeSize(tyIdx)); break; case kTypeArray: { ComputeArrayTypeSizesAligns(ty, tyIdx); break; } case kTypeFArray: case kTypeJArray: { ComputeFArrayOrJArrayTypeSizesAligns(ty, tyIdx); break; } case kTypeUnion: case kTypeStruct: { ComputeStructTypeSizesAligns(ty, tyIdx); break; } case kTypeInterface: { /* interface shouldn't have instance fields */ SetTypeAlign(tyIdx, 0); SetTypeSize(tyIdx, 0); SetStructFieldCount(tyIdx, 0); break; } case kTypeClass: { /* cannot have union or bitfields */ ComputeClassTypeSizesAligns(ty, tyIdx, align); break; } case kTypeByName: case kTypeVoid: default: SetTypeSize(tyIdx, 0); break; } /* there may be passed-in align attribute declared with the symbol */ SetTypeAlign(tyIdx, std::max(GetTypeAlign(tyIdx), align)); } bool BECommon::IsRefField(MIRStructType &structType, FieldID fieldID) const { if (structType.GetKind() == kTypeClass) { CHECK_FATAL(HasJClassLayout(static_cast(structType)), "Cannot found jclass layout information"); const JClassLayout &layout = GetJClassLayout(static_cast(structType)); if (layout.empty()) { ERR(kLncErr, "layout is null in BECommon::IsRefField"); return false; } return layout[fieldID - 1].IsRef(); } return false; } void BECommon::GenFieldOffsetMap(const std::string &className) { MIRType *type = GlobalTables::GetTypeTable().GetOrCreateClassType(className, mirModule); CHECK_FATAL(type != nullptr, "unknown class, type should not be nullptr"); MIRClassType *classType = static_cast(type); for (FieldID i = 1; i <= GetStructFieldCount(classType->GetTypeIndex()); ++i) { FieldID fieldID = i; FieldPair fp = classType->TraverseToFieldRef(fieldID); GStrIdx strIdx = fp.first; if (strIdx == 0u) { continue; } const std::string &fieldName = GlobalTables::GetStrTable().GetStringFromStrIdx(strIdx); TyIdx fieldTyIdx = fp.second.first; uint64 fieldSize = GetTypeSize(fieldTyIdx); MIRType *fieldType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(fieldTyIdx); if ((fieldType->GetKind() == kTypePointer) && (fieldType->GetPrimType() == PTY_a64)) { /* handle class reference field */ fieldSize = RTSupport::GetRTSupportInstance().GetFieldSize(); } std::pair p = GetFieldOffset(*classType, i); CHECK_FATAL(p.second == 0, "expect p.second equals 0"); LogInfo::MapleLogger() << "CLASS_FIELD_OFFSET_MAP(" << className.c_str() << "," << fieldName.c_str() << "," << p.first << "," << fieldSize << ")\n"; } } void BECommon::GenFieldOffsetMap(MIRClassType &classType, FILE &outFile) { const std::string &className = classType.GetName(); /* * We only enumerate fields defined in the current class. There are cases * where a parent classes may define private fields that have the same name as * a field in the current class.This table is generated for the convenience of * C programmers. If the C programmer wants to access parent class fields, * the programmer should access them as `Parent.field`. */ FieldID myEnd = structFieldCountTable.at(classType.GetTypeIndex()); FieldID myBegin = (myEnd - static_cast(classType.GetFieldsSize())) + 1; for (FieldID i = myBegin; i <= myEnd; ++i) { FieldID fieldID = i; FieldPair fp = classType.TraverseToFieldRef(fieldID); GStrIdx strIdx = fp.first; if (strIdx == 0u) { continue; } FieldAttrs attrs = fp.second.second; if (attrs.GetAttr(FLDATTR_static)) { continue; } const std::string &fieldName = GlobalTables::GetStrTable().GetStringFromStrIdx(strIdx); TyIdx fieldTyIdx = fp.second.first; uint64 fieldSize = GetTypeSize(fieldTyIdx); MIRType *fieldType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(fieldTyIdx); if ((fieldType->GetKind() == kTypePointer) && (fieldType->GetPrimType() == PTY_a64)) { /* handle class reference field */ fieldSize = RTSupport::GetRTSupportInstance().GetFieldSize(); ; } std::pair p = GetFieldOffset(classType, i); CHECK_FATAL(p.second == 0, "expect p.second equals 0"); (void)fprintf(&outFile, "__MRT_CLASS_FIELD(%s, %s, %d, %lu)\n", className.c_str(), fieldName.c_str(), p.first, static_cast(fieldSize)); } } void BECommon::GenObjSize(const MIRClassType &classType, FILE &outFile) { const std::string &className = classType.GetName(); uint64_t objSize = GetTypeSize(classType.GetTypeIndex()); if (objSize == 0) { return; } TyIdx parentTypeIdx = classType.GetParentTyIdx(); MIRType *parentType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(parentTypeIdx); const char *parentName = nullptr; if (parentType != nullptr) { MIRClassType *parentClass = static_cast(parentType); parentName = parentClass->GetName().c_str(); } else { parentName = "THIS_IS_ROOT"; } fprintf(&outFile, "__MRT_CLASS(%s, %" PRIu64 ", %s)\n", className.c_str(), objSize, parentName); } FieldInfo BECommon::GetJClassFieldOffset(MIRStructType &classType, FieldID fieldID) const { CHECK_FATAL(fieldID <= GetStructFieldCount(classType.GetTypeIndex()), "GetFieldOFfset: fieldID too large"); if (fieldID == 0) { return {0, 0}; } CHECK_FATAL(HasJClassLayout(static_cast(classType)), "Cannot found jclass layout information"); const JClassLayout &layout = GetJClassLayout(static_cast(classType)); CHECK_FATAL(static_cast(fieldID) - 1 < layout.size(), "subscript out of range"); return {static_cast(layout[static_cast(fieldID) - 1].GetOffset()), 0}; } /* * compute the offset of the field given by fieldID within the structure type * structy; it returns the answer in the pair (byteoffset, bitoffset) such that * if it is a bitfield, byteoffset gives the offset of the container for * extracting the bitfield and bitoffset is with respect to the container */ std::pair BECommon::GetFieldOffset(MIRStructType &structType, FieldID fieldID) { CHECK_FATAL(fieldID <= GetStructFieldCount(structType.GetTypeIndex()), "GetFieldOFfset: fieldID too large"); uint64 allocedSize = 0; uint64 allocedSizeInBits = 0; FieldID curFieldID = 1; if (fieldID == 0) { return std::pair(0, 0); } if (structType.GetKind() == kTypeClass) { CHECK_FATAL(HasJClassLayout(static_cast(structType)), "Cannot found jclass layout information"); const JClassLayout &layout = GetJClassLayout(static_cast(structType)); CHECK_FATAL(static_cast(fieldID) - 1 < layout.size(), "subscript out of range"); return std::pair(static_cast(layout[fieldID - 1].GetOffset()), 0); } /* process the struct fields */ FieldVector fields = structType.GetFields(); auto structPack = static_cast(structType.GetTypeAttrs().GetPack()); for (uint32 j = 0; j < fields.size(); ++j) { TyIdx fieldTyIdx = fields[j].second.first; auto fieldAttr = fields[j].second.second; MIRType *fieldType = GlobalTables::GetTypeTable().GetTypeFromTyIdx(fieldTyIdx); uint32 fieldTypeSize = GetTypeSize(fieldTyIdx); uint64 fieldSizeBits = fieldTypeSize * kBitsPerByte; auto originAlign = GetTypeAlign(fieldTyIdx); auto fieldAlign = fieldAttr.IsPacked() ? 1 : std::min(originAlign, structPack); uint64 fieldAlignBits = static_cast(fieldAlign * kBitsPerByte); CHECK_FATAL(fieldAlign != 0, "fieldAlign should not equal 0"); if (structType.GetKind() != kTypeUnion) { if (fieldType->GetKind() == kTypeBitField) { uint32 fieldSize = static_cast(fieldType)->GetFieldSize(); /* * Is this field is crossing the align boundary of its base type? Or, * is field a zero-with bit field? * Refer to C99 standard (ยง6.7.2.1) : * > As a special case, a bit-field structure member with a width of 0 indicates that no further * > bit-field is to be packed into the unit in which the previous bit-field, if any, was placed. * * We know that A zero-width bit field can cause the next field to be aligned on the next container * boundary where the container is the same size as the underlying type of the bit field. */ CHECK_FATAL(allocedSizeInBits <= UINT64_MAX - fieldSize, "must not be zero"); DEBUG_ASSERT(allocedSizeInBits + fieldSize >= 1, "allocedSizeInBits + fieldSize - 1u must be unsigned"); if ((!structType.GetTypeAttrs().IsPacked() && ((allocedSizeInBits / fieldSizeBits) != ((allocedSizeInBits + fieldSize - 1u) / fieldSizeBits))) || fieldSize == 0) { /* * the field is crossing the align boundary of its base type; * align alloced_size_in_bits to fieldAlign */ allocedSizeInBits = RoundUp(allocedSizeInBits, fieldSizeBits); } /* allocate the bitfield */ if (curFieldID == fieldID) { return std::pair((allocedSizeInBits / fieldAlignBits) * fieldAlign, allocedSizeInBits % fieldAlignBits); } else { ++curFieldID; } allocedSizeInBits += fieldSize; allocedSize = std::max(allocedSize, RoundUp(allocedSizeInBits, fieldAlignBits) / kBitsPerByte); } else { bool leftOverBits = false; uint64 offset = 0; if (allocedSizeInBits == allocedSize * k8BitSize) { allocedSize = RoundUp(allocedSize, fieldAlign); offset = allocedSize; } else { /* still some leftover bits on allocated words, we calculate things based on bits then. */ if (allocedSizeInBits / fieldAlignBits != (allocedSizeInBits + fieldSizeBits - k1BitSize) / fieldAlignBits) { /* the field is crossing the align boundary of its base type */ allocedSizeInBits = RoundUp(allocedSizeInBits, fieldAlignBits); } allocedSize = RoundUp(allocedSize, fieldAlign); offset = static_cast((allocedSizeInBits / fieldAlignBits) * fieldAlign); leftOverBits = true; } if (curFieldID == fieldID) { return std::pair(offset, 0); } else { MIRStructType *subStructType = fieldType->EmbeddedStructType(); if (subStructType == nullptr) { ++curFieldID; } else { if ((curFieldID + GetStructFieldCount(subStructType->GetTypeIndex())) < fieldID) { curFieldID += GetStructFieldCount(subStructType->GetTypeIndex()) + 1; } else { std::pair result = GetFieldOffset(*subStructType, fieldID - curFieldID); return std::pair(result.first + allocedSize, result.second); } } } if (leftOverBits) { allocedSizeInBits += fieldSizeBits; allocedSize = std::max(allocedSize, RoundUp(allocedSizeInBits, fieldAlignBits) / kBitsPerByte); } else { allocedSize += fieldTypeSize; allocedSizeInBits = allocedSize * kBitsPerByte; } } } else { /* for unions, bitfields are treated as non-bitfields */ if (curFieldID == fieldID) { return std::pair(0, 0); } else { MIRStructType *subStructType = fieldType->EmbeddedStructType(); if (subStructType == nullptr) { curFieldID++; } else { if ((curFieldID + GetStructFieldCount(subStructType->GetTypeIndex())) < fieldID) { curFieldID += GetStructFieldCount(subStructType->GetTypeIndex()) + 1; } else { return GetFieldOffset(*subStructType, fieldID - curFieldID); } } } } } CHECK_FATAL(false, "GetFieldOffset() fails to find field"); return std::pair(0, 0); } bool BECommon::TyIsInSizeAlignTable(const MIRType &ty) const { if (typeSizeTable.size() != typeAlignTable.size()) { return false; } return ty.GetTypeIndex() < typeSizeTable.size(); } void BECommon::AddAndComputeSizeAlign(MIRType &ty) { FinalizeTypeTable(ty); typeAlignTable.emplace_back(mirModule.IsCModule()); typeSizeTable.emplace_back(0); ComputeTypeSizesAligns(ty); } void BECommon::AddElementToJClassLayout(MIRClassType &klass, JClassFieldInfo info) { JClassLayout &layout = *(jClassLayoutTable.at(&klass)); layout.emplace_back(info); } void BECommon::AddElementToFuncReturnType(MIRFunction &func, const TyIdx tyIdx) { funcReturnType[&func] = tyIdx; } MIRType *BECommon::BeGetOrCreatePointerType(const MIRType &pointedType) { MIRType *newType = GlobalTables::GetTypeTable().GetOrCreatePointerType(pointedType, GetLoweredPtrType()); if (TyIsInSizeAlignTable(*newType)) { return newType; } AddAndComputeSizeAlign(*newType); return newType; } MIRType *BECommon::BeGetOrCreateFunctionType(TyIdx tyIdx, const std::vector &vecTy, const std::vector &vecAt) { MIRType *newType = GlobalTables::GetTypeTable().GetOrCreateFunctionType(tyIdx, vecTy, vecAt); if (TyIsInSizeAlignTable(*newType)) { return newType; } AddAndComputeSizeAlign(*newType); return newType; } void BECommon::FinalizeTypeTable(const MIRType &ty) { if (ty.GetTypeIndex() > GetSizeOfTypeSizeTable()) { if (mirModule.GetSrcLang() == kSrcLangC) { for (uint32 i = GetSizeOfTypeSizeTable(); i < ty.GetTypeIndex(); ++i) { MIRType *tyTmp = GlobalTables::GetTypeTable().GetTypeFromTyIdx(i); AddAndComputeSizeAlign(*tyTmp); } } else { CHECK_FATAL(ty.GetTypeIndex() == typeSizeTable.size(), "make sure the ty idx is exactly the table size"); } } } BaseNode *BECommon::GetAddressOfNode(const BaseNode &node) { switch (node.GetOpCode()) { case OP_dread: { const DreadNode &dNode = static_cast(node); const StIdx &index = dNode.GetStIdx(); DEBUG_ASSERT(mirModule.CurFunction() != nullptr, "curFunction should not be nullptr"); DEBUG_ASSERT(mirModule.CurFunction()->GetLocalOrGlobalSymbol(index) != nullptr, "nullptr check"); return mirModule.GetMIRBuilder()->CreateAddrof(*mirModule.CurFunction()->GetLocalOrGlobalSymbol(index)); } case OP_iread: { const IreadNode &iNode = static_cast(node); if (iNode.GetFieldID() == 0) { return iNode.Opnd(0); } uint32 index = static_cast(GlobalTables::GetTypeTable().GetTypeTable().at(iNode.GetTyIdx())) ->GetPointedTyIdx(); MIRType *pointedType = GlobalTables::GetTypeTable().GetTypeTable().at(index); std::pair byteBitOffset = GetFieldOffset(static_cast(*pointedType), iNode.GetFieldID()); return mirModule.GetMIRBuilder()->CreateExprBinary( OP_add, *GlobalTables::GetTypeTable().GetPrimType(GetAddressPrimType()), static_cast(iNode.Opnd(0)), mirModule.GetMIRBuilder()->CreateIntConst(byteBitOffset.first, PTY_u32)); } default: return nullptr; } } bool BECommon::CallIsOfAttr(FuncAttrKind attr, const StmtNode *narynode) const { (void)attr; (void)narynode; return false; /* For now, all 64x1_t types object are not propagated to become pregs by mplme, so the following is not needed for now. We need to revisit this later when types are enhanced with attributes */ #if TO_BE_RESURRECTED bool attrFunc = false; if (narynode->GetOpCode() == OP_call) { CallNode *callNode = static_cast(narynode); MIRFunction *func = GlobalTables::GetFunctionTable().GetFunctionFromPuidx(callNode->GetPUIdx()); attrFunc = (mirModule.GetSrcLang() == kSrcLangC && func->GetAttr(attr)) ? true : false; } else if (narynode->GetOpCode() == OP_icall) { IcallNode *icallNode = static_cast(narynode); BaseNode *fNode = icallNode->Opnd(0); MIRFuncType *fType = nullptr; MIRPtrType *pType = nullptr; if (fNode->GetOpCode() == OP_dread) { DreadNode *dNode = static_cast(fNode); MIRSymbol *symbol = mirModule.CurFunction()->GetLocalOrGlobalSymbol(dNode->GetStIdx()); pType = static_cast(symbol->GetType()); MIRType *ty = pType; if (dNode->GetFieldID() != 0) { DEBUG_ASSERT(ty->GetKind() == kTypeStruct || ty->GetKind() == kTypeClass, ""); FieldPair thepair; if (ty->GetKind() == kTypeStruct) { thepair = static_cast(ty)->TraverseToField(dNode->GetFieldID()); } else { thepair = static_cast(ty)->TraverseToField(dNode->GetFieldID()); } pType = static_cast(GlobalTables::GetTypeTable().GetTypeFromTyIdx(thepair.second.first)); } fType = static_cast(pType->GetPointedType()); } else if (fNode->GetOpCode() == OP_iread) { IreadNode *iNode = static_cast(fNode); MIRPtrType *pointerty = static_cast(GlobalTables::GetTypeTable().GetTypeFromTyIdx(iNode->GetTyIdx())); MIRType *pointedType = pointerty->GetPointedType(); if (iNode->GetFieldID() != 0) { pointedType = static_cast(pointedType)->GetFieldType(iNode->GetFieldID()); } if (pointedType->GetKind() == kTypeFunction) { fType = static_cast(pointedType); } else if (pointedType->GetKind() == kTypePointer) { return false; /* assert? */ } } else if (fNode->GetOpCode() == OP_select) { TernaryNode *sNode = static_cast(fNode); BaseNode *expr = sNode->Opnd(1); // both function ptrs under select should have the same signature, chk op1 only AddroffuncNode *afNode = static_cast(expr); MIRFunction *func = GlobalTables::GetFunctionTable().GetFunctionFromPuidx(afNode->GetPUIdx()); attrFunc = mirModule.GetSrcLang() == kSrcLangC && func->GetAttr(attr); } else if (fNode->GetOpCode() == OP_regread) { RegreadNode *rNode = static_cast(fNode); PregIdx pregidx = rNode->GetRegIdx(); MIRPreg *preg = mirModule.CurFunction()->GetPregTab()->PregFromPregIdx(pregidx); MIRType *type = preg->GetMIRType(); if (type == nullptr) { return false; } MIRPtrType *pType = static_cast(type); type = pType->GetPointedType(); if (type == nullptr) { return false; } } else if (fNode->GetOpCode() == OP_retype) { RetypeNode *rNode = static_cast(fNode); pType = static_cast(GlobalTables::GetTypeTable().GetTypeFromTyIdx(rNode->GetTyIdx())); fType = static_cast(pType->GetPointedType()); } else { return false; /* assert? */ } } return attrFunc; #endif } } /* namespace maplebe */