//===--- CodeGenTypes.cpp - Type translation for LLVM CodeGen -------------===// // // The LLVM Compiler Infrastructure // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// // // This is the code that handles AST -> LLVM type lowering. // //===----------------------------------------------------------------------===// #include "CodeGenTypes.h" #include "clang/Basic/TargetInfo.h" #include "clang/AST/AST.h" #include "llvm/DerivedTypes.h" #include "llvm/Module.h" #include "llvm/Target/TargetData.h" using namespace clang; using namespace CodeGen; namespace { /// RecordOrganizer - This helper class, used by CGRecordLayout, layouts /// structs and unions. It manages transient information used during layout. /// FIXME : At the moment assume /// - one to one mapping between AST FieldDecls and /// llvm::StructType elements. /// - Ignore bit fields /// - Ignore field aligments /// - Ignore packed structs class RecordOrganizer { public: explicit RecordOrganizer(CodeGenTypes &Types) : CGT(Types), STy(NULL), llvmFieldNo(0), Cursor(0), llvmSize(0) {} /// addField - Add new field. void addField(const FieldDecl *FD); /// addLLVMField - Add llvm struct field that corresponds to llvm type Ty. /// Increment field count. void addLLVMField(const llvm::Type *Ty); /// addPaddingFields - Current cursor is not suitable place to add next /// field. Add required padding fields. void addPaddingFields(unsigned WaterMark); /// layoutStructFields - Do the actual work and lay out all fields. Create /// corresponding llvm struct type. This should be invoked only after /// all fields are added. void layoutStructFields(const ASTRecordLayout &RL); /// layoutUnionFields - Do the actual work and lay out all fields. Create /// corresponding llvm struct type. This should be invoked only after /// all fields are added. void layoutUnionFields(); /// getLLVMType - Return associated llvm struct type. This may be NULL /// if fields are not laid out. llvm::Type *getLLVMType() const { return STy; } /// placeBitField - Find a place for FD, which is a bit-field. void placeBitField(const FieldDecl *FD); private: CodeGenTypes &CGT; llvm::Type *STy; unsigned llvmFieldNo; uint64_t Cursor; uint64_t llvmSize; llvm::SmallVector FieldDecls; std::vector LLVMFields; llvm::SmallVector Offsets; }; } CodeGenTypes::CodeGenTypes(ASTContext &Ctx, llvm::Module& M, const llvm::TargetData &TD) : Context(Ctx), Target(Ctx.Target), TheModule(M), TheTargetData(TD) { } CodeGenTypes::~CodeGenTypes() { for(llvm::DenseMap::iterator I = CGRecordLayouts.begin(), E = CGRecordLayouts.end(); I != E; ++I) delete I->second; CGRecordLayouts.clear(); } /// isOpaqueTypeDefinition - Return true if LT is a llvm::OpaqueType /// and T is tag definition. This helper routine does not check /// relationship between T and LT. static bool isOpaqueTypeDefinition(QualType T, llvm::Type *LT) { if (!isa(LT)) return false; const clang::Type &Ty = *T.getCanonicalType(); if (Ty.getTypeClass() == Type::Tagged) { const TagType &TT = cast(Ty); const TagDecl *TD = TT.getDecl(); if (TD->isDefinition()) return true; } return false; } /// ConvertType - Convert the specified type to its LLVM form. const llvm::Type *CodeGenTypes::ConvertType(QualType T) { // See if type is already cached. llvm::DenseMap::iterator I = TypeHolderMap.find(T.getTypePtr()); // If type is found in map and this is not a definition for a opaque // place holder type then use it. Otherwise convert type T. if (I != TypeHolderMap.end() && !isOpaqueTypeDefinition(T, I->second.get())) return I->second.get(); const llvm::Type *ResultType = ConvertNewType(T); TypeHolderMap.insert(std::make_pair(T.getTypePtr(), llvm::PATypeHolder(ResultType))); return ResultType; } /// ConvertTypeForMem - Convert type T into a llvm::Type. Maintain and use /// type cache through TypeHolderMap. This differs from ConvertType in that /// it is used to convert to the memory representation for a type. For /// example, the scalar representation for _Bool is i1, but the memory /// representation is usually i8 or i32, depending on the target. const llvm::Type *CodeGenTypes::ConvertTypeForMem(QualType T) { const llvm::Type *R = ConvertType(T); // If this is a non-bool type, don't map it. if (R != llvm::Type::Int1Ty) return R; // Otherwise, return an integer of the target-specified size. unsigned BoolWidth = (unsigned)Context.getTypeSize(T, SourceLocation()); return llvm::IntegerType::get(BoolWidth); } const llvm::Type *CodeGenTypes::ConvertNewType(QualType T) { const clang::Type &Ty = *T.getCanonicalType(); switch (Ty.getTypeClass()) { case Type::TypeName: // typedef isn't canonical. case Type::TypeOfExp: // typeof isn't canonical. case Type::TypeOfTyp: // typeof isn't canonical. assert(0 && "Non-canonical type, shouldn't happen"); case Type::Builtin: { switch (cast(Ty).getKind()) { case BuiltinType::Void: // LLVM void type can only be used as the result of a function call. Just // map to the same as char. return llvm::IntegerType::get(8); case BuiltinType::Bool: // Note that we always return bool as i1 for use as a scalar type. return llvm::Type::Int1Ty; case BuiltinType::Char_S: case BuiltinType::Char_U: case BuiltinType::SChar: case BuiltinType::UChar: case BuiltinType::Short: case BuiltinType::UShort: case BuiltinType::Int: case BuiltinType::UInt: case BuiltinType::Long: case BuiltinType::ULong: case BuiltinType::LongLong: case BuiltinType::ULongLong: return llvm::IntegerType::get( static_cast(Context.getTypeSize(T, SourceLocation()))); case BuiltinType::Float: return llvm::Type::FloatTy; case BuiltinType::Double: return llvm::Type::DoubleTy; case BuiltinType::LongDouble: // FIXME: mapping long double onto double. return llvm::Type::DoubleTy; } break; } case Type::Complex: { std::vector Elts; Elts.push_back(ConvertType(cast(Ty).getElementType())); Elts.push_back(Elts[0]); return llvm::StructType::get(Elts); } case Type::Pointer: { const PointerType &P = cast(Ty); return llvm::PointerType::getUnqual(ConvertType(P.getPointeeType())); } case Type::Reference: { const ReferenceType &R = cast(Ty); return llvm::PointerType::getUnqual(ConvertType(R.getReferenceeType())); } case Type::VariableArray: { const VariableArrayType &A = cast(Ty); assert(A.getSizeModifier() == ArrayType::Normal && A.getIndexTypeQualifier() == 0 && "FIXME: We only handle trivial array types so far!"); if (A.getSizeExpr() == 0) { // int X[] -> [0 x int] return llvm::ArrayType::get(ConvertType(A.getElementType()), 0); } else { assert(0 && "FIXME: VLAs not implemented yet!"); } } case Type::ConstantArray: { const ConstantArrayType &A = cast(Ty); const llvm::Type *EltTy = ConvertType(A.getElementType()); return llvm::ArrayType::get(EltTy, A.getSize().getZExtValue()); } case Type::OCUVector: case Type::Vector: { const VectorType &VT = cast(Ty); return llvm::VectorType::get(ConvertType(VT.getElementType()), VT.getNumElements()); } case Type::FunctionNoProto: case Type::FunctionProto: { const FunctionType &FP = cast(Ty); const llvm::Type *ResultType; if (FP.getResultType()->isVoidType()) ResultType = llvm::Type::VoidTy; // Result of function uses llvm void. else ResultType = ConvertType(FP.getResultType()); // FIXME: Convert argument types. bool isVarArg; std::vector ArgTys; // Struct return passes the struct byref. if (!ResultType->isFirstClassType() && ResultType != llvm::Type::VoidTy) { const llvm::Type *RType = llvm::PointerType::getUnqual(ResultType); QualType RTy = Context.getPointerType(FP.getResultType()); TypeHolderMap.insert(std::make_pair(RTy.getTypePtr(), llvm::PATypeHolder(RType))); ArgTys.push_back(RType); ResultType = llvm::Type::VoidTy; } if (const FunctionTypeProto *FTP = dyn_cast(&FP)) { DecodeArgumentTypes(*FTP, ArgTys); isVarArg = FTP->isVariadic(); } else { isVarArg = true; } return llvm::FunctionType::get(ResultType, ArgTys, isVarArg); } case Type::ObjCInterface: assert(0 && "FIXME: add missing functionality here"); break; case Type::ObjCQualifiedInterface: assert(0 && "FIXME: add missing functionality here"); break; case Type::ObjCQualifiedId: assert(0 && "FIXME: add missing functionality here"); break; case Type::Tagged: const TagType &TT = cast(Ty); const TagDecl *TD = TT.getDecl(); llvm::Type *&ResultType = TagDeclTypes[TD]; // If corresponding llvm type is not a opaque struct type // then use it. if (ResultType && !isOpaqueTypeDefinition(T, ResultType)) return ResultType; if (!TD->isDefinition()) { ResultType = llvm::OpaqueType::get(); } else if (TD->getKind() == Decl::Enum) { return ConvertType(cast(TD)->getIntegerType()); } else if (TD->getKind() == Decl::Struct) { const RecordDecl *RD = cast(TD); // If this is nested record and this RecordDecl is already under // process then return associated OpaqueType for now. llvm::DenseMap::iterator OpaqueI = RecordTypesToResolve.find(RD); if (OpaqueI != RecordTypesToResolve.end()) return OpaqueI->second; llvm::OpaqueType *OpaqueTy = NULL; if (ResultType) OpaqueTy = dyn_cast(ResultType); if (!OpaqueTy) { // Create new OpaqueType now for later use. // FIXME: This creates a lot of opaque types, most of them are not // needed. Reevaluate this when performance analyis finds tons of // opaque types. OpaqueTy = llvm::OpaqueType::get(); TypeHolderMap.insert(std::make_pair(T.getTypePtr(), llvm::PATypeHolder(OpaqueTy))); } RecordTypesToResolve[RD] = OpaqueTy; // Layout fields. RecordOrganizer RO(*this); for (unsigned i = 0, e = RD->getNumMembers(); i != e; ++i) RO.addField(RD->getMember(i)); const ASTRecordLayout &RL = Context.getASTRecordLayout(RD, SourceLocation()); RO.layoutStructFields(RL); // Get llvm::StructType. CGRecordLayout *RLI = new CGRecordLayout(RO.getLLVMType()); ResultType = RLI->getLLVMType(); CGRecordLayouts[ResultType] = RLI; // Refine any OpaqueType associated with this RecordDecl. OpaqueTy->refineAbstractTypeTo(ResultType); OpaqueI = RecordTypesToResolve.find(RD); assert (OpaqueI != RecordTypesToResolve.end() && "Expected RecordDecl in RecordTypesToResolve"); RecordTypesToResolve.erase(OpaqueI); } else if (TD->getKind() == Decl::Union) { const RecordDecl *RD = cast(TD); // Just use the largest element of the union, breaking ties with the // highest aligned member. if (RD->getNumMembers() != 0) { RecordOrganizer RO(*this); for (unsigned i = 0, e = RD->getNumMembers(); i != e; ++i) RO.addField(RD->getMember(i)); RO.layoutUnionFields(); // Get llvm::StructType. CGRecordLayout *RLI = new CGRecordLayout(RO.getLLVMType()); ResultType = RLI->getLLVMType(); CGRecordLayouts[ResultType] = RLI; } else { std::vector Fields; ResultType = llvm::StructType::get(Fields); } } else { assert(0 && "FIXME: Implement tag decl kind!"); } std::string TypeName(TD->getKindName()); TypeName += '.'; // Name the codegen type after the typedef name // if there is no tag type name available if (TD->getIdentifier() == 0) { if (T->getTypeClass() == Type::TypeName) { const TypedefType *TdT = cast(T); TypeName += TdT->getDecl()->getName(); } else TypeName += "anon"; } else TypeName += TD->getName(); TheModule.addTypeName(TypeName, ResultType); return ResultType; } // FIXME: implement. return llvm::OpaqueType::get(); } void CodeGenTypes::DecodeArgumentTypes(const FunctionTypeProto &FTP, std::vector &ArgTys) { for (unsigned i = 0, e = FTP.getNumArgs(); i != e; ++i) { const llvm::Type *Ty = ConvertType(FTP.getArgType(i)); if (Ty->isFirstClassType()) ArgTys.push_back(Ty); else { QualType PTy = Context.getPointerType(FTP.getArgType(i)); const llvm::Type *PtrTy = llvm::PointerType::getUnqual(Ty); TypeHolderMap.insert(std::make_pair(PTy.getTypePtr(), llvm::PATypeHolder(PtrTy))); ArgTys.push_back(PtrTy); } } } /// getLLVMFieldNo - Return llvm::StructType element number /// that corresponds to the field FD. unsigned CodeGenTypes::getLLVMFieldNo(const FieldDecl *FD) { llvm::DenseMap::iterator I = FieldInfo.find(FD); assert (I != FieldInfo.end() && "Unable to find field info"); return I->second; } /// addFieldInfo - Assign field number to field FD. void CodeGenTypes::addFieldInfo(const FieldDecl *FD, unsigned No) { FieldInfo[FD] = No; } /// getBitFieldInfo - Return the BitFieldInfo that corresponds to the field FD. CodeGenTypes::BitFieldInfo CodeGenTypes::getBitFieldInfo(const FieldDecl *FD) { llvm::DenseMap::iterator I = BitFields.find(FD); assert (I != BitFields.end() && "Unable to find bitfield info"); return I->second; } /// addBitFieldInfo - Assign a start bit and a size to field FD. void CodeGenTypes::addBitFieldInfo(const FieldDecl *FD, unsigned Begin, unsigned Size) { BitFields.insert(std::make_pair(FD, BitFieldInfo(Begin, Size))); } /// getCGRecordLayout - Return record layout info for the given llvm::Type. const CGRecordLayout * CodeGenTypes::getCGRecordLayout(const llvm::Type* Ty) const { llvm::DenseMap::iterator I = CGRecordLayouts.find(Ty); assert (I != CGRecordLayouts.end() && "Unable to find record layout information for type"); return I->second; } /// addField - Add new field. void RecordOrganizer::addField(const FieldDecl *FD) { assert (!STy && "Record fields are already laid out"); FieldDecls.push_back(FD); } /// layoutStructFields - Do the actual work and lay out all fields. Create /// corresponding llvm struct type. This should be invoked only after /// all fields are added. /// FIXME : At the moment assume /// - one to one mapping between AST FieldDecls and /// llvm::StructType elements. /// - Ignore bit fields /// - Ignore field aligments /// - Ignore packed structs void RecordOrganizer::layoutStructFields(const ASTRecordLayout &RL) { // FIXME : Use SmallVector llvmSize = 0; llvmFieldNo = 0; Cursor = 0; LLVMFields.clear(); Offsets.clear(); for (llvm::SmallVector::iterator I = FieldDecls.begin(), E = FieldDecls.end(); I != E; ++I) { const FieldDecl *FD = *I; if (FD->isBitField()) placeBitField(FD); else { const llvm::Type *Ty = CGT.ConvertType(FD->getType()); addLLVMField(Ty); CGT.addFieldInfo(FD, llvmFieldNo - 1); Cursor = llvmSize; } } unsigned StructAlign = RL.getAlignment(); if (llvmSize % StructAlign) { unsigned StructPadding = StructAlign - (llvmSize % StructAlign); addPaddingFields(llvmSize + StructPadding); } STy = llvm::StructType::get(LLVMFields); } /// addPaddingFields - Current cursor is not suitable place to add next field. /// Add required padding fields. void RecordOrganizer::addPaddingFields(unsigned WaterMark) { unsigned RequiredBits = WaterMark - llvmSize; unsigned RequiredBytes = (RequiredBits + 7) / 8; for (unsigned i = 0; i != RequiredBytes; ++i) addLLVMField(llvm::Type::Int8Ty); } /// addLLVMField - Add llvm struct field that corresponds to llvm type Ty. /// Increment field count. void RecordOrganizer::addLLVMField(const llvm::Type *Ty) { unsigned AlignmentInBits = CGT.getTargetData().getABITypeAlignment(Ty) * 8; if (llvmSize % AlignmentInBits) { // At the moment, insert padding fields even if target specific llvm // type alignment enforces implict padding fields for FD. Later on, // optimize llvm fields by removing implicit padding fields and // combining consequetive padding fields. unsigned Padding = AlignmentInBits - (llvmSize % AlignmentInBits); addPaddingFields(llvmSize + Padding); } unsigned TySize = CGT.getTargetData().getABITypeSizeInBits(Ty); Offsets.push_back(llvmSize); llvmSize += TySize; LLVMFields.push_back(Ty); ++llvmFieldNo; } /// layoutUnionFields - Do the actual work and lay out all fields. Create /// corresponding llvm struct type. This should be invoked only after /// all fields are added. void RecordOrganizer::layoutUnionFields() { unsigned PrimaryEltNo = 0; std::pair PrimaryElt = CGT.getContext().getTypeInfo(FieldDecls[0]->getType(), SourceLocation()); CGT.addFieldInfo(FieldDecls[0], 0); unsigned Size = FieldDecls.size(); for(unsigned i = 1; i != Size; ++i) { const FieldDecl *FD = FieldDecls[i]; assert (!FD->isBitField() && "Bit fields are not yet supported"); std::pair EltInfo = CGT.getContext().getTypeInfo(FD->getType(), SourceLocation()); // Use largest element, breaking ties with the hightest aligned member. if (EltInfo.first > PrimaryElt.first || (EltInfo.first == PrimaryElt.first && EltInfo.second > PrimaryElt.second)) { PrimaryElt = EltInfo; PrimaryEltNo = i; } // In union, each field gets first slot. CGT.addFieldInfo(FD, 0); } std::vector Fields; const llvm::Type *Ty = CGT.ConvertType(FieldDecls[PrimaryEltNo]->getType()); Fields.push_back(Ty); STy = llvm::StructType::get(Fields); } /// placeBitField - Find a place for FD, which is a bit-field. /// This function searches for the last aligned field. If the bit-field fits in /// it, it is reused. Otherwise, the bit-field is placed in a new field. void RecordOrganizer::placeBitField(const FieldDecl *FD) { assert (FD->isBitField() && "FD is not a bit-field"); Expr *BitWidth = FD->getBitWidth(); llvm::APSInt FieldSize(32); bool isBitField = BitWidth->isIntegerConstantExpr(FieldSize, CGT.getContext()); assert (isBitField && "Invalid BitField size expression"); uint64_t BitFieldSize = FieldSize.getZExtValue(); bool FoundPrevField = false; unsigned TotalOffsets = Offsets.size(); const llvm::Type *Ty = CGT.ConvertType(FD->getType()); uint64_t TySize = CGT.getTargetData().getABITypeSizeInBits(Ty); if (!TotalOffsets) { // Special case: the first field. CGT.addFieldInfo(FD, llvmFieldNo); CGT.addBitFieldInfo(FD, 0, BitFieldSize); addPaddingFields(BitFieldSize); Cursor = BitFieldSize; return; } // Search for the last aligned field. for (unsigned i = TotalOffsets; i != 0; --i) { uint64_t O = Offsets[i - 1]; if (O % TySize == 0) { FoundPrevField = true; if (TySize - (Cursor - O) >= BitFieldSize) { // The bitfield fits in the last aligned field. // This is : struct { char a; int CurrentField:10;}; // where 'CurrentField' shares first field with 'a'. addPaddingFields(Cursor + BitFieldSize); CGT.addFieldInfo(FD, i); CGT.addBitFieldInfo(FD, Cursor, BitFieldSize); Cursor += BitFieldSize; } else { // Place the bitfield in a new LLVM field. // This is : struct { char a; short CurrentField:10;}; // where 'CurrentField' needs a new llvm field. addPaddingFields(O + TySize); CGT.addFieldInfo(FD, llvmFieldNo); CGT.addBitFieldInfo(FD, 0, BitFieldSize); addPaddingFields(O + TySize + BitFieldSize); Cursor = O + TySize + BitFieldSize; } break; } } assert(FoundPrevField && "Unable to find a place for bitfield in struct layout"); }