1 //===- lib/MC/MCAssembler.cpp - Assembler Backend Implementation ----------===//
3 // The LLVM Compiler Infrastructure
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
8 //===----------------------------------------------------------------------===//
10 #define DEBUG_TYPE "assembler"
11 #include "llvm/MC/MCAssembler.h"
12 #include "llvm/MC/MCAsmLayout.h"
13 #include "llvm/MC/MCCodeEmitter.h"
14 #include "llvm/MC/MCExpr.h"
15 #include "llvm/MC/MCObjectWriter.h"
16 #include "llvm/MC/MCSymbol.h"
17 #include "llvm/MC/MCValue.h"
18 #include "llvm/ADT/OwningPtr.h"
19 #include "llvm/ADT/Statistic.h"
20 #include "llvm/ADT/StringExtras.h"
21 #include "llvm/ADT/Twine.h"
22 #include "llvm/Support/Debug.h"
23 #include "llvm/Support/ErrorHandling.h"
24 #include "llvm/Support/raw_ostream.h"
25 #include "llvm/Target/TargetRegistry.h"
26 #include "llvm/Target/TargetAsmBackend.h"
33 STATISTIC(EmittedFragments, "Number of emitted assembler fragments");
34 STATISTIC(EvaluateFixup, "Number of evaluated fixups");
35 STATISTIC(FragmentLayouts, "Number of fragment layouts");
36 STATISTIC(ObjectBytes, "Number of emitted object file bytes");
37 STATISTIC(RelaxationSteps, "Number of assembler layout and relaxation steps");
38 STATISTIC(RelaxedInstructions, "Number of relaxed instructions");
39 STATISTIC(SectionLayouts, "Number of section layouts");
43 // FIXME FIXME FIXME: There are number of places in this file where we convert
44 // what is a 64-bit assembler value used for computation into a value in the
45 // object file, which may truncate it. We should detect that truncation where
46 // invalid and report errors back.
50 void MCAsmLayout::UpdateForSlide(MCFragment *F, int SlideAmount) {
51 // We shouldn't have to do anything special to support negative slides, and it
52 // is a perfectly valid thing to do as long as other parts of the system can
53 // guarantee convergence.
54 assert(SlideAmount >= 0 && "Negative slides not yet supported");
56 // Update the layout by simply recomputing the layout for the entire
57 // file. This is trivially correct, but very slow.
59 // FIXME-PERF: This is O(N^2), but will be eliminated once we get smarter.
61 // Layout the concrete sections and fragments.
62 MCAssembler &Asm = getAssembler();
64 for (MCAssembler::iterator it = Asm.begin(), ie = Asm.end(); it != ie; ++it) {
65 // Skip virtual sections.
66 if (Asm.getBackend().isVirtualSection(it->getSection()))
69 // Layout the section fragments and its size.
70 Address = Asm.LayoutSection(*it, *this, Address);
73 // Layout the virtual sections.
74 for (MCAssembler::iterator it = Asm.begin(), ie = Asm.end(); it != ie; ++it) {
75 if (!Asm.getBackend().isVirtualSection(it->getSection()))
78 // Layout the section fragments and its size.
79 Address = Asm.LayoutSection(*it, *this, Address);
83 uint64_t MCAsmLayout::getFragmentAddress(const MCFragment *F) const {
84 assert(F->getParent() && "Missing section()!");
85 return getSectionAddress(F->getParent()) + getFragmentOffset(F);
88 uint64_t MCAsmLayout::getFragmentEffectiveSize(const MCFragment *F) const {
89 assert(F->EffectiveSize != ~UINT64_C(0) && "Address not set!");
90 return F->EffectiveSize;
93 void MCAsmLayout::setFragmentEffectiveSize(MCFragment *F, uint64_t Value) {
94 F->EffectiveSize = Value;
97 uint64_t MCAsmLayout::getFragmentOffset(const MCFragment *F) const {
98 assert(F->Offset != ~UINT64_C(0) && "Address not set!");
102 void MCAsmLayout::setFragmentOffset(MCFragment *F, uint64_t Value) {
106 uint64_t MCAsmLayout::getSymbolAddress(const MCSymbolData *SD) const {
107 assert(SD->getFragment() && "Invalid getAddress() on undefined symbol!");
108 return getFragmentAddress(SD->getFragment()) + SD->getOffset();
111 uint64_t MCAsmLayout::getSectionAddress(const MCSectionData *SD) const {
112 assert(SD->Address != ~UINT64_C(0) && "Address not set!");
116 void MCAsmLayout::setSectionAddress(MCSectionData *SD, uint64_t Value) {
120 uint64_t MCAsmLayout::getSectionSize(const MCSectionData *SD) const {
121 assert(SD->Size != ~UINT64_C(0) && "File size not set!");
124 void MCAsmLayout::setSectionSize(MCSectionData *SD, uint64_t Value) {
128 uint64_t MCAsmLayout::getSectionFileSize(const MCSectionData *SD) const {
129 assert(SD->FileSize != ~UINT64_C(0) && "File size not set!");
132 void MCAsmLayout::setSectionFileSize(MCSectionData *SD, uint64_t Value) {
133 SD->FileSize = Value;
138 MCFragment::MCFragment() : Kind(FragmentType(~0)) {
141 MCFragment::MCFragment(FragmentType _Kind, MCSectionData *_Parent)
142 : Kind(_Kind), Parent(_Parent), Atom(0), EffectiveSize(~UINT64_C(0))
145 Parent->getFragmentList().push_back(this);
148 MCFragment::~MCFragment() {
153 MCSectionData::MCSectionData() : Section(0) {}
155 MCSectionData::MCSectionData(const MCSection &_Section, MCAssembler *A)
156 : Section(&_Section),
158 Address(~UINT64_C(0)),
160 FileSize(~UINT64_C(0)),
161 HasInstructions(false)
164 A->getSectionList().push_back(this);
169 MCSymbolData::MCSymbolData() : Symbol(0) {}
171 MCSymbolData::MCSymbolData(const MCSymbol &_Symbol, MCFragment *_Fragment,
172 uint64_t _Offset, MCAssembler *A)
173 : Symbol(&_Symbol), Fragment(_Fragment), Offset(_Offset),
174 IsExternal(false), IsPrivateExtern(false),
175 CommonSize(0), CommonAlign(0), Flags(0), Index(0)
178 A->getSymbolList().push_back(this);
183 MCAssembler::MCAssembler(MCContext &_Context, TargetAsmBackend &_Backend,
184 MCCodeEmitter &_Emitter, raw_ostream &_OS)
185 : Context(_Context), Backend(_Backend), Emitter(_Emitter),
186 OS(_OS), RelaxAll(false), SubsectionsViaSymbols(false)
190 MCAssembler::~MCAssembler() {
193 static bool isScatteredFixupFullyResolvedSimple(const MCAssembler &Asm,
194 const MCAsmFixup &Fixup,
195 const MCValue Target,
196 const MCSection *BaseSection) {
197 // The effective fixup address is
198 // addr(atom(A)) + offset(A)
199 // - addr(atom(B)) - offset(B)
200 // - addr(<base symbol>) + <fixup offset from base symbol>
201 // and the offsets are not relocatable, so the fixup is fully resolved when
202 // addr(atom(A)) - addr(atom(B)) - addr(<base symbol>)) == 0.
204 // The simple (Darwin, except on x86_64) way of dealing with this was to
205 // assume that any reference to a temporary symbol *must* be a temporary
206 // symbol in the same atom, unless the sections differ. Therefore, any PCrel
207 // relocation to a temporary symbol (in the same section) is fully
208 // resolved. This also works in conjunction with absolutized .set, which
209 // requires the compiler to use .set to absolutize the differences between
210 // symbols which the compiler knows to be assembly time constants, so we don't
211 // need to worry about considering symbol differences fully resolved.
213 // Non-relative fixups are only resolved if constant.
215 return Target.isAbsolute();
217 // Otherwise, relative fixups are only resolved if not a difference and the
218 // target is a temporary in the same section.
219 if (Target.isAbsolute() || Target.getSymB())
222 const MCSymbol *A = &Target.getSymA()->getSymbol();
223 if (!A->isTemporary() || !A->isInSection() ||
224 &A->getSection() != BaseSection)
230 static bool isScatteredFixupFullyResolved(const MCAssembler &Asm,
231 const MCAsmLayout &Layout,
232 const MCAsmFixup &Fixup,
233 const MCValue Target,
234 const MCSymbolData *BaseSymbol) {
235 // The effective fixup address is
236 // addr(atom(A)) + offset(A)
237 // - addr(atom(B)) - offset(B)
238 // - addr(BaseSymbol) + <fixup offset from base symbol>
239 // and the offsets are not relocatable, so the fixup is fully resolved when
240 // addr(atom(A)) - addr(atom(B)) - addr(BaseSymbol) == 0.
242 // Note that "false" is almost always conservatively correct (it means we emit
243 // a relocation which is unnecessary), except when it would force us to emit a
244 // relocation which the target cannot encode.
246 const MCSymbolData *A_Base = 0, *B_Base = 0;
247 if (const MCSymbolRefExpr *A = Target.getSymA()) {
248 // Modified symbol references cannot be resolved.
249 if (A->getKind() != MCSymbolRefExpr::VK_None)
252 A_Base = Asm.getAtom(Layout, &Asm.getSymbolData(A->getSymbol()));
257 if (const MCSymbolRefExpr *B = Target.getSymB()) {
258 // Modified symbol references cannot be resolved.
259 if (B->getKind() != MCSymbolRefExpr::VK_None)
262 B_Base = Asm.getAtom(Layout, &Asm.getSymbolData(B->getSymbol()));
267 // If there is no base, A and B have to be the same atom for this fixup to be
270 return A_Base == B_Base;
272 // Otherwise, B must be missing and A must be the base.
273 return !B_Base && BaseSymbol == A_Base;
276 bool MCAssembler::isSymbolLinkerVisible(const MCSymbolData *SD) const {
277 // Non-temporary labels should always be visible to the linker.
278 if (!SD->getSymbol().isTemporary())
281 // Absolute temporary labels are never visible.
282 if (!SD->getFragment())
285 // Otherwise, check if the section requires symbols even for temporary labels.
286 return getBackend().doesSectionRequireSymbols(
287 SD->getFragment()->getParent()->getSection());
290 const MCSymbolData *MCAssembler::getAtom(const MCAsmLayout &Layout,
291 const MCSymbolData *SD) const {
292 // Linker visible symbols define atoms.
293 if (isSymbolLinkerVisible(SD))
296 // Absolute and undefined symbols have no defining atom.
297 if (!SD->getFragment())
300 // Otherwise, return the atom for the containing fragment.
301 return SD->getFragment()->getAtom();
304 bool MCAssembler::EvaluateFixup(const MCAsmLayout &Layout,
305 const MCAsmFixup &Fixup, const MCFragment *DF,
306 MCValue &Target, uint64_t &Value) const {
307 ++stats::EvaluateFixup;
309 if (!Fixup.Value->EvaluateAsRelocatable(Target, &Layout))
310 report_fatal_error("expected relocatable expression");
312 // FIXME: How do non-scattered symbols work in ELF? I presume the linker
313 // doesn't support small relocations, but then under what criteria does the
314 // assembler allow symbol differences?
316 Value = Target.getConstant();
319 Emitter.getFixupKindInfo(Fixup.Kind).Flags & MCFixupKindInfo::FKF_IsPCRel;
320 bool IsResolved = true;
321 if (const MCSymbolRefExpr *A = Target.getSymA()) {
322 if (A->getSymbol().isDefined())
323 Value += Layout.getSymbolAddress(&getSymbolData(A->getSymbol()));
327 if (const MCSymbolRefExpr *B = Target.getSymB()) {
328 if (B->getSymbol().isDefined())
329 Value -= Layout.getSymbolAddress(&getSymbolData(B->getSymbol()));
334 // If we are using scattered symbols, determine whether this value is actually
335 // resolved; scattering may cause atoms to move.
336 if (IsResolved && getBackend().hasScatteredSymbols()) {
337 if (getBackend().hasReliableSymbolDifference()) {
338 // If this is a PCrel relocation, find the base atom (identified by its
339 // symbol) that the fixup value is relative to.
340 const MCSymbolData *BaseSymbol = 0;
342 BaseSymbol = DF->getAtom();
348 IsResolved = isScatteredFixupFullyResolved(*this, Layout, Fixup, Target,
351 const MCSection *BaseSection = 0;
353 BaseSection = &DF->getParent()->getSection();
355 IsResolved = isScatteredFixupFullyResolvedSimple(*this, Fixup, Target,
361 Value -= Layout.getFragmentAddress(DF) + Fixup.Offset;
366 uint64_t MCAssembler::LayoutSection(MCSectionData &SD,
368 uint64_t StartAddress) {
369 bool IsVirtual = getBackend().isVirtualSection(SD.getSection());
371 ++stats::SectionLayouts;
373 // Align this section if necessary by adding padding bytes to the previous
374 // section. It is safe to adjust this out-of-band, because no symbol or
375 // fragment is allowed to point past the end of the section at any time.
376 if (uint64_t Pad = OffsetToAlignment(StartAddress, SD.getAlignment())) {
377 // Unless this section is virtual (where we are allowed to adjust the offset
378 // freely), the padding goes in the previous section.
380 // Find the previous non-virtual section.
382 assert(it != begin() && "Invalid initial section address!");
383 for (--it; getBackend().isVirtualSection(it->getSection()); --it) ;
384 Layout.setSectionFileSize(&*it, Layout.getSectionFileSize(&*it) + Pad);
390 // Set the aligned section address.
391 Layout.setSectionAddress(&SD, StartAddress);
393 uint64_t Address = StartAddress;
394 for (MCSectionData::iterator it = SD.begin(), ie = SD.end(); it != ie; ++it) {
397 ++stats::FragmentLayouts;
399 uint64_t FragmentOffset = Address - StartAddress;
400 Layout.setFragmentOffset(&F, FragmentOffset);
402 // Evaluate fragment size.
403 uint64_t EffectiveSize = 0;
404 switch (F.getKind()) {
405 case MCFragment::FT_Align: {
406 MCAlignFragment &AF = cast<MCAlignFragment>(F);
408 EffectiveSize = OffsetToAlignment(Address, AF.getAlignment());
409 if (EffectiveSize > AF.getMaxBytesToEmit())
414 case MCFragment::FT_Data:
415 EffectiveSize = cast<MCDataFragment>(F).getContents().size();
418 case MCFragment::FT_Fill: {
419 MCFillFragment &FF = cast<MCFillFragment>(F);
420 EffectiveSize = FF.getValueSize() * FF.getCount();
424 case MCFragment::FT_Inst:
425 EffectiveSize = cast<MCInstFragment>(F).getInstSize();
428 case MCFragment::FT_Org: {
429 MCOrgFragment &OF = cast<MCOrgFragment>(F);
431 int64_t TargetLocation;
432 if (!OF.getOffset().EvaluateAsAbsolute(TargetLocation, &Layout))
433 report_fatal_error("expected assembly-time absolute expression");
435 // FIXME: We need a way to communicate this error.
436 int64_t Offset = TargetLocation - FragmentOffset;
438 report_fatal_error("invalid .org offset '" + Twine(TargetLocation) +
439 "' (at offset '" + Twine(FragmentOffset) + "'");
441 EffectiveSize = Offset;
445 case MCFragment::FT_ZeroFill: {
446 MCZeroFillFragment &ZFF = cast<MCZeroFillFragment>(F);
448 // Align the fragment offset; it is safe to adjust the offset freely since
449 // this is only in virtual sections.
451 // FIXME: We shouldn't be doing this here.
452 Address = RoundUpToAlignment(Address, ZFF.getAlignment());
453 Layout.setFragmentOffset(&F, Address - StartAddress);
455 EffectiveSize = ZFF.getSize();
460 Layout.setFragmentEffectiveSize(&F, EffectiveSize);
461 Address += EffectiveSize;
464 // Set the section sizes.
465 Layout.setSectionSize(&SD, Address - StartAddress);
467 Layout.setSectionFileSize(&SD, 0);
469 Layout.setSectionFileSize(&SD, Address - StartAddress);
474 /// WriteFragmentData - Write the \arg F data to the output file.
475 static void WriteFragmentData(const MCAssembler &Asm, const MCAsmLayout &Layout,
476 const MCFragment &F, MCObjectWriter *OW) {
477 uint64_t Start = OW->getStream().tell();
480 ++stats::EmittedFragments;
482 // FIXME: Embed in fragments instead?
483 uint64_t FragmentSize = Layout.getFragmentEffectiveSize(&F);
484 switch (F.getKind()) {
485 case MCFragment::FT_Align: {
486 MCAlignFragment &AF = cast<MCAlignFragment>(F);
487 uint64_t Count = FragmentSize / AF.getValueSize();
489 // FIXME: This error shouldn't actually occur (the front end should emit
490 // multiple .align directives to enforce the semantics it wants), but is
491 // severe enough that we want to report it. How to handle this?
492 if (Count * AF.getValueSize() != FragmentSize)
493 report_fatal_error("undefined .align directive, value size '" +
494 Twine(AF.getValueSize()) +
495 "' is not a divisor of padding size '" +
496 Twine(FragmentSize) + "'");
498 // See if we are aligning with nops, and if so do that first to try to fill
499 // the Count bytes. Then if that did not fill any bytes or there are any
500 // bytes left to fill use the the Value and ValueSize to fill the rest.
501 // If we are aligning with nops, ask that target to emit the right data.
502 if (AF.getEmitNops()) {
503 if (!Asm.getBackend().WriteNopData(Count, OW))
504 report_fatal_error("unable to write nop sequence of " +
505 Twine(Count) + " bytes");
509 // Otherwise, write out in multiples of the value size.
510 for (uint64_t i = 0; i != Count; ++i) {
511 switch (AF.getValueSize()) {
513 assert(0 && "Invalid size!");
514 case 1: OW->Write8 (uint8_t (AF.getValue())); break;
515 case 2: OW->Write16(uint16_t(AF.getValue())); break;
516 case 4: OW->Write32(uint32_t(AF.getValue())); break;
517 case 8: OW->Write64(uint64_t(AF.getValue())); break;
523 case MCFragment::FT_Data: {
524 MCDataFragment &DF = cast<MCDataFragment>(F);
525 assert(FragmentSize == DF.getContents().size() && "Invalid size!");
526 OW->WriteBytes(DF.getContents().str());
530 case MCFragment::FT_Fill: {
531 MCFillFragment &FF = cast<MCFillFragment>(F);
532 for (uint64_t i = 0, e = FF.getCount(); i != e; ++i) {
533 switch (FF.getValueSize()) {
535 assert(0 && "Invalid size!");
536 case 1: OW->Write8 (uint8_t (FF.getValue())); break;
537 case 2: OW->Write16(uint16_t(FF.getValue())); break;
538 case 4: OW->Write32(uint32_t(FF.getValue())); break;
539 case 8: OW->Write64(uint64_t(FF.getValue())); break;
545 case MCFragment::FT_Inst:
546 llvm_unreachable("unexpected inst fragment after lowering");
549 case MCFragment::FT_Org: {
550 MCOrgFragment &OF = cast<MCOrgFragment>(F);
552 for (uint64_t i = 0, e = FragmentSize; i != e; ++i)
553 OW->Write8(uint8_t(OF.getValue()));
558 case MCFragment::FT_ZeroFill: {
559 assert(0 && "Invalid zero fill fragment in concrete section!");
564 assert(OW->getStream().tell() - Start == FragmentSize);
567 void MCAssembler::WriteSectionData(const MCSectionData *SD,
568 const MCAsmLayout &Layout,
569 MCObjectWriter *OW) const {
570 uint64_t SectionSize = Layout.getSectionSize(SD);
571 uint64_t SectionFileSize = Layout.getSectionFileSize(SD);
573 // Ignore virtual sections.
574 if (getBackend().isVirtualSection(SD->getSection())) {
575 assert(SectionFileSize == 0 && "Invalid size for section!");
579 uint64_t Start = OW->getStream().tell();
582 for (MCSectionData::const_iterator it = SD->begin(),
583 ie = SD->end(); it != ie; ++it)
584 WriteFragmentData(*this, Layout, *it, OW);
586 // Add section padding.
587 assert(SectionFileSize >= SectionSize && "Invalid section sizes!");
588 OW->WriteZeros(SectionFileSize - SectionSize);
590 assert(OW->getStream().tell() - Start == SectionFileSize);
593 void MCAssembler::Finish() {
594 DEBUG_WITH_TYPE("mc-dump", {
595 llvm::errs() << "assembler backend - pre-layout\n--\n";
598 // Assign section and fragment ordinals, all subsequent backend code is
599 // responsible for updating these in place.
600 unsigned SectionIndex = 0;
601 unsigned FragmentIndex = 0;
602 for (MCAssembler::iterator it = begin(), ie = end(); it != ie; ++it) {
603 it->setOrdinal(SectionIndex++);
605 for (MCSectionData::iterator it2 = it->begin(),
606 ie2 = it->end(); it2 != ie2; ++it2)
607 it2->setOrdinal(FragmentIndex++);
610 // Layout until everything fits.
611 MCAsmLayout Layout(*this);
612 while (LayoutOnce(Layout))
615 DEBUG_WITH_TYPE("mc-dump", {
616 llvm::errs() << "assembler backend - post-relaxation\n--\n";
619 // Finalize the layout, including fragment lowering.
620 FinishLayout(Layout);
622 DEBUG_WITH_TYPE("mc-dump", {
623 llvm::errs() << "assembler backend - final-layout\n--\n";
626 uint64_t StartOffset = OS.tell();
627 llvm::OwningPtr<MCObjectWriter> Writer(getBackend().createObjectWriter(OS));
629 report_fatal_error("unable to create object writer!");
631 // Allow the object writer a chance to perform post-layout binding (for
632 // example, to set the index fields in the symbol data).
633 Writer->ExecutePostLayoutBinding(*this);
635 // Evaluate and apply the fixups, generating relocation entries as necessary.
636 for (MCAssembler::iterator it = begin(), ie = end(); it != ie; ++it) {
637 for (MCSectionData::iterator it2 = it->begin(),
638 ie2 = it->end(); it2 != ie2; ++it2) {
639 MCDataFragment *DF = dyn_cast<MCDataFragment>(it2);
643 for (MCDataFragment::fixup_iterator it3 = DF->fixup_begin(),
644 ie3 = DF->fixup_end(); it3 != ie3; ++it3) {
645 MCAsmFixup &Fixup = *it3;
647 // Evaluate the fixup.
650 if (!EvaluateFixup(Layout, Fixup, DF, Target, FixedValue)) {
651 // The fixup was unresolved, we need a relocation. Inform the object
652 // writer of the relocation, and give it an opportunity to adjust the
653 // fixup value if need be.
654 Writer->RecordRelocation(*this, Layout, DF, Fixup, Target,FixedValue);
657 getBackend().ApplyFixup(Fixup, *DF, FixedValue);
662 // Write the object file.
663 Writer->WriteObject(*this, Layout);
666 stats::ObjectBytes += OS.tell() - StartOffset;
669 bool MCAssembler::FixupNeedsRelaxation(const MCAsmFixup &Fixup,
670 const MCFragment *DF,
671 const MCAsmLayout &Layout) const {
675 // If we cannot resolve the fixup value, it requires relaxation.
678 if (!EvaluateFixup(Layout, Fixup, DF, Target, Value))
681 // Otherwise, relax if the value is too big for a (signed) i8.
683 // FIXME: This is target dependent!
684 return int64_t(Value) != int64_t(int8_t(Value));
687 bool MCAssembler::FragmentNeedsRelaxation(const MCInstFragment *IF,
688 const MCAsmLayout &Layout) const {
689 // If this inst doesn't ever need relaxation, ignore it. This occurs when we
690 // are intentionally pushing out inst fragments, or because we relaxed a
691 // previous instruction to one that doesn't need relaxation.
692 if (!getBackend().MayNeedRelaxation(IF->getInst(), IF->getFixups()))
695 for (MCInstFragment::const_fixup_iterator it = IF->fixup_begin(),
696 ie = IF->fixup_end(); it != ie; ++it)
697 if (FixupNeedsRelaxation(*it, IF, Layout))
703 bool MCAssembler::LayoutOnce(MCAsmLayout &Layout) {
704 ++stats::RelaxationSteps;
706 // Layout the concrete sections and fragments.
707 uint64_t Address = 0;
708 for (iterator it = begin(), ie = end(); it != ie; ++it) {
709 // Skip virtual sections.
710 if (getBackend().isVirtualSection(it->getSection()))
713 // Layout the section fragments and its size.
714 Address = LayoutSection(*it, Layout, Address);
717 // Layout the virtual sections.
718 for (iterator it = begin(), ie = end(); it != ie; ++it) {
719 if (!getBackend().isVirtualSection(it->getSection()))
722 // Layout the section fragments and its size.
723 Address = LayoutSection(*it, Layout, Address);
726 // Scan for fragments that need relaxation.
727 bool WasRelaxed = false;
728 for (iterator it = begin(), ie = end(); it != ie; ++it) {
729 MCSectionData &SD = *it;
731 for (MCSectionData::iterator it2 = SD.begin(),
732 ie2 = SD.end(); it2 != ie2; ++it2) {
733 // Check if this is an instruction fragment that needs relaxation.
734 MCInstFragment *IF = dyn_cast<MCInstFragment>(it2);
735 if (!IF || !FragmentNeedsRelaxation(IF, Layout))
738 ++stats::RelaxedInstructions;
740 // FIXME-PERF: We could immediately lower out instructions if we can tell
741 // they are fully resolved, to avoid retesting on later passes.
743 // Relax the fragment.
746 getBackend().RelaxInstruction(IF, Relaxed);
748 // Encode the new instruction.
750 // FIXME-PERF: If it matters, we could let the target do this. It can
751 // probably do so more efficiently in many cases.
752 SmallVector<MCFixup, 4> Fixups;
753 SmallString<256> Code;
754 raw_svector_ostream VecOS(Code);
755 getEmitter().EncodeInstruction(Relaxed, VecOS, Fixups);
758 // Update the instruction fragment.
759 int SlideAmount = Code.size() - IF->getInstSize();
760 IF->setInst(Relaxed);
761 IF->getCode() = Code;
762 IF->getFixups().clear();
763 for (unsigned i = 0, e = Fixups.size(); i != e; ++i) {
764 MCFixup &F = Fixups[i];
765 IF->getFixups().push_back(MCAsmFixup(F.getOffset(), *F.getValue(),
769 // Update the layout, and remember that we relaxed. If we are relaxing
770 // everything, we can skip this step since nothing will depend on updating
773 Layout.UpdateForSlide(IF, SlideAmount);
781 void MCAssembler::FinishLayout(MCAsmLayout &Layout) {
782 // Lower out any instruction fragments, to simplify the fixup application and
785 // FIXME-PERF: We don't have to do this, but the assumption is that it is
786 // cheap (we will mostly end up eliminating fragments and appending on to data
787 // fragments), so the extra complexity downstream isn't worth it. Evaluate
789 for (iterator it = begin(), ie = end(); it != ie; ++it) {
790 MCSectionData &SD = *it;
792 for (MCSectionData::iterator it2 = SD.begin(),
793 ie2 = SD.end(); it2 != ie2; ++it2) {
794 MCInstFragment *IF = dyn_cast<MCInstFragment>(it2);
798 // Create a new data fragment for the instruction.
800 // FIXME-PERF: Reuse previous data fragment if possible.
801 MCDataFragment *DF = new MCDataFragment();
802 SD.getFragmentList().insert(it2, DF);
804 // Update the data fragments layout data.
806 // FIXME: Add MCAsmLayout utility for this.
807 DF->setParent(IF->getParent());
808 DF->setAtom(IF->getAtom());
809 DF->setOrdinal(IF->getOrdinal());
810 Layout.setFragmentOffset(DF, Layout.getFragmentOffset(IF));
811 Layout.setFragmentEffectiveSize(DF, Layout.getFragmentEffectiveSize(IF));
813 // Copy in the data and the fixups.
814 DF->getContents().append(IF->getCode().begin(), IF->getCode().end());
815 for (unsigned i = 0, e = IF->getFixups().size(); i != e; ++i)
816 DF->getFixups().push_back(IF->getFixups()[i]);
818 // Delete the instruction fragment and update the iterator.
819 SD.getFragmentList().erase(IF);
829 raw_ostream &operator<<(raw_ostream &OS, const MCAsmFixup &AF) {
830 OS << "<MCAsmFixup" << " Offset:" << AF.Offset << " Value:" << *AF.Value
831 << " Kind:" << AF.Kind << ">";
837 void MCFragment::dump() {
838 raw_ostream &OS = llvm::errs();
840 OS << "<MCFragment " << (void*) this << " Offset:" << Offset
841 << " EffectiveSize:" << EffectiveSize;
846 void MCAlignFragment::dump() {
847 raw_ostream &OS = llvm::errs();
849 OS << "<MCAlignFragment ";
850 this->MCFragment::dump();
852 OS << " Alignment:" << getAlignment()
853 << " Value:" << getValue() << " ValueSize:" << getValueSize()
854 << " MaxBytesToEmit:" << getMaxBytesToEmit() << ">";
857 void MCDataFragment::dump() {
858 raw_ostream &OS = llvm::errs();
860 OS << "<MCDataFragment ";
861 this->MCFragment::dump();
864 for (unsigned i = 0, e = getContents().size(); i != e; ++i) {
866 OS << hexdigit((Contents[i] >> 4) & 0xF) << hexdigit(Contents[i] & 0xF);
868 OS << "] (" << getContents().size() << " bytes)";
870 if (!getFixups().empty()) {
873 for (fixup_iterator it = fixup_begin(), ie = fixup_end(); it != ie; ++it) {
874 if (it != fixup_begin()) OS << ",\n ";
883 void MCFillFragment::dump() {
884 raw_ostream &OS = llvm::errs();
886 OS << "<MCFillFragment ";
887 this->MCFragment::dump();
889 OS << " Value:" << getValue() << " ValueSize:" << getValueSize()
890 << " Count:" << getCount() << ">";
893 void MCInstFragment::dump() {
894 raw_ostream &OS = llvm::errs();
896 OS << "<MCInstFragment ";
897 this->MCFragment::dump();
900 getInst().dump_pretty(OS);
904 void MCOrgFragment::dump() {
905 raw_ostream &OS = llvm::errs();
907 OS << "<MCOrgFragment ";
908 this->MCFragment::dump();
910 OS << " Offset:" << getOffset() << " Value:" << getValue() << ">";
913 void MCZeroFillFragment::dump() {
914 raw_ostream &OS = llvm::errs();
916 OS << "<MCZeroFillFragment ";
917 this->MCFragment::dump();
919 OS << " Size:" << getSize() << " Alignment:" << getAlignment() << ">";
922 void MCSectionData::dump() {
923 raw_ostream &OS = llvm::errs();
925 OS << "<MCSectionData";
926 OS << " Alignment:" << getAlignment() << " Address:" << Address
927 << " Size:" << Size << " FileSize:" << FileSize
928 << " Fragments:[\n ";
929 for (iterator it = begin(), ie = end(); it != ie; ++it) {
930 if (it != begin()) OS << ",\n ";
936 void MCSymbolData::dump() {
937 raw_ostream &OS = llvm::errs();
939 OS << "<MCSymbolData Symbol:" << getSymbol()
940 << " Fragment:" << getFragment() << " Offset:" << getOffset()
941 << " Flags:" << getFlags() << " Index:" << getIndex();
943 OS << " (common, size:" << getCommonSize()
944 << " align: " << getCommonAlignment() << ")";
947 if (isPrivateExtern())
948 OS << " (private extern)";
952 void MCAssembler::dump() {
953 raw_ostream &OS = llvm::errs();
955 OS << "<MCAssembler\n";
956 OS << " Sections:[\n ";
957 for (iterator it = begin(), ie = end(); it != ie; ++it) {
958 if (it != begin()) OS << ",\n ";
964 for (symbol_iterator it = symbol_begin(), ie = symbol_end(); it != ie; ++it) {
965 if (it != symbol_begin()) OS << ",\n ";