1 //===- InstrProfWriter.cpp - Instrumented profiling writer ----------------===// 2 // 3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4 // See https://llvm.org/LICENSE.txt for license information. 5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This file contains support for writing profiling data for clang's 10 // instrumentation based PGO and coverage. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "llvm/ProfileData/InstrProfWriter.h" 15 #include "llvm/ADT/STLExtras.h" 16 #include "llvm/ADT/StringRef.h" 17 #include "llvm/IR/ProfileSummary.h" 18 #include "llvm/ProfileData/InstrProf.h" 19 #include "llvm/ProfileData/MemProf.h" 20 #include "llvm/ProfileData/ProfileCommon.h" 21 #include "llvm/Support/Endian.h" 22 #include "llvm/Support/EndianStream.h" 23 #include "llvm/Support/Error.h" 24 #include "llvm/Support/MemoryBuffer.h" 25 #include "llvm/Support/OnDiskHashTable.h" 26 #include "llvm/Support/raw_ostream.h" 27 #include <cstdint> 28 #include <memory> 29 #include <string> 30 #include <tuple> 31 #include <utility> 32 #include <vector> 33 34 using namespace llvm; 35 36 // A struct to define how the data stream should be patched. For Indexed 37 // profiling, only uint64_t data type is needed. 38 struct PatchItem { 39 uint64_t Pos; // Where to patch. 40 uint64_t *D; // Pointer to an array of source data. 41 int N; // Number of elements in \c D array. 42 }; 43 44 namespace llvm { 45 46 // A wrapper class to abstract writer stream with support of bytes 47 // back patching. 48 class ProfOStream { 49 public: 50 ProfOStream(raw_fd_ostream &FD) 51 : IsFDOStream(true), OS(FD), LE(FD, support::little) {} 52 ProfOStream(raw_string_ostream &STR) 53 : IsFDOStream(false), OS(STR), LE(STR, support::little) {} 54 55 uint64_t tell() { return OS.tell(); } 56 void write(uint64_t V) { LE.write<uint64_t>(V); } 57 58 // \c patch can only be called when all data is written and flushed. 59 // For raw_string_ostream, the patch is done on the target string 60 // directly and it won't be reflected in the stream's internal buffer. 61 void patch(PatchItem *P, int NItems) { 62 using namespace support; 63 64 if (IsFDOStream) { 65 raw_fd_ostream &FDOStream = static_cast<raw_fd_ostream &>(OS); 66 const uint64_t LastPos = FDOStream.tell(); 67 for (int K = 0; K < NItems; K++) { 68 FDOStream.seek(P[K].Pos); 69 for (int I = 0; I < P[K].N; I++) 70 write(P[K].D[I]); 71 } 72 // Reset the stream to the last position after patching so that users 73 // don't accidentally overwrite data. This makes it consistent with 74 // the string stream below which replaces the data directly. 75 FDOStream.seek(LastPos); 76 } else { 77 raw_string_ostream &SOStream = static_cast<raw_string_ostream &>(OS); 78 std::string &Data = SOStream.str(); // with flush 79 for (int K = 0; K < NItems; K++) { 80 for (int I = 0; I < P[K].N; I++) { 81 uint64_t Bytes = endian::byte_swap<uint64_t, little>(P[K].D[I]); 82 Data.replace(P[K].Pos + I * sizeof(uint64_t), sizeof(uint64_t), 83 (const char *)&Bytes, sizeof(uint64_t)); 84 } 85 } 86 } 87 } 88 89 // If \c OS is an instance of \c raw_fd_ostream, this field will be 90 // true. Otherwise, \c OS will be an raw_string_ostream. 91 bool IsFDOStream; 92 raw_ostream &OS; 93 support::endian::Writer LE; 94 }; 95 96 class InstrProfRecordWriterTrait { 97 public: 98 using key_type = StringRef; 99 using key_type_ref = StringRef; 100 101 using data_type = const InstrProfWriter::ProfilingData *const; 102 using data_type_ref = const InstrProfWriter::ProfilingData *const; 103 104 using hash_value_type = uint64_t; 105 using offset_type = uint64_t; 106 107 support::endianness ValueProfDataEndianness = support::little; 108 InstrProfSummaryBuilder *SummaryBuilder; 109 InstrProfSummaryBuilder *CSSummaryBuilder; 110 111 InstrProfRecordWriterTrait() = default; 112 113 static hash_value_type ComputeHash(key_type_ref K) { 114 return IndexedInstrProf::ComputeHash(K); 115 } 116 117 static std::pair<offset_type, offset_type> 118 EmitKeyDataLength(raw_ostream &Out, key_type_ref K, data_type_ref V) { 119 using namespace support; 120 121 endian::Writer LE(Out, little); 122 123 offset_type N = K.size(); 124 LE.write<offset_type>(N); 125 126 offset_type M = 0; 127 for (const auto &ProfileData : *V) { 128 const InstrProfRecord &ProfRecord = ProfileData.second; 129 M += sizeof(uint64_t); // The function hash 130 M += sizeof(uint64_t); // The size of the Counts vector 131 M += ProfRecord.Counts.size() * sizeof(uint64_t); 132 133 // Value data 134 M += ValueProfData::getSize(ProfileData.second); 135 } 136 LE.write<offset_type>(M); 137 138 return std::make_pair(N, M); 139 } 140 141 void EmitKey(raw_ostream &Out, key_type_ref K, offset_type N) { 142 Out.write(K.data(), N); 143 } 144 145 void EmitData(raw_ostream &Out, key_type_ref, data_type_ref V, offset_type) { 146 using namespace support; 147 148 endian::Writer LE(Out, little); 149 for (const auto &ProfileData : *V) { 150 const InstrProfRecord &ProfRecord = ProfileData.second; 151 if (NamedInstrProfRecord::hasCSFlagInHash(ProfileData.first)) 152 CSSummaryBuilder->addRecord(ProfRecord); 153 else 154 SummaryBuilder->addRecord(ProfRecord); 155 156 LE.write<uint64_t>(ProfileData.first); // Function hash 157 LE.write<uint64_t>(ProfRecord.Counts.size()); 158 for (uint64_t I : ProfRecord.Counts) 159 LE.write<uint64_t>(I); 160 161 // Write value data 162 std::unique_ptr<ValueProfData> VDataPtr = 163 ValueProfData::serializeFrom(ProfileData.second); 164 uint32_t S = VDataPtr->getSize(); 165 VDataPtr->swapBytesFromHost(ValueProfDataEndianness); 166 Out.write((const char *)VDataPtr.get(), S); 167 } 168 } 169 }; 170 171 } // end namespace llvm 172 173 InstrProfWriter::InstrProfWriter(bool Sparse) 174 : Sparse(Sparse), InfoObj(new InstrProfRecordWriterTrait()) {} 175 176 InstrProfWriter::~InstrProfWriter() { delete InfoObj; } 177 178 // Internal interface for testing purpose only. 179 void InstrProfWriter::setValueProfDataEndianness( 180 support::endianness Endianness) { 181 InfoObj->ValueProfDataEndianness = Endianness; 182 } 183 184 void InstrProfWriter::setOutputSparse(bool Sparse) { 185 this->Sparse = Sparse; 186 } 187 188 void InstrProfWriter::addRecord(NamedInstrProfRecord &&I, uint64_t Weight, 189 function_ref<void(Error)> Warn) { 190 auto Name = I.Name; 191 auto Hash = I.Hash; 192 addRecord(Name, Hash, std::move(I), Weight, Warn); 193 } 194 195 void InstrProfWriter::overlapRecord(NamedInstrProfRecord &&Other, 196 OverlapStats &Overlap, 197 OverlapStats &FuncLevelOverlap, 198 const OverlapFuncFilters &FuncFilter) { 199 auto Name = Other.Name; 200 auto Hash = Other.Hash; 201 Other.accumulateCounts(FuncLevelOverlap.Test); 202 if (FunctionData.find(Name) == FunctionData.end()) { 203 Overlap.addOneUnique(FuncLevelOverlap.Test); 204 return; 205 } 206 if (FuncLevelOverlap.Test.CountSum < 1.0f) { 207 Overlap.Overlap.NumEntries += 1; 208 return; 209 } 210 auto &ProfileDataMap = FunctionData[Name]; 211 bool NewFunc; 212 ProfilingData::iterator Where; 213 std::tie(Where, NewFunc) = 214 ProfileDataMap.insert(std::make_pair(Hash, InstrProfRecord())); 215 if (NewFunc) { 216 Overlap.addOneMismatch(FuncLevelOverlap.Test); 217 return; 218 } 219 InstrProfRecord &Dest = Where->second; 220 221 uint64_t ValueCutoff = FuncFilter.ValueCutoff; 222 if (!FuncFilter.NameFilter.empty() && Name.contains(FuncFilter.NameFilter)) 223 ValueCutoff = 0; 224 225 Dest.overlap(Other, Overlap, FuncLevelOverlap, ValueCutoff); 226 } 227 228 void InstrProfWriter::addRecord(StringRef Name, uint64_t Hash, 229 InstrProfRecord &&I, uint64_t Weight, 230 function_ref<void(Error)> Warn) { 231 auto &ProfileDataMap = FunctionData[Name]; 232 233 bool NewFunc; 234 ProfilingData::iterator Where; 235 std::tie(Where, NewFunc) = 236 ProfileDataMap.insert(std::make_pair(Hash, InstrProfRecord())); 237 InstrProfRecord &Dest = Where->second; 238 239 auto MapWarn = [&](instrprof_error E) { 240 Warn(make_error<InstrProfError>(E)); 241 }; 242 243 if (NewFunc) { 244 // We've never seen a function with this name and hash, add it. 245 Dest = std::move(I); 246 if (Weight > 1) 247 Dest.scale(Weight, 1, MapWarn); 248 } else { 249 // We're updating a function we've seen before. 250 Dest.merge(I, Weight, MapWarn); 251 } 252 253 Dest.sortValueData(); 254 } 255 256 void InstrProfWriter::addMemProfRecord( 257 const Function::GUID Id, const memprof::IndexedMemProfRecord &Record) { 258 auto Result = MemProfRecordData.insert({Id, Record}); 259 // If we inserted a new record then we are done. 260 if (Result.second) { 261 return; 262 } 263 memprof::IndexedMemProfRecord &Existing = Result.first->second; 264 Existing.merge(Record); 265 } 266 267 bool InstrProfWriter::addMemProfFrame(const memprof::FrameId Id, 268 const memprof::Frame &Frame, 269 function_ref<void(Error)> Warn) { 270 auto Result = MemProfFrameData.insert({Id, Frame}); 271 // If a mapping already exists for the current frame id and it does not 272 // match the new mapping provided then reset the existing contents and bail 273 // out. We don't support the merging of memprof data whose Frame -> Id 274 // mapping across profiles is inconsistent. 275 if (!Result.second && Result.first->second != Frame) { 276 Warn(make_error<InstrProfError>(instrprof_error::malformed, 277 "frame to id mapping mismatch")); 278 return false; 279 } 280 return true; 281 } 282 283 void InstrProfWriter::mergeRecordsFromWriter(InstrProfWriter &&IPW, 284 function_ref<void(Error)> Warn) { 285 for (auto &I : IPW.FunctionData) 286 for (auto &Func : I.getValue()) 287 addRecord(I.getKey(), Func.first, std::move(Func.second), 1, Warn); 288 289 MemProfFrameData.reserve(IPW.MemProfFrameData.size()); 290 for (auto &I : IPW.MemProfFrameData) { 291 // If we weren't able to add the frame mappings then it doesn't make sense 292 // to try to merge the records from this profile. 293 if (!addMemProfFrame(I.first, I.second, Warn)) 294 return; 295 } 296 297 MemProfRecordData.reserve(IPW.MemProfRecordData.size()); 298 for (auto &I : IPW.MemProfRecordData) { 299 addMemProfRecord(I.first, I.second); 300 } 301 } 302 303 bool InstrProfWriter::shouldEncodeData(const ProfilingData &PD) { 304 if (!Sparse) 305 return true; 306 for (const auto &Func : PD) { 307 const InstrProfRecord &IPR = Func.second; 308 if (llvm::any_of(IPR.Counts, [](uint64_t Count) { return Count > 0; })) 309 return true; 310 } 311 return false; 312 } 313 314 static void setSummary(IndexedInstrProf::Summary *TheSummary, 315 ProfileSummary &PS) { 316 using namespace IndexedInstrProf; 317 318 const std::vector<ProfileSummaryEntry> &Res = PS.getDetailedSummary(); 319 TheSummary->NumSummaryFields = Summary::NumKinds; 320 TheSummary->NumCutoffEntries = Res.size(); 321 TheSummary->set(Summary::MaxFunctionCount, PS.getMaxFunctionCount()); 322 TheSummary->set(Summary::MaxBlockCount, PS.getMaxCount()); 323 TheSummary->set(Summary::MaxInternalBlockCount, PS.getMaxInternalCount()); 324 TheSummary->set(Summary::TotalBlockCount, PS.getTotalCount()); 325 TheSummary->set(Summary::TotalNumBlocks, PS.getNumCounts()); 326 TheSummary->set(Summary::TotalNumFunctions, PS.getNumFunctions()); 327 for (unsigned I = 0; I < Res.size(); I++) 328 TheSummary->setEntry(I, Res[I]); 329 } 330 331 Error InstrProfWriter::writeImpl(ProfOStream &OS) { 332 using namespace IndexedInstrProf; 333 334 OnDiskChainedHashTableGenerator<InstrProfRecordWriterTrait> Generator; 335 336 InstrProfSummaryBuilder ISB(ProfileSummaryBuilder::DefaultCutoffs); 337 InfoObj->SummaryBuilder = &ISB; 338 InstrProfSummaryBuilder CSISB(ProfileSummaryBuilder::DefaultCutoffs); 339 InfoObj->CSSummaryBuilder = &CSISB; 340 341 // Populate the hash table generator. 342 for (const auto &I : FunctionData) 343 if (shouldEncodeData(I.getValue())) 344 Generator.insert(I.getKey(), &I.getValue()); 345 346 // Write the header. 347 IndexedInstrProf::Header Header; 348 Header.Magic = IndexedInstrProf::Magic; 349 Header.Version = IndexedInstrProf::ProfVersion::CurrentVersion; 350 if (static_cast<bool>(ProfileKind & InstrProfKind::IRInstrumentation)) 351 Header.Version |= VARIANT_MASK_IR_PROF; 352 if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive)) 353 Header.Version |= VARIANT_MASK_CSIR_PROF; 354 if (static_cast<bool>(ProfileKind & 355 InstrProfKind::FunctionEntryInstrumentation)) 356 Header.Version |= VARIANT_MASK_INSTR_ENTRY; 357 if (static_cast<bool>(ProfileKind & InstrProfKind::SingleByteCoverage)) 358 Header.Version |= VARIANT_MASK_BYTE_COVERAGE; 359 if (static_cast<bool>(ProfileKind & InstrProfKind::FunctionEntryOnly)) 360 Header.Version |= VARIANT_MASK_FUNCTION_ENTRY_ONLY; 361 if (static_cast<bool>(ProfileKind & InstrProfKind::MemProf)) 362 Header.Version |= VARIANT_MASK_MEMPROF; 363 364 Header.Unused = 0; 365 Header.HashType = static_cast<uint64_t>(IndexedInstrProf::HashType); 366 Header.HashOffset = 0; 367 Header.MemProfOffset = 0; 368 int N = sizeof(IndexedInstrProf::Header) / sizeof(uint64_t); 369 370 // Only write out all the fields except 'HashOffset' and 'MemProfOffset'. We 371 // need to remember the offset of these fields to allow back patching later. 372 for (int I = 0; I < N - 2; I++) 373 OS.write(reinterpret_cast<uint64_t *>(&Header)[I]); 374 375 // Save the location of Header.HashOffset field in \c OS. 376 uint64_t HashTableStartFieldOffset = OS.tell(); 377 // Reserve the space for HashOffset field. 378 OS.write(0); 379 380 // Save the location of MemProf profile data. This is stored in two parts as 381 // the schema and as a separate on-disk chained hashtable. 382 uint64_t MemProfSectionOffset = OS.tell(); 383 // Reserve space for the MemProf table field to be patched later if this 384 // profile contains memory profile information. 385 OS.write(0); 386 387 // Reserve space to write profile summary data. 388 uint32_t NumEntries = ProfileSummaryBuilder::DefaultCutoffs.size(); 389 uint32_t SummarySize = Summary::getSize(Summary::NumKinds, NumEntries); 390 // Remember the summary offset. 391 uint64_t SummaryOffset = OS.tell(); 392 for (unsigned I = 0; I < SummarySize / sizeof(uint64_t); I++) 393 OS.write(0); 394 uint64_t CSSummaryOffset = 0; 395 uint64_t CSSummarySize = 0; 396 if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive)) { 397 CSSummaryOffset = OS.tell(); 398 CSSummarySize = SummarySize / sizeof(uint64_t); 399 for (unsigned I = 0; I < CSSummarySize; I++) 400 OS.write(0); 401 } 402 403 // Write the hash table. 404 uint64_t HashTableStart = Generator.Emit(OS.OS, *InfoObj); 405 406 // Write the MemProf profile data if we have it. This includes a simple schema 407 // with the format described below followed by the hashtable: 408 // uint64_t RecordTableOffset = RecordTableGenerator.Emit 409 // uint64_t FramePayloadOffset = Stream offset before emitting the frame table 410 // uint64_t FrameTableOffset = FrameTableGenerator.Emit 411 // uint64_t Num schema entries 412 // uint64_t Schema entry 0 413 // uint64_t Schema entry 1 414 // .... 415 // uint64_t Schema entry N - 1 416 // OnDiskChainedHashTable MemProfRecordData 417 // OnDiskChainedHashTable MemProfFrameData 418 uint64_t MemProfSectionStart = 0; 419 if (static_cast<bool>(ProfileKind & InstrProfKind::MemProf)) { 420 MemProfSectionStart = OS.tell(); 421 OS.write(0ULL); // Reserve space for the memprof record table offset. 422 OS.write(0ULL); // Reserve space for the memprof frame payload offset. 423 OS.write(0ULL); // Reserve space for the memprof frame table offset. 424 425 auto Schema = memprof::PortableMemInfoBlock::getSchema(); 426 OS.write(static_cast<uint64_t>(Schema.size())); 427 for (const auto Id : Schema) { 428 OS.write(static_cast<uint64_t>(Id)); 429 } 430 431 auto RecordWriter = std::make_unique<memprof::RecordWriterTrait>(); 432 RecordWriter->Schema = &Schema; 433 OnDiskChainedHashTableGenerator<memprof::RecordWriterTrait> 434 RecordTableGenerator; 435 for (auto &I : MemProfRecordData) { 436 // Insert the key (func hash) and value (memprof record). 437 RecordTableGenerator.insert(I.first, I.second); 438 } 439 440 uint64_t RecordTableOffset = 441 RecordTableGenerator.Emit(OS.OS, *RecordWriter); 442 443 uint64_t FramePayloadOffset = OS.tell(); 444 445 auto FrameWriter = std::make_unique<memprof::FrameWriterTrait>(); 446 OnDiskChainedHashTableGenerator<memprof::FrameWriterTrait> 447 FrameTableGenerator; 448 for (auto &I : MemProfFrameData) { 449 // Insert the key (frame id) and value (frame contents). 450 FrameTableGenerator.insert(I.first, I.second); 451 } 452 453 uint64_t FrameTableOffset = FrameTableGenerator.Emit(OS.OS, *FrameWriter); 454 455 PatchItem PatchItems[] = { 456 {MemProfSectionStart, &RecordTableOffset, 1}, 457 {MemProfSectionStart + sizeof(uint64_t), &FramePayloadOffset, 1}, 458 {MemProfSectionStart + 2 * sizeof(uint64_t), &FrameTableOffset, 1}, 459 }; 460 OS.patch(PatchItems, 3); 461 } 462 463 // Allocate space for data to be serialized out. 464 std::unique_ptr<IndexedInstrProf::Summary> TheSummary = 465 IndexedInstrProf::allocSummary(SummarySize); 466 // Compute the Summary and copy the data to the data 467 // structure to be serialized out (to disk or buffer). 468 std::unique_ptr<ProfileSummary> PS = ISB.getSummary(); 469 setSummary(TheSummary.get(), *PS); 470 InfoObj->SummaryBuilder = nullptr; 471 472 // For Context Sensitive summary. 473 std::unique_ptr<IndexedInstrProf::Summary> TheCSSummary = nullptr; 474 if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive)) { 475 TheCSSummary = IndexedInstrProf::allocSummary(SummarySize); 476 std::unique_ptr<ProfileSummary> CSPS = CSISB.getSummary(); 477 setSummary(TheCSSummary.get(), *CSPS); 478 } 479 InfoObj->CSSummaryBuilder = nullptr; 480 481 // Now do the final patch: 482 PatchItem PatchItems[] = { 483 // Patch the Header.HashOffset field. 484 {HashTableStartFieldOffset, &HashTableStart, 1}, 485 // Patch the Header.MemProfOffset (=0 for profiles without MemProf data). 486 {MemProfSectionOffset, &MemProfSectionStart, 1}, 487 // Patch the summary data. 488 {SummaryOffset, reinterpret_cast<uint64_t *>(TheSummary.get()), 489 (int)(SummarySize / sizeof(uint64_t))}, 490 {CSSummaryOffset, reinterpret_cast<uint64_t *>(TheCSSummary.get()), 491 (int)CSSummarySize}}; 492 493 OS.patch(PatchItems, sizeof(PatchItems) / sizeof(*PatchItems)); 494 495 for (const auto &I : FunctionData) 496 for (const auto &F : I.getValue()) 497 if (Error E = validateRecord(F.second)) 498 return E; 499 500 return Error::success(); 501 } 502 503 Error InstrProfWriter::write(raw_fd_ostream &OS) { 504 // Write the hash table. 505 ProfOStream POS(OS); 506 return writeImpl(POS); 507 } 508 509 std::unique_ptr<MemoryBuffer> InstrProfWriter::writeBuffer() { 510 std::string Data; 511 raw_string_ostream OS(Data); 512 ProfOStream POS(OS); 513 // Write the hash table. 514 if (Error E = writeImpl(POS)) 515 return nullptr; 516 // Return this in an aligned memory buffer. 517 return MemoryBuffer::getMemBufferCopy(Data); 518 } 519 520 static const char *ValueProfKindStr[] = { 521 #define VALUE_PROF_KIND(Enumerator, Value, Descr) #Enumerator, 522 #include "llvm/ProfileData/InstrProfData.inc" 523 }; 524 525 Error InstrProfWriter::validateRecord(const InstrProfRecord &Func) { 526 for (uint32_t VK = 0; VK <= IPVK_Last; VK++) { 527 uint32_t NS = Func.getNumValueSites(VK); 528 if (!NS) 529 continue; 530 for (uint32_t S = 0; S < NS; S++) { 531 uint32_t ND = Func.getNumValueDataForSite(VK, S); 532 std::unique_ptr<InstrProfValueData[]> VD = Func.getValueForSite(VK, S); 533 bool WasZero = false; 534 for (uint32_t I = 0; I < ND; I++) 535 if ((VK != IPVK_IndirectCallTarget) && (VD[I].Value == 0)) { 536 if (WasZero) 537 return make_error<InstrProfError>(instrprof_error::invalid_prof); 538 WasZero = true; 539 } 540 } 541 } 542 543 return Error::success(); 544 } 545 546 void InstrProfWriter::writeRecordInText(StringRef Name, uint64_t Hash, 547 const InstrProfRecord &Func, 548 InstrProfSymtab &Symtab, 549 raw_fd_ostream &OS) { 550 OS << Name << "\n"; 551 OS << "# Func Hash:\n" << Hash << "\n"; 552 OS << "# Num Counters:\n" << Func.Counts.size() << "\n"; 553 OS << "# Counter Values:\n"; 554 for (uint64_t Count : Func.Counts) 555 OS << Count << "\n"; 556 557 uint32_t NumValueKinds = Func.getNumValueKinds(); 558 if (!NumValueKinds) { 559 OS << "\n"; 560 return; 561 } 562 563 OS << "# Num Value Kinds:\n" << Func.getNumValueKinds() << "\n"; 564 for (uint32_t VK = 0; VK < IPVK_Last + 1; VK++) { 565 uint32_t NS = Func.getNumValueSites(VK); 566 if (!NS) 567 continue; 568 OS << "# ValueKind = " << ValueProfKindStr[VK] << ":\n" << VK << "\n"; 569 OS << "# NumValueSites:\n" << NS << "\n"; 570 for (uint32_t S = 0; S < NS; S++) { 571 uint32_t ND = Func.getNumValueDataForSite(VK, S); 572 OS << ND << "\n"; 573 std::unique_ptr<InstrProfValueData[]> VD = Func.getValueForSite(VK, S); 574 for (uint32_t I = 0; I < ND; I++) { 575 if (VK == IPVK_IndirectCallTarget) 576 OS << Symtab.getFuncNameOrExternalSymbol(VD[I].Value) << ":" 577 << VD[I].Count << "\n"; 578 else 579 OS << VD[I].Value << ":" << VD[I].Count << "\n"; 580 } 581 } 582 } 583 584 OS << "\n"; 585 } 586 587 Error InstrProfWriter::writeText(raw_fd_ostream &OS) { 588 // Check CS first since it implies an IR level profile. 589 if (static_cast<bool>(ProfileKind & InstrProfKind::ContextSensitive)) 590 OS << "# CSIR level Instrumentation Flag\n:csir\n"; 591 else if (static_cast<bool>(ProfileKind & InstrProfKind::IRInstrumentation)) 592 OS << "# IR level Instrumentation Flag\n:ir\n"; 593 594 if (static_cast<bool>(ProfileKind & 595 InstrProfKind::FunctionEntryInstrumentation)) 596 OS << "# Always instrument the function entry block\n:entry_first\n"; 597 InstrProfSymtab Symtab; 598 599 using FuncPair = detail::DenseMapPair<uint64_t, InstrProfRecord>; 600 using RecordType = std::pair<StringRef, FuncPair>; 601 SmallVector<RecordType, 4> OrderedFuncData; 602 603 for (const auto &I : FunctionData) { 604 if (shouldEncodeData(I.getValue())) { 605 if (Error E = Symtab.addFuncName(I.getKey())) 606 return E; 607 for (const auto &Func : I.getValue()) 608 OrderedFuncData.push_back(std::make_pair(I.getKey(), Func)); 609 } 610 } 611 612 llvm::sort(OrderedFuncData, [](const RecordType &A, const RecordType &B) { 613 return std::tie(A.first, A.second.first) < 614 std::tie(B.first, B.second.first); 615 }); 616 617 for (const auto &record : OrderedFuncData) { 618 const StringRef &Name = record.first; 619 const FuncPair &Func = record.second; 620 writeRecordInText(Name, Func.first, Func.second, Symtab, OS); 621 } 622 623 for (const auto &record : OrderedFuncData) { 624 const FuncPair &Func = record.second; 625 if (Error E = validateRecord(Func.second)) 626 return E; 627 } 628 629 return Error::success(); 630 } 631