| //===-- ObjectContainerBSDArchive.cpp -------------------------------------===// |
| // |
| // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| // See https://llvm.org/LICENSE.txt for license information. |
| // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #include "ObjectContainerBSDArchive.h" |
| |
| #if defined(_WIN32) || defined(__ANDROID__) || defined(_AIX) |
| // Defines from ar, missing on Windows |
| #define SARMAG 8 |
| #define ARFMAG "`\n" |
| |
| typedef struct ar_hdr { |
| char ar_name[16]; |
| char ar_date[12]; |
| char ar_uid[6], ar_gid[6]; |
| char ar_mode[8]; |
| char ar_size[10]; |
| char ar_fmag[2]; |
| } ar_hdr; |
| #else |
| #include <ar.h> |
| #endif |
| |
| #include "lldb/Core/Module.h" |
| #include "lldb/Core/ModuleSpec.h" |
| #include "lldb/Core/PluginManager.h" |
| #include "lldb/Host/FileSystem.h" |
| #include "lldb/Symbol/ObjectFile.h" |
| #include "lldb/Utility/ArchSpec.h" |
| #include "lldb/Utility/LLDBLog.h" |
| #include "lldb/Utility/Stream.h" |
| #include "lldb/Utility/Timer.h" |
| |
| #include "llvm/Object/Archive.h" |
| #include "llvm/Support/MemoryBuffer.h" |
| |
| using namespace lldb; |
| using namespace lldb_private; |
| |
| using namespace llvm::object; |
| |
| LLDB_PLUGIN_DEFINE(ObjectContainerBSDArchive) |
| |
| ObjectContainerBSDArchive::Object::Object() : ar_name() {} |
| |
| void ObjectContainerBSDArchive::Object::Clear() { |
| ar_name.Clear(); |
| modification_time = 0; |
| size = 0; |
| file_offset = 0; |
| file_size = 0; |
| } |
| |
| void ObjectContainerBSDArchive::Object::Dump() const { |
| printf("name = \"%s\"\n", ar_name.GetCString()); |
| printf("mtime = 0x%8.8" PRIx32 "\n", modification_time); |
| printf("size = 0x%8.8" PRIx32 " (%" PRIu32 ")\n", size, size); |
| printf("file_offset = 0x%16.16" PRIx64 " (%" PRIu64 ")\n", file_offset, |
| file_offset); |
| printf("file_size = 0x%16.16" PRIx64 " (%" PRIu64 ")\n\n", file_size, |
| file_size); |
| } |
| |
| ObjectContainerBSDArchive::Archive::Archive(const lldb_private::ArchSpec &arch, |
| const llvm::sys::TimePoint<> &time, |
| lldb::offset_t file_offset, |
| lldb_private::DataExtractor &data, |
| ArchiveType archive_type) |
| : m_arch(arch), m_modification_time(time), m_file_offset(file_offset), |
| m_objects(), m_data(data), m_archive_type(archive_type) {} |
| |
| Log *l = GetLog(LLDBLog::Object); |
| ObjectContainerBSDArchive::Archive::~Archive() = default; |
| |
| size_t ObjectContainerBSDArchive::Archive::ParseObjects() { |
| DataExtractor &data = m_data; |
| |
| std::unique_ptr<llvm::MemoryBuffer> mem_buffer = |
| llvm::MemoryBuffer::getMemBuffer( |
| llvm::StringRef((const char *)data.GetDataStart(), |
| data.GetByteSize()), |
| llvm::StringRef(), |
| /*RequiresNullTerminator=*/false); |
| |
| auto exp_ar = llvm::object::Archive::create(mem_buffer->getMemBufferRef()); |
| if (!exp_ar) { |
| LLDB_LOG_ERROR(l, exp_ar.takeError(), "failed to create archive: {0}"); |
| return 0; |
| } |
| auto llvm_archive = std::move(exp_ar.get()); |
| |
| llvm::Error iter_err = llvm::Error::success(); |
| Object obj; |
| for (const auto &child : llvm_archive->children(iter_err)) { |
| obj.Clear(); |
| auto exp_name = child.getName(); |
| if (exp_name) { |
| obj.ar_name = ConstString(exp_name.get()); |
| } else { |
| LLDB_LOG_ERROR(l, exp_name.takeError(), |
| "failed to get archive object name: {0}"); |
| continue; |
| } |
| |
| auto exp_mtime = child.getLastModified(); |
| if (exp_mtime) { |
| obj.modification_time = |
| std::chrono::duration_cast<std::chrono::seconds>( |
| std::chrono::time_point_cast<std::chrono::seconds>( |
| exp_mtime.get()) |
| .time_since_epoch()) |
| .count(); |
| } else { |
| LLDB_LOG_ERROR(l, exp_mtime.takeError(), |
| "failed to get archive object time: {0}"); |
| continue; |
| } |
| |
| auto exp_size = child.getRawSize(); |
| if (exp_size) { |
| obj.size = exp_size.get(); |
| } else { |
| LLDB_LOG_ERROR(l, exp_size.takeError(), |
| "failed to get archive object size: {0}"); |
| continue; |
| } |
| |
| obj.file_offset = child.getDataOffset(); |
| |
| auto exp_file_size = child.getSize(); |
| if (exp_file_size) { |
| obj.file_size = exp_file_size.get(); |
| } else { |
| LLDB_LOG_ERROR(l, exp_file_size.takeError(), |
| "failed to get archive object file size: {0}"); |
| continue; |
| } |
| m_object_name_to_index_map.Append(obj.ar_name, m_objects.size()); |
| m_objects.push_back(obj); |
| } |
| if (iter_err) { |
| LLDB_LOG_ERROR(l, std::move(iter_err), |
| "failed to iterate over archive objects: {0}"); |
| } |
| // Now sort all of the object name pointers |
| m_object_name_to_index_map.Sort(); |
| return m_objects.size(); |
| } |
| |
| ObjectContainerBSDArchive::Object * |
| ObjectContainerBSDArchive::Archive::FindObject( |
| ConstString object_name, const llvm::sys::TimePoint<> &object_mod_time) { |
| const ObjectNameToIndexMap::Entry *match = |
| m_object_name_to_index_map.FindFirstValueForName(object_name); |
| if (!match) |
| return nullptr; |
| if (object_mod_time == llvm::sys::TimePoint<>()) |
| return &m_objects[match->value]; |
| |
| const uint64_t object_modification_date = llvm::sys::toTimeT(object_mod_time); |
| if (m_objects[match->value].modification_time == object_modification_date) |
| return &m_objects[match->value]; |
| |
| const ObjectNameToIndexMap::Entry *next_match = |
| m_object_name_to_index_map.FindNextValueForName(match); |
| while (next_match) { |
| if (m_objects[next_match->value].modification_time == |
| object_modification_date) |
| return &m_objects[next_match->value]; |
| next_match = m_object_name_to_index_map.FindNextValueForName(next_match); |
| } |
| |
| return nullptr; |
| } |
| |
| ObjectContainerBSDArchive::Archive::shared_ptr |
| ObjectContainerBSDArchive::Archive::FindCachedArchive( |
| const FileSpec &file, const ArchSpec &arch, |
| const llvm::sys::TimePoint<> &time, lldb::offset_t file_offset) { |
| std::lock_guard<std::recursive_mutex> guard(Archive::GetArchiveCacheMutex()); |
| shared_ptr archive_sp; |
| Archive::Map &archive_map = Archive::GetArchiveCache(); |
| Archive::Map::iterator pos = archive_map.find(file); |
| // Don't cache a value for "archive_map.end()" below since we might delete an |
| // archive entry... |
| while (pos != archive_map.end() && pos->first == file) { |
| bool match = true; |
| if (arch.IsValid() && |
| !pos->second->GetArchitecture().IsCompatibleMatch(arch)) |
| match = false; |
| else if (file_offset != LLDB_INVALID_OFFSET && |
| pos->second->GetFileOffset() != file_offset) |
| match = false; |
| if (match) { |
| if (pos->second->GetModificationTime() == time) { |
| return pos->second; |
| } else { |
| // We have a file at the same path with the same architecture whose |
| // modification time doesn't match. It doesn't make sense for us to |
| // continue to use this BSD archive since we cache only the object info |
| // which consists of file time info and also the file offset and file |
| // size of any contained objects. Since this information is now out of |
| // date, we won't get the correct information if we go and extract the |
| // file data, so we should remove the old and outdated entry. |
| archive_map.erase(pos); |
| pos = archive_map.find(file); |
| continue; // Continue to next iteration so we don't increment pos |
| // below... |
| } |
| } |
| ++pos; |
| } |
| return archive_sp; |
| } |
| |
| ObjectContainerBSDArchive::Archive::shared_ptr |
| ObjectContainerBSDArchive::Archive::ParseAndCacheArchiveForFile( |
| const FileSpec &file, const ArchSpec &arch, |
| const llvm::sys::TimePoint<> &time, lldb::offset_t file_offset, |
| DataExtractor &data, ArchiveType archive_type) { |
| shared_ptr archive_sp( |
| new Archive(arch, time, file_offset, data, archive_type)); |
| if (archive_sp) { |
| const size_t num_objects = archive_sp->ParseObjects(); |
| if (num_objects > 0) { |
| std::lock_guard<std::recursive_mutex> guard( |
| Archive::GetArchiveCacheMutex()); |
| Archive::GetArchiveCache().insert(std::make_pair(file, archive_sp)); |
| } else { |
| archive_sp.reset(); |
| } |
| } |
| return archive_sp; |
| } |
| |
| ObjectContainerBSDArchive::Archive::Map & |
| ObjectContainerBSDArchive::Archive::GetArchiveCache() { |
| static Archive::Map g_archive_map; |
| return g_archive_map; |
| } |
| |
| std::recursive_mutex & |
| ObjectContainerBSDArchive::Archive::GetArchiveCacheMutex() { |
| static std::recursive_mutex g_archive_map_mutex; |
| return g_archive_map_mutex; |
| } |
| |
| void ObjectContainerBSDArchive::Initialize() { |
| PluginManager::RegisterPlugin(GetPluginNameStatic(), |
| GetPluginDescriptionStatic(), CreateInstance, |
| GetModuleSpecifications); |
| } |
| |
| void ObjectContainerBSDArchive::Terminate() { |
| PluginManager::UnregisterPlugin(CreateInstance); |
| } |
| |
| ObjectContainer *ObjectContainerBSDArchive::CreateInstance( |
| const lldb::ModuleSP &module_sp, DataBufferSP &data_sp, |
| lldb::offset_t data_offset, const FileSpec *file, |
| lldb::offset_t file_offset, lldb::offset_t length) { |
| ConstString object_name(module_sp->GetObjectName()); |
| if (!object_name) |
| return nullptr; |
| |
| if (data_sp) { |
| // We have data, which means this is the first 512 bytes of the file Check |
| // to see if the magic bytes match and if they do, read the entire table of |
| // contents for the archive and cache it |
| DataExtractor data; |
| data.SetData(data_sp, data_offset, length); |
| ArchiveType archive_type = ObjectContainerBSDArchive::MagicBytesMatch(data); |
| if (file && data_sp && archive_type != ArchiveType::Invalid) { |
| LLDB_SCOPED_TIMERF( |
| "ObjectContainerBSDArchive::CreateInstance (module = %s, file = " |
| "%p, file_offset = 0x%8.8" PRIx64 ", file_size = 0x%8.8" PRIx64 ")", |
| module_sp->GetFileSpec().GetPath().c_str(), |
| static_cast<const void *>(file), static_cast<uint64_t>(file_offset), |
| static_cast<uint64_t>(length)); |
| |
| // Map the entire .a file to be sure that we don't lose any data if the |
| // file gets updated by a new build while this .a file is being used for |
| // debugging |
| DataBufferSP archive_data_sp = |
| FileSystem::Instance().CreateDataBuffer(*file, length, file_offset); |
| if (!archive_data_sp) |
| return nullptr; |
| |
| lldb::offset_t archive_data_offset = 0; |
| |
| Archive::shared_ptr archive_sp(Archive::FindCachedArchive( |
| *file, module_sp->GetArchitecture(), module_sp->GetModificationTime(), |
| file_offset)); |
| std::unique_ptr<ObjectContainerBSDArchive> container_up( |
| new ObjectContainerBSDArchive(module_sp, archive_data_sp, |
| archive_data_offset, file, file_offset, |
| length, archive_type)); |
| |
| if (container_up) { |
| if (archive_sp) { |
| // We already have this archive in our cache, use it |
| container_up->SetArchive(archive_sp); |
| return container_up.release(); |
| } else if (container_up->ParseHeader()) |
| return container_up.release(); |
| } |
| } |
| } else { |
| // No data, just check for a cached archive |
| Archive::shared_ptr archive_sp(Archive::FindCachedArchive( |
| *file, module_sp->GetArchitecture(), module_sp->GetModificationTime(), |
| file_offset)); |
| if (archive_sp) { |
| std::unique_ptr<ObjectContainerBSDArchive> container_up( |
| new ObjectContainerBSDArchive(module_sp, data_sp, data_offset, file, |
| file_offset, length, |
| archive_sp->GetArchiveType())); |
| |
| if (container_up) { |
| // We already have this archive in our cache, use it |
| container_up->SetArchive(archive_sp); |
| return container_up.release(); |
| } |
| } |
| } |
| return nullptr; |
| } |
| |
| ArchiveType |
| ObjectContainerBSDArchive::MagicBytesMatch(const DataExtractor &data) { |
| uint32_t offset = 0; |
| const char *armag = |
| (const char *)data.PeekData(offset, sizeof(ar_hdr) + SARMAG); |
| if (armag == nullptr) |
| return ArchiveType::Invalid; |
| ArchiveType result = ArchiveType::Invalid; |
| if (strncmp(armag, ArchiveMagic, SARMAG) == 0) |
| result = ArchiveType::Archive; |
| else if (strncmp(armag, ThinArchiveMagic, SARMAG) == 0) |
| result = ArchiveType::ThinArchive; |
| else |
| return ArchiveType::Invalid; |
| |
| armag += offsetof(struct ar_hdr, ar_fmag) + SARMAG; |
| if (strncmp(armag, ARFMAG, 2) == 0) |
| return result; |
| return ArchiveType::Invalid; |
| } |
| |
| ObjectContainerBSDArchive::ObjectContainerBSDArchive( |
| const lldb::ModuleSP &module_sp, DataBufferSP &data_sp, |
| lldb::offset_t data_offset, const lldb_private::FileSpec *file, |
| lldb::offset_t file_offset, lldb::offset_t size, ArchiveType archive_type) |
| : ObjectContainer(module_sp, file, file_offset, size, data_sp, data_offset), |
| m_archive_sp() { |
| m_archive_type = archive_type; |
| } |
| |
| void ObjectContainerBSDArchive::SetArchive(Archive::shared_ptr &archive_sp) { |
| m_archive_sp = archive_sp; |
| } |
| |
| ObjectContainerBSDArchive::~ObjectContainerBSDArchive() = default; |
| |
| bool ObjectContainerBSDArchive::ParseHeader() { |
| if (m_archive_sp.get() == nullptr) { |
| if (m_data.GetByteSize() > 0) { |
| ModuleSP module_sp(GetModule()); |
| if (module_sp) { |
| m_archive_sp = Archive::ParseAndCacheArchiveForFile( |
| m_file, module_sp->GetArchitecture(), |
| module_sp->GetModificationTime(), m_offset, m_data, m_archive_type); |
| } |
| // Clear the m_data that contains the entire archive data and let our |
| // m_archive_sp hold onto the data. |
| m_data.Clear(); |
| } |
| } |
| return m_archive_sp.get() != nullptr; |
| } |
| |
| FileSpec GetChildFileSpecificationsFromThin(llvm::StringRef childPath, |
| const FileSpec &parentFileSpec) { |
| llvm::SmallString<128> FullPath; |
| if (llvm::sys::path::is_absolute(childPath)) { |
| FullPath = childPath; |
| } else { |
| FullPath = parentFileSpec.GetDirectory().GetStringRef(); |
| llvm::sys::path::append(FullPath, childPath); |
| } |
| FileSpec child = FileSpec(FullPath.str(), llvm::sys::path::Style::posix); |
| return child; |
| } |
| |
| ObjectFileSP ObjectContainerBSDArchive::GetObjectFile(const FileSpec *file) { |
| ModuleSP module_sp(GetModule()); |
| if (module_sp) { |
| if (module_sp->GetObjectName() && m_archive_sp) { |
| Object *object = m_archive_sp->FindObject( |
| module_sp->GetObjectName(), module_sp->GetObjectModificationTime()); |
| if (object) { |
| if (m_archive_type == ArchiveType::ThinArchive) { |
| // Set file to child object file |
| FileSpec child = GetChildFileSpecificationsFromThin( |
| object->ar_name.GetStringRef(), m_file); |
| lldb::offset_t file_offset = 0; |
| lldb::offset_t file_size = object->size; |
| std::shared_ptr<DataBuffer> child_data_sp = |
| FileSystem::Instance().CreateDataBuffer(child, file_size, |
| file_offset); |
| if (!child_data_sp || |
| child_data_sp->GetByteSize() != object->file_size) |
| return ObjectFileSP(); |
| lldb::offset_t data_offset = 0; |
| return ObjectFile::FindPlugin( |
| module_sp, &child, m_offset + object->file_offset, |
| object->file_size, child_data_sp, data_offset); |
| } |
| lldb::offset_t data_offset = object->file_offset; |
| return ObjectFile::FindPlugin( |
| module_sp, file, m_offset + object->file_offset, object->file_size, |
| m_archive_sp->GetData().GetSharedDataBuffer(), data_offset); |
| } |
| } |
| } |
| return ObjectFileSP(); |
| } |
| |
| size_t ObjectContainerBSDArchive::GetModuleSpecifications( |
| const lldb_private::FileSpec &file, lldb::DataBufferSP &data_sp, |
| lldb::offset_t data_offset, lldb::offset_t file_offset, |
| lldb::offset_t file_size, lldb_private::ModuleSpecList &specs) { |
| |
| // We have data, which means this is the first 512 bytes of the file Check to |
| // see if the magic bytes match and if they do, read the entire table of |
| // contents for the archive and cache it |
| DataExtractor data; |
| data.SetData(data_sp, data_offset, data_sp->GetByteSize()); |
| ArchiveType archive_type = ObjectContainerBSDArchive::MagicBytesMatch(data); |
| if (!file || !data_sp || archive_type == ArchiveType::Invalid) |
| return 0; |
| |
| const size_t initial_count = specs.GetSize(); |
| llvm::sys::TimePoint<> file_mod_time = |
| FileSystem::Instance().GetModificationTime(file); |
| Archive::shared_ptr archive_sp( |
| Archive::FindCachedArchive(file, ArchSpec(), file_mod_time, file_offset)); |
| bool set_archive_arch = false; |
| if (!archive_sp) { |
| set_archive_arch = true; |
| data_sp = |
| FileSystem::Instance().CreateDataBuffer(file, file_size, file_offset); |
| if (data_sp) { |
| data.SetData(data_sp, 0, data_sp->GetByteSize()); |
| archive_sp = Archive::ParseAndCacheArchiveForFile( |
| file, ArchSpec(), file_mod_time, file_offset, data, archive_type); |
| } |
| } |
| |
| if (archive_sp) { |
| const size_t num_objects = archive_sp->GetNumObjects(); |
| for (size_t idx = 0; idx < num_objects; ++idx) { |
| const Object *object = archive_sp->GetObjectAtIndex(idx); |
| if (object) { |
| if (archive_sp->GetArchiveType() == ArchiveType::ThinArchive) { |
| if (object->ar_name.IsEmpty()) |
| continue; |
| FileSpec child = GetChildFileSpecificationsFromThin( |
| object->ar_name.GetStringRef(), file); |
| if (ObjectFile::GetModuleSpecifications(child, 0, object->file_size, |
| specs)) { |
| ModuleSpec &spec = |
| specs.GetModuleSpecRefAtIndex(specs.GetSize() - 1); |
| llvm::sys::TimePoint<> object_mod_time( |
| std::chrono::seconds(object->modification_time)); |
| spec.GetObjectName() = object->ar_name; |
| spec.SetObjectOffset(0); |
| spec.SetObjectSize(object->file_size); |
| spec.GetObjectModificationTime() = object_mod_time; |
| } |
| continue; |
| } |
| const lldb::offset_t object_file_offset = |
| file_offset + object->file_offset; |
| if (object->file_offset < file_size && file_size > object_file_offset) { |
| if (ObjectFile::GetModuleSpecifications( |
| file, object_file_offset, file_size - object_file_offset, |
| specs)) { |
| ModuleSpec &spec = |
| specs.GetModuleSpecRefAtIndex(specs.GetSize() - 1); |
| llvm::sys::TimePoint<> object_mod_time( |
| std::chrono::seconds(object->modification_time)); |
| spec.GetObjectName() = object->ar_name; |
| spec.SetObjectOffset(object_file_offset); |
| spec.SetObjectSize(object->file_size); |
| spec.GetObjectModificationTime() = object_mod_time; |
| } |
| } |
| } |
| } |
| } |
| const size_t end_count = specs.GetSize(); |
| size_t num_specs_added = end_count - initial_count; |
| if (set_archive_arch && num_specs_added > 0) { |
| // The archive was created but we didn't have an architecture so we need to |
| // set it |
| for (size_t i = initial_count; i < end_count; ++i) { |
| ModuleSpec module_spec; |
| if (specs.GetModuleSpecAtIndex(i, module_spec)) { |
| if (module_spec.GetArchitecture().IsValid()) { |
| archive_sp->SetArchitecture(module_spec.GetArchitecture()); |
| break; |
| } |
| } |
| } |
| } |
| return num_specs_added; |
| } |