2021-11-21 23:59:44 +00:00
|
|
|
// Copyright (c) Meta Platforms, Inc. and affiliates.
|
2021-04-03 09:10:35 +01:00
|
|
|
// SPDX-License-Identifier: GPL-3.0-or-later
|
Rewrite drgn core in C
The current mixed Python/C implementation works well, but it has a
couple of important limitations:
- It's too slow for some common use cases, like iterating over large
data structures.
- It can't be reused in utilities written in other languages.
This replaces the internals with a new library written in C, libdrgn. It
includes Python bindings with mostly the same public interface as
before, with some important improvements:
- Types are now represented by a single Type class rather than the messy
polymorphism in the Python implementation.
- Qualifiers are a bitmask instead of a set of strings.
- Bit fields are not considered a separate type.
- The lvalue/rvalue terminology is replaced with reference/value.
- Structure, union, and array values are better supported.
- Function objects are supported.
- Program distinguishes between lookups of variables, constants, and
functions.
The C rewrite is about 6x as fast as the original Python when using the
Python bindings, and about 8x when using the C API directly.
Currently, the exposed API in C is fairly conservative. In the future,
the memory reader, type index, and object index APIs will probably be
exposed for more flexibility.
2019-03-22 23:27:46 +00:00
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
#include <assert.h>
|
2021-03-26 00:26:06 +00:00
|
|
|
#include <byteswap.h>
|
2020-09-16 01:42:53 +01:00
|
|
|
#include <elf.h>
|
2019-04-28 20:13:29 +01:00
|
|
|
#include <elfutils/libdw.h>
|
2020-09-16 01:42:53 +01:00
|
|
|
#include <elfutils/libdwelf.h>
|
2021-03-20 01:49:06 +00:00
|
|
|
#include <elfutils/version.h>
|
2020-09-24 00:02:02 +01:00
|
|
|
#include <errno.h>
|
|
|
|
#include <fcntl.h>
|
2020-09-16 01:42:53 +01:00
|
|
|
#include <gelf.h>
|
2020-09-24 00:02:02 +01:00
|
|
|
#include <stdarg.h>
|
|
|
|
#include <stdio.h>
|
2020-09-16 01:42:53 +01:00
|
|
|
#include <stdlib.h>
|
Rewrite drgn core in C
The current mixed Python/C implementation works well, but it has a
couple of important limitations:
- It's too slow for some common use cases, like iterating over large
data structures.
- It can't be reused in utilities written in other languages.
This replaces the internals with a new library written in C, libdrgn. It
includes Python bindings with mostly the same public interface as
before, with some important improvements:
- Types are now represented by a single Type class rather than the messy
polymorphism in the Python implementation.
- Qualifiers are a bitmask instead of a set of strings.
- Bit fields are not considered a separate type.
- The lvalue/rvalue terminology is replaced with reference/value.
- Structure, union, and array values are better supported.
- Function objects are supported.
- Program distinguishes between lookups of variables, constants, and
functions.
The C rewrite is about 6x as fast as the original Python when using the
Python bindings, and about 8x when using the C API directly.
Currently, the exposed API in C is fairly conservative. In the future,
the memory reader, type index, and object index APIs will probably be
exposed for more flexibility.
2019-03-22 23:27:46 +00:00
|
|
|
#include <string.h>
|
2020-09-16 01:42:53 +01:00
|
|
|
#include <unistd.h>
|
Rewrite drgn core in C
The current mixed Python/C implementation works well, but it has a
couple of important limitations:
- It's too slow for some common use cases, like iterating over large
data structures.
- It can't be reused in utilities written in other languages.
This replaces the internals with a new library written in C, libdrgn. It
includes Python bindings with mostly the same public interface as
before, with some important improvements:
- Types are now represented by a single Type class rather than the messy
polymorphism in the Python implementation.
- Qualifiers are a bitmask instead of a set of strings.
- Bit fields are not considered a separate type.
- The lvalue/rvalue terminology is replaced with reference/value.
- Structure, union, and array values are better supported.
- Function objects are supported.
- Program distinguishes between lookups of variables, constants, and
functions.
The C rewrite is about 6x as fast as the original Python when using the
Python bindings, and about 8x when using the C API directly.
Currently, the exposed API in C is fairly conservative. In the future,
the memory reader, type index, and object index APIs will probably be
exposed for more flexibility.
2019-03-22 23:27:46 +00:00
|
|
|
|
2020-09-12 01:41:23 +01:00
|
|
|
#include "debug_info.h"
|
2020-09-24 00:02:02 +01:00
|
|
|
#include "error.h"
|
2020-09-16 01:42:53 +01:00
|
|
|
#include "linux_kernel.h"
|
Associate types with program
I originally envisioned types as dumb descriptors. This mostly works for
C because in C, types are fairly simple. However, even then the
drgn_program_member_info() API is awkward. You should be able to look up
a member directly from a type, but we need the program for caching
purposes. This has also held me back from adding offsetof() or
has_member() APIs.
Things get even messier with C++. C++ template parameters can be objects
(e.g., template <int N>). Such parameters would best be represented by a
drgn object, which we need a drgn program for. Static members are a
similar case.
So, let's reimagine types as being owned by a program. This has a few
parts:
1. In libdrgn, simple types are now created by factory functions,
drgn_foo_type_create().
2. To handle their variable length fields, compound types, enum types,
and function types are constructed with a "builder" API.
3. Simple types are deduplicated.
4. The Python type factory functions are replaced by methods of the
Program class.
5. While we're changing the API, the parameters to pointer_type() and
array_type() are reordered to be more logical (and to allow
pointer_type() to take a default size of None for the program's
default pointer size).
6. Likewise, the type factory methods take qualifiers as a keyword
argument only.
A big part of this change is updating the tests and splitting up large
test cases into smaller ones in a few places.
Signed-off-by: Omar Sandoval <osandov@osandov.com>
2020-07-16 00:34:56 +01:00
|
|
|
#include "program.h"
|
2020-09-24 00:02:02 +01:00
|
|
|
#include "util.h"
|
Rewrite drgn core in C
The current mixed Python/C implementation works well, but it has a
couple of important limitations:
- It's too slow for some common use cases, like iterating over large
data structures.
- It can't be reused in utilities written in other languages.
This replaces the internals with a new library written in C, libdrgn. It
includes Python bindings with mostly the same public interface as
before, with some important improvements:
- Types are now represented by a single Type class rather than the messy
polymorphism in the Python implementation.
- Qualifiers are a bitmask instead of a set of strings.
- Bit fields are not considered a separate type.
- The lvalue/rvalue terminology is replaced with reference/value.
- Structure, union, and array values are better supported.
- Function objects are supported.
- Program distinguishes between lookups of variables, constants, and
functions.
The C rewrite is about 6x as fast as the original Python when using the
Python bindings, and about 8x when using the C API directly.
Currently, the exposed API in C is fairly conservative. In the future,
the memory reader, type index, and object index APIs will probably be
exposed for more flexibility.
2019-03-22 23:27:46 +00:00
|
|
|
|
2020-10-23 00:09:31 +01:00
|
|
|
static const char * const drgn_debug_scn_names[] = {
|
|
|
|
[DRGN_SCN_DEBUG_INFO] = ".debug_info",
|
2020-11-07 00:09:27 +00:00
|
|
|
[DRGN_SCN_DEBUG_TYPES] = ".debug_types",
|
2020-10-23 00:09:31 +01:00
|
|
|
[DRGN_SCN_DEBUG_ABBREV] = ".debug_abbrev",
|
|
|
|
[DRGN_SCN_DEBUG_STR] = ".debug_str",
|
2021-06-15 01:40:52 +01:00
|
|
|
[DRGN_SCN_DEBUG_STR_OFFSETS] = ".debug_str_offsets",
|
2020-10-23 00:09:31 +01:00
|
|
|
[DRGN_SCN_DEBUG_LINE] = ".debug_line",
|
2021-06-15 01:40:52 +01:00
|
|
|
[DRGN_SCN_DEBUG_LINE_STR] = ".debug_line_str",
|
2021-07-07 00:26:31 +01:00
|
|
|
[DRGN_SCN_DEBUG_ADDR] = ".debug_addr",
|
2020-11-13 00:14:30 +00:00
|
|
|
[DRGN_SCN_DEBUG_FRAME] = ".debug_frame",
|
|
|
|
[DRGN_SCN_EH_FRAME] = ".eh_frame",
|
2021-03-16 22:39:37 +00:00
|
|
|
[DRGN_SCN_ORC_UNWIND_IP] = ".orc_unwind_ip",
|
|
|
|
[DRGN_SCN_ORC_UNWIND] = ".orc_unwind",
|
2021-05-05 09:26:49 +01:00
|
|
|
[DRGN_SCN_DEBUG_LOC] = ".debug_loc",
|
2021-07-01 22:02:38 +01:00
|
|
|
[DRGN_SCN_DEBUG_LOCLISTS] = ".debug_loclists",
|
2020-11-13 00:14:30 +00:00
|
|
|
[DRGN_SCN_TEXT] = ".text",
|
|
|
|
[DRGN_SCN_GOT] = ".got",
|
2020-10-23 00:09:31 +01:00
|
|
|
};
|
|
|
|
|
2021-03-10 09:35:23 +00:00
|
|
|
struct drgn_error *
|
|
|
|
drgn_error_debug_info_scn(struct drgn_debug_info_module *module,
|
|
|
|
enum drgn_debug_info_scn scn, const char *ptr,
|
|
|
|
const char *message)
|
2020-10-29 00:35:09 +00:00
|
|
|
{
|
|
|
|
const char *name = dwfl_module_info(module->dwfl_module, NULL, NULL,
|
|
|
|
NULL, NULL, NULL, NULL, NULL);
|
|
|
|
return drgn_error_format(DRGN_ERROR_OTHER, "%s: %s+%#tx: %s",
|
|
|
|
name, drgn_debug_scn_names[scn],
|
2020-11-13 00:14:30 +00:00
|
|
|
ptr - (const char *)module->scn_data[scn]->d_buf,
|
2020-10-29 00:35:09 +00:00
|
|
|
message);
|
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_error *drgn_debug_info_buffer_error(struct binary_buffer *bb,
|
|
|
|
const char *pos,
|
|
|
|
const char *message)
|
|
|
|
{
|
|
|
|
struct drgn_debug_info_buffer *buffer =
|
|
|
|
container_of(bb, struct drgn_debug_info_buffer, bb);
|
2021-03-10 09:35:23 +00:00
|
|
|
return drgn_error_debug_info_scn(buffer->module, buffer->scn, pos,
|
|
|
|
message);
|
2020-10-29 00:35:09 +00:00
|
|
|
}
|
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
DEFINE_VECTOR_FUNCTIONS(drgn_debug_info_module_vector)
|
|
|
|
|
2021-10-11 21:01:07 +01:00
|
|
|
struct drgn_debug_info_module_key {
|
|
|
|
const void *build_id;
|
|
|
|
size_t build_id_len;
|
|
|
|
uint64_t start, end;
|
|
|
|
};
|
|
|
|
|
|
|
|
static inline struct drgn_debug_info_module_key
|
|
|
|
drgn_debug_info_module_key(struct drgn_debug_info_module * const *entry)
|
|
|
|
{
|
|
|
|
return (struct drgn_debug_info_module_key){
|
|
|
|
.build_id = (*entry)->build_id,
|
|
|
|
.build_id_len = (*entry)->build_id_len,
|
|
|
|
.start = (*entry)->start,
|
|
|
|
.end = (*entry)->end,
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
static inline struct hash_pair
|
2020-10-13 00:06:23 +01:00
|
|
|
drgn_debug_info_module_key_hash_pair(const struct drgn_debug_info_module_key *key)
|
2020-09-16 01:42:53 +01:00
|
|
|
{
|
2020-10-13 00:06:23 +01:00
|
|
|
size_t hash = hash_bytes(key->build_id, key->build_id_len);
|
2020-09-16 01:42:53 +01:00
|
|
|
hash = hash_combine(hash, key->start);
|
|
|
|
hash = hash_combine(hash, key->end);
|
|
|
|
return hash_pair_from_avalanching_hash(hash);
|
|
|
|
}
|
|
|
|
static inline bool
|
2020-10-13 00:06:23 +01:00
|
|
|
drgn_debug_info_module_key_eq(const struct drgn_debug_info_module_key *a,
|
|
|
|
const struct drgn_debug_info_module_key *b)
|
2020-09-16 01:42:53 +01:00
|
|
|
{
|
|
|
|
return (a->build_id_len == b->build_id_len &&
|
|
|
|
memcmp(a->build_id, b->build_id, a->build_id_len) == 0 &&
|
|
|
|
a->start == b->start && a->end == b->end);
|
|
|
|
}
|
|
|
|
DEFINE_HASH_TABLE_FUNCTIONS(drgn_debug_info_module_table,
|
2021-10-11 21:01:07 +01:00
|
|
|
drgn_debug_info_module_key,
|
2020-10-13 00:06:23 +01:00
|
|
|
drgn_debug_info_module_key_hash_pair,
|
|
|
|
drgn_debug_info_module_key_eq)
|
2020-09-16 01:42:53 +01:00
|
|
|
|
2021-10-11 21:01:07 +01:00
|
|
|
DEFINE_HASH_SET_FUNCTIONS(c_string_set, c_string_key_hash_pair, c_string_key_eq)
|
2020-09-16 01:42:53 +01:00
|
|
|
|
|
|
|
/**
|
|
|
|
* @c Dwfl_Callbacks::find_elf() implementation.
|
|
|
|
*
|
|
|
|
* Ideally we'd use @c dwfl_report_elf() instead, but that doesn't take an @c
|
|
|
|
* Elf handle, which we need for a couple of reasons:
|
|
|
|
*
|
|
|
|
* - We usually already have the @c Elf handle open in order to identify the
|
|
|
|
* file.
|
|
|
|
* - For kernel modules, we set the section addresses in the @c Elf handle
|
|
|
|
* ourselves instead of using @c Dwfl_Callbacks::section_address().
|
|
|
|
*
|
|
|
|
* Additionally, there's a special case for vmlinux. It is usually an @c ET_EXEC
|
|
|
|
* ELF file, but when KASLR is enabled, it needs to be handled like an @c ET_DYN
|
|
|
|
* file. libdwfl has a hack for this when @c dwfl_report_module() is used, but
|
|
|
|
* @ref dwfl_report_elf() bypasses this hack.
|
|
|
|
*
|
|
|
|
* So, we're stuck using @c dwfl_report_module() and this dummy callback.
|
|
|
|
*/
|
|
|
|
static int drgn_dwfl_find_elf(Dwfl_Module *dwfl_module, void **userdatap,
|
|
|
|
const char *name, Dwarf_Addr base,
|
|
|
|
char **file_name, Elf **elfp)
|
|
|
|
{
|
|
|
|
struct drgn_debug_info_module *module = *userdatap;
|
|
|
|
/*
|
|
|
|
* libdwfl consumes the returned path, file descriptor, and ELF handle,
|
|
|
|
* so clear the fields.
|
|
|
|
*/
|
|
|
|
*file_name = module->path;
|
|
|
|
int fd = module->fd;
|
|
|
|
*elfp = module->elf;
|
|
|
|
module->path = NULL;
|
|
|
|
module->fd = -1;
|
|
|
|
module->elf = NULL;
|
|
|
|
return fd;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Uses drgn_dwfl_find_elf() if the ELF file was reported directly and falls
|
|
|
|
* back to dwfl_linux_proc_find_elf() otherwise.
|
|
|
|
*/
|
|
|
|
static int drgn_dwfl_linux_proc_find_elf(Dwfl_Module *dwfl_module,
|
|
|
|
void **userdatap, const char *name,
|
|
|
|
Dwarf_Addr base, char **file_name,
|
|
|
|
Elf **elfp)
|
|
|
|
{
|
|
|
|
struct drgn_debug_info_module *module = *userdatap;
|
|
|
|
if (module->elf) {
|
|
|
|
return drgn_dwfl_find_elf(dwfl_module, userdatap, name, base,
|
|
|
|
file_name, elfp);
|
|
|
|
}
|
|
|
|
return dwfl_linux_proc_find_elf(dwfl_module, userdatap, name, base,
|
|
|
|
file_name, elfp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Uses drgn_dwfl_find_elf() if the ELF file was reported directly and falls
|
|
|
|
* back to dwfl_build_id_find_elf() otherwise.
|
|
|
|
*/
|
|
|
|
static int drgn_dwfl_build_id_find_elf(Dwfl_Module *dwfl_module,
|
|
|
|
void **userdatap, const char *name,
|
|
|
|
Dwarf_Addr base, char **file_name,
|
|
|
|
Elf **elfp)
|
|
|
|
{
|
|
|
|
struct drgn_debug_info_module *module = *userdatap;
|
|
|
|
if (module->elf) {
|
|
|
|
return drgn_dwfl_find_elf(dwfl_module, userdatap, name, base,
|
|
|
|
file_name, elfp);
|
|
|
|
}
|
|
|
|
return dwfl_build_id_find_elf(dwfl_module, userdatap, name, base,
|
|
|
|
file_name, elfp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* @c Dwfl_Callbacks::section_address() implementation.
|
|
|
|
*
|
|
|
|
* We set the section header @c sh_addr in memory instead of using this, but
|
|
|
|
* libdwfl requires the callback pointer to be non-@c NULL. It will be called
|
|
|
|
* for any sections that still have a zero @c sh_addr, meaning they are not
|
|
|
|
* present in memory.
|
|
|
|
*/
|
|
|
|
static int drgn_dwfl_section_address(Dwfl_Module *module, void **userdatap,
|
|
|
|
const char *name, Dwarf_Addr base,
|
|
|
|
const char *secname, Elf32_Word shndx,
|
|
|
|
const GElf_Shdr *shdr, Dwarf_Addr *addr)
|
|
|
|
{
|
|
|
|
*addr = -1;
|
|
|
|
return DWARF_CB_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const Dwfl_Callbacks drgn_dwfl_callbacks = {
|
|
|
|
.find_elf = drgn_dwfl_find_elf,
|
|
|
|
.find_debuginfo = dwfl_standard_find_debuginfo,
|
|
|
|
.section_address = drgn_dwfl_section_address,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const Dwfl_Callbacks drgn_linux_proc_dwfl_callbacks = {
|
|
|
|
.find_elf = drgn_dwfl_linux_proc_find_elf,
|
|
|
|
.find_debuginfo = dwfl_standard_find_debuginfo,
|
|
|
|
.section_address = drgn_dwfl_section_address,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const Dwfl_Callbacks drgn_userspace_core_dump_dwfl_callbacks = {
|
|
|
|
.find_elf = drgn_dwfl_build_id_find_elf,
|
|
|
|
.find_debuginfo = dwfl_standard_find_debuginfo,
|
|
|
|
.section_address = drgn_dwfl_section_address,
|
|
|
|
};
|
|
|
|
|
|
|
|
static void
|
|
|
|
drgn_debug_info_module_destroy(struct drgn_debug_info_module *module)
|
|
|
|
{
|
|
|
|
if (module) {
|
|
|
|
drgn_error_destroy(module->err);
|
2021-11-08 23:39:39 +00:00
|
|
|
drgn_orc_module_info_deinit(module);
|
2021-11-09 01:05:33 +00:00
|
|
|
drgn_dwarf_module_info_deinit(module);
|
2020-09-16 01:42:53 +01:00
|
|
|
elf_end(module->elf);
|
|
|
|
if (module->fd != -1)
|
|
|
|
close(module->fd);
|
|
|
|
free(module->path);
|
|
|
|
free(module->name);
|
|
|
|
free(module);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
drgn_debug_info_module_finish_indexing(struct drgn_debug_info *dbinfo,
|
|
|
|
struct drgn_debug_info_module *module)
|
|
|
|
{
|
|
|
|
module->state = DRGN_DEBUG_INFO_MODULE_INDEXED;
|
|
|
|
if (module->name) {
|
|
|
|
int ret = c_string_set_insert(&dbinfo->module_names,
|
|
|
|
(const char **)&module->name,
|
|
|
|
NULL);
|
|
|
|
/* drgn_debug_info_update_index() should've reserved enough. */
|
|
|
|
assert(ret != -1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-06-09 23:11:36 +01:00
|
|
|
/*
|
|
|
|
* Wrapper around dwfl_report_end() that works around a libdwfl bug which causes
|
|
|
|
* it to close stdin when it frees some modules that were reported by
|
|
|
|
* dwfl_core_file_report(). This was fixed in elfutils 0.177 by commit
|
|
|
|
* d37f6ea7e3e5 ("libdwfl: Fix fd leak/closing wrong fd after
|
|
|
|
* dwfl_core_file_report()"), but we support older versions.
|
|
|
|
*/
|
|
|
|
static int my_dwfl_report_end(struct drgn_debug_info *dbinfo,
|
|
|
|
int (*removed)(Dwfl_Module *, void *,
|
|
|
|
const char *, Dwarf_Addr, void *),
|
|
|
|
void *arg)
|
|
|
|
{
|
|
|
|
int fd = -1;
|
|
|
|
if ((dbinfo->prog->flags
|
|
|
|
& (DRGN_PROGRAM_IS_LINUX_KERNEL | DRGN_PROGRAM_IS_LIVE)) == 0)
|
|
|
|
fd = dup(0);
|
|
|
|
int ret = dwfl_report_end(dbinfo->dwfl, removed, arg);
|
|
|
|
if (fd != -1) {
|
|
|
|
dup2(fd, 0);
|
|
|
|
close(fd);
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
struct drgn_dwfl_module_removed_arg {
|
|
|
|
struct drgn_debug_info *dbinfo;
|
|
|
|
bool finish_indexing;
|
|
|
|
bool free_all;
|
|
|
|
};
|
|
|
|
|
|
|
|
static int drgn_dwfl_module_removed(Dwfl_Module *dwfl_module, void *userdatap,
|
|
|
|
const char *name, Dwarf_Addr base,
|
|
|
|
void *_arg)
|
|
|
|
{
|
|
|
|
struct drgn_dwfl_module_removed_arg *arg = _arg;
|
|
|
|
/*
|
|
|
|
* userdatap is actually a void ** like for the other libdwfl callbacks,
|
|
|
|
* but dwfl_report_end() has the wrong signature for the removed
|
|
|
|
* callback.
|
|
|
|
*/
|
|
|
|
struct drgn_debug_info_module *module = *(void **)userdatap;
|
|
|
|
if (arg->finish_indexing && module &&
|
|
|
|
module->state == DRGN_DEBUG_INFO_MODULE_INDEXING)
|
|
|
|
drgn_debug_info_module_finish_indexing(arg->dbinfo, module);
|
|
|
|
if (arg->free_all || !module ||
|
|
|
|
module->state != DRGN_DEBUG_INFO_MODULE_INDEXED) {
|
|
|
|
drgn_debug_info_module_destroy(module);
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* The module was already indexed. Report it again so libdwfl
|
|
|
|
* doesn't remove it.
|
|
|
|
*/
|
|
|
|
Dwarf_Addr end;
|
|
|
|
dwfl_module_info(dwfl_module, NULL, NULL, &end, NULL, NULL,
|
|
|
|
NULL, NULL);
|
|
|
|
dwfl_report_module(arg->dbinfo->dwfl, name, base, end);
|
|
|
|
}
|
|
|
|
return DWARF_CB_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void drgn_debug_info_free_modules(struct drgn_debug_info *dbinfo,
|
|
|
|
bool finish_indexing, bool free_all)
|
|
|
|
{
|
|
|
|
for (struct drgn_debug_info_module_table_iterator it =
|
|
|
|
drgn_debug_info_module_table_first(&dbinfo->modules); it.entry; ) {
|
|
|
|
struct drgn_debug_info_module *module = *it.entry;
|
|
|
|
struct drgn_debug_info_module **nextp = it.entry;
|
|
|
|
do {
|
|
|
|
struct drgn_debug_info_module *next = module->next;
|
|
|
|
if (finish_indexing &&
|
|
|
|
module->state == DRGN_DEBUG_INFO_MODULE_INDEXING) {
|
|
|
|
drgn_debug_info_module_finish_indexing(dbinfo,
|
|
|
|
module);
|
|
|
|
}
|
|
|
|
if (free_all ||
|
|
|
|
module->state != DRGN_DEBUG_INFO_MODULE_INDEXED) {
|
|
|
|
if (module == *nextp) {
|
|
|
|
if (nextp == it.entry && !next) {
|
|
|
|
it = drgn_debug_info_module_table_delete_iterator(&dbinfo->modules,
|
|
|
|
it);
|
|
|
|
} else {
|
|
|
|
if (!next)
|
|
|
|
it = drgn_debug_info_module_table_next(it);
|
|
|
|
*nextp = next;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
void **userdatap;
|
|
|
|
dwfl_module_info(module->dwfl_module,
|
|
|
|
&userdatap, NULL, NULL, NULL,
|
|
|
|
NULL, NULL, NULL);
|
|
|
|
*userdatap = NULL;
|
|
|
|
drgn_debug_info_module_destroy(module);
|
|
|
|
} else {
|
|
|
|
if (!next)
|
|
|
|
it = drgn_debug_info_module_table_next(it);
|
|
|
|
nextp = &module->next;
|
|
|
|
}
|
|
|
|
module = next;
|
|
|
|
} while (module);
|
|
|
|
}
|
|
|
|
|
|
|
|
dwfl_report_begin(dbinfo->dwfl);
|
|
|
|
struct drgn_dwfl_module_removed_arg arg = {
|
|
|
|
.dbinfo = dbinfo,
|
|
|
|
.finish_indexing = finish_indexing,
|
|
|
|
.free_all = free_all,
|
|
|
|
};
|
2021-06-09 23:11:36 +01:00
|
|
|
my_dwfl_report_end(dbinfo, drgn_dwfl_module_removed, &arg);
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_error *
|
|
|
|
drgn_debug_info_report_error(struct drgn_debug_info_load_state *load,
|
|
|
|
const char *name, const char *message,
|
|
|
|
struct drgn_error *err)
|
|
|
|
{
|
|
|
|
if (err && err->code == DRGN_ERROR_NO_MEMORY) {
|
|
|
|
/* Always fail hard if we're out of memory. */
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
if (load->num_errors == 0 &&
|
|
|
|
!string_builder_append(&load->errors,
|
|
|
|
"could not get debugging information for:"))
|
|
|
|
goto err;
|
|
|
|
if (load->num_errors < load->max_errors) {
|
|
|
|
if (!string_builder_line_break(&load->errors))
|
|
|
|
goto err;
|
|
|
|
if (name && !string_builder_append(&load->errors, name))
|
|
|
|
goto err;
|
|
|
|
if (name && (message || err) &&
|
|
|
|
!string_builder_append(&load->errors, " ("))
|
|
|
|
goto err;
|
|
|
|
if (message && !string_builder_append(&load->errors, message))
|
|
|
|
goto err;
|
|
|
|
if (message && err &&
|
|
|
|
!string_builder_append(&load->errors, ": "))
|
|
|
|
goto err;
|
|
|
|
if (err && !string_builder_append_error(&load->errors, err))
|
|
|
|
goto err;
|
|
|
|
if (name && (message || err) &&
|
|
|
|
!string_builder_appendc(&load->errors, ')'))
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
load->num_errors++;
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
err:
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
return &drgn_enomem;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
drgn_debug_info_report_module(struct drgn_debug_info_load_state *load,
|
|
|
|
const void *build_id, size_t build_id_len,
|
|
|
|
uint64_t start, uint64_t end, const char *name,
|
|
|
|
Dwfl_Module *dwfl_module, const char *path,
|
|
|
|
int fd, Elf *elf, bool *new_ret)
|
|
|
|
{
|
|
|
|
struct drgn_debug_info *dbinfo = load->dbinfo;
|
|
|
|
struct drgn_error *err;
|
|
|
|
char *path_key = NULL;
|
|
|
|
|
|
|
|
if (new_ret)
|
|
|
|
*new_ret = false;
|
|
|
|
|
|
|
|
struct hash_pair hp;
|
|
|
|
struct drgn_debug_info_module_table_iterator it;
|
|
|
|
if (build_id_len) {
|
|
|
|
struct drgn_debug_info_module_key key = {
|
|
|
|
.build_id = build_id,
|
|
|
|
.build_id_len = build_id_len,
|
|
|
|
.start = start,
|
|
|
|
.end = end,
|
|
|
|
};
|
2020-10-13 00:06:23 +01:00
|
|
|
hp = drgn_debug_info_module_table_hash(&key);
|
2020-09-16 01:42:53 +01:00
|
|
|
it = drgn_debug_info_module_table_search_hashed(&dbinfo->modules,
|
|
|
|
&key, hp);
|
|
|
|
if (it.entry &&
|
|
|
|
(*it.entry)->state == DRGN_DEBUG_INFO_MODULE_INDEXED) {
|
|
|
|
/* We've already indexed this module. */
|
|
|
|
err = NULL;
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!dwfl_module) {
|
|
|
|
path_key = realpath(path, NULL);
|
|
|
|
if (!path_key) {
|
|
|
|
path_key = strdup(path);
|
|
|
|
if (!path_key) {
|
|
|
|
err = &drgn_enomem;
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
dwfl_module = dwfl_report_module(dbinfo->dwfl, path_key, start,
|
|
|
|
end);
|
|
|
|
if (!dwfl_module) {
|
|
|
|
err = drgn_error_libdwfl();
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void **userdatap;
|
|
|
|
dwfl_module_info(dwfl_module, &userdatap, NULL, NULL, NULL, NULL, NULL,
|
|
|
|
NULL);
|
|
|
|
if (*userdatap) {
|
|
|
|
/* We've already reported this file at this offset. */
|
|
|
|
err = NULL;
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
if (new_ret)
|
|
|
|
*new_ret = true;
|
|
|
|
|
2021-03-16 22:39:37 +00:00
|
|
|
struct drgn_debug_info_module *module = calloc(1, sizeof(*module));
|
2020-09-16 01:42:53 +01:00
|
|
|
if (!module) {
|
|
|
|
err = &drgn_enomem;
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
module->state = DRGN_DEBUG_INFO_MODULE_NEW;
|
|
|
|
module->build_id = build_id;
|
|
|
|
module->build_id_len = build_id_len;
|
|
|
|
module->start = start;
|
|
|
|
module->end = end;
|
|
|
|
if (name) {
|
|
|
|
module->name = strdup(name);
|
|
|
|
if (!module->name) {
|
|
|
|
err = &drgn_enomem;
|
|
|
|
free(module);
|
|
|
|
goto free;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
module->dwfl_module = dwfl_module;
|
|
|
|
module->path = path_key;
|
|
|
|
module->fd = fd;
|
|
|
|
module->elf = elf;
|
|
|
|
|
|
|
|
/* path_key, fd and elf are owned by the module now. */
|
|
|
|
|
|
|
|
if (!drgn_debug_info_module_vector_append(&load->new_modules,
|
|
|
|
&module)) {
|
|
|
|
drgn_debug_info_module_destroy(module);
|
|
|
|
return &drgn_enomem;
|
|
|
|
}
|
|
|
|
if (build_id_len) {
|
|
|
|
if (it.entry) {
|
|
|
|
/*
|
|
|
|
* The first module with this build ID is in
|
|
|
|
* new_modules, so insert it after in the list, not
|
|
|
|
* before.
|
|
|
|
*/
|
|
|
|
module->next = (*it.entry)->next;
|
|
|
|
(*it.entry)->next = module;
|
|
|
|
} else if (drgn_debug_info_module_table_insert_searched(&dbinfo->modules,
|
|
|
|
&module,
|
|
|
|
hp,
|
|
|
|
NULL) < 0) {
|
|
|
|
load->new_modules.size--;
|
|
|
|
drgn_debug_info_module_destroy(module);
|
|
|
|
return &drgn_enomem;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
*userdatap = module;
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
free:
|
|
|
|
elf_end(elf);
|
|
|
|
if (fd != -1)
|
|
|
|
close(fd);
|
|
|
|
free(path_key);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_error *
|
|
|
|
drgn_debug_info_report_elf(struct drgn_debug_info_load_state *load,
|
|
|
|
const char *path, int fd, Elf *elf, uint64_t start,
|
|
|
|
uint64_t end, const char *name, bool *new_ret)
|
|
|
|
{
|
|
|
|
|
|
|
|
struct drgn_error *err;
|
|
|
|
const void *build_id;
|
|
|
|
ssize_t build_id_len = dwelf_elf_gnu_build_id(elf, &build_id);
|
|
|
|
if (build_id_len < 0) {
|
|
|
|
err = drgn_debug_info_report_error(load, path, NULL,
|
2021-12-06 09:49:53 +00:00
|
|
|
drgn_error_libelf());
|
2020-09-16 01:42:53 +01:00
|
|
|
elf_end(elf);
|
2021-12-06 09:49:53 +00:00
|
|
|
close(fd);
|
2020-09-16 01:42:53 +01:00
|
|
|
return err;
|
|
|
|
} else if (build_id_len == 0) {
|
|
|
|
build_id = NULL;
|
|
|
|
}
|
|
|
|
return drgn_debug_info_report_module(load, build_id, build_id_len,
|
|
|
|
start, end, name, NULL, path, fd,
|
|
|
|
elf, new_ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int drgn_debug_info_report_dwfl_module(Dwfl_Module *dwfl_module,
|
|
|
|
void **userdatap,
|
|
|
|
const char *name, Dwarf_Addr base,
|
|
|
|
void *arg)
|
|
|
|
{
|
|
|
|
struct drgn_debug_info_load_state *load = arg;
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
if (*userdatap) {
|
|
|
|
/*
|
|
|
|
* This was either reported from drgn_debug_info_report_elf() or
|
|
|
|
* already indexed.
|
|
|
|
*/
|
|
|
|
return DWARF_CB_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
const unsigned char *build_id;
|
|
|
|
GElf_Addr build_id_vaddr;
|
|
|
|
int build_id_len = dwfl_module_build_id(dwfl_module, &build_id,
|
|
|
|
&build_id_vaddr);
|
|
|
|
if (build_id_len < 0) {
|
|
|
|
err = drgn_debug_info_report_error(load, name, NULL,
|
|
|
|
drgn_error_libdwfl());
|
|
|
|
if (err)
|
|
|
|
goto err;
|
|
|
|
} else if (build_id_len == 0) {
|
|
|
|
build_id = NULL;
|
|
|
|
}
|
|
|
|
Dwarf_Addr end;
|
|
|
|
dwfl_module_info(dwfl_module, NULL, NULL, &end, NULL, NULL, NULL, NULL);
|
|
|
|
err = drgn_debug_info_report_module(load, build_id, build_id_len, base,
|
|
|
|
end, NULL, dwfl_module, name, -1,
|
|
|
|
NULL, NULL);
|
|
|
|
if (err)
|
|
|
|
goto err;
|
|
|
|
return DWARF_CB_OK;
|
|
|
|
|
|
|
|
err:
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
return DWARF_CB_ABORT;
|
|
|
|
}
|
|
|
|
|
2021-12-08 19:48:49 +00:00
|
|
|
static struct drgn_error *drgn_get_nt_file(Elf *elf, const char **ret,
|
|
|
|
size_t *len_ret)
|
|
|
|
{
|
|
|
|
size_t phnum;
|
|
|
|
if (elf_getphdrnum(elf, &phnum) != 0)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
for (size_t i = 0; i < phnum; i++) {
|
|
|
|
GElf_Phdr phdr_mem, *phdr = gelf_getphdr(elf, i, &phdr_mem);
|
|
|
|
if (!phdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
if (phdr->p_type == PT_NOTE) {
|
|
|
|
Elf_Data *data = elf_getdata_rawchunk(elf,
|
|
|
|
phdr->p_offset,
|
|
|
|
phdr->p_filesz,
|
|
|
|
note_header_type(phdr->p_align));
|
|
|
|
if (!data)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
GElf_Nhdr nhdr;
|
|
|
|
size_t offset = 0, name_offset, desc_offset;
|
|
|
|
while (offset < data->d_size &&
|
|
|
|
(offset = gelf_getnote(data, offset, &nhdr,
|
|
|
|
&name_offset,
|
|
|
|
&desc_offset))) {
|
|
|
|
const char *name =
|
|
|
|
(char *)data->d_buf + name_offset;
|
|
|
|
if (nhdr.n_namesz == sizeof("CORE") &&
|
|
|
|
memcmp(name, "CORE", sizeof("CORE")) == 0 &&
|
|
|
|
nhdr.n_type == NT_FILE) {
|
|
|
|
*ret = (char *)data->d_buf + desc_offset;
|
|
|
|
*len_ret = nhdr.n_descsz;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
*ret = NULL;
|
|
|
|
*len_ret = 0;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_mapped_file_segment {
|
|
|
|
uint64_t start;
|
|
|
|
uint64_t end;
|
|
|
|
uint64_t file_offset;
|
|
|
|
};
|
|
|
|
|
|
|
|
DEFINE_VECTOR(drgn_mapped_file_segment_vector, struct drgn_mapped_file_segment)
|
|
|
|
|
|
|
|
DEFINE_HASH_MAP(drgn_mapped_files, const char *,
|
|
|
|
struct drgn_mapped_file_segment_vector, c_string_key_hash_pair,
|
|
|
|
c_string_key_eq)
|
|
|
|
|
|
|
|
struct userspace_core_report_state {
|
|
|
|
struct drgn_mapped_files files;
|
|
|
|
char *phdr_buf;
|
|
|
|
size_t phdr_buf_capacity;
|
|
|
|
char *segment_buf;
|
|
|
|
size_t segment_buf_capacity;
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct drgn_error *parse_nt_file_error(struct binary_buffer *bb,
|
|
|
|
const char *pos,
|
|
|
|
const char *message)
|
|
|
|
{
|
|
|
|
return drgn_error_create(DRGN_ERROR_OTHER, "couldn't parse NT_FILE");
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool
|
|
|
|
drgn_mapped_file_segments_contiguous(const struct drgn_mapped_file_segment *segment1,
|
|
|
|
const struct drgn_mapped_file_segment *segment2)
|
|
|
|
{
|
|
|
|
if (segment1->end != segment2->start)
|
|
|
|
return false;
|
|
|
|
uint64_t size = segment1->end - segment1->start;
|
|
|
|
return segment1->file_offset + size == segment2->file_offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
userspace_core_get_mapped_files(struct drgn_debug_info_load_state *load,
|
|
|
|
struct userspace_core_report_state *core,
|
|
|
|
const char *nt_file, size_t nt_file_len)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
GElf_Ehdr ehdr_mem, *ehdr = gelf_getehdr(load->dbinfo->prog->core,
|
|
|
|
&ehdr_mem);
|
|
|
|
if (!ehdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
bool is_64_bit = ehdr->e_ident[EI_CLASS] == ELFCLASS64;
|
|
|
|
bool little_endian = ehdr->e_ident[EI_DATA] == ELFDATA2LSB;
|
|
|
|
|
|
|
|
struct binary_buffer bb;
|
|
|
|
binary_buffer_init(&bb, nt_file, nt_file_len, little_endian,
|
|
|
|
parse_nt_file_error);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* fs/binfmt_elf.c in the Linux kernel source code documents the format
|
|
|
|
* of NT_FILE as:
|
|
|
|
*
|
|
|
|
* long count -- how many files are mapped
|
|
|
|
* long page_size -- units for file_ofs
|
|
|
|
* array of [COUNT] elements of
|
|
|
|
* long start
|
|
|
|
* long end
|
|
|
|
* long file_ofs
|
|
|
|
* followed by COUNT filenames in ASCII: "FILE1" NUL "FILE2" NUL...
|
|
|
|
*/
|
|
|
|
uint64_t count, page_size;
|
|
|
|
if (is_64_bit) {
|
|
|
|
if ((err = binary_buffer_next_u64(&bb, &count)))
|
|
|
|
return err;
|
|
|
|
if (count > UINT64_MAX / 24)
|
|
|
|
return binary_buffer_error(&bb, "count is too large");
|
|
|
|
if ((err = binary_buffer_next_u64(&bb, &page_size)) ||
|
|
|
|
(err = binary_buffer_skip(&bb, count * 24)))
|
|
|
|
return err;
|
|
|
|
} else {
|
|
|
|
if ((err = binary_buffer_next_u32_into_u64(&bb, &count)))
|
|
|
|
return err;
|
|
|
|
if (count > UINT64_MAX / 12)
|
|
|
|
return binary_buffer_error(&bb, "count is too large");
|
|
|
|
if ((err = binary_buffer_next_u32_into_u64(&bb, &page_size)) ||
|
|
|
|
(err = binary_buffer_skip(&bb, count * 12)))
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (uint64_t i = 0; i < count; i++) {
|
|
|
|
struct drgn_mapped_file_segment segment;
|
|
|
|
if (is_64_bit) {
|
|
|
|
memcpy(&segment, nt_file + 16 + i * 24, 24);
|
|
|
|
if (bb.bswap) {
|
|
|
|
segment.start = bswap_64(segment.start);
|
|
|
|
segment.end = bswap_64(segment.end);
|
|
|
|
segment.file_offset = bswap_64(segment.file_offset);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
struct {
|
|
|
|
uint32_t start;
|
|
|
|
uint32_t end;
|
|
|
|
uint32_t file_offset;
|
|
|
|
} segment32;
|
|
|
|
memcpy(&segment32, nt_file + 8 + i * 12, 12);
|
|
|
|
if (bb.bswap) {
|
|
|
|
segment.start = bswap_32(segment32.start);
|
|
|
|
segment.end = bswap_32(segment32.end);
|
|
|
|
segment.file_offset = bswap_32(segment32.file_offset);
|
|
|
|
} else {
|
|
|
|
segment.start = segment32.start;
|
|
|
|
segment.end = segment32.end;
|
|
|
|
segment.file_offset = segment32.file_offset;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
segment.file_offset *= page_size;
|
|
|
|
|
|
|
|
struct drgn_mapped_files_entry entry = {
|
|
|
|
.key = bb.pos,
|
|
|
|
};
|
|
|
|
if ((err = binary_buffer_skip_string(&bb)))
|
|
|
|
return err;
|
|
|
|
struct drgn_mapped_files_iterator it;
|
|
|
|
int r = drgn_mapped_files_insert(&core->files, &entry, &it);
|
|
|
|
if (r < 0)
|
|
|
|
return &drgn_enomem;
|
|
|
|
if (r == 1)
|
|
|
|
drgn_mapped_file_segment_vector_init(&it.entry->value);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The Linux kernel creates separate entries for contiguous
|
|
|
|
* mappings with different memory protections even though the
|
|
|
|
* protection is not included in NT_FILE. Merge them if we can.
|
|
|
|
*/
|
|
|
|
if (it.entry->value.size > 0 &&
|
|
|
|
drgn_mapped_file_segments_contiguous(&it.entry->value.data[it.entry->value.size - 1],
|
|
|
|
&segment))
|
|
|
|
it.entry->value.data[it.entry->value.size - 1].end = segment.end;
|
|
|
|
else if (!drgn_mapped_file_segment_vector_append(&it.entry->value,
|
|
|
|
&segment))
|
|
|
|
return &drgn_enomem;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool build_id_matches(Elf *elf, const void *build_id,
|
|
|
|
size_t build_id_len)
|
|
|
|
{
|
|
|
|
const void *elf_build_id;
|
|
|
|
ssize_t elf_build_id_len = dwelf_elf_gnu_build_id(elf, &elf_build_id);
|
|
|
|
if (elf_build_id_len < 0)
|
|
|
|
return false;
|
|
|
|
return (elf_build_id_len == build_id_len &&
|
|
|
|
memcmp(elf_build_id, build_id, build_id_len) == 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
userspace_core_elf_address_range(uint16_t e_type, size_t phnum,
|
|
|
|
struct drgn_error *(*get_phdr)(void *, size_t, GElf_Phdr *),
|
|
|
|
void *arg,
|
|
|
|
const struct drgn_mapped_file_segment *segments,
|
|
|
|
size_t num_segments,
|
|
|
|
const struct drgn_mapped_file_segment *ehdr_segment,
|
|
|
|
uint64_t *bias_ret, uint64_t *start_ret,
|
|
|
|
uint64_t *end_ret)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* First, find the virtual address of the ELF header so that we can
|
|
|
|
* calculate the bias.
|
|
|
|
*/
|
|
|
|
uint64_t ehdr_vaddr;
|
|
|
|
size_t i;
|
|
|
|
for (i = 0; i < phnum; i++) {
|
|
|
|
GElf_Phdr phdr;
|
|
|
|
err = get_phdr(arg, i, &phdr);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
if (phdr.p_type == PT_LOAD) {
|
|
|
|
uint64_t align = phdr.p_align ? phdr.p_align : 1;
|
|
|
|
if ((phdr.p_offset & -align) == 0) {
|
|
|
|
ehdr_vaddr = phdr.p_vaddr & -align;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (i >= phnum) {
|
|
|
|
/*
|
|
|
|
* No loadable segments contain the ELF header. This can't be
|
|
|
|
* our file.
|
|
|
|
*/
|
|
|
|
*bias_ret = 0;
|
|
|
|
not_loaded:
|
|
|
|
*start_ret = *end_ret = 0;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
*bias_ret = ehdr_segment->start - ehdr_vaddr;
|
|
|
|
if (*bias_ret != 0 && e_type == ET_EXEC) {
|
|
|
|
/* The executable is not loaded at the correct address. */
|
|
|
|
goto not_loaded;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Now check all of the program headers to (1) get the module address
|
|
|
|
* range and (2) make sure that they are mapped as expected. If we're
|
|
|
|
* lucky, this can detect a file that was mmap'd and not actually loaded
|
|
|
|
* by the kernel or dynamic loader. This could also be the wrong file.
|
|
|
|
*/
|
|
|
|
const struct drgn_mapped_file_segment *segment = segments;
|
|
|
|
const struct drgn_mapped_file_segment *end_segment =
|
|
|
|
segments + num_segments;
|
|
|
|
uint64_t start = 0, end = 0;
|
|
|
|
bool first = true;
|
|
|
|
for (i = 0; i < phnum; i++) {
|
|
|
|
GElf_Phdr phdr;
|
|
|
|
err = get_phdr(arg, i, &phdr);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
if (phdr.p_type != PT_LOAD)
|
|
|
|
continue;
|
|
|
|
uint64_t vaddr = phdr.p_vaddr + *bias_ret;
|
|
|
|
if (phdr.p_filesz != 0) {
|
|
|
|
/*
|
|
|
|
* Advance to the mapped segment containing the start
|
|
|
|
* address.
|
|
|
|
*/
|
|
|
|
while (vaddr >= segment->end) {
|
|
|
|
if (++segment == end_segment)
|
|
|
|
goto not_loaded;
|
|
|
|
if (vaddr < segment->start)
|
|
|
|
goto not_loaded;
|
|
|
|
}
|
|
|
|
if (segment->file_offset + (vaddr - segment->start) !=
|
|
|
|
phdr.p_offset) {
|
|
|
|
/*
|
|
|
|
* The address in the core dump does not map to
|
|
|
|
* the segment's file offset.
|
|
|
|
*/
|
|
|
|
goto not_loaded;
|
|
|
|
}
|
|
|
|
if (phdr.p_filesz > segment->end - vaddr) {
|
|
|
|
/* Part of the segment is not mapped. */
|
|
|
|
goto not_loaded;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (first) {
|
|
|
|
uint64_t align = phdr.p_align ? phdr.p_align : 1;
|
|
|
|
start = vaddr & -align;
|
|
|
|
first = false;
|
|
|
|
}
|
|
|
|
end = vaddr + phdr.p_memsz;
|
|
|
|
}
|
|
|
|
if (start >= end)
|
|
|
|
goto not_loaded;
|
|
|
|
*start_ret = start;
|
|
|
|
*end_ret = end;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool alloc_or_reuse(char **buf, size_t *capacity, uint64_t size)
|
|
|
|
{
|
|
|
|
if (size > *capacity) {
|
|
|
|
if (size > SIZE_MAX)
|
|
|
|
return false;
|
|
|
|
free(*buf);
|
|
|
|
*buf = malloc(size);
|
|
|
|
if (!*buf) {
|
|
|
|
*capacity = 0;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
*capacity = size;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* ehdr_buf must be aligned as Elf64_Ehdr. */
|
|
|
|
static void read_ehdr(const void *ehdr_buf, GElf_Ehdr *ret, bool *is_64_bit_ret,
|
|
|
|
bool *bswap_ret)
|
|
|
|
{
|
|
|
|
*is_64_bit_ret = ((unsigned char *)ehdr_buf)[EI_CLASS] == ELFCLASS64;
|
|
|
|
bool little_endian =
|
|
|
|
((unsigned char *)ehdr_buf)[EI_DATA] == ELFDATA2LSB;
|
|
|
|
*bswap_ret = little_endian != HOST_LITTLE_ENDIAN;
|
|
|
|
if (*is_64_bit_ret) {
|
|
|
|
const Elf64_Ehdr *ehdr64 = ehdr_buf;
|
|
|
|
if (*bswap_ret) {
|
|
|
|
memcpy(ret->e_ident, ehdr64->e_ident, EI_NIDENT);
|
|
|
|
ret->e_type = bswap_16(ehdr64->e_type);
|
|
|
|
ret->e_machine = bswap_16(ehdr64->e_machine);
|
|
|
|
ret->e_version = bswap_32(ehdr64->e_version);
|
|
|
|
ret->e_entry = bswap_64(ehdr64->e_entry);
|
|
|
|
ret->e_phoff = bswap_64(ehdr64->e_phoff);
|
|
|
|
ret->e_shoff = bswap_64(ehdr64->e_shoff);
|
|
|
|
ret->e_flags = bswap_32(ehdr64->e_flags);
|
|
|
|
ret->e_ehsize = bswap_16(ehdr64->e_ehsize);
|
|
|
|
ret->e_phentsize = bswap_16(ehdr64->e_phentsize);
|
|
|
|
ret->e_phnum = bswap_16(ehdr64->e_phnum);
|
|
|
|
ret->e_shentsize = bswap_16(ehdr64->e_shentsize);
|
|
|
|
ret->e_shnum = bswap_16(ehdr64->e_shnum);
|
|
|
|
ret->e_shstrndx = bswap_16(ehdr64->e_shstrndx);
|
|
|
|
} else {
|
|
|
|
*ret = *ehdr64;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
const Elf32_Ehdr *ehdr32 = ehdr_buf;
|
|
|
|
memcpy(ret->e_ident, ehdr32->e_ident, EI_NIDENT);
|
|
|
|
if (*bswap_ret) {
|
|
|
|
ret->e_type = bswap_16(ehdr32->e_type);
|
|
|
|
ret->e_machine = bswap_16(ehdr32->e_machine);
|
|
|
|
ret->e_version = bswap_32(ehdr32->e_version);
|
|
|
|
ret->e_entry = bswap_32(ehdr32->e_entry);
|
|
|
|
ret->e_phoff = bswap_32(ehdr32->e_phoff);
|
|
|
|
ret->e_shoff = bswap_32(ehdr32->e_shoff);
|
|
|
|
ret->e_flags = bswap_32(ehdr32->e_flags);
|
|
|
|
ret->e_ehsize = bswap_16(ehdr32->e_ehsize);
|
|
|
|
ret->e_phentsize = bswap_16(ehdr32->e_phentsize);
|
|
|
|
ret->e_phnum = bswap_16(ehdr32->e_phnum);
|
|
|
|
ret->e_shentsize = bswap_16(ehdr32->e_shentsize);
|
|
|
|
ret->e_shnum = bswap_16(ehdr32->e_shnum);
|
|
|
|
ret->e_shstrndx = bswap_16(ehdr32->e_shstrndx);
|
|
|
|
} else {
|
|
|
|
ret->e_type = ehdr32->e_type;
|
|
|
|
ret->e_machine = ehdr32->e_machine;
|
|
|
|
ret->e_version = ehdr32->e_version;
|
|
|
|
ret->e_entry = ehdr32->e_entry;
|
|
|
|
ret->e_phoff = ehdr32->e_phoff;
|
|
|
|
ret->e_shoff = ehdr32->e_shoff;
|
|
|
|
ret->e_flags = ehdr32->e_flags;
|
|
|
|
ret->e_ehsize = ehdr32->e_ehsize;
|
|
|
|
ret->e_phentsize = ehdr32->e_phentsize;
|
|
|
|
ret->e_phnum = ehdr32->e_phnum;
|
|
|
|
ret->e_shentsize = ehdr32->e_shentsize;
|
|
|
|
ret->e_shnum = ehdr32->e_shnum;
|
|
|
|
ret->e_shstrndx = ehdr32->e_shstrndx;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* phdr_buf must be aligned as Elf64_Phdr. */
|
|
|
|
static void read_phdr(const void *phdr_buf, size_t i, bool is_64_bit,
|
|
|
|
bool bswap, GElf_Phdr *ret)
|
|
|
|
{
|
|
|
|
if (is_64_bit) {
|
|
|
|
const Elf64_Phdr *phdr64 = (Elf64_Phdr *)phdr_buf + i;
|
|
|
|
if (bswap) {
|
|
|
|
ret->p_type = bswap_32(phdr64->p_type);
|
|
|
|
ret->p_flags = bswap_32(phdr64->p_flags);
|
|
|
|
ret->p_offset = bswap_64(phdr64->p_offset);
|
|
|
|
ret->p_vaddr = bswap_64(phdr64->p_vaddr);
|
|
|
|
ret->p_paddr = bswap_64(phdr64->p_paddr);
|
|
|
|
ret->p_filesz = bswap_64(phdr64->p_filesz);
|
|
|
|
ret->p_memsz = bswap_64(phdr64->p_memsz);
|
|
|
|
ret->p_align = bswap_64(phdr64->p_align);
|
|
|
|
} else {
|
|
|
|
*ret = *phdr64;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
const Elf32_Phdr *phdr32 = (Elf32_Phdr *)phdr_buf + i;
|
|
|
|
if (bswap) {
|
|
|
|
ret->p_type = bswap_32(phdr32->p_type);
|
|
|
|
ret->p_offset = bswap_32(phdr32->p_offset);
|
|
|
|
ret->p_vaddr = bswap_32(phdr32->p_vaddr);
|
|
|
|
ret->p_paddr = bswap_32(phdr32->p_paddr);
|
|
|
|
ret->p_filesz = bswap_32(phdr32->p_filesz);
|
|
|
|
ret->p_memsz = bswap_32(phdr32->p_memsz);
|
|
|
|
ret->p_flags = bswap_32(phdr32->p_flags);
|
|
|
|
ret->p_align = bswap_32(phdr32->p_align);
|
|
|
|
} else {
|
|
|
|
ret->p_type = phdr32->p_type;
|
|
|
|
ret->p_offset = phdr32->p_offset;
|
|
|
|
ret->p_vaddr = phdr32->p_vaddr;
|
|
|
|
ret->p_paddr = phdr32->p_paddr;
|
|
|
|
ret->p_filesz = phdr32->p_filesz;
|
|
|
|
ret->p_memsz = phdr32->p_memsz;
|
|
|
|
ret->p_flags = phdr32->p_flags;
|
|
|
|
ret->p_align = phdr32->p_align;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static const char *read_build_id(const char *buf, size_t buf_len,
|
|
|
|
uint64_t align, bool bswap,
|
|
|
|
size_t *len_ret)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Build IDs are usually 16 or 20 bytes (MD5 or SHA-1, respectively), so
|
|
|
|
* these arbitrary limits are generous.
|
|
|
|
*/
|
|
|
|
static const uint32_t build_id_min_size = 2;
|
|
|
|
static const uint32_t build_id_max_size = 1024;
|
|
|
|
/* Elf32_Nhdr is the same as Elf64_Nhdr. */
|
|
|
|
Elf64_Nhdr nhdr;
|
|
|
|
const char *p = buf;
|
|
|
|
while (buf + buf_len - p >= sizeof(nhdr)) {
|
|
|
|
memcpy(&nhdr, p, sizeof(nhdr));
|
|
|
|
if (bswap) {
|
|
|
|
nhdr.n_namesz = bswap_32(nhdr.n_namesz);
|
|
|
|
nhdr.n_descsz = bswap_32(nhdr.n_descsz);
|
|
|
|
nhdr.n_type = bswap_32(nhdr.n_type);
|
|
|
|
}
|
|
|
|
p += sizeof(nhdr);
|
|
|
|
|
|
|
|
uint64_t namesz = (nhdr.n_namesz + align - 1) & ~(align - 1);
|
|
|
|
if (namesz > buf + buf_len - p)
|
|
|
|
return NULL;
|
|
|
|
const char *name = p;
|
|
|
|
p += namesz;
|
|
|
|
|
|
|
|
if (nhdr.n_namesz == sizeof("GNU") &&
|
|
|
|
memcmp(name, "GNU", sizeof("GNU")) == 0 &&
|
|
|
|
nhdr.n_type == NT_GNU_BUILD_ID &&
|
|
|
|
nhdr.n_descsz >= build_id_min_size &&
|
|
|
|
nhdr.n_descsz <= build_id_max_size) {
|
|
|
|
if (nhdr.n_descsz > buf + buf_len - p)
|
|
|
|
return NULL;
|
|
|
|
*len_ret = nhdr.n_descsz;
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t descsz = (nhdr.n_descsz + align - 1) & ~(align - 1);
|
|
|
|
if (descsz > buf + buf_len - p)
|
|
|
|
return NULL;
|
|
|
|
p += descsz;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct core_get_phdr_arg {
|
|
|
|
const void *phdr_buf;
|
|
|
|
bool is_64_bit;
|
|
|
|
bool bswap;
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
core_get_phdr(void *arg_, size_t i, GElf_Phdr *ret)
|
|
|
|
{
|
|
|
|
struct core_get_phdr_arg *arg = arg_;
|
|
|
|
read_phdr(arg->phdr_buf, i, arg->is_64_bit, arg->bswap, ret);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct userspace_core_identified_file {
|
|
|
|
const void *build_id;
|
|
|
|
size_t build_id_len;
|
|
|
|
uint64_t start, end;
|
|
|
|
bool ignore;
|
|
|
|
bool have_address_range;
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
userspace_core_identify_file(struct drgn_program *prog,
|
|
|
|
struct userspace_core_report_state *core,
|
|
|
|
const struct drgn_mapped_file_segment *segments,
|
|
|
|
size_t num_segments,
|
|
|
|
const struct drgn_mapped_file_segment *ehdr_segment,
|
|
|
|
struct userspace_core_identified_file *ret)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
Elf64_Ehdr ehdr_buf;
|
|
|
|
err = drgn_program_read_memory(prog, &ehdr_buf, ehdr_segment->start,
|
|
|
|
sizeof(ehdr_buf), false);
|
|
|
|
if (err) {
|
|
|
|
if (err->code == DRGN_ERROR_FAULT) {
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
err = NULL;
|
|
|
|
}
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
if (memcmp(&ehdr_buf, ELFMAG, SELFMAG) != 0) {
|
|
|
|
ret->ignore = true;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
GElf_Ehdr ehdr;
|
|
|
|
struct core_get_phdr_arg arg;
|
|
|
|
read_ehdr(&ehdr_buf, &ehdr, &arg.is_64_bit, &arg.bswap);
|
|
|
|
if (ehdr.e_type == ET_CORE ||
|
|
|
|
ehdr.e_phnum == 0 ||
|
|
|
|
ehdr.e_phentsize !=
|
|
|
|
(arg.is_64_bit ? sizeof(Elf64_Phdr) : sizeof(Elf32_Phdr))) {
|
|
|
|
ret->ignore = true;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ehdr.e_phnum > SIZE_MAX / ehdr.e_phentsize ||
|
|
|
|
!alloc_or_reuse(&core->phdr_buf, &core->phdr_buf_capacity,
|
|
|
|
ehdr.e_phnum * ehdr.e_phentsize))
|
|
|
|
return &drgn_enomem;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check whether the mapped segment containing the file header also
|
|
|
|
* contains the program headers. This seems to be the case in practice.
|
|
|
|
*/
|
|
|
|
uint64_t ehdr_segment_file_end =
|
|
|
|
(ehdr_segment->file_offset +
|
|
|
|
(ehdr_segment->end - ehdr_segment->start));
|
|
|
|
if (ehdr_segment_file_end < ehdr.e_phoff ||
|
|
|
|
ehdr_segment_file_end - ehdr.e_phoff <
|
|
|
|
ehdr.e_phnum * ehdr.e_phentsize)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
err = drgn_program_read_memory(prog, core->phdr_buf,
|
|
|
|
ehdr_segment->start + ehdr.e_phoff,
|
|
|
|
ehdr.e_phnum * ehdr.e_phentsize, false);
|
|
|
|
if (err) {
|
|
|
|
if (err->code == DRGN_ERROR_FAULT) {
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
err = NULL;
|
|
|
|
}
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
arg.phdr_buf = core->phdr_buf;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* In theory, if the program has a huge number of program headers, they
|
|
|
|
* may not all be dumped. However, the largest binary I was able to find
|
|
|
|
* still had all program headers within 1k.
|
|
|
|
*
|
|
|
|
* It'd be more reliable to determine the bias based on the headers that
|
|
|
|
* were saved, use that to read the build ID, use that to find the ELF
|
|
|
|
* file, and then determine the address range directly from the ELF
|
|
|
|
* file. However, we need the address range to report the build ID to
|
|
|
|
* libdwfl, so we do it this way.
|
|
|
|
*/
|
|
|
|
uint64_t bias;
|
|
|
|
err = userspace_core_elf_address_range(ehdr.e_type, ehdr.e_phnum,
|
|
|
|
core_get_phdr, &arg, segments,
|
|
|
|
num_segments, ehdr_segment,
|
|
|
|
&bias, &ret->start, &ret->end);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
if (ret->start >= ret->end) {
|
|
|
|
ret->ignore = true;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
ret->have_address_range = true;
|
|
|
|
|
|
|
|
for (uint16_t i = 0; i < ehdr.e_phnum; i++) {
|
|
|
|
GElf_Phdr phdr;
|
|
|
|
core_get_phdr(&arg, i, &phdr);
|
|
|
|
if (phdr.p_type == PT_NOTE) {
|
|
|
|
if (!alloc_or_reuse(&core->segment_buf,
|
|
|
|
&core->segment_buf_capacity,
|
|
|
|
phdr.p_filesz))
|
|
|
|
return &drgn_enomem;
|
|
|
|
err = drgn_program_read_memory(prog, core->segment_buf,
|
|
|
|
phdr.p_vaddr + bias,
|
|
|
|
phdr.p_filesz, false);
|
|
|
|
if (err) {
|
|
|
|
if (err->code == DRGN_ERROR_FAULT) {
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
continue;
|
|
|
|
} else {
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ret->build_id = read_build_id(core->segment_buf,
|
|
|
|
phdr.p_filesz,
|
|
|
|
phdr.p_align, arg.bswap,
|
|
|
|
&ret->build_id_len);
|
|
|
|
if (ret->build_id)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *elf_file_get_phdr(void *arg, size_t i,
|
|
|
|
GElf_Phdr *phdr)
|
|
|
|
{
|
|
|
|
if (!gelf_getphdr(arg, i, phdr))
|
|
|
|
return drgn_error_libelf();
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
userspace_core_maybe_report_file(struct drgn_debug_info_load_state *load,
|
|
|
|
struct userspace_core_report_state *core,
|
|
|
|
const char *path,
|
|
|
|
const struct drgn_mapped_file_segment *segments,
|
|
|
|
size_t num_segments)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
struct drgn_program *prog = load->dbinfo->prog;
|
|
|
|
for (size_t ehdr_idx = 0; ehdr_idx < num_segments; ehdr_idx++) {
|
|
|
|
const struct drgn_mapped_file_segment *ehdr_segment =
|
|
|
|
&segments[ehdr_idx];
|
|
|
|
/*
|
|
|
|
* There should always be a full page mapped, so even if it's a
|
|
|
|
* 32-bit file, we can read the 64-bit size.
|
|
|
|
*/
|
|
|
|
if (ehdr_segment->file_offset != 0 ||
|
|
|
|
ehdr_segment->end - ehdr_segment->start < sizeof(Elf64_Ehdr))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This logic is complicated because we're dealing with two data
|
|
|
|
* sources that we can't completely trust: the memory in the
|
|
|
|
* core dump and the file at the path found in the core dump.
|
|
|
|
*
|
|
|
|
* First, we try to identify the mapped file contents in the
|
|
|
|
* core dump. Ideally, this will find a build ID. However, this
|
|
|
|
* can fail for a few reasons:
|
|
|
|
*
|
|
|
|
* 1. The file is not an ELF file.
|
|
|
|
* 2. The ELF file is not an executable or library.
|
|
|
|
* 3. The ELF file does not have a build ID.
|
|
|
|
* 4. The file header was not dumped to the core dump, in which
|
|
|
|
* case we can't tell whether this is an ELF file. Dumping
|
|
|
|
* the first page of an executable file has been the default
|
|
|
|
* behavior since Linux kernel commit 895021552d6f
|
|
|
|
* ("coredump: default
|
|
|
|
* CONFIG_CORE_DUMP_DEFAULT_ELF_HEADERS=y") (in v2.6.37), but
|
|
|
|
* it can be disabled at kernel build time or toggled at
|
|
|
|
* runtime.
|
|
|
|
* 5. The build ID or the necessary ELF metadata were not dumped
|
|
|
|
* in the core dump. This can happen if the necessary program
|
|
|
|
* headers or note segment were not in the first page of the
|
|
|
|
* file.
|
|
|
|
* 6. The file is mapped but not actually loaded into the
|
|
|
|
* program (e.g., if the program is a tool like a profiler or
|
|
|
|
* a debugger that mmaps binaries [like drgn itself!]).
|
|
|
|
*
|
|
|
|
* In cases 1 and 2, we can simply ignore the file. In cases
|
|
|
|
* 3-5, we blindly trust the path in the core dump. We can
|
|
|
|
* sometimes detect case 6 in
|
|
|
|
* userspace_core_elf_address_range().
|
|
|
|
*
|
|
|
|
* There is also the possibility that the program modified or
|
|
|
|
* corrupted the ELF metadata in memory (more likely if the file
|
|
|
|
* was explicitly mmap'd, since the metadata will usually be
|
|
|
|
* read-only if it was loaded properly). We don't deal with that
|
|
|
|
* yet.
|
|
|
|
*/
|
|
|
|
struct userspace_core_identified_file identity = {};
|
|
|
|
err = userspace_core_identify_file(prog, core, segments,
|
|
|
|
num_segments, ehdr_segment,
|
|
|
|
&identity);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
if (identity.ignore)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
#define CLEAR_ELF() do { \
|
|
|
|
elf = NULL; \
|
|
|
|
fd = -1; \
|
|
|
|
} while (0)
|
|
|
|
#define CLOSE_ELF() do { \
|
|
|
|
elf_end(elf); \
|
|
|
|
close(fd); \
|
|
|
|
CLEAR_ELF(); \
|
|
|
|
} while (0)
|
|
|
|
int fd;
|
|
|
|
Elf *elf;
|
|
|
|
/*
|
|
|
|
* There are a few things that can go wrong here:
|
|
|
|
*
|
|
|
|
* 1. The path no longer exists.
|
|
|
|
* 2. The path refers to a different ELF file than was in the
|
|
|
|
* core dump.
|
|
|
|
* 3. The path refers to something which isn't a valid ELF file.
|
|
|
|
*/
|
|
|
|
err = open_elf_file(path, &fd, &elf);
|
|
|
|
if (err) {
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
CLEAR_ELF();
|
|
|
|
} else if (identity.build_id_len > 0) {
|
|
|
|
if (!build_id_matches(elf, identity.build_id,
|
|
|
|
identity.build_id_len))
|
|
|
|
CLOSE_ELF();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (elf && !identity.have_address_range) {
|
|
|
|
GElf_Ehdr ehdr_mem, *ehdr;
|
|
|
|
size_t phnum;
|
|
|
|
if ((ehdr = gelf_getehdr(elf, &ehdr_mem)) &&
|
|
|
|
(elf_getphdrnum(elf, &phnum) == 0)) {
|
|
|
|
uint64_t bias;
|
|
|
|
err = userspace_core_elf_address_range(ehdr->e_type,
|
|
|
|
phnum,
|
|
|
|
elf_file_get_phdr,
|
|
|
|
elf,
|
|
|
|
segments,
|
|
|
|
num_segments,
|
|
|
|
ehdr_segment,
|
|
|
|
&bias,
|
|
|
|
&identity.start,
|
|
|
|
&identity.end);
|
|
|
|
if (err || identity.start >= identity.end) {
|
|
|
|
drgn_error_destroy(err);
|
|
|
|
CLOSE_ELF();
|
|
|
|
} else {
|
|
|
|
identity.have_address_range = true;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
CLOSE_ELF();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (elf) {
|
|
|
|
err = drgn_debug_info_report_elf(load, path, fd, elf,
|
|
|
|
identity.start,
|
|
|
|
identity.end, NULL,
|
|
|
|
NULL);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
} else {
|
|
|
|
if (!identity.have_address_range)
|
|
|
|
identity.start = identity.end = 0;
|
|
|
|
Dwfl_Module *dwfl_module =
|
|
|
|
dwfl_report_module(load->dbinfo->dwfl, path,
|
|
|
|
identity.start,
|
|
|
|
identity.end);
|
|
|
|
if (!dwfl_module)
|
|
|
|
return drgn_error_libdwfl();
|
|
|
|
if (identity.build_id_len > 0 &&
|
|
|
|
dwfl_module_report_build_id(dwfl_module,
|
|
|
|
identity.build_id,
|
|
|
|
identity.build_id_len,
|
|
|
|
0))
|
|
|
|
return drgn_error_libdwfl();
|
|
|
|
}
|
|
|
|
#undef CLOSE_ELF
|
|
|
|
#undef CLEAR_ELF
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
userspace_core_report_mapped_files(struct drgn_debug_info_load_state *load,
|
|
|
|
struct userspace_core_report_state *core)
|
|
|
|
{
|
|
|
|
|
|
|
|
struct drgn_error *err;
|
|
|
|
for (struct drgn_mapped_files_iterator it =
|
|
|
|
drgn_mapped_files_first(&core->files);
|
|
|
|
it.entry; it = drgn_mapped_files_next(it)) {
|
|
|
|
err = userspace_core_maybe_report_file(load, core,
|
|
|
|
it.entry->key,
|
|
|
|
it.entry->value.data,
|
|
|
|
it.entry->value.size);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
userspace_core_report_debug_info(struct drgn_debug_info_load_state *load,
|
|
|
|
const char *nt_file, size_t nt_file_len)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
struct userspace_core_report_state core = {
|
|
|
|
.files = HASH_TABLE_INIT,
|
|
|
|
};
|
|
|
|
err = userspace_core_get_mapped_files(load, &core, nt_file,
|
|
|
|
nt_file_len);
|
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
err = userspace_core_report_mapped_files(load, &core);
|
|
|
|
out:
|
|
|
|
free(core.segment_buf);
|
|
|
|
free(core.phdr_buf);
|
|
|
|
for (struct drgn_mapped_files_iterator it =
|
|
|
|
drgn_mapped_files_first(&core.files);
|
|
|
|
it.entry; it = drgn_mapped_files_next(it))
|
|
|
|
drgn_mapped_file_segment_vector_deinit(&it.entry->value);
|
|
|
|
drgn_mapped_files_deinit(&core.files);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
libdrgn: debug_info: get address range of reported ET_EXEC files
When explicitly reporting a debugging information file for a userspace
program, userspace_report_debug_info() currently always reports it with
a load address range of [0, 0) (i.e., not actually loaded into the
program). This is because for ET_DYN and ET_REL files, we have to
determine the address range by inspecting the core dump or program
state, which is a bit involved.
However, ET_EXEC is much easier: we can get the address range from the
segment headers. In fact, we already implemented this for vmlinux files,
so we can reuse that with a modification to make it more permissive.
ET_CORE debug info files don't make much sense, but libdwfl seems to
treat a reported ET_CORE file the same as ET_EXEC (see
dwfl_report_elf()), so we do, too.
Unfortunately, most executables on modern Linux distributions are
ET_DYN, but this will at least make testing easier.
Signed-off-by: Omar Sandoval <osandov@osandov.com>
2021-11-19 22:19:32 +00:00
|
|
|
static struct drgn_error *
|
|
|
|
userspace_report_elf_file(struct drgn_debug_info_load_state *load,
|
|
|
|
const char *path)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
int fd;
|
|
|
|
Elf *elf;
|
|
|
|
err = open_elf_file(path, &fd, &elf);
|
|
|
|
if (err)
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
GElf_Ehdr ehdr_mem, *ehdr;
|
|
|
|
ehdr = gelf_getehdr(elf, &ehdr_mem);
|
|
|
|
if (!ehdr) {
|
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto err_close;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* We haven't implemented a way to get the load address for dynamically
|
|
|
|
* loaded or relocatable files, so for now we report those as unloaded.
|
|
|
|
*/
|
|
|
|
uint64_t start = 0, end = 0;
|
|
|
|
if (ehdr->e_type == ET_EXEC || ehdr->e_type == ET_CORE) {
|
|
|
|
err = elf_address_range(elf, 0, &start, &end);
|
|
|
|
if (err)
|
|
|
|
goto err_close;
|
|
|
|
}
|
|
|
|
|
|
|
|
return drgn_debug_info_report_elf(load, path, fd, elf, start, end, NULL,
|
|
|
|
NULL);
|
|
|
|
|
|
|
|
err_close:
|
|
|
|
elf_end(elf);
|
|
|
|
close(fd);
|
|
|
|
err:
|
|
|
|
return drgn_debug_info_report_error(load, path, NULL, err);
|
|
|
|
}
|
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
static struct drgn_error *
|
|
|
|
userspace_report_debug_info(struct drgn_debug_info_load_state *load)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
for (size_t i = 0; i < load->num_paths; i++) {
|
libdrgn: debug_info: get address range of reported ET_EXEC files
When explicitly reporting a debugging information file for a userspace
program, userspace_report_debug_info() currently always reports it with
a load address range of [0, 0) (i.e., not actually loaded into the
program). This is because for ET_DYN and ET_REL files, we have to
determine the address range by inspecting the core dump or program
state, which is a bit involved.
However, ET_EXEC is much easier: we can get the address range from the
segment headers. In fact, we already implemented this for vmlinux files,
so we can reuse that with a modification to make it more permissive.
ET_CORE debug info files don't make much sense, but libdwfl seems to
treat a reported ET_CORE file the same as ET_EXEC (see
dwfl_report_elf()), so we do, too.
Unfortunately, most executables on modern Linux distributions are
ET_DYN, but this will at least make testing easier.
Signed-off-by: Omar Sandoval <osandov@osandov.com>
2021-11-19 22:19:32 +00:00
|
|
|
err = userspace_report_elf_file(load, load->paths[i]);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (load->load_default) {
|
|
|
|
Dwfl *dwfl = load->dbinfo->dwfl;
|
|
|
|
struct drgn_program *prog = load->dbinfo->prog;
|
|
|
|
if (prog->flags & DRGN_PROGRAM_IS_LIVE) {
|
|
|
|
int ret = dwfl_linux_proc_report(dwfl, prog->pid);
|
|
|
|
if (ret == -1) {
|
|
|
|
return drgn_error_libdwfl();
|
|
|
|
} else if (ret) {
|
|
|
|
return drgn_error_create_os("dwfl_linux_proc_report",
|
|
|
|
ret, NULL);
|
|
|
|
}
|
2021-12-08 19:48:49 +00:00
|
|
|
} else {
|
|
|
|
const char *nt_file;
|
|
|
|
size_t nt_file_len;
|
|
|
|
char *env = getenv("DRGN_USE_LIBDWFL_REPORT");
|
|
|
|
if (env && atoi(env)) {
|
|
|
|
nt_file = NULL;
|
|
|
|
nt_file_len = 0;
|
|
|
|
} else {
|
|
|
|
err = drgn_get_nt_file(prog->core, &nt_file,
|
|
|
|
&nt_file_len);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
if (nt_file) {
|
|
|
|
err = userspace_core_report_debug_info(load,
|
|
|
|
nt_file,
|
|
|
|
nt_file_len);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
} else if (dwfl_core_file_report(dwfl, prog->core,
|
|
|
|
NULL) == -1) {
|
|
|
|
return drgn_error_libdwfl();
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
static struct drgn_error *relocate_elf_section(Elf_Scn *scn, Elf_Scn *reloc_scn,
|
|
|
|
Elf_Scn *symtab_scn,
|
|
|
|
const uint64_t *sh_addrs,
|
|
|
|
size_t shdrnum,
|
|
|
|
const struct drgn_platform *platform)
|
2020-09-16 01:42:53 +01:00
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
bool is_64_bit = drgn_platform_is_64_bit(platform);
|
|
|
|
bool bswap = drgn_platform_bswap(platform);
|
|
|
|
apply_elf_rela_fn *apply_elf_rela = platform->arch->apply_elf_rela;
|
|
|
|
|
|
|
|
Elf_Data *data, *reloc_data, *symtab_data;
|
2020-09-16 01:42:53 +01:00
|
|
|
err = read_elf_section(scn, &data);
|
|
|
|
if (err)
|
|
|
|
return err;
|
2021-03-26 00:26:06 +00:00
|
|
|
|
|
|
|
struct drgn_relocating_section relocating = {
|
|
|
|
.buf = data->d_buf,
|
|
|
|
.buf_size = data->d_size,
|
2021-03-26 17:28:10 +00:00
|
|
|
.addr = sh_addrs[elf_ndxscn(scn)],
|
2021-03-26 00:26:06 +00:00
|
|
|
.bswap = bswap,
|
|
|
|
};
|
|
|
|
|
|
|
|
err = read_elf_section(reloc_scn, &reloc_data);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (err)
|
|
|
|
return err;
|
2021-03-26 00:26:06 +00:00
|
|
|
const void *relocs = reloc_data->d_buf;
|
|
|
|
size_t reloc_size = is_64_bit ? sizeof(Elf64_Rela) : sizeof(Elf32_Rela);
|
|
|
|
size_t num_relocs = reloc_data->d_size / reloc_size;
|
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
err = read_elf_section(symtab_scn, &symtab_data);
|
|
|
|
if (err)
|
|
|
|
return err;
|
2021-03-26 00:26:06 +00:00
|
|
|
const void *syms = symtab_data->d_buf;
|
|
|
|
size_t sym_size = is_64_bit ? sizeof(Elf64_Sym) : sizeof(Elf32_Sym);
|
|
|
|
size_t num_syms = symtab_data->d_size / sym_size;
|
|
|
|
|
|
|
|
for (size_t i = 0; i < num_relocs; i++) {
|
|
|
|
uint64_t r_offset;
|
|
|
|
uint32_t r_sym;
|
|
|
|
uint32_t r_type;
|
|
|
|
int64_t r_addend;
|
|
|
|
if (is_64_bit) {
|
|
|
|
Elf64_Rela *rela = (Elf64_Rela *)relocs + i;
|
|
|
|
uint64_t r_info;
|
|
|
|
memcpy(&r_offset, &rela->r_offset, sizeof(r_offset));
|
|
|
|
memcpy(&r_info, &rela->r_info, sizeof(r_info));
|
|
|
|
memcpy(&r_addend, &rela->r_addend, sizeof(r_addend));
|
|
|
|
if (bswap) {
|
|
|
|
r_offset = bswap_64(r_offset);
|
|
|
|
r_info = bswap_64(r_info);
|
|
|
|
r_addend = bswap_64(r_addend);
|
|
|
|
}
|
|
|
|
r_sym = ELF64_R_SYM(r_info);
|
|
|
|
r_type = ELF64_R_TYPE(r_info);
|
|
|
|
} else {
|
|
|
|
Elf32_Rela *rela32 = (Elf32_Rela *)relocs + i;
|
|
|
|
uint32_t r_offset32;
|
|
|
|
uint32_t r_info32;
|
|
|
|
int32_t r_addend32;
|
|
|
|
memcpy(&r_offset32, &rela32->r_offset, sizeof(r_offset32));
|
|
|
|
memcpy(&r_info32, &rela32->r_info, sizeof(r_info32));
|
|
|
|
memcpy(&r_addend32, &rela32->r_addend, sizeof(r_addend32));
|
|
|
|
if (bswap) {
|
|
|
|
r_offset32 = bswap_32(r_offset32);
|
|
|
|
r_info32 = bswap_32(r_info32);
|
|
|
|
r_addend32 = bswap_32(r_addend32);
|
|
|
|
}
|
|
|
|
r_offset = r_offset32;
|
|
|
|
r_sym = ELF32_R_SYM(r_info32);
|
|
|
|
r_type = ELF32_R_TYPE(r_info32);
|
|
|
|
r_addend = r_addend32;
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
if (r_sym >= num_syms) {
|
|
|
|
return drgn_error_create(DRGN_ERROR_OTHER,
|
2021-03-26 00:26:06 +00:00
|
|
|
"invalid ELF relocation symbol");
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
2021-03-26 00:26:06 +00:00
|
|
|
uint16_t st_shndx;
|
|
|
|
uint64_t st_value;
|
|
|
|
if (is_64_bit) {
|
|
|
|
const Elf64_Sym *sym = (Elf64_Sym *)syms + r_sym;
|
|
|
|
memcpy(&st_shndx, &sym->st_shndx, sizeof(st_shndx));
|
|
|
|
memcpy(&st_value, &sym->st_value, sizeof(st_value));
|
|
|
|
if (bswap) {
|
|
|
|
st_shndx = bswap_16(st_shndx);
|
|
|
|
st_value = bswap_64(st_value);
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
} else {
|
2021-03-26 00:26:06 +00:00
|
|
|
const Elf32_Sym *sym = (Elf32_Sym *)syms + r_sym;
|
|
|
|
memcpy(&st_shndx, &sym->st_shndx, sizeof(st_shndx));
|
|
|
|
uint32_t st_value32;
|
|
|
|
memcpy(&st_value32, &sym->st_value, sizeof(st_value32));
|
|
|
|
if (bswap) {
|
|
|
|
st_shndx = bswap_16(st_shndx);
|
|
|
|
st_value32 = bswap_32(st_value32);
|
|
|
|
}
|
|
|
|
st_value = st_value32;
|
|
|
|
}
|
|
|
|
if (st_shndx >= shdrnum) {
|
2020-09-16 01:42:53 +01:00
|
|
|
return drgn_error_create(DRGN_ERROR_OTHER,
|
2021-03-26 00:26:06 +00:00
|
|
|
"invalid ELF symbol section index");
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
2021-03-26 00:26:06 +00:00
|
|
|
|
|
|
|
err = apply_elf_rela(&relocating, r_offset, r_type, r_addend,
|
|
|
|
sh_addrs[st_shndx] + st_value);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Mark the relocation section as empty so that libdwfl doesn't try to
|
|
|
|
* apply it again.
|
|
|
|
*/
|
2021-03-26 00:26:06 +00:00
|
|
|
GElf_Shdr *shdr, shdr_mem;
|
|
|
|
shdr = gelf_getshdr(reloc_scn, &shdr_mem);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (!shdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
shdr->sh_size = 0;
|
2021-03-26 00:26:06 +00:00
|
|
|
if (!gelf_update_shdr(reloc_scn, shdr))
|
2020-09-16 01:42:53 +01:00
|
|
|
return drgn_error_libelf();
|
2021-03-26 00:26:06 +00:00
|
|
|
reloc_data->d_size = 0;
|
2020-09-16 01:42:53 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Before the debugging information in a relocatable ELF file (e.g., Linux
|
|
|
|
* kernel module) can be used, it must have ELF relocations applied. This is
|
|
|
|
* usually done by libdwfl. However, libdwfl is relatively slow at it. This is a
|
2021-03-26 00:26:06 +00:00
|
|
|
* much faster implementation.
|
2020-09-16 01:42:53 +01:00
|
|
|
*/
|
2021-03-26 00:26:06 +00:00
|
|
|
static struct drgn_error *relocate_elf_file(Elf *elf)
|
2020-09-16 01:42:53 +01:00
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
GElf_Ehdr ehdr_mem, *ehdr;
|
2020-09-16 01:42:53 +01:00
|
|
|
ehdr = gelf_getehdr(elf, &ehdr_mem);
|
|
|
|
if (!ehdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
if (ehdr->e_type != ET_REL) {
|
|
|
|
/* Not a relocatable file. */
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_platform platform;
|
|
|
|
drgn_platform_from_elf(ehdr, &platform);
|
|
|
|
if (!platform.arch->apply_elf_rela) {
|
2020-09-16 01:42:53 +01:00
|
|
|
/* Unsupported; fall back to libdwfl. */
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
size_t shdrnum;
|
2020-09-16 01:42:53 +01:00
|
|
|
if (elf_getshdrnum(elf, &shdrnum))
|
|
|
|
return drgn_error_libelf();
|
2021-03-26 00:26:06 +00:00
|
|
|
uint64_t *sh_addrs = calloc(shdrnum, sizeof(sh_addrs[0]));
|
|
|
|
if (!sh_addrs && shdrnum > 0)
|
|
|
|
return &drgn_enomem;
|
2020-09-16 01:42:53 +01:00
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
Elf_Scn *scn = NULL;
|
|
|
|
while ((scn = elf_nextscn(elf, scn))) {
|
|
|
|
GElf_Shdr *shdr, shdr_mem;
|
|
|
|
shdr = gelf_getshdr(scn, &shdr_mem);
|
|
|
|
if (!shdr) {
|
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto out;
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
2021-03-26 00:26:06 +00:00
|
|
|
sh_addrs[elf_ndxscn(scn)] = shdr->sh_addr;
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
size_t shstrndx;
|
2020-09-16 01:42:53 +01:00
|
|
|
if (elf_getshdrstrndx(elf, &shstrndx)) {
|
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
Elf_Scn *reloc_scn = NULL;
|
|
|
|
while ((reloc_scn = elf_nextscn(elf, reloc_scn))) {
|
2020-09-16 01:42:53 +01:00
|
|
|
GElf_Shdr *shdr, shdr_mem;
|
2021-03-26 00:26:06 +00:00
|
|
|
shdr = gelf_getshdr(reloc_scn, &shdr_mem);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (!shdr) {
|
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto out;
|
|
|
|
}
|
2021-03-26 00:26:06 +00:00
|
|
|
/* We don't support any architectures that use SHT_REL yet. */
|
2020-09-16 01:42:53 +01:00
|
|
|
if (shdr->sh_type != SHT_RELA)
|
|
|
|
continue;
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
const char *scnname = elf_strptr(elf, shstrndx, shdr->sh_name);
|
2021-03-26 21:28:16 +00:00
|
|
|
if (!scnname) {
|
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto out;
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
|
2021-03-16 22:39:37 +00:00
|
|
|
if (strstartswith(scnname, ".rela.debug_") ||
|
|
|
|
strstartswith(scnname, ".rela.orc_")) {
|
2021-03-26 00:26:06 +00:00
|
|
|
Elf_Scn *scn = elf_getscn(elf, shdr->sh_info);
|
|
|
|
if (!scn) {
|
2020-09-16 01:42:53 +01:00
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
Elf_Scn *symtab_scn = elf_getscn(elf, shdr->sh_link);
|
|
|
|
if (!symtab_scn) {
|
2020-09-16 01:42:53 +01:00
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2021-03-26 00:26:06 +00:00
|
|
|
err = relocate_elf_section(scn, reloc_scn, symtab_scn,
|
|
|
|
sh_addrs, shdrnum,
|
|
|
|
&platform);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
2021-11-30 10:35:59 +00:00
|
|
|
err = NULL;
|
2020-09-16 01:42:53 +01:00
|
|
|
out:
|
|
|
|
free(sh_addrs);
|
2021-11-30 10:35:59 +00:00
|
|
|
return err;
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
2020-11-13 00:14:30 +00:00
|
|
|
drgn_debug_info_find_sections(struct drgn_debug_info_module *module)
|
2020-09-16 01:42:53 +01:00
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
if (module->elf) {
|
2021-03-26 00:26:06 +00:00
|
|
|
err = relocate_elf_file(module->elf);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Note: not dwfl_module_getelf(), because then libdwfl applies
|
|
|
|
* ELF relocations to all sections, not just debug sections.
|
|
|
|
*/
|
|
|
|
Dwarf_Addr bias;
|
2021-12-09 20:37:14 +00:00
|
|
|
Dwarf *dwarf;
|
|
|
|
#pragma omp critical(drgn_dwfl_module_getdwarf)
|
|
|
|
dwarf = dwfl_module_getdwarf(module->dwfl_module, &bias);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (!dwarf)
|
|
|
|
return drgn_error_libdwfl();
|
|
|
|
Elf *elf = dwarf_getelf(dwarf);
|
|
|
|
if (!elf)
|
|
|
|
return drgn_error_libdw();
|
2020-11-28 20:59:44 +00:00
|
|
|
GElf_Ehdr ehdr_mem, *ehdr = gelf_getehdr(elf, &ehdr_mem);
|
|
|
|
if (!ehdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
drgn_platform_from_elf(ehdr, &module->platform);
|
2020-09-16 01:42:53 +01:00
|
|
|
|
|
|
|
size_t shstrndx;
|
|
|
|
if (elf_getshdrstrndx(elf, &shstrndx))
|
|
|
|
return drgn_error_libelf();
|
|
|
|
|
|
|
|
Elf_Scn *scn = NULL;
|
|
|
|
while ((scn = elf_nextscn(elf, scn))) {
|
|
|
|
GElf_Shdr shdr_mem;
|
|
|
|
GElf_Shdr *shdr = gelf_getshdr(scn, &shdr_mem);
|
|
|
|
if (!shdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
|
2020-11-13 00:14:30 +00:00
|
|
|
if (shdr->sh_type != SHT_PROGBITS)
|
2020-09-16 01:42:53 +01:00
|
|
|
continue;
|
|
|
|
const char *scnname = elf_strptr(elf, shstrndx, shdr->sh_name);
|
|
|
|
if (!scnname)
|
2021-03-26 21:28:16 +00:00
|
|
|
return drgn_error_libelf();
|
2020-09-16 01:42:53 +01:00
|
|
|
|
2020-10-23 00:09:31 +01:00
|
|
|
for (size_t i = 0; i < DRGN_NUM_DEBUG_SCNS; i++) {
|
|
|
|
if (!module->scns[i] &&
|
|
|
|
strcmp(scnname, drgn_debug_scn_names[i]) == 0) {
|
2020-11-13 00:14:30 +00:00
|
|
|
module->scns[i] = scn;
|
2020-10-23 00:09:31 +01:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
2021-12-07 21:13:52 +00:00
|
|
|
|
|
|
|
Dwarf *altdwarf = dwarf_getalt(dwarf);
|
|
|
|
if (altdwarf) {
|
|
|
|
elf = dwarf_getelf(altdwarf);
|
|
|
|
if (!elf)
|
|
|
|
return drgn_error_libdw();
|
|
|
|
if (elf_getshdrstrndx(elf, &shstrndx))
|
|
|
|
return drgn_error_libelf();
|
|
|
|
|
|
|
|
scn = NULL;
|
|
|
|
while ((scn = elf_nextscn(elf, scn))) {
|
|
|
|
GElf_Shdr shdr_mem;
|
|
|
|
GElf_Shdr *shdr = gelf_getshdr(scn, &shdr_mem);
|
|
|
|
if (!shdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
|
|
|
|
if (shdr->sh_type != SHT_PROGBITS)
|
|
|
|
continue;
|
|
|
|
const char *scnname = elf_strptr(elf, shstrndx, shdr->sh_name);
|
|
|
|
if (!scnname)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
|
|
|
|
/*
|
|
|
|
* TODO: save more sections and support imported units.
|
|
|
|
*/
|
|
|
|
if (strcmp(scnname, ".debug_info") == 0 &&
|
|
|
|
!module->alt_debug_info)
|
|
|
|
module->alt_debug_info = scn;
|
|
|
|
else if (strcmp(scnname, ".debug_str") == 0 &&
|
|
|
|
!module->alt_debug_str)
|
|
|
|
module->alt_debug_str = scn;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-11-13 00:14:30 +00:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2021-06-15 01:40:52 +01:00
|
|
|
static void truncate_null_terminated_section(Elf_Data *data)
|
|
|
|
{
|
|
|
|
if (data) {
|
|
|
|
const char *buf = data->d_buf;
|
|
|
|
const char *nul = memrchr(buf, '\0', data->d_size);
|
|
|
|
if (nul)
|
|
|
|
data->d_size = nul - buf + 1;
|
|
|
|
else
|
|
|
|
data->d_size = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-11-13 00:14:30 +00:00
|
|
|
static struct drgn_error *
|
|
|
|
drgn_debug_info_precache_sections(struct drgn_debug_info_module *module)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
for (size_t i = 0; i < DRGN_NUM_DEBUG_SCN_DATA_PRECACHE; i++) {
|
|
|
|
if (module->scns[i]) {
|
|
|
|
err = read_elf_section(module->scns[i],
|
|
|
|
&module->scn_data[i]);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
}
|
2021-12-07 21:13:52 +00:00
|
|
|
if (module->alt_debug_info) {
|
|
|
|
err = read_elf_section(module->alt_debug_info,
|
|
|
|
&module->alt_debug_info_data);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
if (module->alt_debug_str) {
|
|
|
|
err = read_elf_section(module->alt_debug_str,
|
|
|
|
&module->alt_debug_str_data);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Truncate any extraneous bytes so that we can assume that a pointer
|
2021-06-15 01:40:52 +01:00
|
|
|
* within .debug_{,line_}str is always null-terminated.
|
2020-09-16 01:42:53 +01:00
|
|
|
*/
|
2021-06-15 01:40:52 +01:00
|
|
|
truncate_null_terminated_section(module->scn_data[DRGN_SCN_DEBUG_STR]);
|
|
|
|
truncate_null_terminated_section(module->scn_data[DRGN_SCN_DEBUG_LINE_STR]);
|
2021-12-07 21:13:52 +00:00
|
|
|
truncate_null_terminated_section(module->alt_debug_str_data);
|
2020-09-16 01:42:53 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2021-11-08 23:39:39 +00:00
|
|
|
struct drgn_error *
|
2021-06-02 17:43:39 +01:00
|
|
|
drgn_debug_info_module_cache_section(struct drgn_debug_info_module *module,
|
|
|
|
enum drgn_debug_info_scn scn)
|
|
|
|
{
|
|
|
|
if (module->scn_data[scn])
|
|
|
|
return NULL;
|
|
|
|
return read_elf_section(module->scns[scn], &module->scn_data[scn]);
|
|
|
|
}
|
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
static struct drgn_error *
|
|
|
|
drgn_debug_info_read_module(struct drgn_debug_info_load_state *load,
|
2021-11-12 09:35:27 +00:00
|
|
|
struct drgn_dwarf_index_state *index,
|
2020-09-16 01:42:53 +01:00
|
|
|
struct drgn_debug_info_module *head)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
struct drgn_debug_info_module *module;
|
|
|
|
for (module = head; module; module = module->next) {
|
2020-11-13 00:14:30 +00:00
|
|
|
err = drgn_debug_info_find_sections(module);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (err) {
|
|
|
|
module->err = err;
|
|
|
|
continue;
|
|
|
|
}
|
2020-10-23 00:09:31 +01:00
|
|
|
if (module->scns[DRGN_SCN_DEBUG_INFO] &&
|
|
|
|
module->scns[DRGN_SCN_DEBUG_ABBREV]) {
|
2020-11-13 00:14:30 +00:00
|
|
|
err = drgn_debug_info_precache_sections(module);
|
|
|
|
if (err) {
|
|
|
|
module->err = err;
|
|
|
|
continue;
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
module->state = DRGN_DEBUG_INFO_MODULE_INDEXING;
|
2021-11-12 09:35:27 +00:00
|
|
|
return drgn_dwarf_index_read_module(index,
|
2021-05-06 22:52:24 +01:00
|
|
|
module);
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* We checked all of the files and didn't find debugging information.
|
|
|
|
* Report why for each one.
|
|
|
|
*
|
|
|
|
* (If we did find debugging information, we discard errors on the
|
|
|
|
* unused files.)
|
|
|
|
*/
|
|
|
|
err = NULL;
|
|
|
|
#pragma omp critical(drgn_debug_info_read_module_error)
|
|
|
|
for (module = head; module; module = module->next) {
|
|
|
|
const char *name =
|
|
|
|
dwfl_module_info(module->dwfl_module, NULL, NULL, NULL,
|
|
|
|
NULL, NULL, NULL, NULL);
|
|
|
|
if (module->err) {
|
|
|
|
err = drgn_debug_info_report_error(load, name, NULL,
|
|
|
|
module->err);
|
|
|
|
module->err = NULL;
|
|
|
|
} else {
|
|
|
|
err = drgn_debug_info_report_error(load, name,
|
|
|
|
"no debugging information",
|
|
|
|
NULL);
|
|
|
|
}
|
|
|
|
if (err)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
drgn_debug_info_update_index(struct drgn_debug_info_load_state *load)
|
|
|
|
{
|
|
|
|
if (!load->new_modules.size)
|
|
|
|
return NULL;
|
|
|
|
struct drgn_debug_info *dbinfo = load->dbinfo;
|
|
|
|
if (!c_string_set_reserve(&dbinfo->module_names,
|
|
|
|
c_string_set_size(&dbinfo->module_names) +
|
|
|
|
load->new_modules.size))
|
|
|
|
return &drgn_enomem;
|
2021-05-06 22:52:24 +01:00
|
|
|
|
2021-11-12 09:35:27 +00:00
|
|
|
struct drgn_dwarf_index_state index;
|
|
|
|
if (!drgn_dwarf_index_state_init(&index, dbinfo))
|
2021-05-06 22:52:24 +01:00
|
|
|
return &drgn_enomem;
|
|
|
|
struct drgn_error *err = NULL;
|
|
|
|
#pragma omp parallel for schedule(dynamic)
|
2020-09-16 01:42:53 +01:00
|
|
|
for (size_t i = 0; i < load->new_modules.size; i++) {
|
2021-05-06 22:52:24 +01:00
|
|
|
if (err)
|
2020-09-16 01:42:53 +01:00
|
|
|
continue;
|
|
|
|
struct drgn_error *module_err =
|
2021-11-12 09:35:27 +00:00
|
|
|
drgn_debug_info_read_module(load, &index,
|
2020-09-16 01:42:53 +01:00
|
|
|
load->new_modules.data[i]);
|
2021-05-06 22:52:24 +01:00
|
|
|
if (module_err) {
|
|
|
|
#pragma omp critical(drgn_debug_info_update_index_error)
|
|
|
|
if (err)
|
|
|
|
drgn_error_destroy(module_err);
|
|
|
|
else
|
|
|
|
err = module_err;
|
|
|
|
}
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
2021-05-06 22:52:24 +01:00
|
|
|
if (!err)
|
2021-11-12 09:35:27 +00:00
|
|
|
err = drgn_dwarf_info_update_index(&index);
|
|
|
|
drgn_dwarf_index_state_deinit(&index);
|
2021-05-06 22:52:24 +01:00
|
|
|
if (!err)
|
|
|
|
drgn_debug_info_free_modules(dbinfo, true, false);
|
|
|
|
return err;
|
2020-09-16 01:42:53 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_error *
|
|
|
|
drgn_debug_info_report_flush(struct drgn_debug_info_load_state *load)
|
|
|
|
{
|
|
|
|
struct drgn_debug_info *dbinfo = load->dbinfo;
|
2021-06-09 23:11:36 +01:00
|
|
|
my_dwfl_report_end(dbinfo, NULL, NULL);
|
2020-09-16 01:42:53 +01:00
|
|
|
struct drgn_error *err = drgn_debug_info_update_index(load);
|
|
|
|
dwfl_report_begin_add(dbinfo->dwfl);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
load->new_modules.size = 0;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drgn_error *
|
|
|
|
drgn_debug_info_report_finalize_errors(struct drgn_debug_info_load_state *load)
|
|
|
|
{
|
|
|
|
if (load->num_errors > load->max_errors &&
|
|
|
|
(!string_builder_line_break(&load->errors) ||
|
|
|
|
!string_builder_appendf(&load->errors, "... %u more",
|
|
|
|
load->num_errors - load->max_errors))) {
|
|
|
|
free(load->errors.str);
|
|
|
|
return &drgn_enomem;
|
|
|
|
}
|
|
|
|
if (load->num_errors) {
|
|
|
|
return drgn_error_from_string_builder(DRGN_ERROR_MISSING_DEBUG_INFO,
|
|
|
|
&load->errors);
|
|
|
|
} else {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_error *drgn_debug_info_load(struct drgn_debug_info *dbinfo,
|
|
|
|
const char **paths, size_t n,
|
|
|
|
bool load_default, bool load_main)
|
|
|
|
{
|
|
|
|
struct drgn_program *prog = dbinfo->prog;
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
if (load_default)
|
|
|
|
load_main = true;
|
|
|
|
|
|
|
|
const char *max_errors = getenv("DRGN_MAX_DEBUG_INFO_ERRORS");
|
|
|
|
struct drgn_debug_info_load_state load = {
|
|
|
|
.dbinfo = dbinfo,
|
|
|
|
.paths = paths,
|
|
|
|
.num_paths = n,
|
|
|
|
.load_default = load_default,
|
|
|
|
.load_main = load_main,
|
|
|
|
.new_modules = VECTOR_INIT,
|
|
|
|
.max_errors = max_errors ? atoi(max_errors) : 5,
|
|
|
|
};
|
|
|
|
dwfl_report_begin_add(dbinfo->dwfl);
|
|
|
|
if (prog->flags & DRGN_PROGRAM_IS_LINUX_KERNEL)
|
|
|
|
err = linux_kernel_report_debug_info(&load);
|
|
|
|
else
|
|
|
|
err = userspace_report_debug_info(&load);
|
2021-06-09 23:11:36 +01:00
|
|
|
my_dwfl_report_end(dbinfo, NULL, NULL);
|
2020-09-16 01:42:53 +01:00
|
|
|
if (err)
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* userspace_report_debug_info() reports the main debugging information
|
|
|
|
* directly with libdwfl, so we need to report it to dbinfo.
|
|
|
|
*/
|
|
|
|
if (!(prog->flags & DRGN_PROGRAM_IS_LINUX_KERNEL) && load_main &&
|
|
|
|
dwfl_getmodules(dbinfo->dwfl, drgn_debug_info_report_dwfl_module,
|
|
|
|
&load, 0)) {
|
|
|
|
err = &drgn_enomem;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = drgn_debug_info_update_index(&load);
|
|
|
|
if (err)
|
|
|
|
goto err;
|
|
|
|
|
2021-12-08 06:19:53 +00:00
|
|
|
/*
|
|
|
|
* TODO: for core dumps, we need to add memory reader segments for
|
|
|
|
* read-only segments of the loaded binaries since those aren't saved in
|
|
|
|
* the core dump.
|
|
|
|
*/
|
|
|
|
|
2020-09-16 01:42:53 +01:00
|
|
|
/*
|
|
|
|
* If this fails, it's too late to roll back. This can only fail with
|
|
|
|
* enomem, so it's not a big deal.
|
|
|
|
*/
|
|
|
|
err = drgn_debug_info_report_finalize_errors(&load);
|
|
|
|
out:
|
|
|
|
drgn_debug_info_module_vector_deinit(&load.new_modules);
|
|
|
|
return err;
|
|
|
|
|
|
|
|
err:
|
|
|
|
drgn_debug_info_free_modules(dbinfo, false, false);
|
|
|
|
free(load.errors.str);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool drgn_debug_info_is_indexed(struct drgn_debug_info *dbinfo,
|
|
|
|
const char *name)
|
|
|
|
{
|
|
|
|
return c_string_set_search(&dbinfo->module_names, &name).entry != NULL;
|
|
|
|
}
|
|
|
|
|
2021-11-09 01:05:33 +00:00
|
|
|
struct drgn_error *drgn_debug_info_create(struct drgn_program *prog,
|
|
|
|
struct drgn_debug_info **ret)
|
2021-07-07 00:26:31 +01:00
|
|
|
{
|
2021-11-09 01:05:33 +00:00
|
|
|
struct drgn_debug_info *dbinfo = malloc(sizeof(*dbinfo));
|
|
|
|
if (!dbinfo)
|
|
|
|
return &drgn_enomem;
|
|
|
|
dbinfo->prog = prog;
|
|
|
|
const Dwfl_Callbacks *dwfl_callbacks;
|
|
|
|
if (prog->flags & DRGN_PROGRAM_IS_LINUX_KERNEL)
|
|
|
|
dwfl_callbacks = &drgn_dwfl_callbacks;
|
|
|
|
else if (prog->flags & DRGN_PROGRAM_IS_LIVE)
|
|
|
|
dwfl_callbacks = &drgn_linux_proc_dwfl_callbacks;
|
|
|
|
else
|
|
|
|
dwfl_callbacks = &drgn_userspace_core_dump_dwfl_callbacks;
|
|
|
|
dbinfo->dwfl = dwfl_begin(dwfl_callbacks);
|
|
|
|
if (!dbinfo->dwfl) {
|
|
|
|
free(dbinfo);
|
|
|
|
return drgn_error_libdwfl();
|
2021-07-07 00:26:31 +01:00
|
|
|
}
|
2021-11-09 01:05:33 +00:00
|
|
|
drgn_debug_info_module_table_init(&dbinfo->modules);
|
|
|
|
c_string_set_init(&dbinfo->module_names);
|
|
|
|
drgn_dwarf_info_init(dbinfo);
|
|
|
|
*ret = dbinfo;
|
2021-07-07 00:26:31 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2021-11-09 01:05:33 +00:00
|
|
|
void drgn_debug_info_destroy(struct drgn_debug_info *dbinfo)
|
2021-07-07 00:26:31 +01:00
|
|
|
{
|
2021-11-09 01:05:33 +00:00
|
|
|
if (!dbinfo)
|
|
|
|
return;
|
|
|
|
drgn_dwarf_info_deinit(dbinfo);
|
|
|
|
c_string_set_deinit(&dbinfo->module_names);
|
|
|
|
drgn_debug_info_free_modules(dbinfo, false, true);
|
|
|
|
assert(drgn_debug_info_module_table_empty(&dbinfo->modules));
|
|
|
|
drgn_debug_info_module_table_deinit(&dbinfo->modules);
|
|
|
|
dwfl_end(dbinfo->dwfl);
|
|
|
|
free(dbinfo);
|
2021-07-07 00:26:31 +01:00
|
|
|
}
|
|
|
|
|
2021-11-09 01:05:33 +00:00
|
|
|
struct drgn_error *
|
|
|
|
drgn_debug_info_module_find_cfi(struct drgn_program *prog,
|
|
|
|
struct drgn_debug_info_module *module,
|
|
|
|
uint64_t pc, struct drgn_cfi_row **row_ret,
|
|
|
|
bool *interrupted_ret,
|
|
|
|
drgn_register_number *ret_addr_regno_ret)
|
2021-07-01 22:02:38 +01:00
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
2021-11-09 01:05:33 +00:00
|
|
|
Dwarf_Addr bias;
|
|
|
|
dwfl_module_info(module->dwfl_module, NULL, NULL, NULL, &bias, NULL,
|
|
|
|
NULL, NULL);
|
|
|
|
uint64_t unbiased_pc = pc - bias;
|
2021-07-01 22:02:38 +01:00
|
|
|
|
2021-11-09 01:05:33 +00:00
|
|
|
if (prog->prefer_orc_unwinder) {
|
|
|
|
err = drgn_debug_info_find_orc_cfi(module, unbiased_pc, row_ret,
|
|
|
|
interrupted_ret,
|
|
|
|
ret_addr_regno_ret);
|
|
|
|
if (err != &drgn_not_found)
|
|
|
|
return err;
|
|
|
|
return drgn_debug_info_find_dwarf_cfi(module, unbiased_pc,
|
|
|
|
row_ret, interrupted_ret,
|
|
|
|
ret_addr_regno_ret);
|
2021-07-01 22:02:38 +01:00
|
|
|
} else {
|
2021-11-09 01:05:33 +00:00
|
|
|
err = drgn_debug_info_find_dwarf_cfi(module, unbiased_pc,
|
|
|
|
row_ret, interrupted_ret,
|
|
|
|
ret_addr_regno_ret);
|
|
|
|
if (err != &drgn_not_found)
|
2021-07-01 22:02:38 +01:00
|
|
|
return err;
|
2021-11-09 01:05:33 +00:00
|
|
|
return drgn_debug_info_find_orc_cfi(module, unbiased_pc,
|
|
|
|
row_ret, interrupted_ret,
|
|
|
|
ret_addr_regno_ret);
|
2021-07-01 22:02:38 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-11-09 01:05:33 +00:00
|
|
|
#if !_ELFUTILS_PREREQ(0, 175)
|
|
|
|
static Elf *dwelf_elf_begin(int fd)
|
2021-07-01 22:02:38 +01:00
|
|
|
{
|
2021-11-09 01:05:33 +00:00
|
|
|
return elf_begin(fd, ELF_C_READ_MMAP_PRIVATE, NULL);
|
2021-03-10 09:51:33 +00:00
|
|
|
}
|
2021-11-09 01:05:33 +00:00
|
|
|
#endif
|
2021-03-10 09:51:33 +00:00
|
|
|
|
2020-09-24 00:02:02 +01:00
|
|
|
struct drgn_error *open_elf_file(const char *path, int *fd_ret, Elf **elf_ret)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
|
|
|
|
*fd_ret = open(path, O_RDONLY);
|
|
|
|
if (*fd_ret == -1)
|
|
|
|
return drgn_error_create_os("open", errno, path);
|
|
|
|
*elf_ret = dwelf_elf_begin(*fd_ret);
|
|
|
|
if (!*elf_ret) {
|
|
|
|
err = drgn_error_libelf();
|
|
|
|
goto err_fd;
|
|
|
|
}
|
|
|
|
if (elf_kind(*elf_ret) != ELF_K_ELF) {
|
|
|
|
err = drgn_error_create(DRGN_ERROR_OTHER, "not an ELF file");
|
|
|
|
goto err_elf;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
err_elf:
|
|
|
|
elf_end(*elf_ret);
|
|
|
|
err_fd:
|
|
|
|
close(*fd_ret);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_error *find_elf_file(char **path_ret, int *fd_ret, Elf **elf_ret,
|
|
|
|
const char * const *path_formats, ...)
|
|
|
|
{
|
|
|
|
struct drgn_error *err;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
for (i = 0; path_formats[i]; i++) {
|
|
|
|
va_list ap;
|
|
|
|
int ret;
|
|
|
|
char *path;
|
|
|
|
int fd;
|
|
|
|
Elf *elf;
|
|
|
|
|
|
|
|
va_start(ap, path_formats);
|
|
|
|
ret = vasprintf(&path, path_formats[i], ap);
|
|
|
|
va_end(ap);
|
|
|
|
if (ret == -1)
|
|
|
|
return &drgn_enomem;
|
|
|
|
fd = open(path, O_RDONLY);
|
|
|
|
if (fd == -1) {
|
|
|
|
free(path);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
elf = dwelf_elf_begin(fd);
|
|
|
|
if (!elf) {
|
|
|
|
close(fd);
|
|
|
|
free(path);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (elf_kind(elf) != ELF_K_ELF) {
|
|
|
|
err = drgn_error_format(DRGN_ERROR_OTHER,
|
|
|
|
"%s: not an ELF file", path);
|
|
|
|
elf_end(elf);
|
|
|
|
close(fd);
|
|
|
|
free(path);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
*path_ret = path;
|
|
|
|
*fd_ret = fd;
|
|
|
|
*elf_ret = elf;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
*path_ret = NULL;
|
|
|
|
*fd_ret = -1;
|
|
|
|
*elf_ret = NULL;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct drgn_error *read_elf_section(Elf_Scn *scn, Elf_Data **ret)
|
|
|
|
{
|
|
|
|
GElf_Shdr shdr_mem, *shdr;
|
|
|
|
Elf_Data *data;
|
|
|
|
|
|
|
|
shdr = gelf_getshdr(scn, &shdr_mem);
|
|
|
|
if (!shdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
if ((shdr->sh_flags & SHF_COMPRESSED) && elf_compress(scn, 0, 0) < 0)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
data = elf_getdata(scn, NULL);
|
|
|
|
if (!data)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
*ret = data;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2021-12-06 21:33:55 +00:00
|
|
|
/*
|
|
|
|
* Get the start address from the first loadable segment and the end address
|
|
|
|
* from the last loadable segment.
|
|
|
|
*
|
|
|
|
* The ELF specification states that loadable segments are sorted on p_vaddr.
|
|
|
|
* However, vmlinux on x86-64 has an out of order segment for .data..percpu, and
|
|
|
|
* Arm has a couple for .vector and .stubs. Thankfully, those are placed in the
|
|
|
|
* middle by the vmlinux linker script, so we can still rely on the first and
|
|
|
|
* last loadable segments.
|
|
|
|
*/
|
2020-09-24 00:02:02 +01:00
|
|
|
struct drgn_error *elf_address_range(Elf *elf, uint64_t bias,
|
|
|
|
uint64_t *start_ret, uint64_t *end_ret)
|
|
|
|
{
|
2021-12-06 21:33:55 +00:00
|
|
|
size_t phnum;
|
2020-09-24 00:02:02 +01:00
|
|
|
if (elf_getphdrnum(elf, &phnum) != 0)
|
|
|
|
return drgn_error_libelf();
|
2021-12-06 21:33:55 +00:00
|
|
|
|
|
|
|
GElf_Phdr phdr_mem, *phdr;
|
|
|
|
size_t i;
|
2020-09-24 00:02:02 +01:00
|
|
|
for (i = 0; i < phnum; i++) {
|
2021-12-06 21:33:55 +00:00
|
|
|
phdr = gelf_getphdr(elf, i, &phdr_mem);
|
|
|
|
if (!phdr)
|
|
|
|
return drgn_error_libelf();
|
|
|
|
if (phdr->p_type == PT_LOAD) {
|
|
|
|
uint64_t align = phdr->p_align ? phdr->p_align : 1;
|
|
|
|
*start_ret = (phdr->p_vaddr & -align) + bias;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (i >= phnum) {
|
|
|
|
/* There were no loadable segments. */
|
|
|
|
*start_ret = *end_ret = 0;
|
|
|
|
return NULL;
|
|
|
|
}
|
2020-09-24 00:02:02 +01:00
|
|
|
|
2021-12-06 21:33:55 +00:00
|
|
|
for (i = phnum; i-- > 0;) {
|
2020-09-24 00:02:02 +01:00
|
|
|
phdr = gelf_getphdr(elf, i, &phdr_mem);
|
|
|
|
if (!phdr)
|
|
|
|
return drgn_error_libelf();
|
2021-12-06 21:33:55 +00:00
|
|
|
if (phdr->p_type == PT_LOAD) {
|
|
|
|
*end_ret = (phdr->p_vaddr + phdr->p_memsz) + bias;
|
|
|
|
if (*start_ret >= *end_ret)
|
|
|
|
*start_ret = *end_ret = 0;
|
|
|
|
return NULL;
|
2020-09-24 00:02:02 +01:00
|
|
|
}
|
|
|
|
}
|
2021-12-06 21:33:55 +00:00
|
|
|
/* We found a loadable segment earlier, so this shouldn't happen. */
|
|
|
|
assert(!"PT_LOAD segment disappeared");
|
|
|
|
*end_ret = 0;
|
2020-09-24 00:02:02 +01:00
|
|
|
return NULL;
|
|
|
|
}
|