Re: [PATCH v2 7/9] x86: Add support for function granular KASLR
From: Kees Cook
Date: Thu May 21 2020 - 17:08:13 EST
On Thu, May 21, 2020 at 09:56:38AM -0700, Kristen Carlson Accardi wrote:
> At boot time, find all the function sections that have separate .text
> sections, shuffle them, and then copy them to new locations. Adjust
> any relocations accordingly.
Commit log length vs "11 files changed, 1159 insertions(+), 15
deletions(-)" implies to me that a lot more detail is needed here. ;)
Perhaps describe what the code pieces are, why the code is being added
are here, etc (I see at least the ELF parsing, the ELF shuffling, the
relocation updates, the symbol list, and the re-sorting of kallsyms,
ORC, and extables. I think the commit log should prepare someone to read
the diff and know what to expect to find. (In the end, I wonder if these
pieces should be split up into logically separate patches, but for now,
let's just go with it -- though I've made some suggestions below about
things that might be worth splitting out.)
More below...
> Signed-off-by: Kristen Carlson Accardi <kristen@xxxxxxxxxxxxxxx>
> Reviewed-by: Tony Luck <tony.luck@xxxxxxxxx>
> Tested-by: Tony Luck <tony.luck@xxxxxxxxx>
> ---
> Documentation/security/fgkaslr.rst | 155 +++++
> Documentation/security/index.rst | 1 +
> arch/x86/boot/compressed/Makefile | 3 +
> arch/x86/boot/compressed/fgkaslr.c | 823 +++++++++++++++++++++++
> arch/x86/boot/compressed/kaslr.c | 4 -
> arch/x86/boot/compressed/misc.c | 109 ++-
> arch/x86/boot/compressed/misc.h | 34 +
> arch/x86/boot/compressed/utils.c | 12 +
> arch/x86/boot/compressed/vmlinux.symbols | 17 +
> arch/x86/include/asm/boot.h | 15 +-
> include/uapi/linux/elf.h | 1 +
> 11 files changed, 1159 insertions(+), 15 deletions(-)
> create mode 100644 Documentation/security/fgkaslr.rst
> create mode 100644 arch/x86/boot/compressed/fgkaslr.c
> create mode 100644 arch/x86/boot/compressed/utils.c
> create mode 100644 arch/x86/boot/compressed/vmlinux.symbols
>
> diff --git a/Documentation/security/fgkaslr.rst b/Documentation/security/fgkaslr.rst
> new file mode 100644
> index 000000000000..94939c62c50d
> --- /dev/null
> +++ b/Documentation/security/fgkaslr.rst
> @@ -0,0 +1,155 @@
> +.. SPDX-License-Identifier: GPL-2.0
> +
> +=====================================================================
> +Function Granular Kernel Address Space Layout Randomization (fgkaslr)
> +=====================================================================
> +
> +:Date: 6 April 2020
> +:Author: Kristen Accardi
> +
> +Kernel Address Space Layout Randomization (KASLR) was merged into the kernel
> +with the objective of increasing the difficulty of code reuse attacks. Code
> +reuse attacks reused existing code snippets to get around existing memory
> +protections. They exploit software bugs which expose addresses of useful code
> +snippets to control the flow of execution for their own nefarious purposes.
> +KASLR as it was originally implemented moves the entire kernel code text as a
> +unit at boot time in order to make addresses less predictable. The order of the
> +code within the segment is unchanged - only the base address is shifted. There
> +are a few shortcomings to this algorithm.
> +
> +1. Low Entropy - there are only so many locations the kernel can fit in. This
> + means an attacker could guess without too much trouble.
> +2. Knowledge of a single address can reveal the offset of the base address,
> + exposing all other locations for a published/known kernel image.
> +3. Info leaks abound.
> +
> +Finer grained ASLR has been proposed as a way to make ASLR more resistant
> +to info leaks. It is not a new concept at all, and there are many variations
> +possible. Function reordering is an implementation of finer grained ASLR
> +which randomizes the layout of an address space on a function level
> +granularity. The term "fgkaslr" is used in this document to refer to the
> +technique of function reordering when used with KASLR, as well as finer grained
> +KASLR in general.
> +
> +The objective of this patch set is to improve a technology that is already
> +merged into the kernel (KASLR). This code will not prevent all code reuse
> +attacks, and should be considered as one of several tools that can be used.
> +
> +Implementation Details
> +======================
> +
> +The over-arching objective of the fgkaslr implementation is incremental
> +improvement over the existing KASLR algorithm. It is designed to work with
> +the existing solution, and there are two main area where code changes occur:
> +Build time, and Load time.
> +
> +Build time
> +----------
> +
> +GCC has had an option to place functions into individual .text sections
> +for many years now (-ffunction-sections). This option is used to implement
> +function reordering at load time. The final compiled vmlinux retains all the
> +section headers, which can be used to help find the address ranges of each
> +function. Using this information and an expanded table of relocation addresses,
> +individual text sections can be shuffled immediately after decompression.
> +Some data tables inside the kernel that have assumptions about order
> +require sorting after the update. In order to modify these tables,
> +a few key symbols from the objcopy symbol stripping process are preserved
> +for use after shuffling the text segments.
> +
> +Load time
> +---------
> +
> +The boot kernel was modified to parse the vmlinux elf file after
> +decompression to check for symbols for modifying data tables, and to
> +look for any .text.* sections to randomize. The sections are then shuffled,
> +and tables are updated or resorted. The existing code which updated relocation
> +addresses was modified to account for not just a fixed delta from the load
> +address, but the offset that the function section was moved to. This requires
> +inspection of each address to see if it was impacted by a randomization.
> +
> +In order to hide the new layout, symbols reported through /proc/kallsyms will
> +be sorted by name alphabetically rather than by address.
> +
> +Performance Impact
> +==================
> +
> +There are two areas where function reordering can impact performance: boot
> +time latency, and run time performance.
> +
> +Boot time latency
> +-----------------
> +
> +This implementation of finer grained KASLR impacts the boot time of the kernel
> +in several places. It requires additional parsing of the kernel ELF file to
> +obtain the section headers of the sections to be randomized. It calls the
> +random number generator for each section to be randomized to determine that
> +section's new memory location. It copies the decompressed kernel into a new
> +area of memory to avoid corruption when laying out the newly randomized
> +sections. It increases the number of relocations the kernel has to perform at
> +boot time vs. standard KASLR, and it also requires a lookup on each address
> +that needs to be relocated to see if it was in a randomized section and needs
> +to be adjusted by a new offset. Finally, it re-sorts a few data tables that
> +are required to be sorted by address.
> +
> +Booting a test VM on a modern, well appointed system showed an increase in
> +latency of approximately 1 second.
> +
> +Run time
> +--------
> +
> +The performance impact at run-time of function reordering varies by workload.
> +Randomly reordering the functions will cause an increase in cache misses
> +for some workloads. Some workloads perform significantly worse under FGKASLR,
> +while others stay the same or even improve. In general, it will depend on the
> +code flow whether or not finer grained KASLR will impact a workload, and how
> +the underlying code was designed. Because the layout changes per boot, each
> +time a system is rebooted the performance of a workload may change.
> +
> +Image Size
> +==========
> +
> +fgkaslr increases the size of the kernel binary due to the extra section
> +headers that are included, as well as the extra relocations that need to
> +be added. You can expect fgkaslr to increase the size of the resulting
> +vmlinux by about 3%, and the compressed image (bzImage) by 15%.
> +
> +Memory Usage
> +============
> +
> +fgkaslr increases the amount of heap that is required at boot time,
> +although this extra memory is released when the kernel has finished
> +decompression. As a result, it may not be appropriate to use this feature
> +on systems without much memory.
> +
> +Building
> +========
> +
> +To enable fine grained KASLR, you need to have the following config options
> +set (including all the ones you would use to build normal KASLR)
> +
> +``CONFIG_FG_KASLR=y``
> +
> +fgkaslr for the kernel is only supported for the X86_64 architecture.
> +
> +Modules
> +=======
> +
> +Modules are randomized similarly to the rest of the kernel by shuffling
> +the sections at load time prior to moving them into memory. The module must
> +also have been build with the -ffunction-sections compiler option.
> +
> +Although fgkaslr for the kernel is only supported for the X86_64 architecture,
> +it is possible to use fgkaslr with modules on other architectures. To enable
> +this feature, select the following config option:
> +
> +``CONFIG_MODULE_FG_KASLR``
> +
> +This option is selected automatically for X86_64 when CONFIG_FG_KASLR is set.
> +
> +Disabling
> +=========
> +
> +Disabling normal kaslr using the nokaslr command line option also disables
> +fgkaslr. In addtion, it is possible to disable fgkaslr separately by booting
> +with fgkaslr=off on the commandline.
Yay documentation! Are you able to include the references to the papers
you discuss in the cover letter in here too?
> diff --git a/Documentation/security/index.rst b/Documentation/security/index.rst
> index fc503dd689a7..5e370c409ad2 100644
> --- a/Documentation/security/index.rst
> +++ b/Documentation/security/index.rst
> @@ -7,6 +7,7 @@ Security Documentation
>
> credentials
> IMA-templates
> + fgkaslr
> keys/index
> lsm
> lsm-development
> diff --git a/arch/x86/boot/compressed/Makefile b/arch/x86/boot/compressed/Makefile
> index 3a5a004498de..0ad5a31f1f0a 100644
> --- a/arch/x86/boot/compressed/Makefile
> +++ b/arch/x86/boot/compressed/Makefile
> @@ -80,10 +80,12 @@ vmlinux-objs-y := $(obj)/vmlinux.lds $(obj)/kernel_info.o $(obj)/head_$(BITS).o
>
> vmlinux-objs-$(CONFIG_EARLY_PRINTK) += $(obj)/early_serial_console.o
> vmlinux-objs-$(CONFIG_RANDOMIZE_BASE) += $(obj)/kaslr.o
> +vmlinux-objs-$(CONFIG_FG_KASLR) += $(obj)/utils.o
> ifdef CONFIG_X86_64
> vmlinux-objs-$(CONFIG_RANDOMIZE_BASE) += $(obj)/kaslr_64.o
> vmlinux-objs-y += $(obj)/mem_encrypt.o
> vmlinux-objs-y += $(obj)/pgtable_64.o
> + vmlinux-objs-$(CONFIG_FG_KASLR) += $(obj)/fgkaslr.o
> endif
CONFIG_FG_KASLR is already gated by CONFIG_X86-64, so I think you can
just put these all on the same line before the ifdef:
vmlinux-objs-$(CONFIG_FG_KASLR) += $(obj)/utils.o $(obj)/fgkaslr.o
>
> vmlinux-objs-$(CONFIG_ACPI) += $(obj)/acpi.o
> @@ -120,6 +122,7 @@ OBJCOPYFLAGS_vmlinux.bin := -R .comment -S
>
> ifdef CONFIG_FG_KASLR
> RELOCS_ARGS += --fg-kaslr
> + OBJCOPYFLAGS += --keep-symbols=$(srctree)/$(src)/vmlinux.symbols
> endif
>
> $(obj)/vmlinux.bin: vmlinux FORCE
> diff --git a/arch/x86/boot/compressed/fgkaslr.c b/arch/x86/boot/compressed/fgkaslr.c
> new file mode 100644
> index 000000000000..451e807de276
> --- /dev/null
> +++ b/arch/x86/boot/compressed/fgkaslr.c
> @@ -0,0 +1,823 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * fgkaslr.c
> + *
> + * This contains the routines needed to reorder the kernel text section
> + * at boot time.
> + */
> +#include "misc.h"
> +#include "error.h"
> +#include "pgtable.h"
> +#include "../string.h"
> +#include "../voffset.h"
> +#include <linux/sort.h>
> +#include <linux/bsearch.h>
> +#include "../../include/asm/extable.h"
> +#include "../../include/asm/orc_types.h"
> +
> +/*
> + * Longest parameter of 'fgkaslr=' is 'off' right now, plus an extra '\0'
> + * for termination.
> + */
> +#define MAX_FGKASLR_ARG_LENGTH 4
> +int nofgkaslr;
This can be static.
> +
> +/*
> + * Use normal definitions of mem*() from string.c. There are already
> + * included header files which expect a definition of memset() and by
> + * the time we define memset macro, it is too late.
> + */
> +#undef memcpy
> +#undef memset
> +#define memzero(s, n) memset((s), 0, (n))
> +#define memmove memmove
> +
> +void *memmove(void *dest, const void *src, size_t n);
> +
> +static unsigned long percpu_start;
> +static unsigned long percpu_end;
> +
> +static long kallsyms_names;
> +static long kallsyms_offsets;
> +static long kallsyms_num_syms;
> +static long kallsyms_relative_base;
> +static long kallsyms_markers;
> +static long __start___ex_table_addr;
> +static long __stop___ex_table_addr;
> +static long _stext;
> +static long _etext;
> +static long _sinittext;
> +static long _einittext;
> +static long __start_orc_unwind_ip;
> +static long __stop_orc_unwind_ip;
> +static long __start_orc_unwind;
> +
> +/* addresses in mapped address space */
> +static int *base;
> +static u8 *names;
> +static unsigned long relative_base;
> +static unsigned int *markers_addr;
> +
> +struct kallsyms_name {
> + u8 len;
> + u8 indecis[256];
> +};
> +
> +struct kallsyms_name *names_table;
> +
> +struct orc_entry *cur_orc_table;
> +int *cur_orc_ip_table;
static?
> +
> +/*
> + * This is an array of pointers to sections headers for randomized sections
> + */
> +Elf64_Shdr **sections;
Given the creation of the Elf_Shdr etc macros in the header, should all
of the Elf64 things in here be updated to use the size-agnostic macros?
(Is anyone actually going to run a 32-bit kernel with fgkaslr some day?)
> +
> +/*
> + * This is an array of all section headers, randomized or otherwise.
Comments nitpick while I'm thinking about it: "This is" tends to be
redundant, and I'd pick a single-sentence-comment punctuation style:
i.e. with or without an ending period. I personally prefer including the
period, but mainly I think it should just be consistent. :)
> + */
> +Elf64_Shdr *sechdrs;
Also, these can be static?
> +
> +/*
> + * The number of elements in the randomized section header array (sections)
> + */
> +int sections_size;
static? (Also, it seems this is related to "sections" above, so maybe
more it closer?)
> +
> +static bool is_text(long addr)
> +{
> + if ((addr >= _stext && addr < _etext) ||
> + (addr >= _sinittext && addr < _einittext))
> + return true;
> + return false;
> +}
> +
> +bool is_percpu_addr(long pc, long offset)
> +{
> + unsigned long ptr;
> + long address;
> +
> + address = pc + offset + 4;
> +
> + ptr = (unsigned long)address;
> +
> + if (ptr >= percpu_start && ptr < percpu_end)
> + return true;
> +
> + return false;
> +}
> +
> +static int cmp_section_addr(const void *a, const void *b)
> +{
> + unsigned long ptr = (unsigned long)a;
> + Elf64_Shdr *s = *(Elf64_Shdr **)b;
> + unsigned long end = s->sh_addr + s->sh_size;
> +
> + if (ptr >= s->sh_addr && ptr < end)
> + return 0;
> +
> + if (ptr < s->sh_addr)
> + return -1;
> +
> + return 1;
> +}
> +
> +/*
> + * Discover if the address is in a randomized section and if so, adjust
> + * by the saved offset.
> + */
> +Elf64_Shdr *adjust_address(long *address)
> +{
> + Elf64_Shdr **s;
> + Elf64_Shdr *shdr;
> +
> + if (nofgkaslr)
> + return NULL;
> +
> + s = bsearch((const void *)*address, sections, sections_size, sizeof(*s),
> + cmp_section_addr);
> + if (s) {
> + shdr = *s;
> + *address += shdr->sh_offset;
> + return shdr;
> + }
> +
> + return NULL;
> +}
> +
> +void adjust_relative_offset(long pc, long *value, Elf64_Shdr *section)
> +{
> + Elf64_Shdr *s;
> + long address;
> +
> + if (nofgkaslr)
> + return;
> +
> + /*
> + * sometimes we are updating a relative offset that would
> + * normally be relative to the next instruction (such as a call).
> + * In this case to calculate the target, you need to add 32bits to
> + * the pc to get the next instruction value. However, sometimes
> + * targets are just data that was stored in a table such as ksymtab
> + * or cpu alternatives. In this case our target is not relative to
> + * the next instruction.
> + */
Excellent and scary comment. ;) Was this found by trial and error? That
sounds "fun" to debug. :P
> +
> + /*
> + * Calculate the address that this offset would call.
> + */
> + if (!is_text(pc))
> + address = pc + *value;
> + else
> + address = pc + *value + 4;
> +
> + /*
> + * if the address is in section that was randomized,
> + * we need to adjust the offset.
> + */
> + s = adjust_address(&address);
> + if (s)
> + *value += s->sh_offset;
> +
> + /*
> + * If the PC that this offset was calculated for was in a section
> + * that has been randomized, the value needs to be adjusted by the
> + * same amount as the randomized section was adjusted from it's original
> + * location.
> + */
> + if (section)
> + *value -= section->sh_offset;
> +}
> +
> +static void kallsyms_swp(void *a, void *b, int size)
> +{
> + int idx1, idx2;
> + int temp;
> + struct kallsyms_name name_a;
> +
> + /* determine our index into the array */
> + idx1 = (int *)a - base;
> + idx2 = (int *)b - base;
> + temp = base[idx1];
> + base[idx1] = base[idx2];
> + base[idx2] = temp;
> +
> + /* also swap the names table */
> + memcpy(&name_a, &names_table[idx1], sizeof(name_a));
> + memcpy(&names_table[idx1], &names_table[idx2],
> + sizeof(struct kallsyms_name));
> + memcpy(&names_table[idx2], &name_a, sizeof(struct kallsyms_name));
> +}
> +
> +static int kallsyms_cmp(const void *a, const void *b)
> +{
> + int addr_a, addr_b;
> + unsigned long uaddr_a, uaddr_b;
> +
> + addr_a = *(int *)a;
> + addr_b = *(int *)b;
> +
> + if (addr_a >= 0)
> + uaddr_a = addr_a;
> + if (addr_b >= 0)
> + uaddr_b = addr_b;
> +
> + if (addr_a < 0)
> + uaddr_a = relative_base - 1 - addr_a;
> + if (addr_b < 0)
> + uaddr_b = relative_base - 1 - addr_b;
> +
> + if (uaddr_b > uaddr_a)
> + return -1;
> +
> + return 0;
> +}
> +
> +static void deal_with_names(int num_syms)
> +{
> + int num_bytes;
> + int i, j;
> + int offset;
> +
> + /* we should have num_syms kallsyms_name entries */
> + num_bytes = num_syms * sizeof(*names_table);
> + names_table = malloc(num_syms * sizeof(*names_table));
> + if (!names_table) {
> + debug_putstr("\nbytes requested: ");
> + debug_puthex(num_bytes);
> + error("\nunable to allocate space for names table\n");
> + }
> +
> + /* read all the names entries */
> + offset = 0;
> + for (i = 0; i < num_syms; i++) {
> + names_table[i].len = names[offset];
> + offset++;
> + for (j = 0; j < names_table[i].len; j++) {
> + names_table[i].indecis[j] = names[offset];
> + offset++;
> + }
> + }
> +}
> +
> +static void write_sorted_names(int num_syms)
> +{
> + int i, j;
> + int offset = 0;
> + unsigned int *markers;
> +
> + /*
> + * we are going to need to regenerate the markers table, which is a
> + * table of offsets into the compressed stream every 256 symbols.
> + * this code copied almost directly from scripts/kallsyms.c
> + */
Can any of this kallsyms sorting code be reasonably reused instead
of copy/pasting? Or is this mostly novel in that it's taking the output
of that earlier sort, which isn't the input format scripts/kallsyms.c
uses?
> + markers = malloc(sizeof(unsigned int) * ((num_syms + 255) / 256));
> + if (!markers) {
> + debug_putstr("\nfailed to allocate heap space of ");
> + debug_puthex(((num_syms + 255) / 256));
> + debug_putstr(" bytes\n");
> + error("Unable to allocate space for markers table");
> + }
> +
> + for (i = 0; i < num_syms; i++) {
> + if ((i & 0xFF) == 0)
> + markers[i >> 8] = offset;
> +
> + names[offset] = (u8)names_table[i].len;
> + offset++;
> + for (j = 0; j < names_table[i].len; j++) {
> + names[offset] = (u8)names_table[i].indecis[j];
> + offset++;
> + }
> + }
> +
> + /* write new markers table over old one */
> + for (i = 0; i < ((num_syms + 255) >> 8); i++)
> + markers_addr[i] = markers[i];
> +
> + free(markers);
> + free(names_table);
> +}
> +
> +static void sort_kallsyms(unsigned long map)
> +{
> + int num_syms;
> + int i;
> +
> + debug_putstr("\nRe-sorting kallsyms...\n");
> +
> + num_syms = *(int *)(kallsyms_num_syms + map);
> + base = (int *)(kallsyms_offsets + map);
> + relative_base = *(unsigned long *)(kallsyms_relative_base + map);
> + markers_addr = (unsigned int *)(kallsyms_markers + map);
> + names = (u8 *)(kallsyms_names + map);
> +
> + /*
> + * the kallsyms table was generated prior to any randomization.
> + * it is a bunch of offsets from "relative base". In order for
> + * us to check if a symbol has an address that was in a randomized
> + * section, we need to reconstruct the address to it's original
> + * value prior to handle_relocations.
> + */
> + for (i = 0; i < num_syms; i++) {
> + unsigned long addr;
> + int new_base;
> +
> + /*
> + * according to kernel/kallsyms.c, positive offsets are absolute
> + * values and negative offsets are relative to the base.
> + */
> + if (base[i] >= 0)
> + addr = base[i];
> + else
> + addr = relative_base - 1 - base[i];
> +
> + if (adjust_address(&addr)) {
> + /* here we need to recalcuate the offset */
> + new_base = relative_base - 1 - addr;
> + base[i] = new_base;
> + }
> + }
> +
> + /*
> + * here we need to read in all the kallsyms_names info
> + * so that we can regenerate it.
> + */
> + deal_with_names(num_syms);
> +
> + sort(base, num_syms, sizeof(int), kallsyms_cmp, kallsyms_swp);
> +
> + /* write the newly sorted names table over the old one */
> + write_sorted_names(num_syms);
> +}
> +
> +#include "../../../../lib/extable.c"
Now that the earlier linking glitches have been sorted out, I wonder if
it might be nicer to add this as a separate compilation unit, similar to
how early_serial_console.c is done? (Or, I guess, more specifically, why
can't this be in utils.c?)
> +
> +static inline unsigned long
> +ex_fixup_handler(const struct exception_table_entry *x)
> +{
> + return ((unsigned long)&x->handler + x->handler);
> +}
> +
> +static inline unsigned long
> +ex_fixup_addr(const struct exception_table_entry *x)
> +{
> + return ((unsigned long)&x->fixup + x->fixup);
> +}
> +
> +static void update_ex_table(unsigned long map)
> +{
> + struct exception_table_entry *start_ex_table =
> + (struct exception_table_entry *)(__start___ex_table_addr + map);
> + struct exception_table_entry *stop_ex_table =
> + (struct exception_table_entry *)(__stop___ex_table_addr + map);
> + int num_entries =
> + (__stop___ex_table_addr - __start___ex_table_addr) /
> + sizeof(struct exception_table_entry);
> + int i;
> +
> + debug_putstr("\nUpdating exception table...");
> + for (i = 0; i < num_entries; i++) {
> + unsigned long insn = ex_to_insn(&start_ex_table[i]);
> + unsigned long fixup = ex_fixup_addr(&start_ex_table[i]);
> + unsigned long handler = ex_fixup_handler(&start_ex_table[i]);
> + unsigned long addr;
> + Elf64_Shdr *s;
> +
> + /* check each address to see if it needs adjusting */
> + addr = insn - map;
> + s = adjust_address(&addr);
> + if (s)
> + start_ex_table[i].insn += s->sh_offset;
> +
> + addr = fixup - map;
> + s = adjust_address(&addr);
> + if (s)
> + start_ex_table[i].fixup += s->sh_offset;
> +
> + addr = handler - map;
> + s = adjust_address(&addr);
> + if (s)
> + start_ex_table[i].handler += s->sh_offset;
> + }
> +}
> +
> +static void sort_ex_table(unsigned long map)
> +{
> + struct exception_table_entry *start_ex_table =
> + (struct exception_table_entry *)(__start___ex_table_addr + map);
> + struct exception_table_entry *stop_ex_table =
> + (struct exception_table_entry *)(__stop___ex_table_addr + map);
> +
> + debug_putstr("\nRe-sorting exception table...");
> +
> + sort_extable(start_ex_table, stop_ex_table);
> +}
> +
> +static inline unsigned long orc_ip(const int *ip)
> +{
> + return (unsigned long)ip + *ip;
> +}
> +
> +static void orc_sort_swap(void *_a, void *_b, int size)
> +{
> + struct orc_entry *orc_a, *orc_b;
> + struct orc_entry orc_tmp;
> + int *a = _a, *b = _b, tmp;
> + int delta = _b - _a;
> +
> + /* Swap the .orc_unwind_ip entries: */
> + tmp = *a;
> + *a = *b + delta;
> + *b = tmp - delta;
> +
> + /* Swap the corresponding .orc_unwind entries: */
> + orc_a = cur_orc_table + (a - cur_orc_ip_table);
> + orc_b = cur_orc_table + (b - cur_orc_ip_table);
> + orc_tmp = *orc_a;
> + *orc_a = *orc_b;
> + *orc_b = orc_tmp;
> +}
> +
> +static int orc_sort_cmp(const void *_a, const void *_b)
> +{
> + struct orc_entry *orc_a;
> + const int *a = _a, *b = _b;
> + unsigned long a_val = orc_ip(a);
> + unsigned long b_val = orc_ip(b);
> +
> + if (a_val > b_val)
> + return 1;
> + if (a_val < b_val)
> + return -1;
> +
> + /*
> + * The "weak" section terminator entries need to always be on the left
> + * to ensure the lookup code skips them in favor of real entries.
> + * These terminator entries exist to handle any gaps created by
> + * whitelisted .o files which didn't get objtool generation.
> + */
> + orc_a = cur_orc_table + (a - cur_orc_ip_table);
> + return orc_a->sp_reg == ORC_REG_UNDEFINED && !orc_a->end ? -1 : 1;
> +}
> +
> +static void sort_orc_table(unsigned long map)
> +{
> + int num_entries =
> + (__stop_orc_unwind_ip - __start_orc_unwind_ip) / sizeof(int);
> +
> + cur_orc_ip_table = (int *)(__start_orc_unwind_ip + map);
> + cur_orc_table = (struct orc_entry *)(__start_orc_unwind + map);
> +
> + debug_putstr("\nRe-sorting orc tables...\n");
> + sort(cur_orc_ip_table, num_entries, sizeof(int), orc_sort_cmp,
> + orc_sort_swap);
> +}
> +
> +void post_relocations_cleanup(unsigned long map)
> +{
> + if (!nofgkaslr) {
> + update_ex_table(map);
> + sort_ex_table(map);
> + sort_orc_table(map);
> + }
> +
> + /*
> + * maybe one day free will do something. So, we "free" this memory
> + * in either case
> + */
> + free(sections);
> + free(sechdrs);
> +}
> +
> +void pre_relocations_cleanup(unsigned long map)
> +{
> + if (nofgkaslr)
> + return;
> +
> + sort_kallsyms(map);
> +}
> +
> +static void shuffle_sections(int *list, int size)
> +{
> + int i;
> + unsigned long j;
> + int temp;
> +
> + for (i = size - 1; i > 0; i--) {
> + j = kaslr_get_random_long(NULL) % (i + 1);
> +
> + temp = list[i];
> + list[i] = list[j];
> + list[j] = temp;
> + }
> +}
> +
> +static void move_text(int num_sections, char *secstrings, Elf64_Shdr *text,
> + void *source, void *dest, Elf64_Phdr *phdr)
> +{
> + unsigned long adjusted_addr;
> + int copy_bytes;
> + void *stash;
> + Elf64_Shdr **sorted_sections;
> + int *index_list;
> + int i, j;
> +
> + memmove(dest, source + text->sh_offset, text->sh_size);
> + copy_bytes = text->sh_size;
> + dest += text->sh_size;
> + adjusted_addr = text->sh_addr + text->sh_size;
> +
> + /*
> + * we leave the sections sorted in their original order
> + * by s->sh_addr, but shuffle the indexes in a random
> + * order for copying.
> + */
> + index_list = malloc(sizeof(int) * num_sections);
> + if (!index_list)
> + error("Failed to allocate space for index list");
> +
> + for (i = 0; i < num_sections; i++)
> + index_list[i] = i;
> +
> + shuffle_sections(index_list, num_sections);
> +
> + /*
> + * to avoid overwriting earlier sections before they can get
> + * copied to dest, stash everything into a buffer first.
> + * this will cause our source address to be off by
> + * phdr->p_offset though, so we'll adjust s->sh_offset below.
> + *
> + * TBD: ideally we'd simply decompress higher up so that our
> + * copy wasn't in danger of overwriting anything important.
> + */
> + stash = malloc(phdr->p_filesz);
Yeah, I'm glad this is captured in a comment. I'm fine leaving this
as it is for now, but it's certainly an area I'd like to look at again
in the future. I think we could gain some performance by removing this
copy. The KASLR code could choose two kernel-image-sized slots, decompress
into one (which would effectively be "stash" above) and the fgkaslr code
could write the resulting image into the second slot. For now, though,
this doesn't really matter: we just get a double-size image area and
copy the kernel around an extra time. :)
> + if (!stash)
> + error("Failed to allocate space for text stash");
> +
> + memcpy(stash, source + phdr->p_offset, phdr->p_filesz);
> +
> + /* now we'd walk through the sections. */
> + for (j = 0; j < num_sections; j++) {
> + unsigned long aligned_addr;
> + Elf64_Shdr *s;
> + const char *sname;
> + void *src;
> + int pad_bytes;
> +
> + s = sections[index_list[j]];
> +
> + sname = secstrings + s->sh_name;
> +
> + /* align addr for this section */
> + aligned_addr = ALIGN(adjusted_addr, s->sh_addralign);
> +
> + /*
> + * copy out of stash, so adjust offset
> + */
> + src = stash + s->sh_offset - phdr->p_offset;
> +
> + /*
> + * Fill any space between sections with int3
> + */
> + pad_bytes = aligned_addr - adjusted_addr;
> + memset(dest, 0xcc, pad_bytes);
> +
> + dest = (void *)ALIGN((unsigned long)dest, s->sh_addralign);
> +
> + memmove(dest, src, s->sh_size);
> +
> + dest += s->sh_size;
> + copy_bytes += s->sh_size + pad_bytes;
> + adjusted_addr = aligned_addr + s->sh_size;
> +
> + /* we can blow away sh_offset for our own uses */
> + s->sh_offset = aligned_addr - s->sh_addr;
> + }
> +
> + free(index_list);
> +
> + /*
> + * move remainder of text segment. Ok to just use original source
> + * here since this area is untouched.
> + */
> + memmove(dest, source + text->sh_offset + copy_bytes,
> + phdr->p_filesz - copy_bytes);
> + free(stash);
> +}
> +
> +#define GET_SYM(name) \
> + do { \
> + if (!name) { \
> + if (strcmp(#name, strtab + sym->st_name) == 0) {\
> + name = sym->st_value; \
> + continue; \
> + } \
> + } \
> + } while (0)
> +
> +static void parse_symtab(Elf64_Sym *symtab, char *strtab, long num_syms)
> +{
> + Elf64_Sym *sym;
> +
> + if (!symtab || !strtab)
> + return;
> +
> + debug_putstr("\nLooking for symbols... ");
> +
> + /*
> + * walk through the symbol table looking for the symbols
> + * that we care about.
> + */
> + for (sym = symtab; --num_syms >= 0; sym++) {
> + if (!sym->st_name)
> + continue;
> +
> + GET_SYM(kallsyms_num_syms);
> + GET_SYM(kallsyms_offsets);
> + GET_SYM(kallsyms_relative_base);
> + GET_SYM(kallsyms_names);
> + GET_SYM(kallsyms_markers);
> + GET_SYM(_stext);
> + GET_SYM(_etext);
> + GET_SYM(_sinittext);
> + GET_SYM(_einittext);
> + GET_SYM(__start_orc_unwind_ip);
> + GET_SYM(__stop_orc_unwind_ip);
> + GET_SYM(__start_orc_unwind);
> +
> + /*
> + * the GET_SYM macro can't be used here because these have
> + * to be renamed due to the inclusion of lib/extable.c
> + */
Perhaps just globally prefix all the needed image symbols with "loader_"
or "fgkaslr_" or "img_", and then you don't need to open-code this one
with a different target name. Something like this:
#define GET_SYM(name) \
do { \
if (!img_ ## name) { \
if (strcmp(#name, strtab + sym->st_name) == 0) {\
img_ ## name = sym->st_value; \
continue; \
} \
} \
} while (0)
Then there is also a consistent naming convention within the code here
to easily see where addresses came from, and you avoid any future
collisions like are being avoided for the ex_table symbols.
> + if (!__start___ex_table_addr) {
> + if (strcmp("__start___ex_table",
> + strtab + sym->st_name) == 0) {
> + __start___ex_table_addr = sym->st_value;
> + continue;
> + }
> + }
> +
> + if (!__stop___ex_table_addr) {
> + if (strcmp("__stop___ex_table",
> + strtab + sym->st_name) == 0) {
> + __stop___ex_table_addr = sym->st_value;
> + continue;
> + }
> + }
> + }
> +}
> +
> +void parse_sections_headers(void *output, Elf64_Ehdr *ehdr, Elf64_Phdr *phdrs)
> +{
> + Elf64_Phdr *phdr;
> + Elf64_Shdr *s;
> + Elf64_Shdr *text = NULL;
> + Elf64_Shdr *percpu = NULL;
> + char *secstrings;
> + const char *sname;
> + int num_sections = 0;
> + Elf64_Sym *symtab = NULL;
> + char *strtab = NULL;
> + long num_syms = 0;
> + void *dest;
> + int i;
> + char arg[MAX_FGKASLR_ARG_LENGTH];
> + Elf64_Shdr shdr;
> + unsigned long shnum;
> + unsigned int shstrndx;
> +
> + debug_putstr("\nParsing ELF section headers... ");
> +
> + /*
> + * Even though fgkaslr may have been disabled, we still
> + * need to parse through the section headers to get the
> + * start and end of the percpu section. This is because
> + * if we were built with CONFIG_FG_KASLR, there are more
> + * relative relocations present in vmlinux.relocs than
> + * just the percpu, and only the percpu relocs need to be
> + * adjusted when using just normal base address kaslr.
> + */
> + if (cmdline_find_option("fgkaslr", arg, sizeof(arg)) == 3 &&
> + !strncmp(arg, "off", 3)) {
> + warn("FG_KASLR disabled on cmdline.");
> + nofgkaslr = 1;
> + }
> +
> + /* read the first section header */
> + shnum = ehdr->e_shnum;
> + shstrndx = ehdr->e_shstrndx;
> + if (shnum == SHN_UNDEF || shstrndx == SHN_XINDEX) {
> + memcpy(&shdr, output + ehdr->e_shoff, sizeof(shdr));
> + if (shnum == SHN_UNDEF)
> + shnum = shdr.sh_size;
> + if (shstrndx == SHN_XINDEX)
> + shstrndx = shdr.sh_link;
> + }
> +
> + /* we are going to need to allocate space for the section headers */
> + sechdrs = malloc(sizeof(*sechdrs) * shnum);
> + if (!sechdrs)
> + error("Failed to allocate space for shdrs");
> +
> + sections = malloc(sizeof(*sections) * shnum);
> + if (!sections)
> + error("Failed to allocate space for section pointers");
> +
> + memcpy(sechdrs, output + ehdr->e_shoff,
> + sizeof(*sechdrs) * shnum);
> +
> + /* we need to allocate space for the section string table */
> + s = &sechdrs[shstrndx];
> +
> + secstrings = malloc(s->sh_size);
> + if (!secstrings)
> + error("Failed to allocate space for shstr");
> +
> + memcpy(secstrings, output + s->sh_offset, s->sh_size);
> +
> + /*
> + * now we need to walk through the section headers and collect the
> + * sizes of the .text sections to be randomized.
> + */
> + for (i = 0; i < shnum; i++) {
> + s = &sechdrs[i];
> + sname = secstrings + s->sh_name;
> +
> + if (s->sh_type == SHT_SYMTAB) {
> + /* only one symtab per image */
I would assert this claim in the code as well: check symtab is NULL here?
> + symtab = malloc(s->sh_size);
> + if (!symtab)
> + error("Failed to allocate space for symtab");
> +
> + memcpy(symtab, output + s->sh_offset, s->sh_size);
> + num_syms = s->sh_size / sizeof(*symtab);
> + continue;
> + }
> +
> + if (s->sh_type == SHT_STRTAB && i != ehdr->e_shstrndx) {
> + strtab = malloc(s->sh_size);
Similar robustness check?
> + if (!strtab)
> + error("Failed to allocate space for strtab");
> +
> + memcpy(strtab, output + s->sh_offset, s->sh_size);
> + }
> +
> + if (!strcmp(sname, ".text")) {
> + text = s;
> + continue;
> + }
Check text is still NULL here?
Also, why the continue? This means the section isn't included in the
sections[] array? (Obviously things still work, but I don't understand
why.)
> +
> + if (!strcmp(sname, ".data..percpu")) {
> + /* get start addr for later */
> + percpu = s;
Similar, check percpu is still NULL here?
Also, is a "continue" intended here? (This is kind of the reverse of
the "continue" question above.) I think you get the same result during
the next "if", but I was expecting this block to look like the .text
test above.
> + }
> +
> + if (!(s->sh_flags & SHF_ALLOC) ||
> + !(s->sh_flags & SHF_EXECINSTR) ||
> + !(strstarts(sname, ".text")))
> + continue;
> +
> + sections[num_sections] = s;
> +
> + num_sections++;
> + }
> + sections[num_sections] = NULL;
> + sections_size = num_sections;
> +
> + parse_symtab(symtab, strtab, num_syms);
> +
> + for (i = 0; i < ehdr->e_phnum; i++) {
> + phdr = &phdrs[i];
> +
> + switch (phdr->p_type) {
> + case PT_LOAD:
> + if ((phdr->p_align % 0x200000) != 0)
> + error("Alignment of LOAD segment isn't multiple of 2MB");
> + dest = output;
> + dest += (phdr->p_paddr - LOAD_PHYSICAL_ADDR);
> + if (!nofgkaslr &&
> + (text && phdr->p_offset == text->sh_offset)) {
> + move_text(num_sections, secstrings, text,
> + output, dest, phdr);
> + } else {
> + if (percpu &&
> + phdr->p_offset == percpu->sh_offset) {
> + percpu_start = percpu->sh_addr;
> + percpu_end = percpu_start +
> + phdr->p_filesz;
> + }
> + memmove(dest, output + phdr->p_offset,
> + phdr->p_filesz);
> + }
> + break;
> + default: /* Ignore other PT_* */
> + break;
> + }
> + }
> +
> + /* we need to keep the section info to redo relocs */
> + free(secstrings);
> +
> + free(phdrs);
> +}
> +
> diff --git a/arch/x86/boot/compressed/kaslr.c b/arch/x86/boot/compressed/kaslr.c
> index d7408af55738..6f596bd5b6e5 100644
> --- a/arch/x86/boot/compressed/kaslr.c
> +++ b/arch/x86/boot/compressed/kaslr.c
> @@ -39,10 +39,6 @@
> #include <generated/utsrelease.h>
> #include <asm/efi.h>
>
> -/* Macros used by the included decompressor code below. */
> -#define STATIC
> -#include <linux/decompress/mm.h>
> -
> #ifdef CONFIG_X86_5LEVEL
> unsigned int __pgtable_l5_enabled;
> unsigned int pgdir_shift __ro_after_init = 39;
> diff --git a/arch/x86/boot/compressed/misc.c b/arch/x86/boot/compressed/misc.c
> index 9652d5c2afda..5f08922fd12a 100644
> --- a/arch/x86/boot/compressed/misc.c
> +++ b/arch/x86/boot/compressed/misc.c
> @@ -26,9 +26,6 @@
> * it is not safe to place pointers in static structures.
> */
>
> -/* Macros used by the included decompressor code below. */
> -#define STATIC static
Can you split the STATIC macro rearrangement out into a separate patch?
I think it deserves a dedicated commit log to explain why it's changing
the way it is (i.e. we end up with "STATIC" no longer meaning "static"
in misc.c now...)
> -
> /*
> * Use normal definitions of mem*() from string.c. There are already
> * included header files which expect a definition of memset() and by
> @@ -49,6 +46,10 @@ struct boot_params *boot_params;
>
> memptr free_mem_ptr;
> memptr free_mem_end_ptr;
> +#ifdef CONFIG_FG_KASLR
> +unsigned long malloc_ptr;
> +int malloc_count;
> +#endif
>
> static char *vidmem;
> static int vidport;
> @@ -203,10 +204,21 @@ static void handle_relocations(void *output, unsigned long output_len,
> if (IS_ENABLED(CONFIG_X86_64))
> delta = virt_addr - LOAD_PHYSICAL_ADDR;
>
> - if (!delta) {
> - debug_putstr("No relocation needed... ");
> - return;
> + /*
> + * it is possible to have delta be zero and
> + * still have enabled fg kaslr. We need to perform relocations
> + * for fgkaslr regardless of whether the base address has moved.
> + */
Please reflow these sentences (long whitespace at the end of line one).
> + if (!IS_ENABLED(CONFIG_FG_KASLR) ||
> + cmdline_find_option_bool("nokaslr")) {
> + if (!delta) {
> + debug_putstr("No relocation needed... ");
> + return;
> + }
> }
> +
> + pre_relocations_cleanup(map);
> +
> debug_putstr("Performing relocations... ");
>
> /*
> @@ -230,35 +242,106 @@ static void handle_relocations(void *output, unsigned long output_len,
> */
> for (reloc = output + output_len - sizeof(*reloc); *reloc; reloc--) {
> long extended = *reloc;
> + long value;
> +
> + /*
> + * if using fgkaslr, we might have moved the address
> + * of the relocation. Check it to see if it needs adjusting
> + * from the original address.
> + */
> + (void)adjust_address(&extended);
I don't think "(void)" needed here?
> +
> extended += map;
>
> ptr = (unsigned long)extended;
> if (ptr < min_addr || ptr > max_addr)
> error("32-bit relocation outside of kernel!\n");
>
> - *(uint32_t *)ptr += delta;
> + value = *(int32_t *)ptr;
> +
> + /*
> + * If using fgkaslr, the value of the relocation
> + * might need to be changed because it referred
> + * to an address that has moved.
> + */
> + adjust_address(&value);
> +
> + value += delta;
> +
> + *(uint32_t *)ptr = value;
> }
> #ifdef CONFIG_X86_64
> while (*--reloc) {
> long extended = *reloc;
> + long value;
> + long oldvalue;
> + Elf64_Shdr *s;
> +
> + /*
> + * if using fgkaslr, we might have moved the address
> + * of the relocation. Check it to see if it needs adjusting
> + * from the original address.
> + */
> + s = adjust_address(&extended);
> +
> extended += map;
>
> ptr = (unsigned long)extended;
> if (ptr < min_addr || ptr > max_addr)
> error("inverse 32-bit relocation outside of kernel!\n");
>
> - *(int32_t *)ptr -= delta;
> + value = *(int32_t *)ptr;
> + oldvalue = value;
> +
> + /*
> + * If using fgkaslr, these relocs will contain
> + * relative offsets which might need to be
> + * changed because it referred
> + * to an address that has moved.
> + */
> + adjust_relative_offset(*reloc, &value, s);
> +
> + /*
> + * only percpu symbols need to have their values adjusted for
> + * base address kaslr since relative offsets within the .text
> + * and .text.* sections are ok wrt each other.
> + */
> + if (is_percpu_addr(*reloc, oldvalue))
> + value -= delta;
> +
> + *(int32_t *)ptr = value;
> }
> for (reloc--; *reloc; reloc--) {
> long extended = *reloc;
> + long value;
> +
> + /*
> + * if using fgkaslr, we might have moved the address
> + * of the relocation. Check it to see if it needs adjusting
> + * from the original address.
> + */
> + (void)adjust_address(&extended);
> +
> extended += map;
>
> ptr = (unsigned long)extended;
> if (ptr < min_addr || ptr > max_addr)
> error("64-bit relocation outside of kernel!\n");
>
> - *(uint64_t *)ptr += delta;
> + value = *(int64_t *)ptr;
> +
> + /*
> + * If using fgkaslr, the value of the relocation
> + * might need to be changed because it referred
> + * to an address that has moved.
> + */
> + (void)adjust_address(&value);
> +
> + value += delta;
> +
> + *(uint64_t *)ptr = value;
> }
> + post_relocations_cleanup(map);
> #endif
> }
> #else
> @@ -296,6 +379,14 @@ static void parse_elf(void *output)
>
> memcpy(phdrs, output + ehdr.e_phoff, sizeof(*phdrs) * ehdr.e_phnum);
>
> + if (IS_ENABLED(CONFIG_FG_KASLR)) {
> + if (!cmdline_find_option_bool("nokaslr")) {
> + parse_sections_headers(output, &ehdr, phdrs);
> + return;
Hmmm. There is repeated logic between parse_sections_headers() and the
rest of parse_elf() here. I think if they should stay distinct, I would
split the remaining code in parse_elf() into a separate function, and
then parse_elf() becomes more readable as:
void parse_elf(...)
{
if (fgkaslr)
layout_fgkaslr(...)
else
layout_normal(...)
}
Does "layout" seem to be a good name here? Maybe "parse_elf" needs to
be "layout_image"? Yay name bikesheds. My main point is that the
"return" above feels abrupt, and begs the question, "why aren't we doing
this other stuff?" But if it's a boolean between two layout functions,
that's easy to understand, IMO.
> + }
> + warn("FG_KASLR disabled: 'nokaslr' on cmdline.");
> + }
> +
> for (i = 0; i < ehdr.e_phnum; i++) {
> phdr = &phdrs[i];
>
> diff --git a/arch/x86/boot/compressed/misc.h b/arch/x86/boot/compressed/misc.h
> index 726e264410ff..e8e45f263eaf 100644
> --- a/arch/x86/boot/compressed/misc.h
> +++ b/arch/x86/boot/compressed/misc.h
> @@ -39,6 +39,12 @@
> /* misc.c */
> extern memptr free_mem_ptr;
> extern memptr free_mem_end_ptr;
> +#define STATIC
> +#ifdef CONFIG_FG_KASLR
> +#define STATIC_RW_DATA extern
> +#endif
> +#include <linux/decompress/mm.h>
> +
> extern struct boot_params *boot_params;
> void __putstr(const char *s);
> void __puthex(unsigned long value);
> @@ -74,6 +80,34 @@ struct mem_vector {
> unsigned long long size;
> };
>
> +#ifdef CONFIG_X86_64
> +#define Elf_Ehdr Elf64_Ehdr
> +#define Elf_Phdr Elf64_Phdr
> +#define Elf_Shdr Elf64_Shdr
> +#else
> +#define Elf_Ehdr Elf32_Ehdr
> +#define Elf_Phdr Elf32_Phdr
> +#define Elf_Shdr Elf32_Shdr
> +#endif
> +
> +#if CONFIG_FG_KASLR
> +void parse_sections_headers(void *output, Elf_Ehdr *ehdr, Elf_Phdr *phdrs);
> +void pre_relocations_cleanup(unsigned long map);
> +void post_relocations_cleanup(unsigned long map);
> +Elf_Shdr *adjust_address(long *address);
> +void adjust_relative_offset(long pc, long *value, Elf_Shdr *section);
> +bool is_percpu_addr(long pc, long offset);
> +#else
> +static inline void parse_sections_headers(void *output, Elf_Ehdr *ehdr,
> + Elf_Phdr *phdrs) { }
> +static inline void pre_relocations_cleanup(unsigned long map) { }
> +static inline void post_relocations_cleanup(unsigned long map) { }
> +static inline Elf_Shdr *adjust_address(long *address) { return NULL; }
> +static inline void adjust_relative_offset(long pc, long *value,
> + Elf_Shdr *section) { }
> +static inline bool is_percpu_addr(long pc, long offset) { return true; }
> +#endif /* CONFIG_FG_KASLR */
> +
> #if CONFIG_RANDOMIZE_BASE
> /* kaslr.c */
> void choose_random_location(unsigned long input,
> diff --git a/arch/x86/boot/compressed/utils.c b/arch/x86/boot/compressed/utils.c
> new file mode 100644
> index 000000000000..ceefc58d7c71
> --- /dev/null
> +++ b/arch/x86/boot/compressed/utils.c
> @@ -0,0 +1,12 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * utils.c
> + *
> + * This contains various libraries that are needed for fgkaslr
> + */
> +#define __DISABLE_EXPORTS
> +#define _LINUX_KPROBES_H
> +#define NOKPROBE_SYMBOL(fname)
> +#include "../../../../lib/sort.c"
> +#include "../../../../lib/bsearch.c"
> +
> diff --git a/arch/x86/boot/compressed/vmlinux.symbols b/arch/x86/boot/compressed/vmlinux.symbols
> new file mode 100644
> index 000000000000..cc86e79a2a3d
> --- /dev/null
> +++ b/arch/x86/boot/compressed/vmlinux.symbols
> @@ -0,0 +1,17 @@
> +kallsyms_offsets
> +kallsyms_addresses
> +kallsyms_num_syms
> +kallsyms_relative_base
> +kallsyms_names
> +kallsyms_token_table
> +kallsyms_token_index
> +kallsyms_markers
> +__start___ex_table
> +__stop___ex_table
> +_sinittext
> +_einittext
> +_stext
> +_etext
> +__start_orc_unwind_ip
> +__stop_orc_unwind_ip
> +__start_orc_unwind
> diff --git a/arch/x86/include/asm/boot.h b/arch/x86/include/asm/boot.h
> index 680c320363db..6918d33eb5ef 100644
> --- a/arch/x86/include/asm/boot.h
> +++ b/arch/x86/include/asm/boot.h
> @@ -26,8 +26,19 @@
>
> #ifdef CONFIG_KERNEL_BZIP2
> # define BOOT_HEAP_SIZE 0x400000
> -#else /* !CONFIG_KERNEL_BZIP2 */
> -# define BOOT_HEAP_SIZE 0x10000
> +#elif CONFIG_FG_KASLR
> +/*
> + * We need extra boot heap when using fgkaslr because we make a copy
> + * of the original decompressed kernel to avoid issues with writing
> + * over ourselves when shuffling the sections. We also need extra
> + * space for resorting kallsyms after shuffling. This value could
> + * be decreased if free() would release memory properly, or if we
> + * could avoid the kernel copy. It would need to be increased if we
> + * find additional tables that need to be resorted.
> + */
> +# define BOOT_HEAP_SIZE 0x4000000
> +#else /* !CONFIG_KERNEL_BZIP2 && !CONFIG_FG_KASLR */
> +# define BOOT_HEAP_SIZE 0x10000
> #endif
>
> #ifdef CONFIG_X86_64
> diff --git a/include/uapi/linux/elf.h b/include/uapi/linux/elf.h
> index 34c02e4290fe..a85d1792d5a8 100644
> --- a/include/uapi/linux/elf.h
> +++ b/include/uapi/linux/elf.h
> @@ -298,6 +298,7 @@ typedef struct elf64_phdr {
> #define SHN_LIVEPATCH 0xff20
> #define SHN_ABS 0xfff1
> #define SHN_COMMON 0xfff2
> +#define SHN_XINDEX 0xffff
> #define SHN_HIRESERVE 0xffff
>
> typedef struct elf32_shdr {
> --
> 2.20.1
>
Very exciting! :) I'm already looking forward to v3. :)
--
Kees Cook