Message ID | 1447672998-20981-6-git-send-email-ard.biesheuvel@linaro.org |
---|---|
State | New |
Headers | show |
On Mon, Nov 16, 2015 at 12:23:16PM +0100, Ard Biesheuvel wrote: > diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S > index 23cfc08fc8ba..d3e4b5d6a8d2 100644 > --- a/arch/arm64/kernel/head.S > +++ b/arch/arm64/kernel/head.S [...] > @@ -210,7 +210,15 @@ section_table: > ENTRY(stext) > bl preserve_boot_args > bl el2_setup // Drop to EL1, w20=cpu_boot_mode > - adrp x24, __PHYS_OFFSET > + > + /* > + * Before the linear mapping has been set up, __va() translations will > + * not produce usable virtual addresses unless we tweak PHYS_OFFSET to > + * compensate for the offset between the kernel mapping and the base of > + * the linear mapping. We will undo this in map_mem(). > + */ Minor typo in comment: I guess you meant "__pa() translations will not produce usable...". > diff --git a/arch/arm64/mm/dump.c b/arch/arm64/mm/dump.c > index 5a22a119a74c..f6272f450688 100644 > --- a/arch/arm64/mm/dump.c > +++ b/arch/arm64/mm/dump.c > @@ -63,7 +63,8 @@ static struct addr_marker address_markers[] = { > { PCI_IO_END, "PCI I/O end" }, > { MODULES_VADDR, "Modules start" }, > { MODULES_END, "Modules end" }, > - { PAGE_OFFSET, "Kernel Mapping" }, > + { KIMAGE_VADDR, "Kernel Mapping" }, > + { PAGE_OFFSET, "Linear Mapping" }, > { -1, NULL }, > }; Apart from this, please change the pr_notice() in mem_init() to show the linear mapping at the end (keep them in ascending order). -- Catalin _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
On 7 December 2015 at 13:26, Catalin Marinas <catalin.marinas@arm.com> wrote: > On Mon, Nov 16, 2015 at 12:23:16PM +0100, Ard Biesheuvel wrote: >> diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S >> index 23cfc08fc8ba..d3e4b5d6a8d2 100644 >> --- a/arch/arm64/kernel/head.S >> +++ b/arch/arm64/kernel/head.S > [...] >> @@ -210,7 +210,15 @@ section_table: >> ENTRY(stext) >> bl preserve_boot_args >> bl el2_setup // Drop to EL1, w20=cpu_boot_mode >> - adrp x24, __PHYS_OFFSET >> + >> + /* >> + * Before the linear mapping has been set up, __va() translations will >> + * not produce usable virtual addresses unless we tweak PHYS_OFFSET to >> + * compensate for the offset between the kernel mapping and the base of >> + * the linear mapping. We will undo this in map_mem(). >> + */ > > Minor typo in comment: I guess you meant "__pa() translations will not > produce usable...". > No, not quite. __va() translations will normally produce addresses in the linear mapping, which will not be set up when we first start using it in create_mapping(). So until that time, we have to redirect __va() translations into the kernel mapping, where swapper_pg_dir is shadowed. I am hoping that Mark's planned changes to create_mapping() will make this unnecessary, but I haven't seen any of his code yet. As far as __pa() is concerned, that translation is actually tweaked so it will always produce usable addresses, regardless of whether the bias is still set or not. The reason is that va-to-pa translations are always unambiguous. >> diff --git a/arch/arm64/mm/dump.c b/arch/arm64/mm/dump.c >> index 5a22a119a74c..f6272f450688 100644 >> --- a/arch/arm64/mm/dump.c >> +++ b/arch/arm64/mm/dump.c >> @@ -63,7 +63,8 @@ static struct addr_marker address_markers[] = { >> { PCI_IO_END, "PCI I/O end" }, >> { MODULES_VADDR, "Modules start" }, >> { MODULES_END, "Modules end" }, >> - { PAGE_OFFSET, "Kernel Mapping" }, >> + { KIMAGE_VADDR, "Kernel Mapping" }, >> + { PAGE_OFFSET, "Linear Mapping" }, >> { -1, NULL }, >> }; > > Apart from this, please change the pr_notice() in mem_init() to show the > linear mapping at the end (keep them in ascending order). > OK _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
On 7 December 2015 at 13:33, Ard Biesheuvel <ard.biesheuvel@linaro.org> wrote: > On 7 December 2015 at 13:26, Catalin Marinas <catalin.marinas@arm.com> wrote: >> On Mon, Nov 16, 2015 at 12:23:16PM +0100, Ard Biesheuvel wrote: >>> diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S >>> index 23cfc08fc8ba..d3e4b5d6a8d2 100644 >>> --- a/arch/arm64/kernel/head.S >>> +++ b/arch/arm64/kernel/head.S >> [...] >>> @@ -210,7 +210,15 @@ section_table: >>> ENTRY(stext) >>> bl preserve_boot_args >>> bl el2_setup // Drop to EL1, w20=cpu_boot_mode >>> - adrp x24, __PHYS_OFFSET >>> + >>> + /* >>> + * Before the linear mapping has been set up, __va() translations will >>> + * not produce usable virtual addresses unless we tweak PHYS_OFFSET to >>> + * compensate for the offset between the kernel mapping and the base of >>> + * the linear mapping. We will undo this in map_mem(). >>> + */ >> >> Minor typo in comment: I guess you meant "__pa() translations will not >> produce usable...". >> > > No, not quite. __va() translations will normally produce addresses in > the linear mapping, which will not be set up when we first start using > it in create_mapping(). So until that time, we have to redirect __va() > translations into the kernel mapping, where swapper_pg_dir is > shadowed. I am hoping that Mark's planned changes to create_mapping() > will make this unnecessary, but I haven't seen any of his code yet. > > As far as __pa() is concerned, that translation is actually tweaked so > it will always produce usable addresses, regardless of whether the > bias is still set or not. The reason is that va-to-pa translations are > always unambiguous. > ... so of course, the comment is still wrong, -> s/virtual/physical/ addresses >>> diff --git a/arch/arm64/mm/dump.c b/arch/arm64/mm/dump.c >>> index 5a22a119a74c..f6272f450688 100644 >>> --- a/arch/arm64/mm/dump.c >>> +++ b/arch/arm64/mm/dump.c >>> @@ -63,7 +63,8 @@ static struct addr_marker address_markers[] = { >>> { PCI_IO_END, "PCI I/O end" }, >>> { MODULES_VADDR, "Modules start" }, >>> { MODULES_END, "Modules end" }, >>> - { PAGE_OFFSET, "Kernel Mapping" }, >>> + { KIMAGE_VADDR, "Kernel Mapping" }, >>> + { PAGE_OFFSET, "Linear Mapping" }, >>> { -1, NULL }, >>> }; >> >> Apart from this, please change the pr_notice() in mem_init() to show the >> linear mapping at the end (keep them in ascending order). >> > > OK _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
On Mon, Dec 07, 2015 at 01:34:19PM +0100, Ard Biesheuvel wrote: > On 7 December 2015 at 13:33, Ard Biesheuvel <ard.biesheuvel@linaro.org> wrote: > > On 7 December 2015 at 13:26, Catalin Marinas <catalin.marinas@arm.com> wrote: > >> On Mon, Nov 16, 2015 at 12:23:16PM +0100, Ard Biesheuvel wrote: > >>> diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S > >>> index 23cfc08fc8ba..d3e4b5d6a8d2 100644 > >>> --- a/arch/arm64/kernel/head.S > >>> +++ b/arch/arm64/kernel/head.S > >> [...] > >>> @@ -210,7 +210,15 @@ section_table: > >>> ENTRY(stext) > >>> bl preserve_boot_args > >>> bl el2_setup // Drop to EL1, w20=cpu_boot_mode > >>> - adrp x24, __PHYS_OFFSET > >>> + > >>> + /* > >>> + * Before the linear mapping has been set up, __va() translations will > >>> + * not produce usable virtual addresses unless we tweak PHYS_OFFSET to > >>> + * compensate for the offset between the kernel mapping and the base of > >>> + * the linear mapping. We will undo this in map_mem(). > >>> + */ > >> > >> Minor typo in comment: I guess you meant "__pa() translations will not > >> produce usable...". > > > > No, not quite. __va() translations will normally produce addresses in > > the linear mapping, which will not be set up when we first start using > > it in create_mapping(). So until that time, we have to redirect __va() > > translations into the kernel mapping, where swapper_pg_dir is > > shadowed. I guessed what you meant and I remember the reason based on past discussions, only that to me "linear mapping" sounds like something in virtual space while __va() generates a linear mapping -> physical translation (just some wording, nothing serious). > > I am hoping that Mark's planned changes to create_mapping() > > will make this unnecessary, but I haven't seen any of his code yet. Not sure, I haven't seen the details yet. > > As far as __pa() is concerned, that translation is actually tweaked so > > it will always produce usable addresses, regardless of whether the > > bias is still set or not. The reason is that va-to-pa translations are > > always unambiguous. Only that very early during boot memstart_addr is still based on the kernel load address rather than memblock_start_of_DRAM(), that's why I thought you meant __pa(). > ... so of course, the comment is still wrong, -> s/virtual/physical/ > addresses This would do. -- Catalin _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
diff --git a/arch/arm64/include/asm/boot.h b/arch/arm64/include/asm/boot.h index 81151b67b26b..092d1096ce9a 100644 --- a/arch/arm64/include/asm/boot.h +++ b/arch/arm64/include/asm/boot.h @@ -11,4 +11,11 @@ #define MIN_FDT_ALIGN 8 #define MAX_FDT_SIZE SZ_2M +/* + * arm64 requires the kernel image to be 2 MB aligned and + * not exceed 64 MB in size. + */ +#define MIN_KIMG_ALIGN SZ_2M +#define MAX_KIMG_SIZE SZ_64M + #endif diff --git a/arch/arm64/include/asm/kernel-pgtable.h b/arch/arm64/include/asm/kernel-pgtable.h index a459714ee29e..daa8a7b9917a 100644 --- a/arch/arm64/include/asm/kernel-pgtable.h +++ b/arch/arm64/include/asm/kernel-pgtable.h @@ -70,8 +70,9 @@ /* * Initial memory map attributes. */ -#define SWAPPER_PTE_FLAGS (PTE_TYPE_PAGE | PTE_AF | PTE_SHARED) -#define SWAPPER_PMD_FLAGS (PMD_TYPE_SECT | PMD_SECT_AF | PMD_SECT_S) +#define SWAPPER_PTE_FLAGS (PTE_TYPE_PAGE | PTE_AF | PTE_SHARED | PTE_UXN) +#define SWAPPER_PMD_FLAGS (PMD_TYPE_SECT | PMD_SECT_AF | PMD_SECT_S | \ + PMD_SECT_UXN) #if ARM64_SWAPPER_USES_SECTION_MAPS #define SWAPPER_MM_MMUFLAGS (PMD_ATTRINDX(MT_NORMAL) | SWAPPER_PMD_FLAGS) diff --git a/arch/arm64/include/asm/memory.h b/arch/arm64/include/asm/memory.h index 853953cd1f08..3148691bc80a 100644 --- a/arch/arm64/include/asm/memory.h +++ b/arch/arm64/include/asm/memory.h @@ -24,6 +24,7 @@ #include <linux/compiler.h> #include <linux/const.h> #include <linux/types.h> +#include <asm/boot.h> #include <asm/sizes.h> /* @@ -39,7 +40,12 @@ #define PCI_IO_SIZE SZ_16M /* - * PAGE_OFFSET - the virtual address of the start of the kernel image (top + * Offset below PAGE_OFFSET where to map the kernel Image. + */ +#define KIMAGE_OFFSET MAX_KIMG_SIZE + +/* + * PAGE_OFFSET - the virtual address of the base of the linear mapping (top * (VA_BITS - 1)) * VA_BITS - the maximum number of bits for virtual addresses. * VA_START - the first kernel virtual address. @@ -51,7 +57,8 @@ #define VA_BITS (CONFIG_ARM64_VA_BITS) #define VA_START (UL(0xffffffffffffffff) << VA_BITS) #define PAGE_OFFSET (UL(0xffffffffffffffff) << (VA_BITS - 1)) -#define MODULES_END (PAGE_OFFSET) +#define KIMAGE_VADDR (PAGE_OFFSET - KIMAGE_OFFSET) +#define MODULES_END KIMAGE_VADDR #define MODULES_VADDR (MODULES_END - SZ_64M) #define PCI_IO_END (MODULES_VADDR - SZ_2M) #define PCI_IO_START (PCI_IO_END - PCI_IO_SIZE) @@ -75,7 +82,11 @@ * private definitions which should NOT be used outside memory.h * files. Use virt_to_phys/phys_to_virt/__pa/__va instead. */ -#define __virt_to_phys(x) (((phys_addr_t)(x) - PAGE_OFFSET + PHYS_OFFSET)) +#define __virt_to_phys(x) ({ \ + long __x = (long)(x) - PAGE_OFFSET; \ + __x >= 0 ? (phys_addr_t)(__x + PHYS_OFFSET) : \ + (phys_addr_t)(__x + PHYS_OFFSET + kernel_va_offset); }) + #define __phys_to_virt(x) ((unsigned long)((x) - PHYS_OFFSET + PAGE_OFFSET)) /* @@ -106,6 +117,8 @@ extern phys_addr_t memstart_addr; /* PHYS_OFFSET - the physical address of the start of memory. */ #define PHYS_OFFSET ({ memstart_addr; }) +extern u64 kernel_va_offset; + /* * The maximum physical address that the linear direct mapping * of system RAM can cover. (PAGE_OFFSET can be interpreted as @@ -113,6 +126,7 @@ extern phys_addr_t memstart_addr; * maximum size of the linear mapping.) */ #define MAX_MEMBLOCK_ADDR ({ memstart_addr - PAGE_OFFSET - 1; }) +#define MIN_MEMBLOCK_ADDR __pa(KIMAGE_VADDR) /* * PFNs are used to describe any physical page; this means diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S index 23cfc08fc8ba..d3e4b5d6a8d2 100644 --- a/arch/arm64/kernel/head.S +++ b/arch/arm64/kernel/head.S @@ -38,8 +38,6 @@ #include <asm/thread_info.h> #include <asm/virt.h> -#define __PHYS_OFFSET (KERNEL_START - TEXT_OFFSET) - #if (TEXT_OFFSET & 0xfff) != 0 #error TEXT_OFFSET must be at least 4KB aligned #elif (PAGE_OFFSET & 0x1fffff) != 0 @@ -50,6 +48,8 @@ #define KERNEL_START _text #define KERNEL_END _end +#define KERNEL_BASE (KERNEL_START - TEXT_OFFSET) + /* * Kernel startup entry point. @@ -210,7 +210,15 @@ section_table: ENTRY(stext) bl preserve_boot_args bl el2_setup // Drop to EL1, w20=cpu_boot_mode - adrp x24, __PHYS_OFFSET + + /* + * Before the linear mapping has been set up, __va() translations will + * not produce usable virtual addresses unless we tweak PHYS_OFFSET to + * compensate for the offset between the kernel mapping and the base of + * the linear mapping. We will undo this in map_mem(). + */ + adrp x24, KERNEL_BASE + KIMAGE_OFFSET + bl set_cpu_boot_mode_flag bl __create_page_tables // x25=TTBR0, x26=TTBR1 /* @@ -389,10 +397,10 @@ __create_page_tables: * Map the kernel image (starting with PHYS_OFFSET). */ mov x0, x26 // swapper_pg_dir - mov x5, #PAGE_OFFSET + ldr x5, =KERNEL_BASE create_pgd_entry x0, x5, x3, x6 ldr x6, =KERNEL_END // __va(KERNEL_END) - mov x3, x24 // phys offset + adrp x3, KERNEL_BASE // real PHYS_OFFSET create_block_map x0, x7, x3, x5, x6 /* diff --git a/arch/arm64/kernel/vmlinux.lds.S b/arch/arm64/kernel/vmlinux.lds.S index 63fca196c09e..84f780e6b039 100644 --- a/arch/arm64/kernel/vmlinux.lds.S +++ b/arch/arm64/kernel/vmlinux.lds.S @@ -7,6 +7,7 @@ #include <asm-generic/vmlinux.lds.h> #include <asm/kernel-pgtable.h> #include <asm/thread_info.h> +#include <asm/boot.h> #include <asm/memory.h> #include <asm/page.h> #include <asm/pgtable.h> @@ -99,7 +100,7 @@ SECTIONS *(.discard.*) } - . = PAGE_OFFSET + TEXT_OFFSET; + . = KIMAGE_VADDR + TEXT_OFFSET; .head.text : { _text = .; @@ -207,4 +208,10 @@ ASSERT(SIZEOF(.pgdir) <= ALIGNOF(.pgdir), ".pgdir size exceeds its alignment") /* * If padding is applied before .head.text, virt<->phys conversions will fail. */ -ASSERT(_text == (PAGE_OFFSET + TEXT_OFFSET), "HEAD is misaligned") +ASSERT(_text == (KIMAGE_VADDR + TEXT_OFFSET), "HEAD is misaligned") + +/* + * Make sure the memory footprint of the kernel Image does not exceed the limit. + */ +ASSERT(_end - _text + TEXT_OFFSET <= MAX_KIMG_SIZE, + "Kernel Image memory footprint exceeds MAX_KIMG_SIZE") diff --git a/arch/arm64/mm/dump.c b/arch/arm64/mm/dump.c index 5a22a119a74c..f6272f450688 100644 --- a/arch/arm64/mm/dump.c +++ b/arch/arm64/mm/dump.c @@ -63,7 +63,8 @@ static struct addr_marker address_markers[] = { { PCI_IO_END, "PCI I/O end" }, { MODULES_VADDR, "Modules start" }, { MODULES_END, "Modules end" }, - { PAGE_OFFSET, "Kernel Mapping" }, + { KIMAGE_VADDR, "Kernel Mapping" }, + { PAGE_OFFSET, "Linear Mapping" }, { -1, NULL }, }; diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c index 81bb49eaa1a3..c7ba171951c8 100644 --- a/arch/arm64/mm/mmu.c +++ b/arch/arm64/mm/mmu.c @@ -51,6 +51,9 @@ u64 idmap_t0sz = TCR_T0SZ(VA_BITS); struct page *empty_zero_page; EXPORT_SYMBOL(empty_zero_page); +u64 kernel_va_offset __read_mostly; +EXPORT_SYMBOL(kernel_va_offset); + pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t vma_prot) { @@ -479,6 +482,9 @@ static unsigned long __init bootstrap_region(struct bootstrap_pgtables *reg, * Bootstrap the linear ranges that cover the start of DRAM and swapper_pg_dir * so that the statically allocated page tables as well as newly allocated ones * are accessible via the linear mapping. + * Since at this point, PHYS_OFFSET is still biased to redirect __va() + * translations into the kernel text mapping, we need to apply an + * explicit va_offset to calculate linear virtual addresses. */ static void __init bootstrap_linear_mapping(unsigned long va_offset) { @@ -513,7 +519,10 @@ static void __init map_mem(void) { struct memblock_region *reg; - bootstrap_linear_mapping(0); + bootstrap_linear_mapping(KIMAGE_OFFSET); + + kernel_va_offset = KIMAGE_OFFSET; + memstart_addr -= KIMAGE_OFFSET; /* map all the memory banks */ for_each_memblock(memory, reg) { @@ -535,12 +544,12 @@ static void __init fixup_executable(void) #ifdef CONFIG_DEBUG_RODATA /* now that we are actually fully mapped, make the start/end more fine grained */ if (!IS_ALIGNED((unsigned long)_stext, SWAPPER_BLOCK_SIZE)) { - unsigned long aligned_start = round_down(__pa(_stext), + unsigned long aligned_start = round_down((unsigned long)_stext, SWAPPER_BLOCK_SIZE); - create_mapping(aligned_start, __phys_to_virt(aligned_start), - __pa(_stext) - aligned_start, - PAGE_KERNEL); + create_mapping(__pa(_stext), aligned_start, + (unsigned long)_stext - aligned_start, + PAGE_KERNEL); } if (!IS_ALIGNED((unsigned long)__init_end, SWAPPER_BLOCK_SIZE)) {
This moves the primary mapping of the kernel Image out of the linear region. This is a preparatory step towards allowing the kernel Image to reside anywhere in physical memory without affecting the ability to map all of it efficiently. Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> --- arch/arm64/include/asm/boot.h | 7 +++++++ arch/arm64/include/asm/kernel-pgtable.h | 5 +++-- arch/arm64/include/asm/memory.h | 20 +++++++++++++++++--- arch/arm64/kernel/head.S | 18 +++++++++++++----- arch/arm64/kernel/vmlinux.lds.S | 11 +++++++++-- arch/arm64/mm/dump.c | 3 ++- arch/arm64/mm/mmu.c | 19 ++++++++++++++----- 7 files changed, 65 insertions(+), 18 deletions(-) -- 1.9.1 _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel