diff options
Diffstat (limited to 'source/k/patches-revert-i686')
-rw-r--r-- | source/k/patches-revert-i686/0001-x86-ioremap-Fix-page-aligned-size-calculation-in-__i.patch | 53 | ||||
-rw-r--r-- | source/k/patches-revert-i686/README | 5 |
2 files changed, 58 insertions, 0 deletions
diff --git a/source/k/patches-revert-i686/0001-x86-ioremap-Fix-page-aligned-size-calculation-in-__i.patch b/source/k/patches-revert-i686/0001-x86-ioremap-Fix-page-aligned-size-calculation-in-__i.patch new file mode 100644 index 000000000..c8767a252 --- /dev/null +++ b/source/k/patches-revert-i686/0001-x86-ioremap-Fix-page-aligned-size-calculation-in-__i.patch @@ -0,0 +1,53 @@ +From 2e479b3b82c49bfb9422274c0a9c155a41caecb7 Mon Sep 17 00:00:00 2001 +From: Michael Kelley <mikelley@microsoft.com> +Date: Wed, 16 Nov 2022 10:41:24 -0800 +Subject: [PATCH] x86/ioremap: Fix page aligned size calculation in + __ioremap_caller() + +commit 4dbd6a3e90e03130973688fd79e19425f720d999 upstream. + +Current code re-calculates the size after aligning the starting and +ending physical addresses on a page boundary. But the re-calculation +also embeds the masking of high order bits that exceed the size of +the physical address space (via PHYSICAL_PAGE_MASK). If the masking +removes any high order bits, the size calculation results in a huge +value that is likely to immediately fail. + +Fix this by re-calculating the page-aligned size first. Then mask any +high order bits using PHYSICAL_PAGE_MASK. + +Fixes: ffa71f33a820 ("x86, ioremap: Fix incorrect physical address handling in PAE mode") +Signed-off-by: Michael Kelley <mikelley@microsoft.com> +Signed-off-by: Borislav Petkov <bp@suse.de> +Acked-by: Dave Hansen <dave.hansen@linux.intel.com> +Cc: <stable@kernel.org> +Link: https://lore.kernel.org/r/1668624097-14884-2-git-send-email-mikelley@microsoft.com +Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org> +--- + arch/x86/mm/ioremap.c | 8 +++++++- + 1 file changed, 7 insertions(+), 1 deletion(-) + +diff --git a/arch/x86/mm/ioremap.c b/arch/x86/mm/ioremap.c +index 1ad0228f8ceb..19058d746695 100644 +--- a/arch/x86/mm/ioremap.c ++++ b/arch/x86/mm/ioremap.c +@@ -216,9 +216,15 @@ __ioremap_caller(resource_size_t phys_addr, unsigned long size, + * Mappings have to be page-aligned + */ + offset = phys_addr & ~PAGE_MASK; +- phys_addr &= PHYSICAL_PAGE_MASK; ++ phys_addr &= PAGE_MASK; + size = PAGE_ALIGN(last_addr+1) - phys_addr; + ++ /* ++ * Mask out any bits not part of the actual physical ++ * address, like memory encryption bits. ++ */ ++ phys_addr &= PHYSICAL_PAGE_MASK; ++ + retval = memtype_reserve(phys_addr, (u64)phys_addr + size, + pcm, &new_pcm); + if (retval) { +-- +2.39.0 + diff --git a/source/k/patches-revert-i686/README b/source/k/patches-revert-i686/README new file mode 100644 index 000000000..de85ec00a --- /dev/null +++ b/source/k/patches-revert-i686/README @@ -0,0 +1,5 @@ +We need to revert this patch which causes boot failure with 32-bit PAE on +*some* machines, such as: LENOVO 20QV000GUS Intel i7-9850H + +Non-PAE is not affected, but we'll let this get reverted there too since +it's not part of the code path anyway. |