From: Ahmad Fatoum <a.fatoum@pengutronix.de>
To: barebox@lists.infradead.org
Cc: Ahmad Fatoum <a.fatoum@pengutronix.de>
Subject: [PATCH 3/4] ARM: mmu: ensure PTE is written at once
Date: Wed, 21 May 2025 19:41:03 +0200 [thread overview]
Message-ID: <20250521174104.1570239-4-a.fatoum@pengutronix.de> (raw)
In-Reply-To: <20250521174104.1570239-1-a.fatoum@pengutronix.de>
Writing the PTE with a normal non-volatile access is a bad idea, because
speculation may already use a partially written entry.
Use WRITE_ONCE to turn this into a volatile write via a new set_pte
helper. The extra helper is useful for type enforcement and for future
extension when we start to consistently apply break-before-make[1].
[1]: https://lore.kernel.org/all/87a5btrcyr.wl-maz@kernel.org/
Signed-off-by: Ahmad Fatoum <a.fatoum@pengutronix.de>
---
arch/arm/cpu/mmu_32.c | 40 +++++++++++++++++++++++++++++-----------
arch/arm/cpu/mmu_64.c | 20 +++++++++++++-------
2 files changed, 42 insertions(+), 18 deletions(-)
diff --git a/arch/arm/cpu/mmu_32.c b/arch/arm/cpu/mmu_32.c
index ec6bd27da4e1..9f50194c7c2b 100644
--- a/arch/arm/cpu/mmu_32.c
+++ b/arch/arm/cpu/mmu_32.c
@@ -65,6 +65,11 @@ static bool pgd_type_table(u32 pgd)
#define PTE_SIZE (PTRS_PER_PTE * sizeof(u32))
+static void set_pte(uint32_t *pt, uint32_t val)
+{
+ WRITE_ONCE(*pt, val);
+}
+
#ifdef __PBL__
static uint32_t *alloc_pte(void)
{
@@ -141,13 +146,14 @@ static u32 *arm_create_pte(unsigned long virt, unsigned long phys,
ttb_idx = pgd_index(virt);
for (i = 0; i < PTRS_PER_PTE; i++) {
- table[i] = phys | PTE_TYPE_SMALL | flags;
+ set_pte(&table[i], phys | PTE_TYPE_SMALL | flags);
virt += PAGE_SIZE;
phys += PAGE_SIZE;
}
dma_flush_range(table, PTRS_PER_PTE * sizeof(u32));
- ttb[ttb_idx] = (unsigned long)table | PMD_TYPE_TABLE;
+ // TODO break-before-make missing
+ set_pte(&ttb[ttb_idx], (unsigned long)table | PMD_TYPE_TABLE);
dma_flush_range(&ttb[ttb_idx], sizeof(u32));
return table;
@@ -264,14 +270,17 @@ static void __arch_remap_range(void *_virt_addr, phys_addr_t phys_addr, size_t s
if (size >= PGDIR_SIZE && pgdir_size_aligned &&
IS_ALIGNED(phys_addr, PGDIR_SIZE) &&
!pgd_type_table(*pgd)) {
+ u32 val;
/*
* TODO: Add code to discard a page table and
* replace it with a section
*/
chunk = PGDIR_SIZE;
- *pgd = phys_addr | pmd_flags;
+ val = phys_addr | pmd_flags;
if (map_type != MAP_FAULT)
- *pgd |= PMD_TYPE_SECT;
+ val |= PMD_TYPE_SECT;
+ // TODO break-before-make missing
+ set_pte(pgd, val);
dma_flush_range(pgd, sizeof(*pgd));
} else {
unsigned int num_ptes;
@@ -310,10 +319,15 @@ static void __arch_remap_range(void *_virt_addr, phys_addr_t phys_addr, size_t s
}
for (i = 0; i < num_ptes; i++) {
- pte[i] = phys_addr + i * PAGE_SIZE;
- pte[i] |= pte_flags;
+ u32 val;
+
+ val = phys_addr + i * PAGE_SIZE;
+ val |= pte_flags;
if (map_type != MAP_FAULT)
- pte[i] |= PTE_TYPE_SMALL;
+ val |= PTE_TYPE_SMALL;
+
+ // TODO break-before-make missing
+ set_pte(&pte[i], val);
}
dma_flush_range(pte, num_ptes * sizeof(u32));
@@ -350,7 +364,7 @@ static void create_sections(unsigned long first, unsigned long last,
unsigned int i, addr = first;
for (i = ttb_start; i < ttb_end; i++) {
- ttb[i] = addr | flags;
+ set_pte(&ttb[i], addr | flags);
addr += PGDIR_SIZE;
}
}
@@ -366,8 +380,10 @@ void *map_io_sections(unsigned long phys, void *_start, size_t size)
unsigned long start = (unsigned long)_start, sec;
uint32_t *ttb = get_ttb();
- for (sec = start; sec < start + size; sec += PGDIR_SIZE, phys += PGDIR_SIZE)
- ttb[pgd_index(sec)] = phys | get_pmd_flags(MAP_UNCACHED);
+ for (sec = start; sec < start + size; sec += PGDIR_SIZE, phys += PGDIR_SIZE) {
+ // TODO break-before-make missing
+ set_pte(&ttb[pgd_index(sec)], phys | get_pmd_flags(MAP_UNCACHED));
+ }
dma_flush_range(ttb, 0x4000);
tlb_invalidate();
@@ -410,7 +426,9 @@ static void create_vector_table(unsigned long adr)
vectors, adr);
arm_create_pte(adr, adr, get_pte_flags(MAP_UNCACHED));
pte = find_pte(adr);
- *pte = (u32)vectors | PTE_TYPE_SMALL | get_pte_flags(MAP_CACHED);
+ // TODO break-before-make missing
+ set_pte(pte, (u32)vectors | PTE_TYPE_SMALL |
+ get_pte_flags(MAP_CACHED));
}
arm_fixup_vectors();
diff --git a/arch/arm/cpu/mmu_64.c b/arch/arm/cpu/mmu_64.c
index d9b0b74d7314..efd788c93e7b 100644
--- a/arch/arm/cpu/mmu_64.c
+++ b/arch/arm/cpu/mmu_64.c
@@ -31,12 +31,14 @@ static uint64_t *get_ttb(void)
return (uint64_t *)get_ttbr(current_el());
}
+static void set_pte(uint64_t *pt, uint64_t val)
+{
+ WRITE_ONCE(*pt, val);
+}
+
static void set_table(uint64_t *pt, uint64_t *table_addr)
{
- uint64_t val;
-
- val = PTE_TYPE_TABLE | (uint64_t)table_addr;
- *pt = val;
+ set_pte(pt, PTE_TYPE_TABLE | (uint64_t)table_addr);
}
#ifdef __PBL__
@@ -114,14 +116,16 @@ static void split_block(uint64_t *pte, int level)
for (i = 0; i < MAX_PTE_ENTRIES; i++) {
- new_table[i] = old_pte | (i << levelshift);
+ set_pte(&new_table[i], old_pte | (i << levelshift));
/* Level 3 block PTEs have the table type */
if ((level + 1) == 3)
new_table[i] |= PTE_TYPE_TABLE;
}
- /* Set the new table into effect */
+ /* Set the new table into effect
+ * TODO: break-before-make missing
+ */
set_table(pte, new_table);
}
@@ -157,7 +161,9 @@ static void create_sections(uint64_t virt, uint64_t phys, uint64_t size,
IS_ALIGNED(phys, block_size)) {
type = (level == 3) ?
PTE_TYPE_PAGE : PTE_TYPE_BLOCK;
- *pte = phys | attr | type;
+
+ /* TODO: break-before-make missing */
+ set_pte(pte, phys | attr | type);
addr += block_size;
phys += block_size;
size -= block_size;
--
2.39.5
next prev parent reply other threads:[~2025-05-21 17:42 UTC|newest]
Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-05-21 17:41 [PATCH 0/4] " Ahmad Fatoum
2025-05-21 17:41 ` [PATCH 1/4] arch: sync READ_ONCE/WRITE_ONCE with Linux Ahmad Fatoum
2025-05-21 17:41 ` [PATCH 2/4] ARM64: mmu: document granule size calculation Ahmad Fatoum
2025-05-21 17:41 ` Ahmad Fatoum [this message]
2025-05-21 17:41 ` [PATCH 4/4] WIP: break before make and don't invalidate uncached regions being remapped Ahmad Fatoum
2025-05-26 13:51 ` (subset) [PATCH 0/4] ARM: mmu: ensure PTE is written at once Sascha Hauer
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250521174104.1570239-4-a.fatoum@pengutronix.de \
--to=a.fatoum@pengutronix.de \
--cc=barebox@lists.infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox