mirror of
https://github.com/Fishwaldo/linux-bl808.git
synced 2025-03-19 05:24:11 +00:00
arm64: mm: replace 'block_mappings_allowed' with 'page_mappings_only'
In preparation of adding support for contiguous PTE and PMD mappings, let's replace 'block_mappings_allowed' with 'page_mappings_only', which will be a more accurate description of the nature of the setting once we add such contiguous mappings into the mix. Reviewed-by: Mark Rutland <mark.rutland@arm.com> Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
This commit is contained in:
parent
e98216b521
commit
f14c66ce81
3 changed files with 21 additions and 21 deletions
|
@ -34,7 +34,7 @@ extern void __iomem *early_io_map(phys_addr_t phys, unsigned long virt);
|
||||||
extern void init_mem_pgprot(void);
|
extern void init_mem_pgprot(void);
|
||||||
extern void create_pgd_mapping(struct mm_struct *mm, phys_addr_t phys,
|
extern void create_pgd_mapping(struct mm_struct *mm, phys_addr_t phys,
|
||||||
unsigned long virt, phys_addr_t size,
|
unsigned long virt, phys_addr_t size,
|
||||||
pgprot_t prot, bool allow_block_mappings);
|
pgprot_t prot, bool page_mappings_only);
|
||||||
extern void *fixmap_remap_fdt(phys_addr_t dt_phys);
|
extern void *fixmap_remap_fdt(phys_addr_t dt_phys);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -62,8 +62,8 @@ struct screen_info screen_info __section(.data);
|
||||||
int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
|
int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
|
||||||
{
|
{
|
||||||
pteval_t prot_val = create_mapping_protection(md);
|
pteval_t prot_val = create_mapping_protection(md);
|
||||||
bool allow_block_mappings = (md->type != EFI_RUNTIME_SERVICES_CODE &&
|
bool page_mappings_only = (md->type == EFI_RUNTIME_SERVICES_CODE ||
|
||||||
md->type != EFI_RUNTIME_SERVICES_DATA);
|
md->type == EFI_RUNTIME_SERVICES_DATA);
|
||||||
|
|
||||||
if (!PAGE_ALIGNED(md->phys_addr) ||
|
if (!PAGE_ALIGNED(md->phys_addr) ||
|
||||||
!PAGE_ALIGNED(md->num_pages << EFI_PAGE_SHIFT)) {
|
!PAGE_ALIGNED(md->num_pages << EFI_PAGE_SHIFT)) {
|
||||||
|
@ -76,12 +76,12 @@ int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
|
||||||
* from the MMU routines. So avoid block mappings altogether in
|
* from the MMU routines. So avoid block mappings altogether in
|
||||||
* that case.
|
* that case.
|
||||||
*/
|
*/
|
||||||
allow_block_mappings = false;
|
page_mappings_only = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
create_pgd_mapping(mm, md->phys_addr, md->virt_addr,
|
create_pgd_mapping(mm, md->phys_addr, md->virt_addr,
|
||||||
md->num_pages << EFI_PAGE_SHIFT,
|
md->num_pages << EFI_PAGE_SHIFT,
|
||||||
__pgprot(prot_val | PTE_NG), allow_block_mappings);
|
__pgprot(prot_val | PTE_NG), page_mappings_only);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -143,7 +143,7 @@ static void alloc_init_pte(pmd_t *pmd, unsigned long addr,
|
||||||
static void alloc_init_pmd(pud_t *pud, unsigned long addr, unsigned long end,
|
static void alloc_init_pmd(pud_t *pud, unsigned long addr, unsigned long end,
|
||||||
phys_addr_t phys, pgprot_t prot,
|
phys_addr_t phys, pgprot_t prot,
|
||||||
phys_addr_t (*pgtable_alloc)(void),
|
phys_addr_t (*pgtable_alloc)(void),
|
||||||
bool allow_block_mappings)
|
bool page_mappings_only)
|
||||||
{
|
{
|
||||||
pmd_t *pmd;
|
pmd_t *pmd;
|
||||||
unsigned long next;
|
unsigned long next;
|
||||||
|
@ -170,7 +170,7 @@ static void alloc_init_pmd(pud_t *pud, unsigned long addr, unsigned long end,
|
||||||
|
|
||||||
/* try section mapping first */
|
/* try section mapping first */
|
||||||
if (((addr | next | phys) & ~SECTION_MASK) == 0 &&
|
if (((addr | next | phys) & ~SECTION_MASK) == 0 &&
|
||||||
allow_block_mappings) {
|
!page_mappings_only) {
|
||||||
pmd_set_huge(pmd, phys, prot);
|
pmd_set_huge(pmd, phys, prot);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -207,7 +207,7 @@ static inline bool use_1G_block(unsigned long addr, unsigned long next,
|
||||||
static void alloc_init_pud(pgd_t *pgd, unsigned long addr, unsigned long end,
|
static void alloc_init_pud(pgd_t *pgd, unsigned long addr, unsigned long end,
|
||||||
phys_addr_t phys, pgprot_t prot,
|
phys_addr_t phys, pgprot_t prot,
|
||||||
phys_addr_t (*pgtable_alloc)(void),
|
phys_addr_t (*pgtable_alloc)(void),
|
||||||
bool allow_block_mappings)
|
bool page_mappings_only)
|
||||||
{
|
{
|
||||||
pud_t *pud;
|
pud_t *pud;
|
||||||
unsigned long next;
|
unsigned long next;
|
||||||
|
@ -229,7 +229,7 @@ static void alloc_init_pud(pgd_t *pgd, unsigned long addr, unsigned long end,
|
||||||
/*
|
/*
|
||||||
* For 4K granule only, attempt to put down a 1GB block
|
* For 4K granule only, attempt to put down a 1GB block
|
||||||
*/
|
*/
|
||||||
if (use_1G_block(addr, next, phys) && allow_block_mappings) {
|
if (use_1G_block(addr, next, phys) && !page_mappings_only) {
|
||||||
pud_set_huge(pud, phys, prot);
|
pud_set_huge(pud, phys, prot);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -240,7 +240,7 @@ static void alloc_init_pud(pgd_t *pgd, unsigned long addr, unsigned long end,
|
||||||
pud_val(*pud)));
|
pud_val(*pud)));
|
||||||
} else {
|
} else {
|
||||||
alloc_init_pmd(pud, addr, next, phys, prot,
|
alloc_init_pmd(pud, addr, next, phys, prot,
|
||||||
pgtable_alloc, allow_block_mappings);
|
pgtable_alloc, page_mappings_only);
|
||||||
|
|
||||||
BUG_ON(pud_val(old_pud) != 0 &&
|
BUG_ON(pud_val(old_pud) != 0 &&
|
||||||
pud_val(old_pud) != pud_val(*pud));
|
pud_val(old_pud) != pud_val(*pud));
|
||||||
|
@ -255,7 +255,7 @@ static void __create_pgd_mapping(pgd_t *pgdir, phys_addr_t phys,
|
||||||
unsigned long virt, phys_addr_t size,
|
unsigned long virt, phys_addr_t size,
|
||||||
pgprot_t prot,
|
pgprot_t prot,
|
||||||
phys_addr_t (*pgtable_alloc)(void),
|
phys_addr_t (*pgtable_alloc)(void),
|
||||||
bool allow_block_mappings)
|
bool page_mappings_only)
|
||||||
{
|
{
|
||||||
unsigned long addr, length, end, next;
|
unsigned long addr, length, end, next;
|
||||||
pgd_t *pgd = pgd_offset_raw(pgdir, virt);
|
pgd_t *pgd = pgd_offset_raw(pgdir, virt);
|
||||||
|
@ -275,7 +275,7 @@ static void __create_pgd_mapping(pgd_t *pgdir, phys_addr_t phys,
|
||||||
do {
|
do {
|
||||||
next = pgd_addr_end(addr, end);
|
next = pgd_addr_end(addr, end);
|
||||||
alloc_init_pud(pgd, addr, next, phys, prot, pgtable_alloc,
|
alloc_init_pud(pgd, addr, next, phys, prot, pgtable_alloc,
|
||||||
allow_block_mappings);
|
page_mappings_only);
|
||||||
phys += next - addr;
|
phys += next - addr;
|
||||||
} while (pgd++, addr = next, addr != end);
|
} while (pgd++, addr = next, addr != end);
|
||||||
}
|
}
|
||||||
|
@ -304,17 +304,17 @@ static void __init create_mapping_noalloc(phys_addr_t phys, unsigned long virt,
|
||||||
&phys, virt);
|
&phys, virt);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
__create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL, true);
|
__create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
void __init create_pgd_mapping(struct mm_struct *mm, phys_addr_t phys,
|
void __init create_pgd_mapping(struct mm_struct *mm, phys_addr_t phys,
|
||||||
unsigned long virt, phys_addr_t size,
|
unsigned long virt, phys_addr_t size,
|
||||||
pgprot_t prot, bool allow_block_mappings)
|
pgprot_t prot, bool page_mappings_only)
|
||||||
{
|
{
|
||||||
BUG_ON(mm == &init_mm);
|
BUG_ON(mm == &init_mm);
|
||||||
|
|
||||||
__create_pgd_mapping(mm->pgd, phys, virt, size, prot,
|
__create_pgd_mapping(mm->pgd, phys, virt, size, prot,
|
||||||
pgd_pgtable_alloc, allow_block_mappings);
|
pgd_pgtable_alloc, page_mappings_only);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void create_mapping_late(phys_addr_t phys, unsigned long virt,
|
static void create_mapping_late(phys_addr_t phys, unsigned long virt,
|
||||||
|
@ -327,7 +327,7 @@ static void create_mapping_late(phys_addr_t phys, unsigned long virt,
|
||||||
}
|
}
|
||||||
|
|
||||||
__create_pgd_mapping(init_mm.pgd, phys, virt, size, prot,
|
__create_pgd_mapping(init_mm.pgd, phys, virt, size, prot,
|
||||||
NULL, !debug_pagealloc_enabled());
|
NULL, debug_pagealloc_enabled());
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __init __map_memblock(pgd_t *pgd, phys_addr_t start, phys_addr_t end)
|
static void __init __map_memblock(pgd_t *pgd, phys_addr_t start, phys_addr_t end)
|
||||||
|
@ -345,7 +345,7 @@ static void __init __map_memblock(pgd_t *pgd, phys_addr_t start, phys_addr_t end
|
||||||
__create_pgd_mapping(pgd, start, __phys_to_virt(start),
|
__create_pgd_mapping(pgd, start, __phys_to_virt(start),
|
||||||
end - start, PAGE_KERNEL,
|
end - start, PAGE_KERNEL,
|
||||||
early_pgtable_alloc,
|
early_pgtable_alloc,
|
||||||
!debug_pagealloc_enabled());
|
debug_pagealloc_enabled());
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -358,13 +358,13 @@ static void __init __map_memblock(pgd_t *pgd, phys_addr_t start, phys_addr_t end
|
||||||
__phys_to_virt(start),
|
__phys_to_virt(start),
|
||||||
kernel_start - start, PAGE_KERNEL,
|
kernel_start - start, PAGE_KERNEL,
|
||||||
early_pgtable_alloc,
|
early_pgtable_alloc,
|
||||||
!debug_pagealloc_enabled());
|
debug_pagealloc_enabled());
|
||||||
if (kernel_end < end)
|
if (kernel_end < end)
|
||||||
__create_pgd_mapping(pgd, kernel_end,
|
__create_pgd_mapping(pgd, kernel_end,
|
||||||
__phys_to_virt(kernel_end),
|
__phys_to_virt(kernel_end),
|
||||||
end - kernel_end, PAGE_KERNEL,
|
end - kernel_end, PAGE_KERNEL,
|
||||||
early_pgtable_alloc,
|
early_pgtable_alloc,
|
||||||
!debug_pagealloc_enabled());
|
debug_pagealloc_enabled());
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Map the linear alias of the [_text, __init_begin) interval as
|
* Map the linear alias of the [_text, __init_begin) interval as
|
||||||
|
@ -374,7 +374,7 @@ static void __init __map_memblock(pgd_t *pgd, phys_addr_t start, phys_addr_t end
|
||||||
*/
|
*/
|
||||||
__create_pgd_mapping(pgd, kernel_start, __phys_to_virt(kernel_start),
|
__create_pgd_mapping(pgd, kernel_start, __phys_to_virt(kernel_start),
|
||||||
kernel_end - kernel_start, PAGE_KERNEL_RO,
|
kernel_end - kernel_start, PAGE_KERNEL_RO,
|
||||||
early_pgtable_alloc, !debug_pagealloc_enabled());
|
early_pgtable_alloc, debug_pagealloc_enabled());
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __init map_mem(pgd_t *pgd)
|
static void __init map_mem(pgd_t *pgd)
|
||||||
|
@ -424,7 +424,7 @@ static void __init map_kernel_segment(pgd_t *pgd, void *va_start, void *va_end,
|
||||||
BUG_ON(!PAGE_ALIGNED(size));
|
BUG_ON(!PAGE_ALIGNED(size));
|
||||||
|
|
||||||
__create_pgd_mapping(pgd, pa_start, (unsigned long)va_start, size, prot,
|
__create_pgd_mapping(pgd, pa_start, (unsigned long)va_start, size, prot,
|
||||||
early_pgtable_alloc, !debug_pagealloc_enabled());
|
early_pgtable_alloc, debug_pagealloc_enabled());
|
||||||
|
|
||||||
vma->addr = va_start;
|
vma->addr = va_start;
|
||||||
vma->phys_addr = pa_start;
|
vma->phys_addr = pa_start;
|
||||||
|
|
Loading…
Add table
Reference in a new issue