|
|
|
/*
|
|
|
|
* Copyright (c) 2017-2020, Arm Limited and Contributors. All rights reserved.
|
|
|
|
*
|
|
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <arch_helpers.h>
|
|
|
|
#include <assert.h>
|
|
|
|
|
|
|
|
#include <platform_def.h>
|
|
|
|
|
|
|
|
#include <common/debug.h>
|
|
|
|
#include <lib/xlat_tables/xlat_tables_defs.h>
|
|
|
|
#include <lib/xlat_tables/xlat_tables_v2.h>
|
|
|
|
|
|
|
|
#include "xlat_tables_private.h"
|
|
|
|
|
|
|
|
/*
|
|
|
|
* MMU configuration register values for the active translation context. Used
|
|
|
|
* from the MMU assembly helpers.
|
|
|
|
*/
|
|
|
|
uint64_t mmu_cfg_params[MMU_CFG_PARAM_MAX];
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate and initialise the default translation context for the BL image
|
|
|
|
* currently executing.
|
|
|
|
*/
|
|
|
|
REGISTER_XLAT_CONTEXT(tf, MAX_MMAP_REGIONS, MAX_XLAT_TABLES,
|
|
|
|
PLAT_VIRT_ADDR_SPACE_SIZE, PLAT_PHY_ADDR_SPACE_SIZE);
|
|
|
|
|
|
|
|
void mmap_add_region(unsigned long long base_pa, uintptr_t base_va, size_t size,
|
|
|
|
unsigned int attr)
|
|
|
|
{
|
|
|
|
mmap_region_t mm = MAP_REGION(base_pa, base_va, size, attr);
|
|
|
|
|
|
|
|
mmap_add_region_ctx(&tf_xlat_ctx, &mm);
|
|
|
|
}
|
|
|
|
|
|
|
|
void mmap_add(const mmap_region_t *mm)
|
|
|
|
{
|
|
|
|
mmap_add_ctx(&tf_xlat_ctx, mm);
|
|
|
|
}
|
|
|
|
|
|
|
|
void mmap_add_region_alloc_va(unsigned long long base_pa, uintptr_t *base_va,
|
|
|
|
size_t size, unsigned int attr)
|
|
|
|
{
|
|
|
|
mmap_region_t mm = MAP_REGION_ALLOC_VA(base_pa, size, attr);
|
|
|
|
|
|
|
|
mmap_add_region_alloc_va_ctx(&tf_xlat_ctx, &mm);
|
|
|
|
|
|
|
|
*base_va = mm.base_va;
|
|
|
|
}
|
|
|
|
|
|
|
|
void mmap_add_alloc_va(mmap_region_t *mm)
|
|
|
|
{
|
|
|
|
while (mm->granularity != 0U) {
|
|
|
|
assert(mm->base_va == 0U);
|
|
|
|
mmap_add_region_alloc_va_ctx(&tf_xlat_ctx, mm);
|
|
|
|
mm++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#if PLAT_XLAT_TABLES_DYNAMIC
|
|
|
|
|
|
|
|
int mmap_add_dynamic_region(unsigned long long base_pa, uintptr_t base_va,
|
|
|
|
size_t size, unsigned int attr)
|
|
|
|
{
|
|
|
|
mmap_region_t mm = MAP_REGION(base_pa, base_va, size, attr);
|
|
|
|
|
|
|
|
return mmap_add_dynamic_region_ctx(&tf_xlat_ctx, &mm);
|
|
|
|
}
|
|
|
|
|
|
|
|
int mmap_add_dynamic_region_alloc_va(unsigned long long base_pa,
|
|
|
|
uintptr_t *base_va, size_t size,
|
|
|
|
unsigned int attr)
|
|
|
|
{
|
|
|
|
mmap_region_t mm = MAP_REGION_ALLOC_VA(base_pa, size, attr);
|
|
|
|
|
|
|
|
int rc = mmap_add_dynamic_region_alloc_va_ctx(&tf_xlat_ctx, &mm);
|
|
|
|
|
|
|
|
*base_va = mm.base_va;
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
int mmap_remove_dynamic_region(uintptr_t base_va, size_t size)
|
|
|
|
{
|
|
|
|
return mmap_remove_dynamic_region_ctx(&tf_xlat_ctx,
|
|
|
|
base_va, size);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* PLAT_XLAT_TABLES_DYNAMIC */
|
|
|
|
|
|
|
|
void __init init_xlat_tables(void)
|
|
|
|
{
|
|
|
|
assert(tf_xlat_ctx.xlat_regime == EL_REGIME_INVALID);
|
|
|
|
|
xlat: Fix MISRA defects
Fix defects of MISRA C-2012 rules 8.13, 10.1, 10.3, 10.4, 10.8, 11.6,
14.4, 15.7, 17.8, 20.10, 20.12, 21.1 and Directive 4.9.
Change-Id: I7ff61e71733908596dbafe2e99d99b4fce9765bd
Signed-off-by: Antonio Nino Diaz <antonio.ninodiaz@arm.com>
6 years ago
|
|
|
unsigned int current_el = xlat_arch_current_el();
|
|
|
|
|
xlat: Fix MISRA defects
Fix defects of MISRA C-2012 rules 8.13, 10.1, 10.3, 10.4, 10.8, 11.6,
14.4, 15.7, 17.8, 20.10, 20.12, 21.1 and Directive 4.9.
Change-Id: I7ff61e71733908596dbafe2e99d99b4fce9765bd
Signed-off-by: Antonio Nino Diaz <antonio.ninodiaz@arm.com>
6 years ago
|
|
|
if (current_el == 1U) {
|
|
|
|
tf_xlat_ctx.xlat_regime = EL1_EL0_REGIME;
|
|
|
|
} else if (current_el == 2U) {
|
|
|
|
tf_xlat_ctx.xlat_regime = EL2_REGIME;
|
|
|
|
} else {
|
xlat: Fix MISRA defects
Fix defects of MISRA C-2012 rules 8.13, 10.1, 10.3, 10.4, 10.8, 11.6,
14.4, 15.7, 17.8, 20.10, 20.12, 21.1 and Directive 4.9.
Change-Id: I7ff61e71733908596dbafe2e99d99b4fce9765bd
Signed-off-by: Antonio Nino Diaz <antonio.ninodiaz@arm.com>
6 years ago
|
|
|
assert(current_el == 3U);
|
|
|
|
tf_xlat_ctx.xlat_regime = EL3_REGIME;
|
|
|
|
}
|
|
|
|
|
|
|
|
init_xlat_tables_ctx(&tf_xlat_ctx);
|
|
|
|
}
|
|
|
|
|
|
|
|
int xlat_get_mem_attributes(uintptr_t base_va, uint32_t *attr)
|
|
|
|
{
|
|
|
|
return xlat_get_mem_attributes_ctx(&tf_xlat_ctx, base_va, attr);
|
|
|
|
}
|
|
|
|
|
|
|
|
int xlat_change_mem_attributes(uintptr_t base_va, size_t size, uint32_t attr)
|
|
|
|
{
|
|
|
|
return xlat_change_mem_attributes_ctx(&tf_xlat_ctx, base_va, size, attr);
|
|
|
|
}
|
|
|
|
|
|
|
|
#if PLAT_RO_XLAT_TABLES
|
|
|
|
/* Change the memory attributes of the descriptors which resolve the address
|
|
|
|
* range that belongs to the translation tables themselves, which are by default
|
|
|
|
* mapped as part of read-write data in the BL image's memory.
|
|
|
|
*
|
|
|
|
* Since the translation tables map themselves via these level 3 (page)
|
|
|
|
* descriptors, any change applied to them with the MMU on would introduce a
|
|
|
|
* chicken and egg problem because of the break-before-make sequence.
|
|
|
|
* Eventually, it would reach the descriptor that resolves the very table it
|
|
|
|
* belongs to and the invalidation (break step) would cause the subsequent write
|
|
|
|
* (make step) to it to generate an MMU fault. Therefore, the MMU is disabled
|
|
|
|
* before making the change.
|
|
|
|
*
|
|
|
|
* No assumption is made about what data this function needs, therefore all the
|
|
|
|
* caches are flushed in order to ensure coherency. A future optimization would
|
|
|
|
* be to only flush the required data to main memory.
|
|
|
|
*/
|
|
|
|
int xlat_make_tables_readonly(void)
|
|
|
|
{
|
|
|
|
assert(tf_xlat_ctx.initialized == true);
|
|
|
|
#ifdef __aarch64__
|
|
|
|
if (tf_xlat_ctx.xlat_regime == EL1_EL0_REGIME) {
|
|
|
|
disable_mmu_el1();
|
|
|
|
} else if (tf_xlat_ctx.xlat_regime == EL3_REGIME) {
|
|
|
|
disable_mmu_el3();
|
|
|
|
} else {
|
|
|
|
assert(tf_xlat_ctx.xlat_regime == EL2_REGIME);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Flush all caches. */
|
|
|
|
dcsw_op_all(DCCISW);
|
|
|
|
#else /* !__aarch64__ */
|
|
|
|
assert(tf_xlat_ctx.xlat_regime == EL1_EL0_REGIME);
|
|
|
|
/* On AArch32, we flush the caches before disabling the MMU. The reason
|
|
|
|
* for this is that the dcsw_op_all AArch32 function pushes some
|
|
|
|
* registers onto the stack under the assumption that it is writing to
|
|
|
|
* cache, which is not true with the MMU off. This would result in the
|
|
|
|
* stack becoming corrupted and a wrong/junk value for the LR being
|
|
|
|
* restored at the end of the routine.
|
|
|
|
*/
|
|
|
|
dcsw_op_all(DC_OP_CISW);
|
|
|
|
disable_mmu_secure();
|
|
|
|
#endif
|
|
|
|
|
|
|
|
int rc = xlat_change_mem_attributes_ctx(&tf_xlat_ctx,
|
|
|
|
(uintptr_t)tf_xlat_ctx.tables,
|
|
|
|
tf_xlat_ctx.tables_num * XLAT_TABLE_SIZE,
|
|
|
|
MT_RO_DATA | MT_SECURE);
|
|
|
|
|
|
|
|
#ifdef __aarch64__
|
|
|
|
if (tf_xlat_ctx.xlat_regime == EL1_EL0_REGIME) {
|
|
|
|
enable_mmu_el1(0U);
|
|
|
|
} else {
|
|
|
|
assert(tf_xlat_ctx.xlat_regime == EL3_REGIME);
|
|
|
|
enable_mmu_el3(0U);
|
|
|
|
}
|
|
|
|
#else /* !__aarch64__ */
|
|
|
|
enable_mmu_svc_mon(0U);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
if (rc == 0) {
|
|
|
|
tf_xlat_ctx.readonly_tables = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
#endif /* PLAT_RO_XLAT_TABLES */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If dynamic allocation of new regions is disabled then by the time we call the
|
|
|
|
* function enabling the MMU, we'll have registered all the memory regions to
|
|
|
|
* map for the system's lifetime. Therefore, at this point we know the maximum
|
|
|
|
* physical address that will ever be mapped.
|
|
|
|
*
|
|
|
|
* If dynamic allocation is enabled then we can't make any such assumption
|
|
|
|
* because the maximum physical address could get pushed while adding a new
|
|
|
|
* region. Therefore, in this case we have to assume that the whole address
|
|
|
|
* space size might be mapped.
|
|
|
|
*/
|
|
|
|
#if PLAT_XLAT_TABLES_DYNAMIC
|
|
|
|
#define MAX_PHYS_ADDR tf_xlat_ctx.pa_max_address
|
|
|
|
#else
|
|
|
|
#define MAX_PHYS_ADDR tf_xlat_ctx.max_pa
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef __aarch64__
|
|
|
|
|
|
|
|
void enable_mmu_el1(unsigned int flags)
|
|
|
|
{
|
|
|
|
setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
|
|
|
|
tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
|
|
|
|
tf_xlat_ctx.va_max_address, EL1_EL0_REGIME);
|
|
|
|
enable_mmu_direct_el1(flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
void enable_mmu_el2(unsigned int flags)
|
|
|
|
{
|
|
|
|
setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
|
|
|
|
tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
|
|
|
|
tf_xlat_ctx.va_max_address, EL2_REGIME);
|
|
|
|
enable_mmu_direct_el2(flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
void enable_mmu_el3(unsigned int flags)
|
|
|
|
{
|
|
|
|
setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
|
|
|
|
tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
|
|
|
|
tf_xlat_ctx.va_max_address, EL3_REGIME);
|
|
|
|
enable_mmu_direct_el3(flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
void enable_mmu(unsigned int flags)
|
|
|
|
{
|
|
|
|
switch (get_current_el_maybe_constant()) {
|
|
|
|
case 1:
|
|
|
|
enable_mmu_el1(flags);
|
|
|
|
break;
|
|
|
|
case 2:
|
|
|
|
enable_mmu_el2(flags);
|
|
|
|
break;
|
|
|
|
case 3:
|
|
|
|
enable_mmu_el3(flags);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
panic();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#else /* !__aarch64__ */
|
|
|
|
|
|
|
|
void enable_mmu_svc_mon(unsigned int flags)
|
|
|
|
{
|
|
|
|
setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
|
|
|
|
tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
|
|
|
|
tf_xlat_ctx.va_max_address, EL1_EL0_REGIME);
|
|
|
|
enable_mmu_direct_svc_mon(flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
void enable_mmu_hyp(unsigned int flags)
|
|
|
|
{
|
|
|
|
setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
|
|
|
|
tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
|
|
|
|
tf_xlat_ctx.va_max_address, EL2_REGIME);
|
|
|
|
enable_mmu_direct_hyp(flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* __aarch64__ */
|