1
0
Fork 0
zynq-rs/libcortex_a9/src/mmu.rs

424 lines
12 KiB
Rust
Raw Normal View History

2019-06-17 09:32:10 +08:00
use bit_field::BitField;
2020-06-22 08:02:11 +08:00
use super::{regs::*, asm::*, cache::*};
use libregister::RegisterW;
2019-06-17 09:32:10 +08:00
#[derive(Copy, Clone)]
#[repr(u8)]
pub enum AccessDomain {
NoAccess = 0b00,
Client = 0b01,
_Reserved = 0b10,
Manager = 0b11,
}
const ACCESS_DOMAINS_SIZE: usize = 16;
pub struct AccessDomains([AccessDomain; ACCESS_DOMAINS_SIZE]);
impl AccessDomains {
pub fn all_manager() -> Self {
AccessDomains([AccessDomain::Manager; ACCESS_DOMAINS_SIZE])
}
}
impl Into<u32> for AccessDomains {
fn into(self) -> u32 {
let mut result = 0;
for (i, domain) in self.0.iter().enumerate() {
result |= (*domain as u32) << (2 * i);
}
result
}
}
#[derive(Copy, Clone)]
#[repr(u8)]
pub enum AccessPermissions {
PermissionFault = 0,
PrivilegedOnly,
NoUserWrite,
FullAccess,
_Reserved1,
PrivilegedReadOnly,
ReadOnly,
_Reserved2
}
impl AccessPermissions {
2020-06-15 05:50:00 +08:00
fn new(ap: u8, apx: bool) -> Self {
unsafe {
core::mem::transmute(if apx { 0b100 } else { 0 } | ap)
}
}
2019-06-17 09:32:10 +08:00
fn ap(&self) -> u8 {
(*self as u8) & 0b11
}
fn apx(&self) -> bool {
(*self as u8) > (AccessPermissions::FullAccess as u8)
}
}
pub struct L1Section {
pub global: bool,
pub shareable: bool,
pub access: AccessPermissions,
/// Type EXtension
pub tex: u8,
pub domain: u8,
pub exec: bool,
pub cacheable: bool,
pub bufferable: bool,
}
2020-06-15 05:50:00 +08:00
const ENTRY_TYPE_SECTION: u32 = 0b10;
pub const L1_PAGE_SIZE: usize = 0x100000;
2019-06-17 09:32:10 +08:00
#[repr(C)]
#[derive(Clone, Copy)]
pub struct L1Entry(u32);
impl L1Entry {
#[inline(always)]
2020-06-15 05:50:00 +08:00
pub fn from_section(phys_base: u32, section: L1Section) -> Self {
2019-08-11 06:55:27 +08:00
// Must be aligned to 1 MB
2019-06-17 09:32:10 +08:00
assert!(phys_base & 0x000f_ffff == 0);
let mut entry = L1Entry(phys_base);
2020-06-15 05:50:00 +08:00
entry.set_section(section);
entry
}
pub fn get_section(&mut self) -> L1Section {
assert_eq!(self.0.get_bits(0..=1), ENTRY_TYPE_SECTION);
let access = AccessPermissions::new(
self.0.get_bits(10..=11) as u8,
self.0.get_bit(15)
);
L1Section {
global: !self.0.get_bit(17),
shareable: self.0.get_bit(16),
access,
tex: self.0.get_bits(12..=14) as u8,
domain: self.0.get_bits(5..=8) as u8,
exec: !self.0.get_bit(4),
cacheable: self.0.get_bit(3),
bufferable: self.0.get_bit(2),
}
}
pub fn set_section(&mut self, section: L1Section) {
self.0.set_bits(0..=1, ENTRY_TYPE_SECTION);
self.0.set_bit(2, section.bufferable);
self.0.set_bit(3, section.cacheable);
self.0.set_bit(4, !section.exec);
2019-06-17 09:32:10 +08:00
assert!(section.domain < 16);
2020-06-15 05:50:00 +08:00
self.0.set_bits(5..=8, section.domain.into());
self.0.set_bits(10..=11, section.access.ap().into());
2019-06-17 09:32:10 +08:00
assert!(section.tex < 8);
2020-06-15 05:50:00 +08:00
self.0.set_bits(12..=14, section.tex.into());
self.0.set_bit(15, section.access.apx());
self.0.set_bit(16, section.shareable);
self.0.set_bit(17, !section.global);
2019-06-17 09:32:10 +08:00
}
}
2019-06-18 08:22:07 +08:00
const L1_TABLE_SIZE: usize = 4096;
2020-05-01 07:17:53 +08:00
static mut L1_TABLE: L1Table = L1Table {
2019-06-18 08:22:07 +08:00
table: [L1Entry(0); L1_TABLE_SIZE]
};
2019-06-17 09:32:10 +08:00
#[repr(C, align(16384))]
2019-06-17 09:32:10 +08:00
pub struct L1Table {
2019-06-18 08:22:07 +08:00
table: [L1Entry; L1_TABLE_SIZE]
2019-06-17 09:32:10 +08:00
}
impl L1Table {
2019-06-18 08:22:07 +08:00
pub fn get() -> &'static mut Self {
unsafe {
2020-05-01 07:17:53 +08:00
&mut L1_TABLE
2019-06-18 08:22:07 +08:00
}
}
pub fn setup_flat_layout(&mut self) -> &Self {
2019-06-17 09:32:10 +08:00
/* 0x00000000 - 0x00100000 (cacheable) */
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(0, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: true,
access: AccessPermissions::FullAccess,
tex: 0b101,
domain: 0b1111,
exec: true,
cacheable: true,
2019-06-17 09:32:10 +08:00
bufferable: true,
});
/* (DDR cacheable) */
for ddr in 1..=0x3ff {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(ddr, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: true,
access: AccessPermissions::FullAccess,
tex: 0b101,
domain: 0b1111,
exec: true,
cacheable: true,
bufferable: true,
2019-06-17 09:32:10 +08:00
});
}
/* 0x40000000 - 0x7fffffff (FPGA slave0) */
for fpga_slave in 0x400..=0x7ff {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(fpga_slave, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: false,
cacheable: false,
bufferable: false,
});
}
/* 0x80000000 - 0xbfffffff (FPGA slave1) */
for fpga_slave in 0x800..=0xbff {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(fpga_slave, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: false,
cacheable: false,
bufferable: false,
});
}
/* 0xc0000000 - 0xdfffffff (unassigned/reserved). */
for undef in 0xc00..=0xdff {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(undef, L1Section {
2019-06-17 09:32:10 +08:00
global: false,
shareable: false,
access: AccessPermissions::PermissionFault,
tex: 0,
domain: 0,
exec: false,
cacheable: false,
bufferable: false,
});
}
/* 0xe0000000 - 0xe02fffff (Memory mapped devices)
* UART/USB/IIC/SPI/CAN/GEM/GPIO/QSPI/SD/NAND */
for mmapped_dev in 0xe00..=0xe02 {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(mmapped_dev, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: true,
cacheable: false,
bufferable: true,
});
}
/* 0xe0300000 - 0xe0ffffff (unassigned/reserved). */
for undef in 0xe03..=0xe0f {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(undef, L1Section {
2019-06-17 09:32:10 +08:00
global: false,
shareable: false,
access: AccessPermissions::PermissionFault,
tex: 0,
domain: 0,
exec: false,
cacheable: false,
bufferable: false,
});
}
/* 0xe1000000 - 0xe1ffffff (NAND) */
for nand in 0xe10..=0xe1f {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(nand, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: true,
cacheable: false,
bufferable: true,
});
}
/* 0xe2000000 - 0xe3ffffff (NOR) */
for nor in 0xe20..=0xe3f {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(nor, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: true,
cacheable: false,
bufferable: true,
});
}
/* 0xe4000000 - 0xe5ffffff (SRAM) */
for nor in 0xe40..=0xe5f {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(nor, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: true,
cacheable: true,
bufferable: true,
});
}
/* 0xe6000000 - 0xf7ffffff (unassigned/reserved). */
for undef in 0xe60..=0xf7f {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(undef, L1Section {
2019-06-17 09:32:10 +08:00
global: false,
shareable: false,
access: AccessPermissions::PermissionFault,
tex: 0,
domain: 0,
exec: false,
cacheable: false,
bufferable: false,
});
}
/* 0xf8000000 - 0xf8ffffff (AMBA APB Peripherals) */
for apb in 0xf80..=0xf8f {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(apb, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: true,
cacheable: false,
bufferable: true,
});
}
/* 0xf9000000 - 0xfbffffff (unassigned/reserved). */
for undef in 0xf90..=0xfbf {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(undef, L1Section {
2019-06-17 09:32:10 +08:00
global: false,
shareable: false,
access: AccessPermissions::PermissionFault,
tex: 0,
domain: 0,
exec: false,
cacheable: false,
bufferable: false,
});
}
/* 0xfc000000 - 0xfdffffff (Linear QSPI - XIP) */
for qspi in 0xfc0..=0xfdf {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(qspi, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0,
domain: 0,
exec: true,
cacheable: false,
bufferable: true,
});
}
/* 0xfe000000 - 0xffefffff (unassigned/reserved). */
for undef in 0xfe0..=0xffe {
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(undef, L1Section {
2019-06-17 09:32:10 +08:00
global: false,
shareable: false,
access: AccessPermissions::PermissionFault,
tex: 0,
domain: 0,
exec: false,
cacheable: false,
bufferable: false,
});
}
/* 0xfff00000 - 0xffffffff (256K OCM when mapped to high address space) */
2019-06-18 08:22:07 +08:00
self.direct_mapped_section(0xfff, L1Section {
2019-06-17 09:32:10 +08:00
global: true,
shareable: false,
access: AccessPermissions::FullAccess,
tex: 0b100,
domain: 0,
exec: true,
cacheable: true,
bufferable: true,
});
2019-06-18 08:22:07 +08:00
self
2019-06-17 09:32:10 +08:00
}
#[inline(always)]
fn direct_mapped_section(&mut self, index: usize, section: L1Section) {
2019-06-18 08:22:07 +08:00
assert!(index < L1_TABLE_SIZE);
2019-06-17 09:32:10 +08:00
let base = (index as u32) << 20;
2020-06-15 05:50:00 +08:00
self.table[index] = L1Entry::from_section(base, section);
}
pub fn update<T, F, R>(&mut self, ptr: *const T, f: F) -> R
where
F: FnOnce(&'_ mut L1Section) -> R,
{
let index = (ptr as usize) >> 20;
let entry = &mut self.table[index];
let mut section = entry.get_section();
let result = f(&mut section);
entry.set_section(section);
2020-06-22 08:02:11 +08:00
// Flush L1Dcache
dcciall();
// // TODO: L2?
// Invalidate TLB
tlbiall();
// Invalidate all branch predictors
bpiall();
// ensure completion of the BP and TLB invalidation
dsb();
// synchronize context on this processor
isb();
2020-06-15 05:50:00 +08:00
result
2019-06-17 09:32:10 +08:00
}
}
pub fn with_mmu<F: FnMut() -> !>(l1table: &L1Table, mut f: F) -> ! {
let domains = AccessDomains::all_manager();
DACR.write(domains.into());
let table_base = &l1table.table as *const _ as u32;
assert!(table_base & 0x3fff == 0);
TTBR0.write(
TTBR0::zeroed()
.irgn1(true)
.s(true)
// Outer Cacheable Write-Back, no allocate on write.
.rgn(0b11)
.irgn0(true)
.table_base(table_base >> 14)
);
2019-06-18 08:22:07 +08:00
2019-06-17 09:32:10 +08:00
// Enable I-Cache and D-Cache
SCTLR.write(
SCTLR::zeroed()
.m(true)
.a(false)
.c(true)
.i(true)
.unaligned(true)
);
// Synchronization barriers
// Allows MMU to start
2020-06-22 08:02:11 +08:00
dsb();
2019-06-17 09:32:10 +08:00
// Flushes pre-fetch buffer
2020-06-22 08:02:11 +08:00
isb();
2019-06-17 09:32:10 +08:00
f();
}