use bit_field::BitField; use super::{regs::*, asm::*, cache::*}; use libregister::RegisterW; #[derive(Copy, Clone)] #[repr(u8)] pub enum AccessDomain { NoAccess = 0b00, Client = 0b01, _Reserved = 0b10, Manager = 0b11, } const ACCESS_DOMAINS_SIZE: usize = 16; pub struct AccessDomains([AccessDomain; ACCESS_DOMAINS_SIZE]); impl AccessDomains { pub fn all_manager() -> Self { AccessDomains([AccessDomain::Manager; ACCESS_DOMAINS_SIZE]) } } impl Into for AccessDomains { fn into(self) -> u32 { let mut result = 0; for (i, domain) in self.0.iter().enumerate() { result |= (*domain as u32) << (2 * i); } result } } #[derive(Copy, Clone)] #[repr(u8)] pub enum AccessPermissions { PermissionFault = 0, PrivilegedOnly, NoUserWrite, FullAccess, _Reserved1, PrivilegedReadOnly, ReadOnly, _Reserved2 } impl AccessPermissions { fn new(ap: u8, apx: bool) -> Self { unsafe { core::mem::transmute(if apx { 0b100 } else { 0 } | ap) } } fn ap(&self) -> u8 { (*self as u8) & 0b11 } fn apx(&self) -> bool { (*self as u8) > (AccessPermissions::FullAccess as u8) } } pub struct L1Section { pub global: bool, pub shareable: bool, pub access: AccessPermissions, /// Type EXtension pub tex: u8, pub domain: u8, pub exec: bool, pub cacheable: bool, pub bufferable: bool, } const ENTRY_TYPE_SECTION: u32 = 0b10; pub const L1_PAGE_SIZE: usize = 0x100000; #[repr(C)] #[derive(Clone, Copy)] pub struct L1Entry(u32); impl L1Entry { #[inline(always)] pub fn from_section(phys_base: u32, section: L1Section) -> Self { // Must be aligned to 1 MB assert!(phys_base & 0x000f_ffff == 0); let mut entry = L1Entry(phys_base); entry.set_section(section); entry } pub fn get_section(&mut self) -> L1Section { assert_eq!(self.0.get_bits(0..=1), ENTRY_TYPE_SECTION); let access = AccessPermissions::new( self.0.get_bits(10..=11) as u8, self.0.get_bit(15) ); L1Section { global: !self.0.get_bit(17), shareable: self.0.get_bit(16), access, tex: self.0.get_bits(12..=14) as u8, domain: self.0.get_bits(5..=8) as u8, exec: !self.0.get_bit(4), cacheable: self.0.get_bit(3), bufferable: self.0.get_bit(2), } } pub fn set_section(&mut self, section: L1Section) { self.0.set_bits(0..=1, ENTRY_TYPE_SECTION); self.0.set_bit(2, section.bufferable); self.0.set_bit(3, section.cacheable); self.0.set_bit(4, !section.exec); assert!(section.domain < 16); self.0.set_bits(5..=8, section.domain.into()); self.0.set_bits(10..=11, section.access.ap().into()); assert!(section.tex < 8); self.0.set_bits(12..=14, section.tex.into()); self.0.set_bit(15, section.access.apx()); self.0.set_bit(16, section.shareable); self.0.set_bit(17, !section.global); } } const L1_TABLE_SIZE: usize = 4096; static mut L1_TABLE: L1Table = L1Table { table: [L1Entry(0); L1_TABLE_SIZE] }; #[repr(C, align(16384))] pub struct L1Table { table: [L1Entry; L1_TABLE_SIZE] } impl L1Table { pub fn get() -> &'static mut Self { unsafe { &mut L1_TABLE } } pub fn setup_flat_layout(&mut self) -> &Self { /* 0x00000000 - 0x00100000 (cacheable) */ self.direct_mapped_section(0, L1Section { global: true, shareable: true, access: AccessPermissions::FullAccess, tex: 0b101, domain: 0b1111, exec: true, cacheable: true, bufferable: true, }); /* (DDR cacheable) */ for ddr in 1..=0x3ff { self.direct_mapped_section(ddr, L1Section { global: true, shareable: true, access: AccessPermissions::FullAccess, tex: 0b0, domain: 0b1111, exec: true, cacheable: true, bufferable: true, }); } /* 0x40000000 - 0x7fffffff (FPGA slave0) */ for fpga_slave in 0x400..=0x7ff { self.direct_mapped_section(fpga_slave, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: false, }); } /* 0x80000000 - 0xbfffffff (FPGA slave1) */ for fpga_slave in 0x800..=0xbff { self.direct_mapped_section(fpga_slave, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: false, }); } /* 0xc0000000 - 0xdfffffff (unassigned/reserved). */ for undef in 0xc00..=0xdff { self.direct_mapped_section(undef, L1Section { global: false, shareable: false, access: AccessPermissions::PermissionFault, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: false, }); } /* 0xe0000000 - 0xe02fffff (Memory mapped devices) * UART/USB/IIC/SPI/CAN/GEM/GPIO/QSPI/SD/NAND */ for mmapped_dev in 0xe00..=0xe02 { self.direct_mapped_section(mmapped_dev, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: true, }); } /* 0xe0300000 - 0xe0ffffff (unassigned/reserved). */ for undef in 0xe03..=0xe0f { self.direct_mapped_section(undef, L1Section { global: false, shareable: false, access: AccessPermissions::PermissionFault, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: false, }); } /* 0xe1000000 - 0xe1ffffff (NAND) */ for nand in 0xe10..=0xe1f { self.direct_mapped_section(nand, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: true, cacheable: false, bufferable: true, }); } /* 0xe2000000 - 0xe3ffffff (NOR) */ for nor in 0xe20..=0xe3f { self.direct_mapped_section(nor, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: true, cacheable: false, bufferable: true, }); } /* 0xe4000000 - 0xe5ffffff (SRAM) */ for nor in 0xe40..=0xe5f { self.direct_mapped_section(nor, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: true, cacheable: true, bufferable: true, }); } /* 0xe6000000 - 0xf7ffffff (unassigned/reserved). */ for undef in 0xe60..=0xf7f { self.direct_mapped_section(undef, L1Section { global: false, shareable: false, access: AccessPermissions::PermissionFault, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: false, }); } /* 0xf8000000 - 0xf8ffffff (AMBA APB Peripherals) */ for apb in 0xf80..=0xf8f { self.direct_mapped_section(apb, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: true, cacheable: false, bufferable: true, }); } /* 0xf9000000 - 0xfbffffff (unassigned/reserved). */ for undef in 0xf90..=0xfbf { self.direct_mapped_section(undef, L1Section { global: false, shareable: false, access: AccessPermissions::PermissionFault, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: false, }); } /* 0xfc000000 - 0xfdffffff (Linear QSPI - XIP) */ for qspi in 0xfc0..=0xfdf { self.direct_mapped_section(qspi, L1Section { global: true, shareable: false, access: AccessPermissions::FullAccess, tex: 0, domain: 0, exec: true, cacheable: false, bufferable: true, }); } /* 0xfe000000 - 0xffefffff (unassigned/reserved). */ for undef in 0xfe0..=0xffe { self.direct_mapped_section(undef, L1Section { global: false, shareable: false, access: AccessPermissions::PermissionFault, tex: 0, domain: 0, exec: false, cacheable: false, bufferable: false, }); } /* 0xfff00000 - 0xffffffff (256K OCM when mapped to high address space) */ self.direct_mapped_section(0xfff, L1Section { global: true, shareable: true, access: AccessPermissions::FullAccess, tex: 0b100, domain: 0, exec: true, cacheable: true, bufferable: true, }); self } #[inline(always)] fn direct_mapped_section(&mut self, index: usize, section: L1Section) { assert!(index < L1_TABLE_SIZE); let base = (index as u32) << 20; self.table[index] = L1Entry::from_section(base, section); } pub fn update(&mut self, ptr: *const T, f: F) -> R where F: FnOnce(&'_ mut L1Section) -> R, { let index = (ptr as usize) >> 20; let entry = &mut self.table[index]; let mut section = entry.get_section(); let result = f(&mut section); entry.set_section(section); // Flush L1Dcache dcciall(); // // TODO: L2? // Invalidate TLB tlbiall(); // Invalidate all branch predictors bpiall(); // ensure completion of the BP and TLB invalidation dsb(); // synchronize context on this processor isb(); result } } pub fn with_mmu !>(l1table: &L1Table, mut f: F) -> ! { let domains = AccessDomains::all_manager(); DACR.write(domains.into()); let table_base = &l1table.table as *const _ as u32; assert!(table_base & 0x3fff == 0); TTBR0.write( TTBR0::zeroed() .irgn1(true) .s(true) // Outer Cacheable Write-Back, no allocate on write. .rgn(0b11) .irgn0(true) .table_base(table_base >> 14) ); // Enable I-Cache and D-Cache SCTLR.write( SCTLR::zeroed() .m(true) .a(false) .c(true) .i(true) .z(true) .unaligned(true) ); // Synchronization barriers // Allows MMU to start dsb(); // Flushes pre-fetch buffer isb(); f(); }