1
0
Fork 0
zynq-rs/src/eth/mod.rs

488 lines
15 KiB
Rust
Raw Normal View History

2019-05-08 01:28:33 +08:00
use crate::regs::*;
2019-05-25 09:06:39 +08:00
use crate::slcr;
2019-05-08 01:28:33 +08:00
2019-05-30 08:42:42 +08:00
pub mod phy;
2019-05-08 01:28:33 +08:00
mod regs;
pub mod rx;
pub mod tx;
2019-05-08 01:28:33 +08:00
2019-06-22 07:34:17 +08:00
/// Size of all the buffers
pub const MTU: usize = 1536;
2019-06-26 03:50:15 +08:00
pub const IO_PLL: u32 = 1_000;
2019-06-22 07:34:17 +08:00
pub struct Eth<RX, TX> {
regs: &'static mut regs::RegisterBlock,
rx: RX,
tx: TX,
2019-05-08 01:28:33 +08:00
}
impl Eth<(), ()> {
2019-06-09 07:02:10 +08:00
pub fn default(macaddr: [u8; 6]) -> Self {
2019-05-25 09:06:39 +08:00
slcr::RegisterBlock::unlocked(|slcr| {
// Manual example: 0x0000_1280
2019-05-25 09:06:39 +08:00
// MDIO
slcr.mio_pin_53.write(
slcr::MioPin53::zeroed()
.l3_sel(0b100)
.io_type(slcr::IoBufferType::Lvcmos18)
.pullup(true)
);
// MDC
slcr.mio_pin_52.write(
slcr::MioPin52::zeroed()
.l3_sel(0b100)
.io_type(slcr::IoBufferType::Lvcmos18)
.pullup(true)
);
// Manual example: 0x0000_3902
2019-05-25 09:06:39 +08:00
// TX_CLK
slcr.mio_pin_16.write(
slcr::MioPin16::zeroed()
.l0_sel(true)
2019-06-29 06:00:22 +08:00
.speed(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
.disable_rcvr(true)
2019-05-25 09:06:39 +08:00
);
// TX_CTRL
slcr.mio_pin_21.write(
slcr::MioPin21::zeroed()
.l0_sel(true)
2019-06-29 06:00:22 +08:00
.speed(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
.disable_rcvr(true)
2019-05-25 09:06:39 +08:00
);
// TXD3
slcr.mio_pin_20.write(
slcr::MioPin20::zeroed()
.l0_sel(true)
2019-06-29 06:00:22 +08:00
.speed(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
.disable_rcvr(true)
2019-05-25 09:06:39 +08:00
);
// TXD2
slcr.mio_pin_19.write(
slcr::MioPin19::zeroed()
2019-06-29 06:00:22 +08:00
.l0_sel(true)
.speed(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
.disable_rcvr(true)
2019-05-25 09:06:39 +08:00
);
// TXD1
slcr.mio_pin_18.write(
slcr::MioPin18::zeroed()
2019-06-29 06:00:22 +08:00
.l0_sel(true)
.speed(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
.disable_rcvr(true)
2019-05-25 09:06:39 +08:00
);
// TXD0
slcr.mio_pin_17.write(
slcr::MioPin17::zeroed()
.l0_sel(true)
2019-06-29 06:00:22 +08:00
.speed(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
.disable_rcvr(true)
2019-05-25 09:06:39 +08:00
);
// Manual example: 0x0000_1903
2019-05-25 09:06:39 +08:00
// RX_CLK
slcr.mio_pin_22.write(
slcr::MioPin22::zeroed()
.tri_enable(true)
.l0_sel(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
);
// RX_CTRL
slcr.mio_pin_27.write(
slcr::MioPin27::zeroed()
.tri_enable(true)
.l0_sel(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
);
// RXD3
slcr.mio_pin_26.write(
slcr::MioPin26::zeroed()
.tri_enable(true)
.l0_sel(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
);
// RXD2
slcr.mio_pin_25.write(
slcr::MioPin25::zeroed()
.tri_enable(true)
.l0_sel(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
);
// RXD1
slcr.mio_pin_24.write(
slcr::MioPin24::zeroed()
.tri_enable(true)
.l0_sel(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
);
// RXD0
slcr.mio_pin_23.write(
slcr::MioPin23::zeroed()
.tri_enable(true)
.l0_sel(true)
.io_type(slcr::IoBufferType::Hstl)
2019-05-25 09:06:39 +08:00
.pullup(true)
);
// VREF internal generator
slcr.gpiob_ctrl.write(
slcr::GpiobCtrl::zeroed()
.vref_en(true)
);
2019-05-25 09:06:39 +08:00
});
2019-06-09 07:02:10 +08:00
Self::gem0(macaddr)
2019-05-25 09:06:39 +08:00
}
2019-06-09 07:02:10 +08:00
pub fn gem0(macaddr: [u8; 6]) -> Self {
2019-06-26 03:50:15 +08:00
Self::setup_gem0_clock(125);
let regs = regs::RegisterBlock::gem0();
Self::from_regs(regs, macaddr)
}
pub fn gem1(macaddr: [u8; 6]) -> Self {
Self::setup_gem1_clock(125);
let regs = regs::RegisterBlock::gem1();
Self::from_regs(regs, macaddr)
}
fn from_regs(regs: &'static mut regs::RegisterBlock, macaddr: [u8; 6]) -> Self {
let mut eth = Eth {
regs,
rx: (),
tx: (),
}.init();
eth.configure(macaddr);
eth
}
}
impl<RX, TX> Eth<RX, TX> {
pub fn setup_gem0_clock(tx_clock: u32) {
let d0 = (IO_PLL / tx_clock).min(63);
let d1 = (IO_PLL / tx_clock / d0).min(63);
2019-05-30 08:26:19 +08:00
slcr::RegisterBlock::unlocked(|slcr| {
slcr.gem0_clk_ctrl.write(
// 0x0050_0801: 8, 5: 100 Mb/s
2019-06-26 03:50:15 +08:00
// ...: 8, 1: 1000 Mb/s
2019-05-30 08:26:19 +08:00
slcr::ClkCtrl::zeroed()
.clkact(true)
.srcsel(slcr::PllSource::IoPll)
2019-06-26 03:50:15 +08:00
.divisor(d0 as u8)
.divisor1(d1 as u8)
2019-05-30 08:26:19 +08:00
);
2019-06-26 03:50:15 +08:00
// Enable gem0 recv clock
slcr.gem0_rclk_ctrl.write(
// 0x0000_0801
slcr::RclkCtrl::zeroed()
.clkact(true)
);
2019-05-30 08:26:19 +08:00
});
2019-05-08 01:28:33 +08:00
}
2019-06-26 03:50:15 +08:00
pub fn setup_gem1_clock(tx_clock: u32) {
let d0 = (IO_PLL / tx_clock).min(63);
let d1 = (IO_PLL / tx_clock / d0).min(63);
slcr::RegisterBlock::unlocked(|slcr| {
slcr.gem1_clk_ctrl.write(
slcr::ClkCtrl::zeroed()
.clkact(true)
.srcsel(slcr::PllSource::IoPll)
2019-06-26 03:50:15 +08:00
.divisor(d0 as u8)
.divisor1(d1 as u8)
);
2019-06-26 03:50:15 +08:00
// Enable gem1 recv clock
slcr.gem1_rclk_ctrl.write(
2019-06-26 03:50:15 +08:00
// 0x0000_0801
slcr::RclkCtrl::zeroed()
.clkact(true)
);
});
}
2019-06-05 05:48:09 +08:00
fn init(mut self) -> Self {
2019-05-08 01:28:33 +08:00
// Clear the Network Control register.
self.regs.net_ctrl.write(regs::NetCtrl::zeroed());
self.regs.net_ctrl.write(regs::NetCtrl::zeroed().clear_stat_regs(true));
// Clear the Status registers.
self.regs.rx_status.write(
regs::RxStatus::zeroed()
.buffer_not_avail(true)
.frame_recd(true)
.rx_overrun(true)
.hresp_not_ok(true)
);
self.regs.tx_status.write(
regs::TxStatus::zeroed()
.used_bit_read(true)
.collision(true)
.retry_limit_exceeded(true)
.tx_go(true)
.tx_corr_ahb_err(true)
.tx_complete(true)
.tx_under_run(true)
.late_collision(true)
// not in the manual:
2019-05-08 01:28:33 +08:00
.hresp_not_ok(true)
);
// Disable all interrupts.
self.regs.intr_dis.write(
regs::IntrDis::zeroed()
.mgmt_done(true)
.rx_complete(true)
.rx_used_read(true)
.tx_used_read(true)
.tx_underrun(true)
.retry_ex_late_collisn(true)
.tx_corrupt_ahb_err(true)
.tx_complete(true)
.link_chng(true)
.rx_overrun(true)
.hresp_not_ok(true)
.pause_nonzeroq(true)
.pause_zero(true)
.pause_tx(true)
.ex_intr(true)
.autoneg_complete(true)
.partner_pg_rx(true)
.delay_req_rx(true)
.sync_rx(true)
.delay_req_tx(true)
.sync_tx(true)
.pdelay_req_rx(true)
.pdelay_resp_rx(true)
.pdelay_req_tx(true)
.pdelay_resp_tx(true)
.tsu_sec_incr(true)
);
// Clear the buffer queues.
self.regs.rx_qbar.write(
regs::RxQbar::zeroed()
);
self.regs.tx_qbar.write(
regs::TxQbar::zeroed()
);
self
}
2019-05-25 09:06:39 +08:00
2019-06-09 07:02:10 +08:00
fn configure(&mut self, macaddr: [u8; 6]) {
2019-05-25 09:06:39 +08:00
self.regs.net_cfg.write(
regs::NetCfg::zeroed()
.full_duplex(true)
.gige_en(true)
.speed(true)
.no_broadcast(false)
.multi_hash_en(true)
2019-06-22 07:20:18 +08:00
// Promiscuous mode (TODO?)
2019-05-25 09:06:39 +08:00
.copy_all(true)
2019-06-22 07:20:18 +08:00
// Remove 4-byte Frame CheckSum
.fcs_remove(true)
// One of the slower speeds
// TODO: calculate properly
.mdc_clk_div(0b110)
2019-05-25 09:06:39 +08:00
);
2019-06-09 07:02:10 +08:00
let macaddr_msbs =
(u16::from(macaddr[0]) << 8) |
u16::from(macaddr[1]);
let macaddr_lsbs =
(u32::from(macaddr[2]) << 24) |
(u32::from(macaddr[3]) << 16) |
(u32::from(macaddr[4]) << 8) |
u32::from(macaddr[5]);
self.regs.spec_addr1_top.write(
regs::SpecAddrTop::zeroed()
.addr_msbs(macaddr_msbs)
);
self.regs.spec_addr1_bot.write(
regs::SpecAddrBot::zeroed()
.addr_lsbs(macaddr_lsbs)
);
self.regs.dma_cfg.write(
regs::DmaCfg::zeroed()
// 1536 bytes
2019-06-22 07:34:17 +08:00
.ahb_mem_rx_buf_size((MTU >> 6) as u8)
2019-06-09 07:02:10 +08:00
// 8 KB
.rx_pktbuf_memsz_sel(0x3)
// 4 KB
.tx_pktbuf_memsz_sel(true)
// .csum_gen_offload_en(true)
// Little-endian
.ahb_endian_swp_mgmt_en(false)
// INCR16 AHB burst
.ahb_fixed_burst_len(0x10)
);
2019-05-30 08:42:42 +08:00
self.regs.net_ctrl.write(
regs::NetCtrl::zeroed()
.mgmt_port_en(true)
);
}
2019-06-22 07:34:17 +08:00
pub fn start_rx<'rx>(self, rx_list: &'rx mut [rx::DescEntry], rx_buffers: &'rx mut [[u8; MTU]]) -> Eth<rx::DescList<'rx>, TX> {
let new_self = Eth {
regs: self.regs,
2019-06-22 07:34:17 +08:00
rx: rx::DescList::new(rx_list, rx_buffers),
tx: self.tx,
};
let list_addr = new_self.rx.list_addr();
2019-06-09 08:23:37 +08:00
assert!(list_addr & 0b11 == 0);
new_self.regs.rx_qbar.write(
2019-06-09 07:02:10 +08:00
regs::RxQbar::zeroed()
.rx_q_baseaddr(list_addr >> 2)
);
new_self.regs.net_ctrl.modify(|_, w|
w.rx_en(true)
2019-06-10 02:28:33 +08:00
);
new_self
2019-06-09 07:02:10 +08:00
}
2019-06-22 07:34:47 +08:00
pub fn start_tx<'tx>(self, tx_list: &'tx mut [tx::DescEntry], tx_buffers: &'tx mut [[u8; MTU]]) -> Eth<RX, tx::DescList<'tx>> {
let new_self = Eth {
regs: self.regs,
rx: self.rx,
tx: tx::DescList::new(tx_list, tx_buffers),
};
let list_addr = &new_self.tx.list_addr();
assert!(list_addr & 0b11 == 0);
new_self.regs.tx_qbar.write(
regs::TxQbar::zeroed()
.tx_q_baseaddr(list_addr >> 2)
);
new_self.regs.net_ctrl.modify(|_, w|
w.tx_en(true)
);
new_self
}
2019-06-22 07:20:18 +08:00
2019-05-30 08:42:42 +08:00
fn wait_phy_idle(&self) {
2019-06-05 05:48:09 +08:00
while !self.regs.net_status.read().phy_mgmt_idle() {}
2019-05-30 08:42:42 +08:00
}
2019-06-26 03:48:47 +08:00
pub fn reset_phy(&mut self) -> bool {
match phy::Phy::find(self) {
Some(phy) => {
phy.modify_control(self, |control|
control.set_reset(true)
);
while phy.get_control(self).reset() {
println!("Wait for PHY reset");
}
phy.modify_control(self, |control|
control.set_autoneg_enable(true)
.set_restart_autoneg(true)
);
// 125 MHz for 1000base-TX
Self::setup_gem0_clock(125);
true
}
None => false
}
}
2019-05-30 08:42:42 +08:00
}
2019-06-10 08:44:29 +08:00
impl<'rx, TX> Eth<rx::DescList<'rx>, TX> {
2019-06-22 07:34:17 +08:00
pub fn recv_next<'s: 'p, 'p>(&'s mut self) -> Result<Option<rx::PktRef<'p>>, rx::Error> {
2019-06-22 07:20:18 +08:00
let status = self.regs.rx_status.read();
if status.hresp_not_ok() {
// Clear
self.regs.rx_status.write(
regs::RxStatus::zeroed()
.hresp_not_ok(true)
);
return Err(rx::Error::HrespNotOk);
}
if status.rx_overrun() {
// Clear
self.regs.rx_status.write(
regs::RxStatus::zeroed()
.rx_overrun(true)
);
return Err(rx::Error::RxOverrun);
}
if status.buffer_not_avail() {
// Clear
self.regs.rx_status.write(
regs::RxStatus::zeroed()
.buffer_not_avail(true)
);
return Err(rx::Error::BufferNotAvail);
}
if status.frame_recd() {
let result = self.rx.recv_next();
match result {
Ok(None) => {
// No packet, clear status bit
self.regs.rx_status.write(
regs::RxStatus::zeroed()
.frame_recd(true)
);
}
_ => {}
}
result
} else {
Ok(None)
}
2019-06-10 08:44:29 +08:00
}
}
2019-06-22 07:34:47 +08:00
impl<'tx, RX> Eth<RX, tx::DescList<'tx>> {
pub fn send<'s: 'p, 'p>(&'s mut self, length: usize) -> Option<tx::PktRef<'p>> {
self.tx.send(&mut self.regs, length)
}
}
impl<RX, TX> phy::PhyAccess for Eth<RX, TX> {
2019-05-30 08:42:42 +08:00
fn read_phy(&mut self, addr: u8, reg: u8) -> u16 {
self.wait_phy_idle();
self.regs.phy_maint.write(
regs::PhyMaint::zeroed()
.clause_22(true)
.operation(regs::PhyOperation::Read)
.phy_addr(addr)
.reg_addr(reg)
.must_10(0b10)
);
self.wait_phy_idle();
self.regs.phy_maint.read().data()
}
fn write_phy(&mut self, addr: u8, reg: u8, data: u16) {
self.wait_phy_idle();
self.regs.phy_maint.write(
regs::PhyMaint::zeroed()
.clause_22(true)
.operation(regs::PhyOperation::Write)
.phy_addr(addr)
.reg_addr(reg)
.must_10(0b10)
.data(data)
);
self.wait_phy_idle();
2019-05-25 09:06:39 +08:00
}
2019-05-08 01:28:33 +08:00
}