2022-02-18 19:38:48 +01:00
|
|
|
#![allow(clippy::inline_always)]
|
2019-06-13 23:56:59 +02:00
|
|
|
use core::{
|
|
|
|
cell::Cell,
|
|
|
|
sync::atomic::{AtomicBool, Ordering},
|
|
|
|
};
|
2018-11-03 17:02:41 +01:00
|
|
|
|
2020-12-12 23:24:54 +01:00
|
|
|
pub use crate::tq::{NotReady, TimerQueue};
|
2020-10-01 20:01:25 +02:00
|
|
|
pub use bare_metal::CriticalSection;
|
2018-11-03 17:02:41 +01:00
|
|
|
pub use cortex_m::{
|
2021-12-14 21:52:57 +01:00
|
|
|
asm::nop,
|
2019-06-13 23:56:59 +02:00
|
|
|
asm::wfi,
|
|
|
|
interrupt,
|
2021-09-28 10:18:43 +02:00
|
|
|
peripheral::{scb::SystemHandler, DWT, NVIC, SCB, SYST},
|
2019-06-13 23:56:59 +02:00
|
|
|
Peripherals,
|
2018-11-03 17:02:41 +01:00
|
|
|
};
|
2021-08-16 15:37:39 +02:00
|
|
|
pub use heapless::sorted_linked_list::SortedLinkedList;
|
2021-04-03 19:30:34 +02:00
|
|
|
pub use heapless::spsc::Queue;
|
|
|
|
pub use heapless::BinaryHeap;
|
2021-02-18 19:30:59 +01:00
|
|
|
pub use rtic_monotonic as monotonic;
|
2018-11-03 17:02:41 +01:00
|
|
|
|
2021-04-03 19:30:34 +02:00
|
|
|
pub type SCFQ<const N: usize> = Queue<u8, N>;
|
|
|
|
pub type SCRQ<T, const N: usize> = Queue<(T, u8), N>;
|
2018-11-03 17:02:41 +01:00
|
|
|
|
2021-09-14 16:13:28 +02:00
|
|
|
#[cfg(armv7m)]
|
|
|
|
use cortex_m::register::basepri;
|
|
|
|
|
2018-11-03 17:02:41 +01:00
|
|
|
#[cfg(armv7m)]
|
|
|
|
#[inline(always)]
|
2019-04-21 20:02:59 +02:00
|
|
|
pub fn run<F>(priority: u8, f: F)
|
2018-11-03 17:02:41 +01:00
|
|
|
where
|
|
|
|
F: FnOnce(),
|
|
|
|
{
|
2019-04-21 20:02:59 +02:00
|
|
|
if priority == 1 {
|
2020-09-01 19:04:55 +02:00
|
|
|
// If the priority of this interrupt is `1` then BASEPRI can only be `0`
|
2019-04-21 20:02:59 +02:00
|
|
|
f();
|
|
|
|
unsafe { basepri::write(0) }
|
|
|
|
} else {
|
|
|
|
let initial = basepri::read();
|
|
|
|
f();
|
|
|
|
unsafe { basepri::write(initial) }
|
|
|
|
}
|
2018-11-03 17:02:41 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(not(armv7m))]
|
|
|
|
#[inline(always)]
|
2019-04-21 20:02:59 +02:00
|
|
|
pub fn run<F>(_priority: u8, f: F)
|
2018-11-03 17:02:41 +01:00
|
|
|
where
|
|
|
|
F: FnOnce(),
|
|
|
|
{
|
|
|
|
f();
|
|
|
|
}
|
|
|
|
|
2019-06-13 23:56:59 +02:00
|
|
|
pub struct Barrier {
|
|
|
|
inner: AtomicBool,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl Barrier {
|
|
|
|
pub const fn new() -> Self {
|
|
|
|
Barrier {
|
|
|
|
inner: AtomicBool::new(false),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn release(&self) {
|
2022-02-18 19:38:48 +01:00
|
|
|
self.inner.store(true, Ordering::Release);
|
2019-06-13 23:56:59 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
pub fn wait(&self) {
|
|
|
|
while !self.inner.load(Ordering::Acquire) {}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-02-16 00:22:00 +01:00
|
|
|
// Newtype over `Cell` that forbids mutation through a shared reference
|
|
|
|
pub struct Priority {
|
|
|
|
inner: Cell<u8>,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl Priority {
|
2020-10-13 16:16:33 +02:00
|
|
|
/// Create a new Priority
|
|
|
|
///
|
|
|
|
/// # Safety
|
|
|
|
///
|
|
|
|
/// Will overwrite the current Priority
|
2019-02-16 00:22:00 +01:00
|
|
|
#[inline(always)]
|
|
|
|
pub unsafe fn new(value: u8) -> Self {
|
2019-02-16 00:26:07 +01:00
|
|
|
Priority {
|
|
|
|
inner: Cell::new(value),
|
|
|
|
}
|
2019-02-16 00:22:00 +01:00
|
|
|
}
|
|
|
|
|
2020-10-13 16:16:33 +02:00
|
|
|
/// Change the current priority to `value`
|
2020-09-01 19:04:55 +02:00
|
|
|
// These two methods are used by `lock` (see below) but can't be used from the RTIC application
|
2019-02-16 00:22:00 +01:00
|
|
|
#[inline(always)]
|
|
|
|
fn set(&self, value: u8) {
|
2022-02-18 19:38:48 +01:00
|
|
|
self.inner.set(value);
|
2019-02-16 00:22:00 +01:00
|
|
|
}
|
|
|
|
|
2020-10-13 16:16:33 +02:00
|
|
|
/// Get the current priority
|
2019-02-16 00:22:00 +01:00
|
|
|
#[inline(always)]
|
|
|
|
fn get(&self) -> u8 {
|
|
|
|
self.inner.get()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-01-21 21:49:45 +01:00
|
|
|
/// Const helper to check architecture
|
|
|
|
pub const fn is_armv6() -> bool {
|
|
|
|
#[cfg(not(armv6m))]
|
|
|
|
{
|
|
|
|
false
|
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(armv6m)]
|
|
|
|
{
|
|
|
|
true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-11-03 17:02:41 +01:00
|
|
|
#[inline(always)]
|
|
|
|
pub fn assert_send<T>()
|
|
|
|
where
|
|
|
|
T: Send,
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
#[inline(always)]
|
|
|
|
pub fn assert_sync<T>()
|
|
|
|
where
|
|
|
|
T: Sync,
|
|
|
|
{
|
|
|
|
}
|
2018-11-04 18:50:42 +01:00
|
|
|
|
2020-12-10 20:33:13 +01:00
|
|
|
#[inline(always)]
|
|
|
|
pub fn assert_monotonic<T>()
|
|
|
|
where
|
2021-02-18 19:30:59 +01:00
|
|
|
T: monotonic::Monotonic,
|
2020-12-10 20:33:13 +01:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2022-01-21 21:49:45 +01:00
|
|
|
/// Lock implementation using BASEPRI and global Critical Section (CS)
|
2020-10-13 16:16:33 +02:00
|
|
|
///
|
|
|
|
/// # Safety
|
|
|
|
///
|
2022-01-21 21:49:45 +01:00
|
|
|
/// The system ceiling is raised from current to ceiling
|
|
|
|
/// by either
|
|
|
|
/// - raising the BASEPRI to the ceiling value, or
|
|
|
|
/// - disable all interrupts in case we want to
|
|
|
|
/// mask interrupts with maximum priority
|
|
|
|
///
|
|
|
|
/// Dereferencing a raw pointer inside CS
|
|
|
|
///
|
|
|
|
/// The priority.set/priority.get can safely be outside the CS
|
|
|
|
/// as being a context local cell (not affected by preemptions).
|
|
|
|
/// It is merely used in order to omit masking in case current
|
|
|
|
/// priority is current priority >= ceiling.
|
|
|
|
///
|
|
|
|
/// Lock Efficiency:
|
|
|
|
/// Experiments validate (sub)-zero cost for CS implementation
|
|
|
|
/// (Sub)-zero as:
|
|
|
|
/// - Either zero OH (lock optimized out), or
|
|
|
|
/// - Amounting to an optimal assembly implementation
|
|
|
|
/// - The BASEPRI value is folded to a constant at compile time
|
|
|
|
/// - CS entry, single assembly instruction to write BASEPRI
|
|
|
|
/// - CS exit, single assembly instruction to write BASEPRI
|
|
|
|
/// - priority.set/get optimized out (their effect not)
|
|
|
|
/// - On par or better than any handwritten implementation of SRP
|
|
|
|
///
|
|
|
|
/// Limitations:
|
|
|
|
/// The current implementation reads/writes BASEPRI once
|
|
|
|
/// even in some edge cases where this may be omitted.
|
|
|
|
/// Total OH of per task is max 2 clock cycles, negligible in practice
|
|
|
|
/// but can in theory be fixed.
|
|
|
|
///
|
2018-11-04 18:50:42 +01:00
|
|
|
#[cfg(armv7m)]
|
|
|
|
#[inline(always)]
|
2019-04-21 20:02:59 +02:00
|
|
|
pub unsafe fn lock<T, R>(
|
2018-11-04 18:50:42 +01:00
|
|
|
ptr: *mut T,
|
2019-02-16 00:22:00 +01:00
|
|
|
priority: &Priority,
|
2018-11-04 18:50:42 +01:00
|
|
|
ceiling: u8,
|
|
|
|
nvic_prio_bits: u8,
|
2022-01-21 21:49:45 +01:00
|
|
|
_mask: &[u32; 3],
|
2019-04-21 20:02:59 +02:00
|
|
|
f: impl FnOnce(&mut T) -> R,
|
|
|
|
) -> R {
|
2018-11-04 18:50:42 +01:00
|
|
|
let current = priority.get();
|
|
|
|
|
2019-04-21 20:02:59 +02:00
|
|
|
if current < ceiling {
|
2018-11-04 18:50:42 +01:00
|
|
|
if ceiling == (1 << nvic_prio_bits) {
|
2019-06-13 23:56:59 +02:00
|
|
|
priority.set(u8::max_value());
|
2018-11-04 18:50:42 +01:00
|
|
|
let r = interrupt::free(|_| f(&mut *ptr));
|
|
|
|
priority.set(current);
|
|
|
|
r
|
|
|
|
} else {
|
|
|
|
priority.set(ceiling);
|
|
|
|
basepri::write(logical2hw(ceiling, nvic_prio_bits));
|
|
|
|
let r = f(&mut *ptr);
|
|
|
|
basepri::write(logical2hw(current, nvic_prio_bits));
|
|
|
|
priority.set(current);
|
|
|
|
r
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
f(&mut *ptr)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-01-21 21:49:45 +01:00
|
|
|
/// Lock implementation using interrupt masking
|
2020-10-13 16:16:33 +02:00
|
|
|
///
|
|
|
|
/// # Safety
|
|
|
|
///
|
2022-01-21 21:49:45 +01:00
|
|
|
/// The system ceiling is raised from current to ceiling
|
|
|
|
/// by computing a 32 bit `mask` (1 bit per interrupt)
|
|
|
|
/// 1: ceiling >= priority > current
|
|
|
|
/// 0: else
|
|
|
|
///
|
|
|
|
/// On CS entry, `clear_enable_mask(mask)` disables interrupts
|
|
|
|
/// On CS exit, `set_enable_mask(mask)` re-enables interrupts
|
|
|
|
///
|
|
|
|
/// The priority.set/priority.get can safely be outside the CS
|
|
|
|
/// as being a context local cell (not affected by preemptions).
|
|
|
|
/// It is merely used in order to omit masking in case
|
|
|
|
/// current priority >= ceiling.
|
|
|
|
///
|
|
|
|
/// Dereferencing a raw pointer is done safely inside the CS
|
|
|
|
///
|
|
|
|
/// Lock Efficiency:
|
|
|
|
/// Early experiments validate (sub)-zero cost for CS implementation
|
|
|
|
/// (Sub)-zero as:
|
|
|
|
/// - Either zero OH (lock optimized out), or
|
|
|
|
/// - Amounting to an optimal assembly implementation
|
|
|
|
/// - if ceiling == (1 << nvic_prio_bits)
|
|
|
|
/// - we execute the closure in a global critical section (interrupt free)
|
|
|
|
/// - CS entry cost, single write to core register
|
|
|
|
/// - CS exit cost, single write to core register
|
|
|
|
/// else
|
|
|
|
/// - The `mask` value is folded to a constant at compile time
|
|
|
|
/// - CS entry, single write of the 32 bit `mask` to the `icer` register
|
|
|
|
/// - CS exit, single write of the 32 bit `mask` to the `iser` register
|
|
|
|
/// - priority.set/get optimized out (their effect not)
|
|
|
|
/// - On par or better than any hand written implementation of SRP
|
|
|
|
///
|
|
|
|
/// Limitations:
|
|
|
|
/// Current implementation does not allow for tasks with shared resources
|
|
|
|
/// to be bound to exception handlers, as these cannot be masked in HW.
|
|
|
|
///
|
|
|
|
/// Possible solutions:
|
|
|
|
/// - Mask exceptions by global critical sections (interrupt::free)
|
|
|
|
/// - Temporary lower exception priority
|
|
|
|
///
|
|
|
|
/// These possible solutions are set goals for future work
|
2018-11-04 18:50:42 +01:00
|
|
|
#[cfg(not(armv7m))]
|
|
|
|
#[inline(always)]
|
2019-04-21 20:02:59 +02:00
|
|
|
pub unsafe fn lock<T, R>(
|
2018-11-04 18:50:42 +01:00
|
|
|
ptr: *mut T,
|
2019-02-16 00:22:00 +01:00
|
|
|
priority: &Priority,
|
2018-11-04 18:50:42 +01:00
|
|
|
ceiling: u8,
|
|
|
|
_nvic_prio_bits: u8,
|
2022-01-21 21:49:45 +01:00
|
|
|
masks: &[u32; 3],
|
2019-04-21 20:02:59 +02:00
|
|
|
f: impl FnOnce(&mut T) -> R,
|
|
|
|
) -> R {
|
2018-11-04 18:50:42 +01:00
|
|
|
let current = priority.get();
|
2019-04-21 20:02:59 +02:00
|
|
|
if current < ceiling {
|
2022-01-21 21:49:45 +01:00
|
|
|
if ceiling >= 4 {
|
|
|
|
// safe to manipulate outside critical section
|
|
|
|
priority.set(ceiling);
|
|
|
|
// execute closure under protection of raised system ceiling
|
|
|
|
let r = interrupt::free(|_| f(&mut *ptr));
|
|
|
|
// safe to manipulate outside critical section
|
|
|
|
priority.set(current);
|
|
|
|
r
|
|
|
|
} else {
|
|
|
|
// safe to manipulate outside critical section
|
|
|
|
priority.set(ceiling);
|
|
|
|
let mask = compute_mask(current, ceiling, masks);
|
|
|
|
clear_enable_mask(mask);
|
|
|
|
|
|
|
|
// execute closure under protection of raised system ceiling
|
|
|
|
let r = f(&mut *ptr);
|
|
|
|
|
|
|
|
set_enable_mask(mask);
|
|
|
|
|
|
|
|
// safe to manipulate outside critical section
|
|
|
|
priority.set(current);
|
|
|
|
r
|
|
|
|
}
|
2018-11-04 18:50:42 +01:00
|
|
|
} else {
|
2022-01-21 21:49:45 +01:00
|
|
|
// execute closure without raising system ceiling
|
2018-11-04 18:50:42 +01:00
|
|
|
f(&mut *ptr)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-01-21 21:49:45 +01:00
|
|
|
#[cfg(not(armv7m))]
|
|
|
|
#[inline(always)]
|
|
|
|
fn compute_mask(from_prio: u8, to_prio: u8, masks: &[u32; 3]) -> u32 {
|
|
|
|
let mut res = 0;
|
|
|
|
masks[from_prio as usize..to_prio as usize]
|
|
|
|
.iter()
|
|
|
|
.for_each(|m| res |= m);
|
|
|
|
res
|
|
|
|
}
|
|
|
|
|
|
|
|
// enables interrupts
|
|
|
|
#[cfg(not(armv7m))]
|
|
|
|
#[inline(always)]
|
|
|
|
unsafe fn set_enable_mask(mask: u32) {
|
|
|
|
(*NVIC::ptr()).iser[0].write(mask)
|
|
|
|
}
|
|
|
|
|
|
|
|
// disables interrupts
|
|
|
|
#[cfg(not(armv7m))]
|
|
|
|
#[inline(always)]
|
|
|
|
unsafe fn clear_enable_mask(mask: u32) {
|
|
|
|
(*NVIC::ptr()).icer[0].write(mask)
|
|
|
|
}
|
|
|
|
|
2018-11-04 18:50:42 +01:00
|
|
|
#[inline]
|
2022-02-18 19:38:48 +01:00
|
|
|
#[must_use]
|
2019-04-21 20:02:59 +02:00
|
|
|
pub fn logical2hw(logical: u8, nvic_prio_bits: u8) -> u8 {
|
2018-11-04 18:50:42 +01:00
|
|
|
((1 << nvic_prio_bits) - logical) << (8 - nvic_prio_bits)
|
|
|
|
}
|
2022-04-20 10:46:03 +02:00
|
|
|
|
|
|
|
#[cfg(not(armv6m))]
|
|
|
|
pub const fn create_mask<const N: usize>(_: [u32; N]) -> u32 {
|
|
|
|
0
|
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(armv6m)]
|
|
|
|
pub const fn create_mask<const N: usize>(list_of_shifts: [u32; N]) -> u32 {
|
|
|
|
let mut mask = 0;
|
|
|
|
let mut i = 0;
|
|
|
|
|
|
|
|
while i < N {
|
|
|
|
let shift = list_of_shifts[i];
|
|
|
|
i += 1;
|
|
|
|
|
|
|
|
if shift > 31 {
|
|
|
|
panic!("Generating masks for thumbv6 failed! Are you compiling for thumbv6 on an thumbv7 MCU?");
|
|
|
|
}
|
|
|
|
|
|
|
|
mask |= 1 << shift;
|
|
|
|
}
|
|
|
|
|
|
|
|
mask
|
|
|
|
}
|
2022-04-20 13:02:55 +02:00
|
|
|
|
|
|
|
#[cfg(not(armv6m))]
|
|
|
|
pub const fn v6_panic() {
|
|
|
|
// For non-v6 all is fine
|
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(armv6m)]
|
|
|
|
pub const fn v6_panic() {
|
|
|
|
panic!("Exceptions with shared resources are not allowed when compiling for thumbv6. Use local resources or `#[lock_free]` shared resources");
|
|
|
|
}
|