ChanServ changed the topic of #rust-embedded to: Welcome to the Rust Embedded IRC channel! Bridged to #rust-embedded:matrix.org and logged at https://libera.irclog.whitequark.org/rust-embedded, code of conduct at https://www.rust-lang.org/conduct.html
starblue has quit [Ping timeout: 246 seconds]
rom4ik has quit [Quit: Ping timeout (120 seconds)]
rom4ik has joined #rust-embedded
cr1901 has quit [Read error: Connection reset by peer]
cr1901 has joined #rust-embedded
cr1901_ has joined #rust-embedded
sroemer has joined #rust-embedded
cr1901 has quit [Ping timeout: 260 seconds]
cr1901_ is now known as cr1901
sroemer_ has joined #rust-embedded
sroemer__ has joined #rust-embedded
sroemer has quit [Ping timeout: 244 seconds]
sroemer_ has quit [Ping timeout: 252 seconds]
Foxyloxy has quit [Read error: Connection reset by peer]
ivche has joined #rust-embedded
starblue has joined #rust-embedded
pcs38 has joined #rust-embedded
MartinSivk[m] has joined #rust-embedded
<MartinSivk[m]> Good morning folks. I have a question about async GPIO Wait traits from embedded-hal. The short version is - anyone has an idea about how do I make sure there is no race condition between configuration of some device and the call to wait_for_low()? I tried asking in the project, but no answer yet (https://github.com/rust-embedded/embedded-hal/issues/653). Surely this pattern of 1. enable interrupts 2. configure device 3. wait
<MartinSivk[m]> for pin change can't be that unusual.
JamesMunns[m] has joined #rust-embedded
<JamesMunns[m]> Not sure how it would fit into the embedded-hal interface, but `maitake-sync` typically has `subscribe` methods that you can call to ensure the future is "listening" before you perform the triggering action: https://docs.rs/maitake-sync/latest/maitake_sync/wait_queue/struct.Wait.html#method.subscribe
<JamesMunns[m]> So, it might be worth it to have similar methods on e-hal methods to "anchor" the wait like with `subscribe`
danielb[m] has joined #rust-embedded
<danielb[m]> in this case our future behaves somewhat naughtily, though, as you need to poll the driver an unknown amount of times before it actually starts listening
<JamesMunns[m]> oh? Why is that?
<danielb[m]> there is only a single interrupt handler for ALL gpios together, but two cores, and the possibility of user-registered interrupt handlers. The interrupt handler may actually be running on the other core while you start listening for a pin event, and the driver needs to wait for that handler to finish before it reconfigures the pin for the event you actually want
<danielb[m]> it's complicated and boils down to a weird decision where we allow user handlers and async operations on GPIOs
<JamesMunns[m]> so, that might be more appropriate, so you can await as many polls as it takes
<JamesMunns[m]> I think WaitQueue specifically can be done in one poll, maybe that's why we made the switch to a non-async method? I'd have to go back and look tho
<JamesMunns[m]> We also have methods like `wait_for`, which does the right thing in a polling loop, of always re-subscribing before checking the trigger condition: https://docs.rs/maitake-sync/latest/maitake_sync/struct.WaitCell.html#method.wait_for
<JamesMunns[m]> basically like a less-manual version of poll_fn
<danielb[m]> even funnier case is where the task that calls the GPIO wait runs above the priority of the GPIO interrupt handler - so blocking waiting for the handler to finish may be an infinite loop
<JamesMunns[m]> but for that you'd need some kind of counter or something that counts the number of "unyielded triggers", like an interrupt that counts up + wakes every time it happens, and a future that counts down every time the future returns Ready, and only returns Pending if the count is zero
M9names[m] has quit [Quit: Idle timeout reached: 172800s]
<JamesMunns[m]> <danielb[m]> "even funnier case is where the..." <- I feel like this is something you could detect and panic on, honestly.
<JamesMunns[m]> it'd have to be done for like... everything though, which would be tedious and annoying :/
<danielb[m]> but it's not wrong 🤔
<JamesMunns[m]> like, have the driver store its prio level, and in the polling func check the current prio level, if it ever exceeds the stored level, panic
<JamesMunns[m]> danielb[m]: But it's also not ever useful?
<danielb[m]> who am I to say that :D
<JamesMunns[m]> I dunno if esp32 has the ability to see the current execution context (in interrupt, prio level) like cortex-m does, but on cortex-m I think it is possible.
<danielb[m]> ah yeah it's probably incorrect, you should have said that :D
<danielb[m]> wake_by_ref in an interrupt executor isn't a great idea
vollbrecht[m] has joined #rust-embedded
<vollbrecht[m]> also funny when you have a broken edge trigger ( e.g hardware bug) as in the classical esp32 :D E.g you cannot reliable use edge-triggers because under certain conditions the 32bit GPIO status register cannot be updated, if at the same time a interrupt status clear is run. The only workaround is using level-trigger there, wich fundamently works differently, and has its own kinda worms.
<JamesMunns[m]> "userspace" code with higher prio than "interrupt" code is probably isn't ever a great idea :p
<danielb[m]> vollbrecht[m]: wtf
<danielb[m]> the esp32 is such a mess of a chip
<danielb[m]> sometimes the DMA forgets to write back how much it read
<danielb[m]> I think our errata should be about the length of the reference manual
<vollbrecht[m]> * is run. ( This misses edges that happend) The only
<JamesMunns[m]> I wish vendors would start just inlining errata in the reference manual, instead of making them separate docs
<danielb[m]> i wish we had decent documentation
<danielb[m]> 😂
<danielb[m]> technically everything is inlined into esp-idf :D
<JamesMunns[m]> like info/warning boxes, or actually removing things from the RM that were impossible (like the stm32 CAN thing that just never worked)
dngrs[m] has joined #rust-embedded
<dngrs[m]> You'd absolutely love the USB video spec PDF which is just a collection of cross references that aren't hyperlinked
phcoder_cat[m] has quit [Quit: Idle timeout reached: 172800s]
pcs38 has quit [Quit: leaving]
MartinSivk[m]1 has joined #rust-embedded
<MartinSivk[m]1> Depending on the runtime. I actually implemented my own simple no_std runtime that only sets a flag and places the job id into the queue during the wake() call. Fast, no side effects, no synchronization (except disabling interrupts for the atomic set).
<MartinSivk[m]1> Works perfectly fine, at least on single core stm32..
<MartinSivk[m]1> Thanks, this looks like a possible solution. But yes, it is clunky.
<MartinSivk[m]1> I am currently using a simplified version of that style. let wait = pin_w_irq.wait_for_low(); setup(); wait.await; The initial wait method subscribes and returns the future impl. But I was hoping to modernize my code and migrate to the embedded_hal_async style. Alas, this pattern does not currently fit its model.
<MartinSivk[m]1> On a side note.. I wish I had more time to send couple of PRs to the stm32l0x1_hal crate too, it is still based on the 0.2 traits.
sajattack[m]1 has quit [Quit: Idle timeout reached: 172800s]
jason-kairos[m] has quit [Quit: Idle timeout reached: 172800s]
dirbaio[m] has quit [Quit: Idle timeout reached: 172800s]