rom4ik has quit [Quit: Ping timeout (120 seconds)]
rom4ik has joined #rust-embedded
cr1901 has quit [Read error: Connection reset by peer]
cr1901 has joined #rust-embedded
cr1901_ has joined #rust-embedded
sroemer has joined #rust-embedded
cr1901 has quit [Ping timeout: 260 seconds]
cr1901_ is now known as cr1901
sroemer_ has joined #rust-embedded
sroemer__ has joined #rust-embedded
sroemer has quit [Ping timeout: 244 seconds]
sroemer_ has quit [Ping timeout: 252 seconds]
Foxyloxy has quit [Read error: Connection reset by peer]
ivche has joined #rust-embedded
starblue has joined #rust-embedded
pcs38 has joined #rust-embedded
MartinSivk[m] has joined #rust-embedded
<MartinSivk[m]>
Good morning folks. I have a question about async GPIO Wait traits from embedded-hal. The short version is - anyone has an idea about how do I make sure there is no race condition between configuration of some device and the call to wait_for_low()? I tried asking in the project, but no answer yet (https://github.com/rust-embedded/embedded-hal/issues/653). Surely this pattern of 1. enable interrupts 2. configure device 3. wait
<MartinSivk[m]>
for pin change can't be that unusual.
<JamesMunns[m]>
So, it might be worth it to have similar methods on e-hal methods to "anchor" the wait like with `subscribe`
danielb[m] has joined #rust-embedded
<danielb[m]>
in this case our future behaves somewhat naughtily, though, as you need to poll the driver an unknown amount of times before it actually starts listening
<JamesMunns[m]>
oh? Why is that?
<danielb[m]>
there is only a single interrupt handler for ALL gpios together, but two cores, and the possibility of user-registered interrupt handlers. The interrupt handler may actually be running on the other core while you start listening for a pin event, and the driver needs to wait for that handler to finish before it reconfigures the pin for the event you actually want
<danielb[m]>
it's complicated and boils down to a weird decision where we allow user handlers and async operations on GPIOs
<JamesMunns[m]>
so, that might be more appropriate, so you can await as many polls as it takes
<JamesMunns[m]>
I think WaitQueue specifically can be done in one poll, maybe that's why we made the switch to a non-async method? I'd have to go back and look tho
<JamesMunns[m]>
basically like a less-manual version of poll_fn
<danielb[m]>
even funnier case is where the task that calls the GPIO wait runs above the priority of the GPIO interrupt handler - so blocking waiting for the handler to finish may be an infinite loop
<JamesMunns[m]>
but for that you'd need some kind of counter or something that counts the number of "unyielded triggers", like an interrupt that counts up + wakes every time it happens, and a future that counts down every time the future returns Ready, and only returns Pending if the count is zero
M9names[m] has quit [Quit: Idle timeout reached: 172800s]
<JamesMunns[m]>
<danielb[m]> "even funnier case is where the..." <- I feel like this is something you could detect and panic on, honestly.
<JamesMunns[m]>
it'd have to be done for like... everything though, which would be tedious and annoying :/
<danielb[m]>
but it's not wrong 🤔
<JamesMunns[m]>
like, have the driver store its prio level, and in the polling func check the current prio level, if it ever exceeds the stored level, panic
<JamesMunns[m]>
danielb[m]: But it's also not ever useful?
<danielb[m]>
who am I to say that :D
<JamesMunns[m]>
I dunno if esp32 has the ability to see the current execution context (in interrupt, prio level) like cortex-m does, but on cortex-m I think it is possible.
<danielb[m]>
ah yeah it's probably incorrect, you should have said that :D
<danielb[m]>
wake_by_ref in an interrupt executor isn't a great idea
vollbrecht[m] has joined #rust-embedded
<vollbrecht[m]>
also funny when you have a broken edge trigger ( e.g hardware bug) as in the classical esp32 :D E.g you cannot reliable use edge-triggers because under certain conditions the 32bit GPIO status register cannot be updated, if at the same time a interrupt status clear is run. The only workaround is using level-trigger there, wich fundamently works differently, and has its own kinda worms.
<JamesMunns[m]>
"userspace" code with higher prio than "interrupt" code is probably isn't ever a great idea :p
<danielb[m]>
vollbrecht[m]: wtf
<danielb[m]>
the esp32 is such a mess of a chip
<danielb[m]>
sometimes the DMA forgets to write back how much it read
<danielb[m]>
I think our errata should be about the length of the reference manual
<vollbrecht[m]>
* is run. ( This misses edges that happend) The only
<JamesMunns[m]>
I wish vendors would start just inlining errata in the reference manual, instead of making them separate docs
<danielb[m]>
i wish we had decent documentation
<danielb[m]>
😂
<danielb[m]>
technically everything is inlined into esp-idf :D
<JamesMunns[m]>
like info/warning boxes, or actually removing things from the RM that were impossible (like the stm32 CAN thing that just never worked)
dngrs[m] has joined #rust-embedded
<dngrs[m]>
You'd absolutely love the USB video spec PDF which is just a collection of cross references that aren't hyperlinked
phcoder_cat[m] has quit [Quit: Idle timeout reached: 172800s]
pcs38 has quit [Quit: leaving]
MartinSivk[m]1 has joined #rust-embedded
<MartinSivk[m]1>
Depending on the runtime. I actually implemented my own simple no_std runtime that only sets a flag and places the job id into the queue during the wake() call. Fast, no side effects, no synchronization (except disabling interrupts for the atomic set).
<MartinSivk[m]1>
Works perfectly fine, at least on single core stm32..
<MartinSivk[m]1>
Thanks, this looks like a possible solution. But yes, it is clunky.
<MartinSivk[m]1>
I am currently using a simplified version of that style. let wait = pin_w_irq.wait_for_low(); setup(); wait.await; The initial wait method subscribes and returns the future impl. But I was hoping to modernize my code and migrate to the embedded_hal_async style. Alas, this pattern does not currently fit its model.
<MartinSivk[m]1>
On a side note.. I wish I had more time to send couple of PRs to the stm32l0x1_hal crate too, it is still based on the 0.2 traits.
sajattack[m]1 has quit [Quit: Idle timeout reached: 172800s]
jason-kairos[m] has quit [Quit: Idle timeout reached: 172800s]
dirbaio[m] has quit [Quit: Idle timeout reached: 172800s]