Trezor Core event loop
The event loop is implemented in src/trezor/loop.py
and forms the core of the
processing. At boot time, default tasks are started and inserted into an event queue.
Such task will usually run in an endless loop: wait for event, process event, loop back.
Application code is written with async/await
constructs. Low level of the event queue
processes running coroutines via coroutine.send()
and coroutine.throw()
calls.
MicroPython details
MicroPython does not distinguish between coroutines, awaitables, and generators. Some
low-level constructs are using yield
and yield from
constructions.
async def
definition marks the function as a generator, even if it does not contain
await
or yield
expressions. It is thus possible to see async def __iter__
, which
indicates that the function is a generator.
For type-checking purposes, objects usually define an __await__
method that delegates
to __iter__
. The __await__
method is never executed, however.
Low-level API
Function summary
loop.run()
starts the event loop. The call only returns when there are no further
waiting tasks -- so, in usual conditions, never.
loop.schedule(task, value, deadline, finalizer, reschedule)
schedules an awaitable to
be run either as soon as possible, or at a specified time (given as a deadline
in
microseconds since system bootup.)
In addition, when the task finishes processing or is closed externally, the finalizer
callback will be executed, with the task and the return value (or the raised exception)
as a parameter.
If reschedule
is true, the task is first cleared from the scheduled queue -- in
effect, it is rescheduled to run at a different time.
loop.close(task)
removes a previously scheduled task from the list of waiting tasks
and calls its finalizer.
loop.pause(task, interface)
sets the task as waiting for a particular interface:
either reading from or writing to one of the USB interfaces, or waiting for a touch
event.
Implementation details
Trezor Core runs coroutine-based cooperative multitasking, i.e., there is no preemption.
Every task is a coroutine, which means that it runs uninterrupted until it yields a
value (or, in async terms, until it awaits something). In every processing step, the
currently selected coroutine is resumed by sending a value to it (which is returned as a
result of the yield/await, or raised as an exception if it is an instance of
BaseException
). The tasks then runs uninterrupted again, until it yields or exits.
A loop in loop.run()
spins for as long as any tasks are waiting. Two lists of waiting
tasks exist:
-
_queue
is a priority queue where the ordering is defined by real-time deadlines. In most cases, tasks are scheduled for "now", which makes them run one after another in FIFO order. It is also possible to schedule a task to run in the future. -
_paused
is a collection of tasks grouped by the interface for which they are waiting.
In each run of the loop, io.poll
is called to query I/O events. If an event arrives on
an interface, all tasks waiting on that interface are resumed one after another. No
scheduled tasks in _queue
can execute until the waiting tasks yield again.
At most one I/O event is processsed in this phase.
When the I/O phase is done, a task with the highest priority is popped from _queue
and
resumed.
I/O wait
When no tasks are paused on a given interface, events on that interface remain in queue.
When multiple tasks are paused on the same interface, all of them receive every event. However, a waiting task receives at most one event. To receive more, it must pause itself again. Event processing is usually done in an endless loop with a pause call.
If two tasks are attempting to read from the same interface, and one of them re-pauses
itself immediately while the other doesn't (possibly due to use of loop.race
, which
introduces scheduling gaps), the other task might lose some events.
For this reason, you should avoid waiting on the same interface from multiple tasks.
Syscalls
Syscalls bridge the gap between await
-based application code and the coroutine-based
low-level implementation.
Every sequence of await
s will at some point boil down to yielding a Syscall
instance. (Yielding anything else is an error.) When that happens, control returns to
the event loop.
The handle(task)
method is called on the result. This way the syscall gets hold of the
task object, and can schedule()
or pause()
it as appropriate.
As an example, consider pausing on an input event. A running task has no way to call
pause()
on itself. It would need to pass a separate function as a callback.
The wait
syscall can be implemented as a simple wrapper around the pause()
low-level
call:
class wait(Syscall):
def __init__(self, msg_iface: int) -> None:
self.msg_iface = msg_iface
def handle(self, task: Task) -> None:
pause(task, self.msg_iface)
The __init__()
method takes all the arguments of the "call", and handle()
pauses the
task on the given interface.
Calling code will look like this:
event = await loop.wait(io.TOUCH)
The loop.wait(io.TOUCH)
expression instantiates a new Syscall
object. The argument
is passed to the constructor, and stored on the instance. The rest boils down to
event = await some_syscall_instance
which is equivalent to
event = yield from some_syscall_instance.__iter__()
The Syscall.__iter__()
method yields self
, returning control to the event loop. The
event loop invokes some_syscall_instance.handle(task_object)
. The task_object
is
then set to resume when a touch event arrives.
A side-effect of this design is that it is possible to store and reuse syscall instances. That can be advantageous for avoiding unnecessary allocations.
while True:
# every run of the loop allocates a new object
event = await loop.wait(io.TOUCH)
process_event(event)
touch_source = loop.wait(io.TOUCH)
while True:
# same instance is reused
event = await touch_source
process_event(event)
High-level API
Application code should not be using any of the above low-level functions. Awaiting syscalls is the preferred method of writing code.
The following syscalls and constructs are available:
loop.sleep(delay_ms: int)
: Suspend execution until the given delay (in
milliseconds) elapses. Return value is the planned deadline in milliseconds since system
start.
Calling await loop.sleep(0)
yields execution to other tasks, and schedules the current
task for the next tick.
loop.wait(interface)
: Wait indefinitely for an event on the given interface.
Return value is the event.
Upcoming code modification adds a timeout parameter to loop.wait
.
loop.race(*children)
: Schedule each argument to run, and suspend execution until
the first of them finishes.
It is possible to specify wait timeout for loop.wait
by using loop.race
:
result = await loop.race(loop.wait(io.TOUCH), loop.sleep(1000))
This introduces scheduling gaps: every child is treated as a task and scheduled
to run. This means that if the child is a syscall, as in the above example, its action
is not done immediately. Instead, the wait
begins on the next tick (or whenever the
newly created coroutine runs) and the sleep
in the tick afterwards. When nesting
multiple race
s, the child races
also run later.
Also, when a child task is done, another scheduling gap happens, and the parent task is scheduled to run on the next tick.
Upcoming changes may solve this in relevant cases, by inlining syscall operations.
loop.mailbox()
is an unidirectional communication channel, simplification of Go
channels.
It allows to put a value
in the channel using put(value)
on the mailbox instance.
To retrieve the value, use await
on the same mailbox
instance.
loop.spawn(task)
: Start the task asynchronously. Return an object that allows
the caller to await its result, or shut the task down.
Example usage:
task = loop.spawn(some_background_task())
await do_something_here()
result = await task
Unlike other syscalls, loop.spawn
starts the task at instantiation time. await
ing
the same loop.spawn
instance a second time will immediately return the result of the
original run.
If the task is cancelled (usually by calling task.close()
), the awaiter receives a
loop.TaskClosed
exception.
It is also possible to register a synchronous finalizer callback via
task.set_finalizer
. This is used internally to implement workflow management.