By efficiently checking when the next work has to be done, we can
increase performance not having to "busy loop". Time from transcription
-> message to LLM agent is now down to sub 1 millisecond.
ref: N25B-316
Main BDI brain structure implemented. Still some TODOs left, and very
basic implementation (only one belief "user_said(Message)" and every
message is sent straight to a function which is responsible for getting
an LLM response.
ref: N25B-197