[
- Currently implements belief collection from text-based mock agent.
- The beliefs communicated by this agent look like this:
{
"type": "belief_extraction_text",
"beliefs": [
{"user_said": [["hello"],["Can you help me?"],["stop talking to me"],["No"],["Pepper do a dance"]]}
]
}
* have yet to add tests (but we want to merge into Dev. asap)
]
[ref]: N25B-206
body: added the llmAgent class and made it run at the start.
modified the bdi_core to send a test message and recieve an awnser from LLM agent
Added a connection to a local llm via lmstudio.
Tests are Tba.
ref: N25B-207
The communication with other agents has been tested with mock data as the other agents (transcriber and belief collector) are not yet implemented.
ref: N25B-208
Main BDI brain structure implemented. Still some TODOs left, and very
basic implementation (only one belief "user_said(Message)" and every
message is sent straight to a function which is responsible for getting
an LLM response.
ref: N25B-197
Inside the `/message` enpoint we put a message onto the internal event
queue, which gets read by TestAgent. This agent, in turn, logs the
message (temporary behaviour until we implement RI integration).
The name TestAgent is of course temporary, this is just for exploratory
purposes.
ref: N25B-165