15 Commits

Author SHA1 Message Date
JobvAlewijk
e129113c9e chore: forgot copyright 2026-01-29 17:53:57 +01:00
JobvAlewijk
18a4bde4ca test: added tests and docstrings
ref: N25B-397
2026-01-29 17:51:39 +01:00
Twirre Meulenbelt
815fc7bcde feat: publish face detection instead of req/res
ref: N25B-395
2026-01-29 17:18:39 +01:00
JobvAlewijk
3bbe97579d Merge branch 'main' of ssh://git.science.uu.nl/ics/sp/2025/n25b/pepperplus-ri into feat/face-detection 2026-01-29 16:22:42 +01:00
ad58b16559 Merge branch 'dev' into 'main'
Merge dev with main

See merge request ics/sp/2025/n25b/pepperplus-ri!27
2026-01-28 10:54:22 +00:00
fb0d7850cc Merge branch 'main' into dev 2026-01-28 11:53:23 +01:00
JobvAlewijk
4afceccf46 feat: fixed connection !!! 2026-01-19 16:58:01 +01:00
JobvAlewijk
83099a2810 chore: modified into req reply socket on 5559 2026-01-17 14:01:32 +01:00
JobvAlewijk
4e9afbaaf5 Merge branch 'dev' of ssh://git.science.uu.nl/ics/sp/2025/n25b/pepperplus-ri into feat/face-detection 2026-01-16 16:48:51 +01:00
Pim Hutting
da97eb8a1a Merge branch 'feat/robot-speech-agent-force-speech' into 'dev'
feat: implemented forced speech and speech queue

See merge request ics/sp/2025/n25b/pepperplus-ri!23
2026-01-14 14:26:39 +00:00
Luijkx,S.O.H. (Storm)
e51cf8fe65 feat: implemented forced speech and speech queue 2026-01-14 14:26:38 +00:00
JobvAlewijk
49386ef8cd feat: communicate face to CB
Had to do some weird socket stuff

ref: N25B-397
2026-01-12 14:25:10 +01:00
JobvAlewijk
3b470c8f29 feat: fully working face detection
ref: N25B-397
2026-01-07 17:56:21 +01:00
JobvAlewijk
b8f71f6bee feat: base face detection
ref: N25B-397
2026-01-04 18:56:04 +01:00
aad2044b6e chore: add .gitignore 2025-09-27 17:58:12 +02:00
11 changed files with 422 additions and 197 deletions

View File

@@ -17,6 +17,10 @@ class AgentSettings(object):
:vartype video_sender_port: int
:ivar audio_sender_port: Port used for sending audio data, defaults to 5558.
:vartype audio_sender_port: int
:ivar face_detection_port: Port used for sending face detection events, defaults to 5559.
:vartype face_detection_port: int
:ivar face_detection_interval: Time between face detection events, defaults to 1000 ms.
:vartype face_detection_interval: int
"""
def __init__(
self,
@@ -25,12 +29,16 @@ class AgentSettings(object):
main_receiver_port=None,
video_sender_port=None,
audio_sender_port=None,
face_detection_port=None,
face_detection_interval=None,
):
self.control_backend_host = get_config(control_backend_host, "AGENT__CONTROL_BACKEND_HOST", "localhost")
self.actuation_receiver_port = get_config(actuation_receiver_port, "AGENT__ACTUATION_RECEIVER_PORT", 5557, int)
self.main_receiver_port = get_config(main_receiver_port, "AGENT__MAIN_RECEIVER_PORT", 5555, int)
self.video_sender_port = get_config(video_sender_port, "AGENT__VIDEO_SENDER_PORT", 5556, int)
self.audio_sender_port = get_config(audio_sender_port, "AGENT__AUDIO_SENDER_PORT", 5558, int)
self.face_detection_port = get_config(face_detection_port, "AGENT__FACE_DETECTION_PORT", 5559, int)
self.face_detection_interval = get_config(face_detection_interval, "AGENT__FACE_DETECTION_INTERVAL", 1000, int)
class VideoConfig(object):

View File

@@ -1,6 +1,5 @@
from __future__ import unicode_literals # So that we can log texts with Unicode characters
import logging
import time
from threading import Thread
import Queue
@@ -8,7 +7,6 @@ import zmq
from robot_interface.endpoints.receiver_base import ReceiverBase
from robot_interface.state import state
from robot_interface.core.config import settings
from robot_interface.endpoints.gesture_settings import GestureTags
@@ -36,11 +34,8 @@ class ActuationReceiver(ReceiverBase):
self._tts_service = None
self._animation_service = None
self._message_queue = Queue.Queue()
self._gesture_queue = Queue.Queue()
self.message_thread = Thread(target=self._handle_messages)
self.message_thread.start()
self.gesture_thread = Thread(target=self._handle_gestures)
self.gesture_thread.start()
def _handle_speech(self, message):
"""
@@ -67,7 +62,7 @@ class ActuationReceiver(ReceiverBase):
if not self._tts_service:
self._tts_service = state.qi_session.service("ALTextToSpeech")
if (message.get("is_priority")):
if message.get("is_priority"):
# Bypass queue and speak immediately
self.clear_queue()
self._message_queue.put(text)
@@ -75,7 +70,6 @@ class ActuationReceiver(ReceiverBase):
else:
self._message_queue.put(text)
def clear_queue(self):
"""
Safely drains all pending messages from the queue.
@@ -89,21 +83,6 @@ class ActuationReceiver(ReceiverBase):
pass
logging.info("Message queue cleared.")
def clear_gesture_queue(self):
"""
Safely drains all pending gestures from the gesture queue.
"""
logging.info("Gesture queue size: {}".format(self._gesture_queue.qsize()))
try:
while True:
# Remove items one by one without waiting
self._gesture_queue.get_nowait()
except Queue.Empty:
pass
logging.info("Gesture queue cleared.")
logging.info("Gesture queue size: {}".format(self._gesture_queue.qsize()))
def _handle_gesture(self, message, is_single):
"""
Handle a gesture actuation request.
@@ -144,16 +123,12 @@ class ActuationReceiver(ReceiverBase):
# Play the gesture. Pepper comes with predefined animations like "Wave", "Greet", "Clap"
# You can also create custom animations using Choregraphe and upload them to the robot.
if (message.get("is_priority")):
# Clear queue and play
self.clear_gesture_queue()
logging.debug("Force playing gesture immediately: {}".format(gesture))
if is_single:
logging.debug("Adding single gesture to queue: {}".format(gesture))
logging.debug("Playing single gesture: {}".format(gesture))
getattr(qi, "async")(self._animation_service.run, gesture)
else:
logging.debug("Adding tag gesture to queue: {}".format(gesture))
self._gesture_queue.put(gesture)
logging.debug("Playing tag gesture: {}".format(gesture))
getattr(qi, "async")(self._animation_service.runTag, gesture)
def handle_message(self, message):
"""
@@ -173,25 +148,13 @@ class ActuationReceiver(ReceiverBase):
while not state.exit_event.is_set():
try:
text = self._message_queue.get(timeout=0.1)
if not state.is_speaking: print("Started speaking.")
state.is_speaking = True
self._tts_service.say(text)
except Queue.Empty:
if state.is_speaking: print("Finished speaking.")
state.is_speaking = False
except RuntimeError:
logging.warn("Lost connection to Pepper. Please check if you're connected to the local WiFi and restart this application.")
state.exit_event.set()
def _handle_gestures(self):
while not state.exit_event.is_set():
try:
gesture = self._gesture_queue.get(timeout=0.1)
self._animation_service.run(gesture)
except Queue.Empty:
pass
except RuntimeError:
logging.warn("Lost connection to Pepper. Please check if you're connected to the local WiFi and restart this application.")
logging.error("Lost connection to Pepper. Please check if you're connected to the "
"local WiFi and restart this application.")
state.exit_event.set()
def endpoint_description(self):

View File

@@ -89,6 +89,7 @@ class AudioSender(SocketBase):
try:
while not state.exit_event.is_set():
data = stream.read(chunk)
if (state.is_speaking): continue # Do not send audio while the robot is speaking
self.socket.send(data)
except IOError as e:
logger.error("Stopped listening: failed to get audio from microphone.", exc_info=e)

View File

@@ -0,0 +1,93 @@
# -*- coding: utf-8 -*-
"""
This program has been developed by students from the bachelor Computer Science at Utrecht
University within the Software Project course.
© Copyright Utrecht University (Department of Information and Computing Sciences)
"""
from __future__ import unicode_literals
import json
import logging
import threading
import time
import zmq
from robot_interface.endpoints.socket_base import SocketBase
from robot_interface.state import state
from robot_interface.core.config import settings
class FaceDetectionSender(SocketBase):
"""
Face detection endpoint.
Subscribes to and polls ALMemory["FaceDetected"], sends events to CB.
"""
def __init__(self, zmq_context, port=settings.agent_settings.face_detection_port):
super(FaceDetectionSender, self).__init__("face")
self.create_socket(zmq_context, zmq.PUB, port)
self._face_service = None
self._memory_service = None
self._face_thread = None
def start_face_detection(self):
if not state.qi_session:
logging.warning("No Qi session available. Face detection not started.")
return
self._face_service = state.qi_session.service("ALFaceDetection")
self._memory_service = state.qi_session.service("ALMemory")
self._face_service.setTrackingEnabled(False)
self._face_service.setRecognitionEnabled(False)
self._face_service.subscribe(
"FaceDetectionSender",
settings.agent_settings.face_detection_interval,
0.0,
)
self._face_thread = threading.Thread(target=self._face_loop)
self._face_thread.start()
logging.info("Face detection started.")
def _face_loop(self):
"""
Continuously send face detected to the CB, at the interval set in the
``start_face_detection`` method.
"""
while not state.exit_event.is_set():
try:
value = self._memory_service.getData("FaceDetected", 0)
face_present = (
value
and len(value) > 1
and value[1]
and value[1][0]
and len(value[1][0]) > 0
)
self.socket.send(json.dumps({"face_detected": face_present}).encode("utf-8"))
except Exception:
logging.exception("Error reading FaceDetected")
time.sleep(settings.agent_settings.face_detection_interval / 1000.0)
def stop_face_detection(self):
try:
if self._face_service:
self._face_service.unsubscribe("FaceDetectionSender")
self._face_service.setTrackingEnabled(False)
logging.info("Face detection stopped.")
except Exception:
logging.warning("Error during face detection cleanup.")
def close(self):
super(FaceDetectionSender, self).close()
self.stop_face_detection()

View File

@@ -4,6 +4,7 @@ from robot_interface.endpoints.receiver_base import ReceiverBase
from robot_interface.state import state
from robot_interface.core.config import settings
from robot_interface.endpoints.face_detector import FaceDetectionSender
class MainReceiver(ReceiverBase):
@@ -37,6 +38,7 @@ class MainReceiver(ReceiverBase):
"""
return {"endpoint": "ping", "data": message.get("data")}
@staticmethod
def _handle_port_negotiation(message):
"""

View File

@@ -6,6 +6,7 @@ from robot_interface.endpoints.socket_base import SocketBase
from robot_interface.state import state
from robot_interface.core.config import settings
class VideoSender(SocketBase):
"""
Video sender endpoint, responsible for sending video frames.

View File

@@ -12,6 +12,8 @@ from robot_interface.endpoints.video_sender import VideoSender
from robot_interface.state import state
from robot_interface.core.config import settings
from robot_interface.utils.timeblock import TimeBlock
from robot_interface.endpoints.face_detector import FaceDetectionSender
def main_loop(context):
@@ -35,6 +37,12 @@ def main_loop(context):
video_sender.start_video_rcv()
audio_sender.start()
# --- Face detection sender ---
face_sender = FaceDetectionSender(context)
state.sockets.append(face_sender)
face_sender.start_face_detection()
# Sockets that can run on the main thread. These sockets' endpoints should not block for long (say 50 ms at most).
receivers = [main_receiver, actuation_receiver]

View File

@@ -1,13 +1,25 @@
import sys
import time
import mock
import pytest
import zmq
import Queue
from robot_interface.endpoints.actuation_receiver import ActuationReceiver
from robot_interface.endpoints.gesture_settings import GestureTags
@pytest.fixture
def zmq_context():
"""
A pytest fixture that creates and yields a ZMQ context.
:return: An initialized ZeroMQ context.
:rtype: zmq.Context
"""
context = zmq.Context()
yield context
def test_force_speech_clears_queue(mocker):
"""
Tests that a force speech message clears the existing queue
@@ -42,6 +54,7 @@ def test_force_speech_clears_queue(mocker):
queued_item = receiver._message_queue.get()
assert queued_item == "Emergency Notification"
def test_handle_unimplemented_endpoint(mocker):
"""
Tests handling of unknown endpoints.
@@ -57,6 +70,7 @@ def test_handle_unimplemented_endpoint(mocker):
"data": None,
})
def test_speech_message_no_data(mocker):
"""
Tests that if the message data is empty, the receiver returns immediately
@@ -214,7 +228,7 @@ def test_handle_messages_loop(mocker):
receiver._tts_service = mock_tts_service
# This ensures the while loop iterates exactly once
mock_state.exit_event.is_set.side_effect = [False, True, True , True, True]
mock_state.exit_event.is_set.side_effect = [False, True]
# Put an item in the queue
receiver._message_queue.put("Hello World")
@@ -229,24 +243,40 @@ def test_handle_messages_loop(mocker):
assert mock_state.is_speaking is True
def test_handle_gestures_runtime_error(mocker):
def test_handle_messages_queue_empty(mocker):
"""
Tests the Queue.Empty exception handler in the consumer loop.
This covers the logic that resets 'state.is_speaking' to False.
"""
# Prevent the real background thread from starting
mocker.patch("threading.Thread")
# Mock the state object
mock_state = mocker.patch("robot_interface.endpoints.actuation_receiver.state")
# Use a side_effect that returns False then True thereafter
mock_state.exit_event.is_set.side_effect = [False, True, True, True, True]
# Setup 'is_speaking' property mock
# We set return_value=True so the code enters the 'if state.is_speaking:' block.
# We use PropertyMock to track when this attribute is set.
type(mock_state).is_speaking = True
mock_zmq_ctx = mock.Mock()
receiver = ActuationReceiver(mock_zmq_ctx)
# ... rest of your setup ...
mock_anim = mock.Mock()
mock_anim.run.side_effect = RuntimeError("Wifi Lost")
receiver._animation_service = mock_anim
# This ensures the while loop body runs exactly once for our test
mock_state.exit_event.is_set.side_effect = [False, True]
receiver._gesture_queue.put("wave")
# Force get() to raise Queue.Empty immediately (simulate timeout)
# We patch the 'get' method on the specific queue instance of our receiver
#mocker.patch.object(receiver._message_queue, 'get', side_effect=Queue.Empty)
# Run the loop logic manually (synchronously)
receiver._handle_messages()
# Final Assertion: Verify is_speaking was set to False
# The code execution order is: read (returns True) -> print -> set (to False)
# assert_called_with checks the arguments of the LAST call, which is the setter.
assert mock_state.is_speaking is False
receiver._handle_gestures()
def test_handle_messages_runtime_error(mocker):
"""
@@ -266,7 +296,7 @@ def test_handle_messages_runtime_error(mocker):
# Initialize receiver with the mock context
receiver = ActuationReceiver(mock_zmq_ctx)
mock_state.exit_event.is_set.side_effect = [False, True, True, True ]
mock_state.exit_event.is_set.side_effect = [False, True]
receiver._message_queue.put("Test Message")
@@ -304,61 +334,54 @@ def test_clear_queue(mocker):
# Assert the queue is empty
assert receiver._message_queue.qsize() == 0
def test_gesture_no_data(mocker):
mock_zmq = mock.Mock()
receiver = ActuationReceiver(mock_zmq)
def test_gesture_no_data(zmq_context, mocker):
receiver = ActuationReceiver(zmq_context)
receiver._handle_gesture({"endpoint": "actuate/gesture/single", "data": ""}, True)
# Just ensuring no crash
def test_gesture_invalid_data(mocker):
mock_zmq = mock.Mock()
receiver = ActuationReceiver(mock_zmq)
def test_gesture_invalid_data(zmq_context, mocker):
receiver = ActuationReceiver(zmq_context)
receiver._handle_gesture({"endpoint": "actuate/gesture/single", "data": 123}, True)
# No crash expected
def test_gesture_single_not_found(mocker):
mock_zmq = mock.Mock()
def test_gesture_single_not_found(zmq_context, mocker):
mock_tags = mocker.patch("robot_interface.endpoints.actuation_receiver.GestureTags")
mock_tags.single_gestures = ["wave", "bow"] # allowed single gestures
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
receiver._handle_gesture({"endpoint": "actuate/gesture/single", "data": "unknown_gesture"}, True)
# No crash expected
def test_gesture_tag_not_found(mocker):
mock_zmq = mock.Mock()
def test_gesture_tag_not_found(zmq_context, mocker):
mock_tags = mocker.patch("robot_interface.endpoints.actuation_receiver.GestureTags")
mock_tags.tags = ["happy", "sad"]
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
receiver._handle_gesture({"endpoint": "actuate/gesture/tag", "data": "not_a_tag"}, False)
# No crash expected
def test_gesture_no_qi_session( mocker):
mock_zmq = mock.Mock()
def test_gesture_no_qi_session(zmq_context, mocker):
mock_state = mocker.patch("robot_interface.endpoints.actuation_receiver.state")
mock_state.qi_session = None
mock_tags = mocker.patch("robot_interface.endpoints.actuation_receiver.GestureTags")
mock_tags.single_gestures = ["hello"]
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
receiver._handle_gesture({"endpoint": "actuate/gesture/single", "data": "hello"}, True)
# No crash, path returns early
def test_gesture_single_success(mocker):
mock_zmq = mock.Mock()
def test_gesture_single_success(zmq_context, mocker):
mock_state = mocker.patch("robot_interface.endpoints.actuation_receiver.state")
# Allow loops to run
mock_state.exit_event.is_set.return_value = False
mock_qi = mock.Mock()
sys.modules["qi"] = mock_qi
# Setup gesture settings
mock_tags = mocker.patch("robot_interface.endpoints.actuation_receiver.GestureTags")
mock_tags.single_gestures = ["wave"]
@@ -366,25 +389,17 @@ def test_gesture_single_success(mocker):
mock_state.qi_session = mock.Mock()
mock_state.qi_session.service.return_value = mock_animation_service
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
receiver._handle_gesture({"endpoint": "actuate/gesture/single", "data": "wave"}, True)
time.sleep(0.2)
mock_state.qi_session.service.assert_called_with("ALAnimationPlayer")
mock_animation_service.run.assert_called_with("wave")
# CLEANUP: Signal exit AND join threads
mock_state.exit_event.is_set.return_value = True
receiver.message_thread.join(timeout=1.0)
receiver.gesture_thread.join(timeout=1.0)
mock_state.qi_session.service.assert_called_once_with("ALAnimationPlayer")
getattr(mock_qi, "async").assert_called_once()
assert getattr(mock_qi, "async").call_args[0][0] == mock_animation_service.run
assert getattr(mock_qi, "async").call_args[0][1] == "wave"
def test_gesture_tag_success(mocker):
mock_zmq = mock.Mock()
def test_gesture_tag_success(zmq_context, mocker):
mock_state = mocker.patch("robot_interface.endpoints.actuation_receiver.state")
mock_state.exit_event.is_set.return_value = False
mock_qi = mock.Mock()
sys.modules["qi"] = mock_qi
@@ -395,26 +410,21 @@ def test_gesture_tag_success(mocker):
mock_state.qi_session = mock.Mock()
mock_state.qi_session.service.return_value = mock_animation_service
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
receiver._handle_gesture({"endpoint": "actuate/gesture/tag", "data": "greeting"}, False)
time.sleep(0.2)
mock_state.qi_session.service.assert_called_with("ALAnimationPlayer")
mock_animation_service.run.assert_called_with("greeting")
# CLEANUP: Signal exit AND join threads
mock_state.exit_event.is_set.return_value = True
receiver.message_thread.join(timeout=1.0)
receiver.gesture_thread.join(timeout=1.0)
mock_state.qi_session.service.assert_called_once_with("ALAnimationPlayer")
getattr(mock_qi, "async").assert_called_once()
assert getattr(mock_qi, "async").call_args[0][0] == mock_animation_service.runTag
assert getattr(mock_qi, "async").call_args[0][1] == "greeting"
def test_handle_message_all_routes(mocker):
def test_handle_message_all_routes(zmq_context, mocker):
"""
Ensures all handle_message endpoint branches route correctly.
"""
mock_zmq = mock.Mock()
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
mock_speech = mocker.patch.object(receiver, "_handle_speech")
mock_gesture = mocker.patch.object(receiver, "_handle_gesture")
@@ -426,13 +436,12 @@ def test_handle_message_all_routes(mocker):
assert mock_gesture.call_count == 2
def test_endpoint_description(mocker):
mock_zmq = mock.Mock()
def test_endpoint_description(zmq_context, mocker):
mock_tags = mocker.patch("robot_interface.endpoints.actuation_receiver.GestureTags")
mock_tags.tags = ["happy"]
mock_tags.single_gestures = ["wave"]
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
desc = receiver.endpoint_description()
assert "gestures" in desc
@@ -442,20 +451,26 @@ def test_endpoint_description(mocker):
assert desc["single_gestures"] == ["wave"]
def test_gesture_single_real_gesturetags(mocker):
mock_zmq = mock.Mock()
def test_gesture_single_real_gesturetags(zmq_context, mocker):
"""
Uses the real GestureTags (no mocking) to ensure the receiver
references GestureTags.single_gestures correctly.
"""
# Ensure qi session exists so we pass the early return
mock_state = mocker.patch("robot_interface.endpoints.actuation_receiver.state")
mock_state.qi_session = mock.Mock()
mock_state.exit_event.is_set.return_value = False
# Mock qi.async to avoid real async calls
mock_qi = mock.Mock()
sys.modules["qi"] = mock_qi
# Mock animation service
mock_animation_service = mock.Mock()
mock_state.qi_session.service.return_value = mock_animation_service
receiver = ActuationReceiver(mock_zmq)
receiver = ActuationReceiver(zmq_context)
# Pick a real gesture from GestureTags.single_gestures
assert len(GestureTags.single_gestures) > 0, "GestureTags.single_gestures must not be empty"
gesture = GestureTags.single_gestures[0]
@@ -464,85 +479,8 @@ def test_gesture_single_real_gesturetags(mocker):
is_single=True,
)
time.sleep(0.2)
mock_state.qi_session.service.assert_called_with("ALAnimationPlayer")
mock_animation_service.run.assert_called_with(gesture)
mock_state.qi_session.service.assert_called_once_with("ALAnimationPlayer")
getattr(mock_qi, "async").assert_called_once()
assert getattr(mock_qi, "async").call_args[0][0] == mock_animation_service.run
assert getattr(mock_qi, "async").call_args[0][1] == gesture
# CLEANUP: Signal exit AND join threads
mock_state.exit_event.is_set.return_value = True
receiver.message_thread.join(timeout=1.0)
receiver.gesture_thread.join(timeout=1.0)
def test_clear_gesture_queue(mocker):
# Prevent background threads from eating the items
mocker.patch("threading.Thread")
mock_zmq_ctx = mock.Mock()
receiver = ActuationReceiver(mock_zmq_ctx)
# Populate the queue
receiver._gesture_queue.put("gesture1")
receiver._gesture_queue.put("gesture2")
assert receiver._gesture_queue.qsize() == 2
# Clear the queue
receiver.clear_gesture_queue()
# Assert the queue is empty
assert receiver._gesture_queue.qsize() == 0
def test_gesture_priority_clears_queue(mocker):
mocker.patch("threading.Thread")
mock_state = mocker.patch("robot_interface.endpoints.actuation_receiver.state")
# Mock QI and Tags so valid checks pass
mock_qi = mock.Mock()
sys.modules["qi"] = mock_qi
mock_tags = mocker.patch("robot_interface.endpoints.actuation_receiver.GestureTags")
mock_tags.single_gestures = ["urgent_wave"]
# Setup Animation Service
mock_anim = mock.Mock()
mock_state.qi_session.service.return_value = mock_anim
mock_zmq_ctx = mock.Mock()
receiver = ActuationReceiver(mock_zmq_ctx)
# Pre-fill queue with "slow" gestures
receiver._gesture_queue.put("slow_gesture_1")
receiver._gesture_queue.put("slow_gesture_2")
assert receiver._gesture_queue.qsize() == 2
# Send priority gesture
priority_msg = {
"endpoint": "actuate/gesture/single",
"data": "urgent_wave",
"is_priority": True,
}
receiver._handle_gesture(priority_msg, is_single=True)
# Assert old items are gone and only new one remains
assert receiver._gesture_queue.qsize() == 1
assert receiver._gesture_queue.get() == "urgent_wave"
def test_handle_gestures_loop_empty(mocker):
mocker.patch("threading.Thread")
mock_state = mocker.patch("robot_interface.endpoints.actuation_receiver.state")
mock_zmq_ctx = mock.Mock()
receiver = ActuationReceiver(mock_zmq_ctx)
# Run loop exactly once
mock_state.exit_event.is_set.side_effect = [False, True, True, True]
# We don't put anything in the queue, so .get(timeout=0.1) will raise Queue.Empty.
# The code should catch it and pass.
receiver._handle_gestures()
# If we reached here without raising an exception, the test passes.
# We can assert that the queue is still valid/empty.
assert receiver._gesture_queue.empty()

View File

@@ -78,6 +78,7 @@ def test_sending_audio(mocker):
mock_zmq_context = mock.Mock()
send_socket = mock.Mock()
mock_state.is_speaking = False
# If there's something wrong with the microphone, it will raise an IOError when `read`ing.
stream = mock.Mock()
stream.read = _fake_read
@@ -93,6 +94,36 @@ def test_sending_audio(mocker):
send_socket.assert_called()
def test_no_sending_if_speaking(mocker):
"""
Tests the successful sending of audio data over a ZeroMQ socket.
"""
mock_choose_mic = mocker.patch("robot_interface.endpoints.audio_sender.choose_mic")
mock_choose_mic.return_value = {"name": u"Some mic", "index": 0L}
mock_state = mocker.patch("robot_interface.endpoints.audio_sender.state")
mock_state.exit_event.is_set.side_effect = [False, True]
mock_zmq_context = mock.Mock()
send_socket = mock.Mock()
mock_state.is_speaking = True
# If there's something wrong with the microphone, it will raise an IOError when `read`ing.
stream = mock.Mock()
stream.read = _fake_read
sender = AudioSender(mock_zmq_context)
sender.socket.send = send_socket
sender.audio.open = mock.Mock()
sender.audio.open.return_value = stream
sender.start()
sender.wait_until_done()
send_socket.assert_not_called()
def _fake_read_error(num_frames):
"""
Helper function to simulate an I/O error during microphone stream reading.

View File

@@ -0,0 +1,175 @@
# -*- coding: utf-8 -*-
"""
This program has been developed by students from the bachelor Computer Science at Utrecht
University within the Software Project course.
© Copyright Utrecht University (Department of Information and Computing Sciences)
"""
from __future__ import unicode_literals
import json
import mock
import pytest
from robot_interface.endpoints.face_detector import FaceDetectionSender
from robot_interface.state import state
@pytest.fixture(autouse=True)
def initialized_state(monkeypatch):
"""
Fully initialize global state so __getattribute__ allows access.
"""
# Bypass the initialization guard
monkeypatch.setattr(state, "is_initialized", True, raising=False)
# Install a controllable exit_event
exit_event = mock.Mock()
exit_event.is_set = mock.Mock(return_value=True)
monkeypatch.setattr(state, "exit_event", exit_event, raising=False)
# Default qi_session is None unless overridden
monkeypatch.setattr(state, "qi_session", None, raising=False)
yield
def test_start_face_detection_no_qi_session():
"""
Returns early when qi_session is None.
"""
sender = FaceDetectionSender(mock.Mock())
sender.start_face_detection()
assert sender._face_thread is None
assert sender._face_service is None
assert sender._memory_service is None
def test_start_face_detection_happy_path(mocker):
"""
Initializes services and starts background thread.
"""
mock_face = mock.Mock()
mock_memory = mock.Mock()
mock_qi = mock.Mock()
mock_qi.service.side_effect = lambda name: {
"ALFaceDetection": mock_face,
"ALMemory": mock_memory,
}[name]
state.qi_session = mock_qi
fake_thread = mock.Mock()
mocker.patch("threading.Thread", return_value=fake_thread)
sender = FaceDetectionSender(mock.Mock())
sender.start_face_detection()
mock_face.setTrackingEnabled.assert_called_with(False)
mock_face.setRecognitionEnabled.assert_called_with(False)
mock_face.subscribe.assert_called_once()
fake_thread.start.assert_called_once()
def test_face_loop_face_detected_true(mocker):
"""
Sends face_detected=True when face data exists.
"""
sender = FaceDetectionSender(mock.Mock())
sender._memory_service = mock.Mock()
sender._memory_service.getData.return_value = [0, [[1]]]
sender.socket = mock.Mock()
mocker.patch("time.sleep")
state.exit_event.is_set.side_effect = [False, True]
sender._face_loop()
sent = sender.socket.send.call_args[0][0]
payload = json.loads(sent.decode("utf-8"))
assert payload["face_detected"] is True
def test_face_loop_face_detected_false(mocker):
"""
Sends face_detected=False when no face data exists.
"""
sender = FaceDetectionSender(mock.Mock())
sender._memory_service = mock.Mock()
sender._memory_service.getData.return_value = []
sender.socket = mock.Mock()
mocker.patch("time.sleep")
state.exit_event.is_set.side_effect = [False, True]
sender._face_loop()
sent = sender.socket.send.call_args[0][0]
payload = json.loads(sent.decode("utf-8"))
assert not payload["face_detected"]
def test_face_loop_handles_exception(mocker):
"""
Exceptions inside loop are swallowed.
"""
sender = FaceDetectionSender(mock.Mock())
sender._memory_service = mock.Mock()
sender._memory_service.getData.side_effect = Exception("boom")
sender.socket = mock.Mock()
mocker.patch("time.sleep")
state.exit_event.is_set.side_effect = [False, True]
# Must not raise
sender._face_loop()
def test_stop_face_detection_happy_path():
"""
Unsubscribes and disables tracking.
"""
sender = FaceDetectionSender(mock.Mock())
mock_face = mock.Mock()
sender._face_service = mock_face
sender.stop_face_detection()
mock_face.unsubscribe.assert_called_once()
mock_face.setTrackingEnabled.assert_called_with(False)
def test_stop_face_detection_exception():
"""
stop_face_detection swallows service exceptions.
"""
sender = FaceDetectionSender(mock.Mock())
mock_face = mock.Mock()
mock_face.unsubscribe.side_effect = Exception("fail")
sender._face_service = mock_face
sender.stop_face_detection()
def test_close_calls_stop_face_detection(mocker):
"""
close() calls parent close and stop_face_detection().
"""
sender = FaceDetectionSender(mock.Mock())
mocker.patch.object(sender, "stop_face_detection")
mocker.patch(
"robot_interface.endpoints.face_detector.SocketBase.close"
)
sender.close()
sender.stop_face_detection.assert_called_once()

View File

@@ -55,6 +55,9 @@ class DummySender:
def start(self):
self.called = True
def start_face_detection(self):
self.called = True
def close(self):
pass
@@ -108,11 +111,13 @@ def patched_main_components(monkeypatch, fake_sockets, fake_poll):
fake_act = FakeReceiver(act_sock)
video_sender = DummySender()
audio_sender = DummySender()
face_sender = DummySender()
monkeypatch.setattr(main_mod, "MainReceiver", lambda ctx: fake_main)
monkeypatch.setattr(main_mod, "ActuationReceiver", lambda ctx: fake_act)
monkeypatch.setattr(main_mod, "VideoSender", lambda ctx: video_sender)
monkeypatch.setattr(main_mod, "AudioSender", lambda ctx: audio_sender)
monkeypatch.setattr(main_mod, "FaceDetectionSender", lambda ctx: face_sender)
# Register sockets for the fake poller
fake_poll.registered = {main_sock: zmq.POLLIN, act_sock: zmq.POLLIN}