Project Feasibility: Multimodal Medical Robot on STM32H723ZG (Audio/Vision AI + USB Host)
Hello ST Community,
I am working on my final year project: an autonomous "Multimodal Medical Assistant Robot" based on the NUCLEO-H723ZG (Cortex-M7, 550MHz). I would like to have your expert opinion on the feasibility and technical architecture.
Key Features:
100% Offline AI (Edge AI):
Audio: Keyword Spotting (7 commands) using INMP441 (I2S) and CMSIS-DSP (MFCC).
Vision: Fall detection using OV2640 (DCMI) and a light MobileNet model.
Software Stack: X-CUBE-AI, TensorFlow Lite Micro, FatFS.
Innovation (Dynamic Loading): We plan to load the AI models (.tflite) and voice responses (.wav) dynamically from a USB Flash Drive (USB Host MSC) into RAM at boot time.
Specific Questions:
Is the NUCLEO-H723ZG powerful enough to run both Audio and Vision inference concurrently while managing motor control (PWM) and USB Host?
Regarding the USB Host / RAM loading: Does X-CUBE-AI support "relocatable weights" loaded into RAM from an external storage device via FatFS? Any specific memory alignment tips for the H7 AXI SRAM?
Any advice on DMA priorities between DCMI (Vision) and I2S (Audio) to avoid data loss?
Thank you for your help!
