Skip to main content
Associate III
September 19, 2025
Solved

Does STM32MP257 support MMAction2 with TSN or SlowFast behavior recognition? 3D CONV?

  • September 19, 2025
  • 4 replies
  • 841 views

Can I use MMAction2 with TSN or SlowFast behavior recognition models and deploy them to the STM32MP257? Does the STM32MP257 support 3D CONV operators?

Best answer by Julian E.

Hello @fanronghua0123456,

 

The dev team probably have such tools, but I cannot share them.

 

Have a good day,

Julian

4 replies

Julian E.
Technical Moderator
September 22, 2025

Hello @fanronghua0123456,

 

Here is the list of supported layers for the MP2:

https://wiki.st.com/stm32mpu/wiki/STM32MP2_NPU_description#Operation_support

 

There is indeed support for the CONV3D with the operation VSI_NN_OP_CONV3D with NPU in 8bits and on the GPU in 16bits.

 

For the model MMAction, it would be required to quantize the model in fflite or onnx for it to work on the HW. We don't have the description of the model, but it seems similar to this, if you want to take a look:

https://wiki.st.com/stm32mpu/wiki/How_to_deploy_your_NN_model_on_STM32MPU

 

Have a good day,

Julian

​In order to give better visibility on the answered topics, please click on 'Accept as Solution' on the reply which solved your issue or answered your question.
Associate III
September 23, 2025

Hello @Julian E.

 Thanks ,

How do I evaluate the operational efficiency of the model on the development stm32mp25x board? Currently, the model trained with yolov11 has a very large inference speed after quantization. I would like to know how long each network layer of my model takes to infer, so that we can optimize the model structure better.

Julian E.
Technical Moderator
September 23, 2025

Hello @fanronghua0123456,

 

I think the ST Developer Cloud could help you: https://stedgeai-dc.st.com/session

You can import model, select the MP2 and benchmark its inference time on a real board in our board farm

JulianE_0-1758621225634.png

 

You can also do it manually with the validate command: https://stedgeai-dc.st.com/assets/embedded-docs/stm32mpu_command_line_interface.html 

 

Have a good day,

Julian

​In order to give better visibility on the answered topics, please click on 'Accept as Solution' on the reply which solved your issue or answered your question.
Associate III
September 23, 2025

hello,@Julian E.

thanks, so I want to know  details of this duration times . and  how long each network layer of my model takes to infer?

fanronghua0123456_0-1758623607599.png

 thanks.

Julian E.
Technical Moderator
September 23, 2025

Hello @fanronghua0123456,

 

I asked the dev team, but unfortunately, we don't have such tool for the MP2.

 

Have a good day,

Julian

​In order to give better visibility on the answered topics, please click on 'Accept as Solution' on the reply which solved your issue or answered your question.
Associate III
September 24, 2025

Because the performance of each GPU and NPU is different, when we want to deploy the model on embedded devices, the inference speed of each device is different. Therefore, I believe you must have a tool to analyze the inference speed of each layer of the model's network.

Julian E.
Julian E.Best answer
Technical Moderator
September 24, 2025

Hello @fanronghua0123456,

 

The dev team probably have such tools, but I cannot share them.

 

Have a good day,

Julian

​In order to give better visibility on the answered topics, please click on 'Accept as Solution' on the reply which solved your issue or answered your question.