Inference Module Real-Time Evaluation

1. Description

This report presents the real-time performance analysis of the Inference Engine module in the FluxSand project. This module utilizes a pre-trained ONNX model to classify motion or status based on time-series sensor input.

The analysis is based on 50 consecutive inference runs as recorded in the unit test log. The model takes as input a 3D tensor of shape [1, 1500, 8] and produces output of shape [-1, 10].

2. Test Environment

  • Model: ONNX format, loaded from /usr/local/share/FluxSand/model/model.onnx
  • Input Shape: [1, 1500, 8]
  • Output Shape: [*, 10]
  • Framework: ONNX Runtime
  • Platform: ARM-based Linux system with RT kernel
  • Thread Affinity / Priority: Not explicitly set in test, assumed normal user-space context

3. Inference Time Results (from 50 runs)

Run # Inference Time (ms) Result
1 14.979 Unrecognized
2 14.202 Unrecognized
3 14.242 Still
25 10.213 Still
33 15.805 Still
50 12.817 Still

Summary

  • Total Runs: 50
  • Minimum Time: 10.213 ms
  • Maximum Time: 18.018 ms
  • Average Time: 13.701 ms

4. Analysis

  • The inference latency across 50 runs ranged from ~10.2 ms to 18.0 ms, which indicates relatively stable performance with slight variations likely due to thread scheduling or memory fluctuations.
  • The average latency is 13.701 ms, allowing the system to handle inference at approximately 70–100 FPS if inference is the only load.
  • The early runs exhibited slightly higher latency, likely due to initial memory paging or cache warm-up. The majority of inferences settled within 12–15 ms.
  • The result output was dominantly “Still”, indicating correct environment classification consistency once the model was stabilized.

5. Conclusion

The inference module demonstrates stable and acceptable real-time performance. With average response times under 14 ms, the system is capable of responsive and interactive behavior, suitable for continuous or event-driven AI inference tasks in embedded environments.