Evaluation and Postprocessing
Use this guide when you need to interpret benchmark outputs from boxmot eval, Boxmot.val(...), tune, or research.
Core metrics
HOTAfor overall tracking qualityMOTAfor CLEAR-style summary qualityIDF1for identity consistencyAssAandAssRefor association qualityIDSWandIDsfor identity-switch context
Where metrics appear
evalreports benchmark results directlytuneuses validation results to score parameter trialsresearchoptimizes code changes against combined benchmark summaries
For raw runtime summaries from the Python API, evaluate(...) aggregates counts and timings but does not replace TrackEval ground-truth evaluation.
Postprocessing modes
eval supports three postprocessing modes through --postprocessing:
nonegsifor Gaussian-smoothed interpolationgbrcfor gradient-boosting-based reconnection and interpolation
Common commands
boxmot eval --benchmark mot17-ablation --tracker boosttrack
boxmot eval --benchmark mot17-ablation --tracker boosttrack --postprocessing gsi
boxmot eval --benchmark mot17-ablation --tracker boosttrack --postprocessing gbrc
Main outputs
- combined benchmark metrics such as
HOTA,MOTA, andIDF1 - per-sequence summaries
- MOT-style tracker outputs
- reused cache paths and evaluation artifacts in the run directory