You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The goal is to decompose ABTF repository into reusable automation recipes to make the benchmarking, evaluation and training process of ABTF models more deterministic, portable and extensible with new models, frameworks and data sets ....
We need to develop the following CM scripts (automation recipes) to support ABTF benchmarking with loadgen across different platforms, OS and hardware:
Prepare examples of docker containers with CM: see examples
Demos
Prepare demo for live ABTF model evaluation
Download Cognata subset
Show live visualization of predictions
Document
For the next tasks we need more engineering resources.
MLCommons committed to fund CM development with 1 CM engineer until the end of 2024 to modularize and automate MLPerf inference. ABTF colleagues should sync developments with the MLPerf inference WG.
Improve performance
Add performance profiling, analysis and debugging
Current performance on 8-core CPU and Laptop GPU is low (10 sec per frame for 8M model and 3 sec per frame for 3M model on CPU) - need further optimization (quantization, hardware specific optimizations, fine-tuning, etc)
Add C++ harness for loadgen with ABTF model
Develop C++ harness for loadgen with ONNX
Export PyTorch model to TFLite
Develop native C++ harness for loadgen test with TFLite model
Develop C++ harness for loadgen with PyTorch
Support other hardware
PyTorch native
Support ABTF demo on Nvidia GPU via CUDA
Generate Docker container for the demo
Cross-compilation
Samsung Exynos
Requires C++ loadgen harness implementation with cross-compilation
ONNX backend
TFLite backend
Automate ABTF model quantization
TBD
Developers
ABTF model
Radoyeh Shojaei
CM automation for ABTF model
@gfursin has completed a prototype of a CM automation and MLPerf harness for ABTF model in May 2024. Further developments should be done by MLCommons CM inference engineer.
The text was updated successfully, but these errors were encountered:
Current tasks
The goal is to decompose ABTF repository into reusable automation recipes to make the benchmarking, evaluation and training process of ABTF models more deterministic, portable and extensible with new models, frameworks and data sets ....
We need to develop the following CM scripts (automation recipes) to support ABTF benchmarking with loadgen across different platforms, OS and hardware:
See the current CM-ABTF documentation/demos here.
Preparing ABTF demo
@gfursin helped to prepare first CM automation for ABTF and we now plan to delegate further developments to dedicated engineers.
Test inference with ABTF model
Export ABTF model to other formats
Evaluate ABTF model with Cognata sub-set
Automate training of ABTF model with Cognata sub-set
Add Python harness for loadgen with ABTF model
See related CM script and simple Python harness.
Generate/use Docker containers
Demos
For the next tasks we need more engineering resources.
MLCommons committed to fund CM development with 1 CM engineer until the end of 2024 to modularize and automate MLPerf inference. ABTF colleagues should sync developments with the MLPerf inference WG.
Improve performance
Add C++ harness for loadgen with ABTF model
Support other hardware
PyTorch native
Cross-compilation
Automate ABTF model quantization
TBD
Developers
ABTF model
CM automation for ABTF model
The text was updated successfully, but these errors were encountered: