(Note that the above is a screenshot of the benchmark. Please visit the project page for the latest version and an interactive experience.)
Helps to estimate the runtime of algorithms on a different GPU
Measures GPU processing speed independent of GPU memory capacity
Contains adjustable weightings through interactive UIs
This repo contains the timing scripts used in the GPU benchmark. This latency-based benchmark is designed to compare algorithms with runtime reported under different GPUs, and it also serves as a GPU purchasing guide. Please check out the project page for the complete benchmark with detailed descriptions. This page documents instructions on how to run the code and the changelog of the benchmark.
git clone https://github.com/mtli/DLGPUBench.git
cd DLGPUBench
conda env create -f environment.yml
conda activate bench
Download and unpack ImageNet (ILSVRC2012) and MS COCO. For running the detection scripts, you also need to download the pretrained model from this link. Modify the dataset paths in each script you plan to run.
- Update the timing setting for classification by excluding the time spent on GPU-host data transfer, and disabling multi-threading to make timing more stable and faster.
- Update to work with llcv 0.0.9.
- Change the default batch size for classification inference to 64
- Add results for GTX 1080 and RTX A6000.