Skip to content

rkfg/dl4j-benchmark

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Dl4j-Benchmarks

Repository to track Dl4j benchmarks in relation to well known frameworks on cpu and gpu and for dl4j on spark.

Core Benchmarks

  • ModelCompare: compares same structure across all frameworks in this repo
    • MLP: using simple, single layer feed forward with MNIST data
    • Lenet: using common LeNet CNN model with MNIST data
  • Cifar10: compares Dl4j against best structures from each framework
  • Experiment: explores other comparisons and more of storage for drafts and works in progress

Spark Benchmarks

The deeplearning4j-spark-benchmark package contains a number of synthetic benchmarks to test Spark training performance under a variety of situations.

For more details, see the readme [here - TODO]

Core Packages Comparison

Main packages included for comparison so far...

Dl4j (v0.6.0)

  • Install: http://deeplearning4j.org/quickstart
  • Setup packages: add to pom.xml
  • Set GPU: change in pom file under nd4j-backends (native for cpu and cuda-7.5 for gpu)
  • Compile: mvn clean install -P (native or cuda)

Caffe (vr3)

Tensorflow(v0.9.0)

Torch (v7)

To run any of these examples, configure the system to the platform, install this repo and run:

    ./dl4j-core-benchmarks/scripts/model_compare_.sh

Note: If multiple GPUs, control how many used by adding 'export CUDA_VISIBLE_DEVICES=' to .bashrc or .bash_profile and setting empty for CPU and 0,1,2,3 for GPUs (0 if just one and 0,1 if just two)

Benchmark System

Running benchmarks on following system setup:

  • Ubuntu 14.0.4
  • 60GB RAM
  • 32 Intel Xeon E5-2670 CPUs
  • 4 Grid GPUs 4GB RAM
  • gcc & g++ v4.9
  • BLAS: OpenBLAS v1.13 or Cublas v7.5
  • cuDNN v5.1.3

In order to run all examples in core, you need to configure a system for all of the platforms. Each platform differs in requirements and be especially aware of software versions that are not supported.

Package Comparisons

Initial analysis. Consider all numbers hostile with potential to change as we get additional reviews sorted.

Timing for Training Only

MLP Example

Package CPU GPU Multi Accuracy
Dl4j 6m18s 2m40ms 1m10s ~97.5%
Caffe 2m18s 13s 33s ~97.4%
Tensorflow 1m10s 38s 1m11s ~98.3%*
Torch 4m54s 51s 1m34s ~98.0%

Lenet Example w/ cuDNN

Package CPU GPU Multi Accuracy
Dl4j 19m58s 3m03s 1m07s ~99.0%
Caffe 19m49s 52s 1m12s ~99.0%
Tensorflow 5m10s 1m37s 2m36s ~98.6%
Torch 17m59s 6m11s 3m37s ~98.3%
Timing for Full Script Run

MLP Example

Package CPU GPU Multi Accuracy
Dl4j 6m26s 2m50ms 1m20s ~97.5%
Caffe 2m20s 15s 36s ~97.4%
Tensorflow 1m15s 43s 1m18s ~98.3%*
Torch 4m56s 1m03s 1m46s ~98.0%

Lenet Example w/ cuDNN

Package CPU GPU Multi Accuracy
Dl4j 20m08s 3m13s 1m18s ~99.0%
Caffe 19m52s 53s 1m14s ~99.0%
Tensorflow 5m15s 1m44s 2m44s ~98.6%
Torch 18m03s 6m25s 3m50s ~98.3%

Note:

  • Tensorflow required learning rate modification on MLP by 1/10th otherwise accuracy drops to 9%
  • Accuracy varies slighty between cpu, single gpu & multi-gpu (e.g. DL4J multi-gpu is .90 for mlp and .97 for lenet).
  • Timings vary (potentially a couple seconds) for all packages on each run
  • Time to transfer and consolidate data can lead to longer performance times on multi-gpu (larger datasets needed for comparison)
  • Issues getting nccl setup on system for Torch multiple gpu tests; thus, not used

How to Help

Help is welcome to improve comparisons. If you know a better way or see a fix that is needed, please submit a pull request. Top of mind next steps that help would be appreciated:

- Compare LSTMs, Autoencoders, RBMs where available
- Setup Dl4j AlexNet functionality with multiple GPUs for benchmark

About

Repo to track dl4j benchmark code.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Java 54.6%
  • Python 25.8%
  • Lua 17.8%
  • Shell 1.6%
  • Batchfile 0.2%