Caffe con Troll

Abstract
We present Caffe con Troll (CcT), a fully compatible end-to-end version of the popular framework Caffe with rebuilt internals. We built CcT to examine the performance characteristics of training and deploying general-purpose convolutional neural networks across different hardware architectures. We find that, by employing standard batching optimizations for CPU training, we achieve a 6:3× throughput improvement over Caffe on popular networks like CaffeNet. Moreover, with these improvements, the end-to-end training time for CNNs is directly proportional to the FLOPS delivered by the CPU, which enables us to efficiently train hybrid CPU-GPU systems for CNNs.
Funding Information
  • Defense Advanced Research Projects Agency (FA8750-12-2-0335, FA8750-13-2-0039)
  • National Science Foundation (IIS-1353606)
  • National Institutes of Health (U54EB020405)
  • Office of Naval Research (N000141210041, N000141310129)

This publication has 4 references indexed in Scilit: