Models training runtime and memory footprint

The Pasa and the Densenet models were trained on a machine equipped with an 11 GB GeForce GTX 1080 Ti GPU, an 8-core processor, 48 GB of RAM and Debian 10. The Logistic Regression model was trained on a Macbook Pro with an 8-core processor, 32 GB of RAM and macOS Big Sur.

Pasa

  • Training on MC: 2’000 epochs in 2.5 hours, ~2 GB of CPU memory, ~0.75 GB of GPU memory

  • Training on MC-CH: 2’000 epochs in 17 hours, ~2 GB of CPU memory, ~0.75 GB of GPU memory

  • Training on MC-CH-IN: 2’000 epochs in 16.5 hours, ~2 GB of CPU memory, ~0.75 GB of GPU memory

Densenet pretraining

  • Training on NIH CXR14: 10 epochs in 12 hours, ~7.2 GB of CPU memory, ~6.4 GB of GPU memory

Densenet fine-tuning

  • Training on MC: 300 epochs in 0.5 hours, ~2 GB of CPU memory, ~6.4 GB of GPU memory

  • Training on MC-CH: 300 epochs in 2.5 hours, ~2 GB of CPU memory, ~6.4 GB of GPU memory

  • Training on MC-CH-IN: 300 epochs in 3.5 hours, ~2 GB of CPU memory, ~6.4 GB of GPU memory

Logistic Regression

  • Training on MC: 100 epochs in a few seconds, ~17 GB of CPU memory

  • Training on MC-CH: 100 epochs in a few seconds, ~17 GB of CPU memory

  • Training on MC-CH-IN: 100 epochs in a few seconds, ~17 GB of CPU memory