When I train my PyTorch Lightning model on two GPUs on jupyter lab with strategy=“ddp_notebook”, only two CPUs are used and their usages are 100%. How can I overcome this CPU bottleneck?

Edit: I tested with PyTorchProfiler and it was because of old ssds used on the server

  • Spott@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    11 months ago

    Without knowing more, I would expect it is a dataloader issue: your CPUs are bottlenecked trying to get enough data to your GPUs.

    You can add more workers to your dataloader in order to paralyze it, though this can lead to weird parallelization bugs sometimes, so if things start acting weird, that might be a reason.

    • troye888@lemmy.one
      link
      fedilink
      arrow-up
      3
      ·
      11 months ago

      Yup this, if you would like more help we need the code, or at least a minimal viable reproduction scenario.