Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Fix tests for multiprocessing  #47

@starride-teklia

Description

@starride-teklia

The bump to Pytorch 1.13 broke some tests related to multiprocessing on CPU and GPU. We get the following errors:

  • torch.multiprocessing.spawn.ProcessRaisedException
  • AttributeError: 'LightningDistributedDataParallel' object has no attribute '_sync_params'

On these tests:

tests/callbacks/learning_rate_test.py:# TODO: fix test with num_processes=2
tests/callbacks/training_timer_test.py:# TODO: fix test with num_processes=2
tests/loggers/epoch_csv_logger_test.py:# TODO: fix test with num_processes=2
tests/scripts/htr/decode_ctc_test.py:# TODO: fix test with nprocs=2
tests/scripts/htr/netout_test.py:# TODO: fix test with nprocs=2
tests/scripts/htr/train_ctc_test.py:# TODO: fix "ddp_cpu" mode
tests/scripts/htr/train_ctc_test.py:# TODO: fix "ddp" mode
tests/scripts/htr/train_ctc_test.py:# TODO: fix first assertion

I skipped the tests for now, but I need to investigate why we are getting this error and how to fix it.

Metadata

Metadata

Labels

No labels
No labels

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions