Skip to main content

Kimsehun725's group workspace

Timestamps visible
2022-10-20 11:03:35
GPU available: True (cuda), used: True
2022-10-20 11:03:35
TPU available: False, using: 0 TPU cores
2022-10-20 11:03:35
IPU available: False, using: 0 IPUs
2022-10-20 11:03:35
HPU available: False, using: 0 HPUs
2022-10-20 11:03:37
wandb: logging graph, to disable use `wandb.watch(log_graph=False)`
2022-10-20 11:03:41
Initializing distributed: GLOBAL_RANK: 0, MEMBER: 1/4
2022-10-20 11:03:51
----------------------------------------------------------------------------------------------------
2022-10-20 11:03:51
distributed_backend=nccl
2022-10-20 11:03:51
All distributed processes registered. Starting with 4 processes
2022-10-20 11:03:51
----------------------------------------------------------------------------------------------------
2022-10-20 11:03:55
LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3]
2022-10-20 11:03:55
  | Name                   | Type             | Params
2022-10-20 11:03:55
------------------------------------------------------------
2022-10-20 11:03:55
0 | VAE_encoder            | Conv1DEncoder    | 1.8 M
2022-10-20 11:03:55
1 | fc1                    | Linear           | 23.1 K
2022-10-20 11:03:55
2 | fc2                    | Linear           | 23.1 K
2022-10-20 11:03:55
3 | conv_stack             | ConvStack        | 1.6 M
2022-10-20 11:03:55
4 | z_linear               | Sequential       | 282 K
2022-10-20 11:03:55
5 | self_attention_block   | ConformerEncoder | 92.0 K
2022-10-20 11:03:55
6 | frame_tab_output_layer | Sequential       | 24.6 K
2022-10-20 11:03:55
7 | softmax_by_string      | Softmax          | 0
2022-10-20 11:03:55
------------------------------------------------------------
2022-10-20 11:03:55
3.9 M     Trainable params
2022-10-20 11:03:55
0         Non-trainable params
2022-10-20 11:03:55
3.9 M     Total params
2022-10-20 11:03:55
15.573    Total estimated model params size (MB)
2022-10-20 11:03:58
Sanity Checking: 0it [00:00, ?it/s]batch=[None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None, None]
2022-10-20 11:04:03
Traceback (most recent call last):
2022-10-20 11:04:03
  File "src/train.py", line 76, in <module>
2022-10-20 11:04:03
    main(kwargs)
2022-10-20 11:04:03
  File "src/train.py", line 26, in main
2022-10-20 11:04:03
    step3(kwargs)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/src/step3.py", line 235, in step3
2022-10-20 11:04:03
    train(kwargs, use_pretrained_model, pretrained_time, pretrained_epoch, now, test_num, train_data_list,
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/src/step3.py", line 173, in train
2022-10-20 11:04:03
    trainer.fit(model, train_dataloaders=train_loader,
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 696, in fit
2022-10-20 11:04:03
    self._call_and_handle_interrupt(
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 648, in _call_and_handle_interrupt
2022-10-20 11:04:03
    return self.strategy.launcher.launch(trainer_fn, *args, trainer=self, **kwargs)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/strategies/launchers/subprocess_script.py", line 93, in launch
2022-10-20 11:04:03
    return function(*args, **kwargs)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 735, in _fit_impl
2022-10-20 11:04:03
    results = self._run(model, ckpt_path=self.ckpt_path)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 1166, in _run
2022-10-20 11:04:03
    results = self._run_stage()
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 1252, in _run_stage
2022-10-20 11:04:03
    return self._run_train()
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 1274, in _run_train
2022-10-20 11:04:03
    self._run_sanity_check()
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 1343, in _run_sanity_check
2022-10-20 11:04:03
    val_loop.run()
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/loops/loop.py", line 200, in run
2022-10-20 11:04:03
    self.advance(*args, **kwargs)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/loops/dataloader/evaluation_loop.py", line 155, in advance
2022-10-20 11:04:03
    dl_outputs = self.epoch_loop.run(self._data_fetcher, dl_max_batches, kwargs)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/loops/loop.py", line 200, in run
2022-10-20 11:04:03
    self.advance(*args, **kwargs)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/loops/epoch/evaluation_epoch_loop.py", line 127, in advance
2022-10-20 11:04:03
    batch = next(data_fetcher)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/utilities/fetching.py", line 184, in __next__
2022-10-20 11:04:03
    return self.fetching_function()
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/utilities/fetching.py", line 263, in fetching_function
2022-10-20 11:04:03
    self._fetch_next_batch(self.dataloader_iter)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/pytorch_lightning/utilities/fetching.py", line 277, in _fetch_next_batch
2022-10-20 11:04:03
    batch = next(iterator)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 681, in __next__
2022-10-20 11:04:03
    data = self._next_data()
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 1376, in _next_data
2022-10-20 11:04:03
    return self._process_data(data)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 1402, in _process_data
2022-10-20 11:04:03
    data.reraise()
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/torch/_utils.py", line 461, in reraise
2022-10-20 11:04:03
    raise exception
2022-10-20 11:04:03
TypeError: Caught TypeError in DataLoader worker process 0.
2022-10-20 11:04:03
Original Traceback (most recent call last):
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/torch/utils/data/_utils/worker.py", line 302, in _worker_loop
2022-10-20 11:04:03
    data = fetcher.fetch(index)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/venv/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 52, in fetch
2022-10-20 11:04:03
    return self.collate_fn(data)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/src/step3.py", line 46, in __call__
2022-10-20 11:04:03
    return self.pad_collate(batch)
2022-10-20 11:04:03
  File "/data/group1/z44543r/vae_separation/src/step3.py", line 27, in pad_collate
2022-10-20 11:04:03
    for i, (x1, x2, gt) in enumerate(batch):
2022-10-20 11:04:03
TypeError: cannot unpack non-iterable NoneType object