Testing started at ์คํ 5:22 ...
C:\Users\cote\Anaconda3\envs\sc\python.exe "C:\Program Files\JetBrains\PyCharm Community Edition with Anaconda plugin 2019.3.3\plugins\python-ce\helpers\pycharm\_jb_unittest_runner.py" --path C:/Users/cote/PycharmProjects/kospeech2/tests/test_conformer_lstm.py
Launching unittests with arguments python -m unittest C:/Users/cote/PycharmProjects/kospeech2/tests/test_conformer_lstm.py in C:\Users\cote\PycharmProjects\kospeech2\tests
Error
Traceback (most recent call last):
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 59, in testPartExecutor
yield
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 628, in run
testMethod()
File "C:\Users\cote\PycharmProjects\kospeech2\tests\test_conformer_lstm.py", line 56, in test_beam_search
prediction = model(DUMMY_INPUTS, DUMMY_INPUT_LENGTHS)["predictions"]
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\conformer_lstm\model.py", line 109, in forward
return super(ConformerLSTMModel, self).forward(inputs, input_lengths)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\openspeech_encoder_decoder_model.py", line 125, in forward
predictions = self.decoder(encoder_outputs, encoder_output_lengths)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\search\beam_search_lstm.py", line 78, in forward
step_outputs, hidden_states, attn = self.forward_step(inputs, hidden_states, encoder_outputs)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\decoders\lstm_attention_decoder.py", line 140, in forward_step
embedded = self.embedding(input_var)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\sparse.py", line 158, in forward
self.norm_type, self.scale_grad_by_freq, self.sparse)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\functional.py", line 1916, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
RuntimeError: Input, output and indices must be on the current device
Error
Traceback (most recent call last):
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 59, in testPartExecutor
yield
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 628, in run
testMethod()
File "C:\Users\cote\PycharmProjects\kospeech2\tests\test_conformer_lstm.py", line 37, in test_forward
outputs = model(DUMMY_INPUTS, DUMMY_INPUT_LENGTHS)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\conformer_lstm\model.py", line 109, in forward
return super(ConformerLSTMModel, self).forward(inputs, input_lengths)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\openspeech_encoder_decoder_model.py", line 130, in forward
teacher_forcing_ratio=0.0,
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\decoders\lstm_attention_decoder.py", line 220, in forward
attn=attn,
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\decoders\lstm_attention_decoder.py", line 140, in forward_step
embedded = self.embedding(input_var)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\sparse.py", line 158, in forward
self.norm_type, self.scale_grad_by_freq, self.sparse)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\functional.py", line 1916, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
RuntimeError: Input, output and indices must be on the current device
Error
Traceback (most recent call last):
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 59, in testPartExecutor
yield
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 628, in run
testMethod()
File "C:\Users\cote\PycharmProjects\kospeech2\tests\test_conformer_lstm.py", line 103, in test_test_step
batch=(DUMMY_INPUTS, DUMMY_TARGETS, DUMMY_INPUT_LENGTHS, DUMMY_TARGET_LENGTHS), batch_idx=i
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\conformer_lstm\model.py", line 148, in test_step
return super(ConformerLSTMModel, self).test_step(batch, batch_idx)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\openspeech_encoder_decoder_model.py", line 225, in test_step
teacher_forcing_ratio=0.0,
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\decoders\lstm_attention_decoder.py", line 220, in forward
attn=attn,
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\decoders\lstm_attention_decoder.py", line 140, in forward_step
embedded = self.embedding(input_var)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\sparse.py", line 158, in forward
self.norm_type, self.scale_grad_by_freq, self.sparse)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\functional.py", line 1916, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
RuntimeError: Input, output and indices must be on the current device
Error
Traceback (most recent call last):
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 59, in testPartExecutor
yield
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 628, in run
testMethod()
File "C:\Users\cote\PycharmProjects\kospeech2\tests\test_conformer_lstm.py", line 71, in test_training_step
batch=(DUMMY_INPUTS, DUMMY_TARGETS, DUMMY_INPUT_LENGTHS, DUMMY_TARGET_LENGTHS), batch_idx=i
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\conformer_lstm\model.py", line 122, in training_step
return super(ConformerLSTMModel, self).training_step(batch, batch_idx)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\openspeech_encoder_decoder_model.py", line 177, in training_step
target_lengths=target_lengths,
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\openspeech_encoder_decoder_model.py", line 105, in collect_outputs
"learning_rate": self.get_lr(),
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\openspeech_model.py", line 215, in get_lr
for g in self.optimizer.param_groups:
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 948, in __getattr__
type(self).__name__, name))
AttributeError: 'ConformerLSTMModel' object has no attribute 'optimizer'
Error
Traceback (most recent call last):
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 59, in testPartExecutor
yield
File "C:\Users\cote\Anaconda3\Lib\unittest\case.py", line 628, in run
testMethod()
File "C:\Users\cote\PycharmProjects\kospeech2\tests\test_conformer_lstm.py", line 87, in test_validation_step
batch=(DUMMY_INPUTS, DUMMY_TARGETS, DUMMY_INPUT_LENGTHS, DUMMY_TARGET_LENGTHS), batch_idx=i
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\conformer_lstm\model.py", line 135, in validation_step
return super(ConformerLSTMModel, self).validation_step(batch, batch_idx)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\models\openspeech_encoder_decoder_model.py", line 197, in validation_step
teacher_forcing_ratio=0.0,
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\decoders\lstm_attention_decoder.py", line 220, in forward
attn=attn,
File "C:\Users\cote\PycharmProjects\kospeech2\openspeech\decoders\lstm_attention_decoder.py", line 140, in forward_step
embedded = self.embedding(input_var)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\modules\sparse.py", line 158, in forward
self.norm_type, self.scale_grad_by_freq, self.sparse)
File "C:\Users\cote\Anaconda3\envs\sc\lib\site-packages\torch\nn\functional.py", line 1916, in embedding
return torch.embedding(weight, input, padding_idx, scale_grad_by_freq, sparse)
RuntimeError: Input, output and indices must be on the current device
Assertion failed
Ran 5 tests in 24.698s
FAILED (errors=5)
Process finished with exit code 1
Assertion failed
Assertion failed