---------------------------------------------------------------------------
RuntimeError Traceback (most recent call last)
<ipython-input-25-b54aed7e715d> in <module>
----> 1 trainer.train(start_epoch=0, end_epoch=1)
~/amrefenv/lib/python3.6/site-packages/building_footprint_segmentation/trainer.py in train(self, start_epoch, end_epoch, step, bst_vld_loss)
179 to_new_line_data=True,
180 )
--> 181 raise ex
182
183 one_liner.one_line(
~/amrefenv/lib/python3.6/site-packages/building_footprint_segmentation/trainer.py in train(self, start_epoch, end_epoch, step, bst_vld_loss)
93
94 train_loss, train_metric, step, progress_bar = self.state_train(
---> 95 step, progress_bar
96 )
97 progress_bar.close()
~/amrefenv/lib/python3.6/site-packages/building_footprint_segmentation/trainer.py in state_train(self, step, progress_bar)
203 train_data = gpu_variable(train_data)
204
--> 205 prediction = self.model(train_data["images"])
206 calculated_loss = self.criterion(train_data["ground_truth"], prediction)
207 self.optimizer.zero_grad()
~/amrefenv/lib/python3.6/site-packages/torch/nn/modules/module.py in _call_impl(self, *input, **kwargs)
1100 if not (self._backward_hooks or self._forward_hooks or self._forward_pre_hooks or _global_backward_hooks
1101 or _global_forward_hooks or _global_forward_pre_hooks):
-> 1102 return forward_call(*input, **kwargs)
1103 # Do not call functions when jit is used
1104 full_backward_hooks, non_full_backward_hooks = [], []
~/amrefenv/lib/python3.6/site-packages/torch/nn/parallel/data_parallel.py in forward(self, *inputs, **kwargs)
164
165 if len(self.device_ids) == 1:
--> 166 return self.module(*inputs[0], **kwargs[0])
167 replicas = self.replicate(self.module, self.device_ids[:len(inputs)])
168 outputs = self.parallel_apply(replicas, inputs, kwargs)
~/amrefenv/lib/python3.6/site-packages/torch/nn/modules/module.py in _call_impl(self, *input, **kwargs)
1100 if not (self._backward_hooks or self._forward_hooks or self._forward_pre_hooks or _global_backward_hooks
1101 or _global_forward_hooks or _global_forward_pre_hooks):
-> 1102 return forward_call(*input, **kwargs)
1103 # Do not call functions when jit is used
1104 full_backward_hooks, non_full_backward_hooks = [], []
~/amrefenv/lib/python3.6/site-packages/building_footprint_segmentation/seg/binary/models/mfrn.py in forward(self, input_feature)
287
288 transition_up_1 = self.mfrn.decoder.decodertransition1(
--> 289 bottle_neck, skip_connections.pop()
290 )
291 dense_layer_6 = self.mfrn.decoder.decoderdenseblock1(transition_up_1)
~/amrefenv/lib/python3.6/site-packages/torch/nn/modules/module.py in _call_impl(self, *input, **kwargs)
1100 if not (self._backward_hooks or self._forward_hooks or self._forward_pre_hooks or _global_backward_hooks
1101 or _global_forward_hooks or _global_forward_pre_hooks):
-> 1102 return forward_call(*input, **kwargs)
1103 # Do not call functions when jit is used
1104 full_backward_hooks, non_full_backward_hooks = [], []
~/amrefenv/lib/python3.6/site-packages/building_footprint_segmentation/seg/binary/models/mfrn.py in forward(self, x, skip)
114 def forward(self, x, skip):
115 out = self.Transpose(x)
--> 116 out = torch.cat([out, skip], 1)
117 return out
118
RuntimeError: Sizes of tensors must match except in dimension 1. Expected size 92 but got size 93 for tensor number 1 in the list.