I have this edge case where i train a RNN for several batches, but sometimes it happens for one batch that there is no ground truth available (frames are not annotated). However I still want to run the network in order to update the hidden states.
Is it possible to override both backward and optimizer step if loss is None or undefined?