The model checkpoints seem to be hard-coded as the BertForMaskedLM and are unable to load but to the CoCondensor class.
Adding the following attributes in the initialization can surpass the exceptions but all the weights were not loaded.
self._keys_to_ignore_on_save = None
self._keys_to_ignore_on_load_missing = None
Is there a way to resume training after interruptions?
Thanks!