Might be simpler to inherit from RobertaModel rather than PreTrainedModel f0ad7f1 PeteBleackley commited on Oct 9, 2023
Removed a base model that was causing a loop in model initialisation 87535ff PeteBleackley commited on Oct 9, 2023
Further changes for compatibility with HuggingFace Pytorch implementation 5b7a8ed PeteBleackley commited on Oct 9, 2023
PyTorch implementation of HugggingFace PreTrainedModel class does not allow direct setting of base_model. Rejig constructors accordingly 519dfd1 PeteBleackley commited on Oct 9, 2023
Low level RoBERTa layers don't necessarily return what I expect them to 0941a89 PeteBleackley commited on Sep 29, 2023
Making sure RoBERTa layers have all required arguments b2593fa PeteBleackley commited on Sep 25, 2023
Encoder, Decoder and Trainer models (assuming RoBERTa base models) f9c0522 PeteBleackley commited on Sep 5, 2023