from transformers import PretrainedConfig class LUARConfig(PretrainedConfig): model_type = "LUAR" def __init__(self, embedding_size: int = 512, **kwargs, ): self.embedding_size = embedding_size super().__init__(**kwargs)