This is a Spanish GPT-2 model trained from scratch on the large_spanish_corpus aka BETO's corpus with Flax This is part of the Flax/Jax Community Week , organised by HuggingFace and TPU usage sponsored by Google.
The dataset is about 20 GB. 95% of the data was used for training and the rest 5% for validation.