模型:
cosimoiaia/Loquace-410m
Model Card for Loquace-410m
An exclusively Italian speaking, instruction finetuned, Large Language model. ??
The Loquace Italian LLM models are created as a proof-of-concept to evaluate on how language tuning can be achieved using QLoRa by instruct-tunings foundational LLMs using dataset of a specific language.
The QLoRa ( https://github.com/artidoro/qlora ) method of fine-tuning significantly lower the resources requirements compared to any other methods available, this allow to easily execute the process on significanly larger dataset while still using consumers GPUs and still achieve high accuracy.
Loquace-410m is the second smallest model of the Loquace family. It was trained using QLoRa on a large dataset of 102k question/answer pairs exclusively in Italian using pythia-410m as base.
The related code can be found at: https://github.com/cosimoiaia/Loquace
Loquace-410m is part of the big Loquace family:
https://huggingface.co/cosimoiaia/Loquace-70m - Based on pythia-70m https://huggingface.co/cosimoiaia/Loquace-410m - Based on pythia-410m https://huggingface.co/cosimoiaia/Loquace-7B - Based on Falcon-7B. https://huggingface.co/cosimoiaia/Loquace-12B - Based on pythia-12B https://huggingface.co/cosimoiaia/Loquace-20B - Based on gpt-neox-20B
from transformers import ( AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig ) tokenizer = AutoTokenizer.from_pretrained("cosimoiaia/Loquace-410m", padding_side="right", use_fast=True) model = AutoModelForCausalLM.from_pretrained( "cosimoiaia/Loquace-410m", load_in_8bit=True, device_map="auto", quantization_config=BitsAndBytesConfig( load_in_4bit=True, llm_int8_has_fp16_weight=False ) )
Loquace-410m was trained on a conversational dataset comprising 102k question/answer pairs in Italian language. The training data was constructed by putting together translations from the original alpaca Dataset and other sources like the OpenAssistant dataset. The model was trained for only 10000 iterations and took 9 hours on a single RTX 3090, kindly provided by Genesis Cloud. ( https://gnsiscld.co/26qhlf )