模型:
pszemraj/long-t5-tglobal-xl-16384-book-summary-8bit
This is an 8-bit quantized version of the pszemraj/long-t5-tglobal-xl-16384-book-summary model, The model has been compressed using bitsandbytes and can be loaded with low memory usage.
Refer to the original model for all details about the model architecture and training process. For more information on loading 8-bit models, refer to the 4.28.0 release information and the example repository .
To use the model, install or upgrade transformers , accelerate , and bitsandbytes . Make sure to have transformers>=4.28.0 and bitsandbytes>0.37.2 .
pip install -U -q transformers bitsandbytes accelerate
Load the model with AutoTokenizer and AutoModelForSeq2SeqLM :
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM model_name = "pszemraj/long-t5-tglobal-xl-16384-book-summary-8bit" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSeq2SeqLM.from_pretrained(model_name)