模型:
TheBloke/galpaca-30B-GPTQ-4bit-128g
任务:
文本生成许可:
cc-by-nc-4.0Chat & support: my new Discord server
Want to contribute? TheBloke's Patreon page
This is an attempt to create a GPTQ 4-bit version of Galpaca 30B .
I created these files on request. I have no previous experience of Galactica or Galpaca, and have not done much testing to confirm that the output is useful and usable.
You will need 18+ GB VRAM to load these models on a GPU.
galpaca-30B-4bit-128g.no-act-order.pt
galpaca-30B-4bit-128g.pt
The GPTQ code used to create these models can be found at GPTQ-for-LLaMa .
Note that during the GPTQ process, the following warning was seen:
Token indices sequence length is longer than the specified maximum sequence length for this model (1915 > 512). Running this sequence through the model will result in indexing errors
I do not know if this is indicates a potential problem in the GPTQ output, or if it can be ignored. If you know more about this, do let me know.
General model info below is as per the original Galpaca 30B repository .
For more information, example prompts and more, please see the original repository.
GALACTICA 30B fine-tuned on the Alpaca dataset.
The model card from the original Galactica repo can be found here , and the original paper here .
The dataset card for Alpaca can be found here , and the project homepage here . The Alpaca dataset was collected with a modified version of the Self-Instruct Framework , and was built using OpenAI's text-davinci-003 model. As such it is subject to OpenAI's terms of service.
For further support, and discussions on these models and AI in general, join us at:
Thanks to the chirper.ai team!
I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
Patreon special mentions : Aemon Algiz, Dmitriy Samsonov, Nathan LeClaire, Trenton Dambrowitz, Mano Prime, David Flickinger, vamX, Nikolai Manek, senxiiz, Khalefa Al-Ahmad, Illia Dulskyi, Jonathan Leane, Talal Aujan, V. Lukas, Joseph William Delisle, Pyrater, Oscar Rangel, Lone Striker, Luke Pendergrass, Eugene Pentland, Sebastain Graf, Johann-Peter Hartman.
Thank you to all my generous patrons and donaters!
The GALACTICA models are trained on a large-scale scientific corpus and are designed to perform scientific tasks. The Alpaca dataset is a set of 52k instruct-response pairs designed to enhace the instruction following capabilites of pre-trained language models.
The GALACTICA model card specifies that the primary indended users of the GALACTICA models are researchers studying language models applied to the scientific domain, and it cautions against production use of GALACTICA without safeguards due to the potential for the model to produce inaccurate information. The original GALACTICA models are available under a non-commercial CC BY-NC 4.0 license, and the GALPACA model is additionally subject to the OpenAI Terms of Service .