![]() These models will be trained on up to 1.5 trillion tokens. StableLM-Alpha models are trained on the new dataset that build on The Pile, which contains 1.5 trillion tokens, roughly 3x the size of The Pile. Please visit HuggingFace checkpoint for more information about how to combine our delta weights with the original model. StableVicuna's delta weights are released under ( CC BY-NC-SA-4.0). Phung leading the training effort.ĭue to the original non-commercial license of LLaMA, we can only release the weights of our model as deltas over the original model's weights. ![]() This model is developed by StabilityAI's CarperAI team, with Duy V. It is our attempt at creating an open-source RLHF LLM Chatbot. StableVicuna is an RLHF fine-tune of Vicuna-13B v0, which itself is a fine-tune of LLaMA-13B. Try to chat with our 7B model, StableLM-Tuned-Alpha-7B, on Hugging Face Spaces. Base models are released under CC BY-SA-4.0. Released initial set of StableLM-alpha models, with 3B and 7B parameters. Delta weights over the original Llama model is released under ( CC BY-NC-SA-4.0). Released StableVicuna-13B, our RLHF fine-tune of Vicuna-13B v0, which itself is a fine-tune of LLaMA-13B.The following provides an overview of all currently available models. This repository contains Stability AI's ongoing development of the StableLM series of language models and will be continuously updated with new checkpoints. “A Stochastic Parrot, flat design, vector art” - Stable Diffusion XL
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |