LLaMA-7B
This repo contains a low-rank adapter for LLaMA-7b trained on Bulgarian dataset.
This model was introduced in this paper.
Model description
The training data is private Bulgarian dataset.
Intended uses & limitations
This is an instruction-based model similart to ChatGPT but in Bulgarian.
How to use
Here is how to use this model in PyTorch:
>>> git clone https://github.com/tloen/alpaca-lora.git
>>> cd alpaca-lora
>>> pip install -r requirements.txt
>>>
>>> python generate.py \
--load_8bit \
--base_model 'yahma/llama-7b-hf' \
--lora_weights 'rmihaylov/alpaca-lora-bg-7b' \
--share_gradio
This will download both a base model and an adapter from huggingface. Then it will run a gradio interface for chatting.
Example using this model: Colab. You need a Colab Pro because the model needs high ram when loading.
Interface
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model authors have turned it off explicitly.