LLaMA-7B

This repo contains a low-rank adapter for LLaMA-7b trained on Bulgarian dataset.

This model was introduced in this paper.

Model description

The training data is private Bulgarian dataset.

Intended uses & limitations

This is an instruction-based model similart to ChatGPT but in Bulgarian.

How to use

Here is how to use this model in PyTorch:

>>> git clone https://github.com/tloen/alpaca-lora.git
>>> cd alpaca-lora
>>> pip install -r requirements.txt
>>>
>>> python generate.py \
    --load_8bit \
    --base_model 'yahma/llama-7b-hf' \
    --lora_weights 'rmihaylov/alpaca-lora-bg-7b' \
    --share_gradio

This will download both a base model and an adapter from huggingface. Then it will run a gradio interface for chatting.

Example using this model: Colab. You need a Colab Pro because the model needs high ram when loading.

Interface

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model authors have turned it off explicitly.