Text Generation
Transformers
Safetensors
English
llama
conversational
text-generation-inference
Inference Endpoints

πŸ’» Usage

!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "frankenmerger/MiniLlama-1.8b-Chat-v0.1"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
Downloads last month
182
Safetensors
Model size
1.89B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for frankenmerger/MiniLlama-1.8b-Chat-v0.1

Finetuned
(197)
this model
Finetunes
6 models
Quantizations
3 models

Datasets used to train frankenmerger/MiniLlama-1.8b-Chat-v0.1