cuijian0819's picture
Add comprehensive README
29906cf verified
---
tags:
- pytorch
- safetensors
- transformers
- gpt-oss
- multilingual
- text-generation
language:
- en
- es
- fr
- de
- it
- pt
license: apache-2.0
model_type: gpt-oss
pipeline_tag: text-generation
base_model: openai/gpt-oss-20b
---
# GPT-OSS-20B Function Calling Model
This repository contains the GPT-OSS-20B model fine-tuned on function calling data in PyTorch/Safetensors format, ready for use with the Transformers library.
## Model Details
- **Base Model:** openai/gpt-oss-20b
- **Fine-tuning Dataset:** Salesforce/xlam-function-calling-60k (2000 samples)
- **Fine-tuning Method:** LoRA (r=8, alpha=16)
- **Context Length:** 131,072 tokens
- **Model Size:** 20B parameters
## Files
- `model.safetensors`: Model weights in Safetensors format
- `config.json`: Model configuration
- `tokenizer.json`, `tokenizer_config.json`: Tokenizer files
- `generation_config.json`: Generation configuration
## Usage
### With Transformers Library
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained(
"cuijian0819/gpt-oss-20b-function-calling",
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("cuijian0819/gpt-oss-20b-function-calling")
# Generate text
inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_length=100, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
```
### Fine-tuning
This model can be further fine-tuned using standard PyTorch/Transformers workflows:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, Trainer, TrainingArguments
model = AutoModelForCausalLM.from_pretrained("cuijian0819/gpt-oss-20b-function-calling")
tokenizer = AutoTokenizer.from_pretrained("cuijian0819/gpt-oss-20b-function-calling")
# Your fine-tuning code here
```
## GGUF Version
For efficient inference with llama.cpp or Ollama, check out the GGUF version: [cuijian0819/gpt-oss-20b-function-calling-gguf](https://huggingface.co/cuijian0819/gpt-oss-20b-function-calling-gguf)
## Training Details
- **Training Epochs:** 5
- **Learning Rate:** 0.0002
- **Batch Size:** 4
- **Gradient Accumulation:** 4
- **Max Length:** 1024
## License
This model inherits the license from the base openai/gpt-oss-20b model.
## Citation
```bibtex
@misc{gpt-oss-20b-function-calling,
title={GPT-OSS-20B Function Calling Model},
author={cuijian0819},
year={2025},
url={https://huggingface.co/cuijian0819/gpt-oss-20b-function-calling}
}
```