--- tags: - pytorch - safetensors - transformers - gpt-oss - multilingual - text-generation language: - en - es - fr - de - it - pt license: apache-2.0 model_type: gpt-oss pipeline_tag: text-generation base_model: openai/gpt-oss-20b --- # GPT-OSS-20B Function Calling Model This repository contains the GPT-OSS-20B model fine-tuned on function calling data in PyTorch/Safetensors format, ready for use with the Transformers library. ## Model Details - **Base Model:** openai/gpt-oss-20b - **Fine-tuning Dataset:** Salesforce/xlam-function-calling-60k (2000 samples) - **Fine-tuning Method:** LoRA (r=8, alpha=16) - **Context Length:** 131,072 tokens - **Model Size:** 20B parameters ## Files - `model.safetensors`: Model weights in Safetensors format - `config.json`: Model configuration - `tokenizer.json`, `tokenizer_config.json`: Tokenizer files - `generation_config.json`: Generation configuration ## Usage ### With Transformers Library ```python from transformers import AutoModelForCausalLM, AutoTokenizer # Load model and tokenizer model = AutoModelForCausalLM.from_pretrained( "cuijian0819/gpt-oss-20b-function-calling", torch_dtype="auto", device_map="auto" ) tokenizer = AutoTokenizer.from_pretrained("cuijian0819/gpt-oss-20b-function-calling") # Generate text inputs = tokenizer("Your prompt here", return_tensors="pt") outputs = model.generate(**inputs, max_length=100, temperature=0.7) response = tokenizer.decode(outputs[0], skip_special_tokens=True) ``` ### Fine-tuning This model can be further fine-tuned using standard PyTorch/Transformers workflows: ```python from transformers import AutoModelForCausalLM, AutoTokenizer, Trainer, TrainingArguments model = AutoModelForCausalLM.from_pretrained("cuijian0819/gpt-oss-20b-function-calling") tokenizer = AutoTokenizer.from_pretrained("cuijian0819/gpt-oss-20b-function-calling") # Your fine-tuning code here ``` ## GGUF Version For efficient inference with llama.cpp or Ollama, check out the GGUF version: [cuijian0819/gpt-oss-20b-function-calling-gguf](https://huggingface.co/cuijian0819/gpt-oss-20b-function-calling-gguf) ## Training Details - **Training Epochs:** 5 - **Learning Rate:** 0.0002 - **Batch Size:** 4 - **Gradient Accumulation:** 4 - **Max Length:** 1024 ## License This model inherits the license from the base openai/gpt-oss-20b model. ## Citation ```bibtex @misc{gpt-oss-20b-function-calling, title={GPT-OSS-20B Function Calling Model}, author={cuijian0819}, year={2025}, url={https://huggingface.co/cuijian0819/gpt-oss-20b-function-calling} } ```