--- base_model: unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit library_name: peft license: apache-2.0 tags: - oracle - scm - fusion-cloud - adapter - dora - lora language: - en pipeline_tag: text-generation --- # Oracle Fusion Cloud SCM - DoRA Adapter Bu Oracle Fusion Cloud SCM konularında uzmanlaşmış **DoRA (Weight-Decomposed Low-Rank Adaptation)** adapter'ıdır. ## 🎯 Kullanım ### Google Colab'da Merge: ```python # 1. Base model yükle from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel import torch base_model_name = "unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit" adapter_name = "ozkurt7/oracle-deepseek-r1-adapter" # 2. Model ve adapter yükle tokenizer = AutoTokenizer.from_pretrained(base_model_name) base_model = AutoModelForCausalLM.from_pretrained( base_model_name, torch_dtype=torch.float16, device_map="auto" ) model = PeftModel.from_pretrained(base_model, adapter_name) # 3. Merge işlemi merged_model = model.merge_and_unload() # 4. Kaydet merged_model.save_pretrained("./oracle-merged") tokenizer.save_pretrained("./oracle-merged") # 5. Test messages = [ {"role": "system", "content": "You are an Oracle Fusion Cloud SCM expert."}, {"role": "user", "content": "What is Oracle SCM Cloud?"} ] text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) inputs = tokenizer(text, return_tensors="pt") outputs = merged_model.generate(**inputs, max_new_tokens=200, temperature=0.7) response = tokenizer.decode(outputs[0], skip_special_tokens=True) print(response) ``` ### Local kullanım: ```bash # Adapter'ı indir git clone https://huggingface.co/ozkurt7/oracle-deepseek-r1-adapter # Python'da merge python merge_adapter.py ``` ## 📊 Model Details - **Base Model**: unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit - **Technique**: DoRA (Weight-Decomposed Low-Rank Adaptation) - **Domain**: Oracle Fusion Cloud SCM - **Status**: Adapter only (merge required) - **Memory**: ~500MB (adapter only) ## 🚀 Next Steps 1. Google Colab'da bu adapter'ı kullanarak merge yapın 2. Merge edilen modeli yeni repo'ya upload edin 3. GGUF formatına dönüştürün ## 🛠️ Troubleshooting - **Memory Error**: Colab Pro kullanın veya local'de merge yapın - **Loading Error**: `trust_remote_code=True` ekleyin - **CUDA Error**: `device_map="auto"` kullanın **Created by**: Kaggle → Google Colab workflow **Date**: 2025-08-12