Text Generation
Transformers
GGUF
Korean
English
unsloth
phi-3
ollama
conversational

Phi-3-medium-4k-instruct-ko-poc-gguf-v0.1

Model Details

This model converted the daekeun-ml/Phi-3-medium-4k-instruct-ko-poc-v0.1 to gguf 4-bit format.

For detailed instructions, please refer to Microsoft's official repo.

Dataset

The dataset used for training is as follows. To prevent catastrophic forgetting, I included non-Korean corpus as training data. Note that we did not use all of the data, but only sampled some of it. Korean textbooks were converted to Q&A format. The Guanaco dataset has been reformatted to fit the multiturn format like <|user|>\n{Q1}<|end|>\n<|assistant|>\n{A1}<|end|>\n<|user|>\n{Q2}<|end|>\n<|assistant|>\n{A2}<|end|>.

How to Get Started with the Model using Ollama

  1. Install Ollama:
curl -fsSL https://ollama.com/install.sh | sh
  1. Get the Modelfile:
huggingface-cli download daekeun-ml/Phi-3-medium-4k-instruct-ko-poc-gguf-v0.1 Modelfile_q4 --local-dir /path/to/your/local/dir
  1. Build the Ollama Model: Use the Ollama CLI to create your model with the following command:
ollama create phi3-ko -f Modelfile_q4
  1. Run the model:
ollama run phi3-ko What is Machine Learning?

Notes

License

apache 2.0; The license of phi-3 is MIT, but I considered the licensing of the dataset and library used for training.

Caution

This model was created as a personal experiment, unrelated to the organization I work for. The model may not operate correctly because separate verification was not performed. Please be careful unless it is for personal experimentation or PoC (Proof of Concept)!

Downloads last month
13
GGUF
Model size
14B params
Architecture
llama

4-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model authors have turned it off explicitly.

Datasets used to train daekeun-ml/Phi-3-medium-4k-instruct-ko-poc-gguf-v0.1