sayhan's picture
Create README.md
c8faae9 verified
---
license: other
license_name: qwen
license_link: >-
https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT
language:
- en
- zh
tags:
- qwen
- qwen1.5
- qwen2
- llama
inference: false
---
## Description
This repo containst the "LLaMAfied" version of [Qwen1.5-72B-Chat](https://huggingface.co/Qwen/Qwen1.5-72B-Chat) by Alibaba Cloud. I used the amazing [script](https://github.com/Minami-su/character_AI_open/blob/main/llamafy_qwen_v2.py) made by [Minami-su](https://huggingface.co/Minami-su) to LLaMAfy the model.
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
tokenizer = AutoTokenizer.from_pretrained("sayhan/Qwen1.5-72B-Chat-LLaMAfied")
model = AutoModelForCausalLM.from_pretrained("sayhan/Qwen1.5-72B-Chat-LLaMAfied", torch_dtype="auto", device_map="auto")
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
messages = [
{"role": "user", "content": "Who are you?"}
]
inputs = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
inputs = inputs.to("cuda")
generate_ids = model.generate(inputs,max_length=2048, streamer=streamer)
```
## Other LLaMAfied Qwen1.5 Models
The two other sizes of the Qwen1.5 have been LLaMAfied by [Minami-su](https://huggingface.co/Minami-su)
- **0.5B:** [Minami-su/Qwen1.5-0.5B-Chat_llamafy](https://huggingface.co/Minami-su/Qwen1.5-0.5B-Chat_llamafy)
- **7B:** [Minami-su/Qwen1.5-7B-Chat_llamafy](https://huggingface.co/Minami-su/Qwen1.5-7B-Chat_llamafy)