|
--- |
|
license: other |
|
license_name: qwen |
|
license_link: >- |
|
https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT |
|
language: |
|
- en |
|
- zh |
|
tags: |
|
- qwen |
|
- qwen1.5 |
|
- qwen2 |
|
- llama |
|
inference: false |
|
--- |
|
## Description |
|
This repo containst the "LLaMAfied" version of [Qwen1.5-72B-Chat](https://huggingface.co/Qwen/Qwen1.5-72B-Chat) by Alibaba Cloud. I used the amazing [script](https://github.com/Minami-su/character_AI_open/blob/main/llamafy_qwen_v2.py) made by [Minami-su](https://huggingface.co/Minami-su) to LLaMAfy the model. |
|
|
|
## Usage |
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer |
|
tokenizer = AutoTokenizer.from_pretrained("sayhan/Qwen1.5-72B-Chat-LLaMAfied") |
|
model = AutoModelForCausalLM.from_pretrained("sayhan/Qwen1.5-72B-Chat-LLaMAfied", torch_dtype="auto", device_map="auto") |
|
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) |
|
|
|
messages = [ |
|
{"role": "user", "content": "Who are you?"} |
|
] |
|
inputs = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt") |
|
inputs = inputs.to("cuda") |
|
generate_ids = model.generate(inputs,max_length=2048, streamer=streamer) |
|
``` |
|
|
|
## Other LLaMAfied Qwen1.5 Models |
|
The two other sizes of the Qwen1.5 have been LLaMAfied by [Minami-su](https://huggingface.co/Minami-su) |
|
- **0.5B:** [Minami-su/Qwen1.5-0.5B-Chat_llamafy](https://huggingface.co/Minami-su/Qwen1.5-0.5B-Chat_llamafy) |
|
- **7B:** [Minami-su/Qwen1.5-7B-Chat_llamafy](https://huggingface.co/Minami-su/Qwen1.5-7B-Chat_llamafy) |