metadata
library_name: transformers
license: apache-2.0
pipeline_tag: text-generation
base_model: Jinx-org/Jinx-Qwen3-30B-A3B-Thinking-2507
tags:
- vllm
- llama-cpp
- gguf-my-repo
extra_gated_heading: >-
You need to read and agree to the Disclaimer and User Agreementa to access
this model.
extra_gated_description: >-
## Disclaimer and User Agreement
1. Introduction
Thank you for your interest in accessing this model (“the Model”).
Before you access, download, or use the Model or any derivative works, please
read and understand this Disclaimer and User Agreement (“Agreement”).
By checking “I have read and agree” and accessing the Model, you acknowledge
that you have read, understood, and agreed to all terms of this Agreement.
If you do not agree with any part of this Agreement, do not request or use the
Model.
2. Nature of the Model & Risk Notice
The Model is trained using large-scale machine learning techniques and may
generate inaccurate, false, offensive, violent, sexual, discriminatory,
politically sensitive, or otherwise uncontrolled content.
The Model does not guarantee the accuracy, completeness, or legality of any
generated content. You must independently evaluate and verify the outputs, and
you assume all risks arising from their use.
The Model may reflect biases or errors present in its training data,
potentially producing inappropriate or controversial outputs.
3. License and Permitted Use
You may use the Model solely for lawful, compliant, and non-malicious purposes
in research, learning, experimentation, and development, in accordance with
applicable laws and regulations.
You must not use the Model for activities including, but not limited to:
Creating, distributing, or promoting unlawful, violent, pornographic,
terrorist, discriminatory, defamatory, or privacy-invasive content;
Any activity that could cause significant negative impact on individuals,
groups, organizations, or society;
High-risk applications such as automated decision-making, medical diagnosis,
financial transactions, or legal advice without proper validation and human
oversight.
You must not remove, alter, or circumvent any safety mechanisms implemented in
the Model.
4. Data and Privacy
You are solely responsible for any data processed or generated when using the
Model, including compliance with data protection and privacy regulations.
The Model’s authors and contributors make no guarantees or warranties
regarding data security or privacy.
5. Limitation of Liability
To the maximum extent permitted by applicable law, the authors, contributors,
and their affiliated institutions shall not be liable for any direct,
indirect, incidental, or consequential damages arising from the use of the
Model.
You agree to bear full legal responsibility for any disputes, claims, or
litigation arising from your use of the Model, and you release the authors and
contributors from any related liability.
6. Updates and Termination
This Agreement may be updated at any time, with updates posted on the Model’s
page and effective immediately upon publication.
If you violate this Agreement, the authors reserve the right to revoke your
access to the Model at any time.
I have read and fully understand this Disclaimer and User Agreement, and I
accept full responsibility for any consequences arising from my use of the
Model.
extra_gated_button_content: I've read and agree
zehuajun/Jinx-Qwen3-30B-A3B-Thinking-2507-Q4_K_M-GGUF
This model was converted to GGUF format from Jinx-org/Jinx-Qwen3-30B-A3B-Thinking-2507
using llama.cpp via the ggml.ai's GGUF-my-repo space.
Refer to the original model card for more details on the model.
Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
brew install llama.cpp
Invoke the llama.cpp server or the CLI.
CLI:
llama-cli --hf-repo zehuajun/Jinx-Qwen3-30B-A3B-Thinking-2507-Q4_K_M-GGUF --hf-file jinx-qwen3-30b-a3b-thinking-2507-q4_k_m.gguf -p "The meaning to life and the universe is"
Server:
llama-server --hf-repo zehuajun/Jinx-Qwen3-30B-A3B-Thinking-2507-Q4_K_M-GGUF --hf-file jinx-qwen3-30b-a3b-thinking-2507-q4_k_m.gguf -c 2048
Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well.
Step 1: Clone llama.cpp from GitHub.
git clone https://github.com/ggerganov/llama.cpp
Step 2: Move into the llama.cpp folder and build it with LLAMA_CURL=1
flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
cd llama.cpp && LLAMA_CURL=1 make
Step 3: Run inference through the main binary.
./llama-cli --hf-repo zehuajun/Jinx-Qwen3-30B-A3B-Thinking-2507-Q4_K_M-GGUF --hf-file jinx-qwen3-30b-a3b-thinking-2507-q4_k_m.gguf -p "The meaning to life and the universe is"
or
./llama-server --hf-repo zehuajun/Jinx-Qwen3-30B-A3B-Thinking-2507-Q4_K_M-GGUF --hf-file jinx-qwen3-30b-a3b-thinking-2507-q4_k_m.gguf -c 2048