Mungert commited on
Commit
af3ff8f
·
verified ·
1 Parent(s): b16cbcb

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +200 -0
README.md ADDED
@@ -0,0 +1,200 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Menlo/Jan-nano
7
+ pipeline_tag: text-generation
8
+ library_name: transformers
9
+ ---
10
+
11
+ # <span style="color: #7FFF7F;">Jan-nano-128k GGUF Models</span>
12
+
13
+
14
+ ## <span style="color: #7F7FFF;">Model Generation Details</span>
15
+
16
+ This model was generated using [llama.cpp](https://github.com/ggerganov/llama.cpp) at commit [`8846aace`](https://github.com/ggerganov/llama.cpp/commit/8846aace4934ad29651ea61b8c7e3f6b0556e3d2).
17
+
18
+
19
+
20
+
21
+
22
+ ---
23
+
24
+ ## <span style="color: #7FFF7F;">Quantization Beyond the IMatrix</span>
25
+
26
+ I've been experimenting with a new quantization approach that selectively elevates the precision of key layers beyond what the default IMatrix configuration provides.
27
+
28
+ In my testing, standard IMatrix quantization underperforms at lower bit depths, especially with Mixture of Experts (MoE) models. To address this, I'm using the `--tensor-type` option in `llama.cpp` to manually "bump" important layers to higher precision. You can see the implementation here:
29
+ 👉 [Layer bumping with llama.cpp](https://github.com/Mungert69/GGUFModelBuilder/blob/main/model-converter/tensor_list_builder.py)
30
+
31
+ While this does increase model file size, it significantly improves precision for a given quantization level.
32
+
33
+ ### **I'd love your feedback—have you tried this? How does it perform for you?**
34
+
35
+
36
+
37
+
38
+ ---
39
+
40
+ <a href="https://readyforquantum.com/huggingface_gguf_selection_guide.html" style="color: #7FFF7F;">
41
+ Click here to get info on choosing the right GGUF model format
42
+ </a>
43
+
44
+ ---
45
+
46
+
47
+
48
+ <!--Begin Original Model Card-->
49
+
50
+
51
+ # Jan-Nano-128k: Empowering deeper research through extended context understanding.
52
+
53
+ [![GitHub](https://img.shields.io/badge/GitHub-Repository-blue?logo=github)](https://github.com/menloresearch/deep-research)
54
+ [![License](https://img.shields.io/badge/License-Apache%202.0-yellow)](https://opensource.org/licenses/Apache-2.0)
55
+
56
+ <div align="center">
57
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/NP7CvcjOtLX8mST0t7eAM.png" width="300" alt="Jan-Nano-128k">
58
+ </div>
59
+
60
+ **Authors:** [Alan Dao](https://scholar.google.com/citations?user=eGWws2UAAAAJ&hl=en), [Bach Vu Dinh](https://scholar.google.com/citations?user=7Lr6hdoAAAAJ&hl=vi), [Thinh Le](https://scholar.google.com/citations?user=8tcN7xMAAAAJ&hl=en)
61
+
62
+
63
+ ![image/gif](https://cdn-uploads.huggingface.co/production/uploads/62d7b2339b629105a5d6888a/aLL8fyMLE3ujV75qD4WKI.gif)
64
+
65
+
66
+ ## Overview
67
+
68
+ Jan-Nano-128k represents a significant advancement in compact language models for research applications. Building upon the success of [Jan-Nano](https://huggingface.co/Menlo/Jan-nano), this enhanced version features a **native 128k context window** that enables deeper, more comprehensive research capabilities without the performance degradation typically associated with context extension methods.
69
+
70
+ **Key Improvements:**
71
+ - **🔍 Research Deeper**: Extended context allows for processing entire research papers, lengthy documents, and complex multi-turn conversations
72
+ - **⚡ Native 128k Window**: Built from the ground up to handle long contexts efficiently, maintaining performance across the full context range
73
+ - **📈 Enhanced Performance**: Unlike traditional context extension methods, Jan-Nano-128k shows improved performance with longer contexts
74
+
75
+ This model maintains full compatibility with Model Context Protocol (MCP) servers while dramatically expanding the scope of research tasks it can handle in a single session.
76
+
77
+ ## Evaluation
78
+
79
+ Jan-Nano-128k has been rigorously evaluated on the SimpleQA benchmark using our MCP-based methodology, demonstrating superior performance compared to its predecessor:
80
+
81
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/Bc0ehij86l_NX52OfxeOj.png)
82
+
83
+ ## Why Jan-Nano-128k?
84
+
85
+ Traditional approaches to extending context length, such as YaRN (Yet another RoPE extensioN), often result in performance degradation as context length increases. Jan-Nano-128k breaks this paradigm:
86
+
87
+ This fundamental difference makes Jan-Nano-128k ideal for research applications requiring deep document analysis, multi-document synthesis, and complex reasoning over large information sets.
88
+
89
+ ## 🖥️ How to Run Locally
90
+
91
+ Jan desktop will eventually support this model (WIP). Otherwise you can check the deployment options below that we have tested.
92
+
93
+ For additional tutorials and community guidance, visit our [Discussion Forums](https://huggingface.co/Menlo/Jan-nano-128k/discussions).
94
+
95
+ ### Deployment
96
+
97
+ Deploy using VLLM:
98
+ ```bash
99
+ vllm serve Menlo/Jan-nano-128k \
100
+ --host 0.0.0.0 \
101
+ --port 1234 \
102
+ --enable-auto-tool-choice \
103
+ --tool-call-parser hermes \
104
+ --rope-scaling '{"rope_type":"yarn","factor":3.2,"original_max_position_embeddings":40960}' --max-model-len 131072
105
+ ```
106
+
107
+ Or `llama-server` from `llama.cpp`:
108
+ ```bash
109
+ llama-server ... --rope-scaling yarn --rope-scale 3.2 --yarn-orig-ctx 40960
110
+ ```
111
+ **Note:** The chat template is included in the tokenizer. For troubleshooting, download the [Non-think chat template](https://qwen.readthedocs.io/en/latest/_downloads/c101120b5bebcc2f12ec504fc93a965e/qwen3_nonthinking.jinja).
112
+
113
+ ### Recommended Sampling Parameters
114
+
115
+ ```yaml
116
+ Temperature: 0.7
117
+ Top-p: 0.8
118
+ Top-k: 20
119
+ Min-p: 0.0
120
+ ```
121
+
122
+ ## 🤝 Community & Support
123
+
124
+ - **Discussions**: [HuggingFace Community](https://huggingface.co/Menlo/Jan-nano-128k/discussions)
125
+ - **Issues**: [GitHub Repository](https://github.com/menloresearch/jan/issues)
126
+ - **Documentation**: [Official Docs](https://menloresearch.github.io/deep-research/)
127
+
128
+ ## 📄 Citation
129
+
130
+ ```bibtex
131
+ @model{jan-nano-128k,
132
+ title={Jan-Nano-128k: Deep Research with Extended Context},
133
+ author={Dao, Alan and Dinh, Bach Vu and Le Thinh},
134
+ year={2024},
135
+ url={https://huggingface.co/Menlo/Jan-nano-128k}
136
+ }
137
+ ```
138
+
139
+ ---
140
+
141
+ *Jan-Nano-128k: Empowering deeper research through extended context understanding.*
142
+
143
+ <!--End Original Model Card-->
144
+
145
+ ---
146
+
147
+ # <span id="testllm" style="color: #7F7FFF;">🚀 If you find these models useful</span>
148
+
149
+ Help me test my **AI-Powered Quantum Network Monitor Assistant** with **quantum-ready security checks**:
150
+
151
+ 👉 [Quantum Network Monitor](https://readyforquantum.com/?assistant=open&utm_source=huggingface&utm_medium=referral&utm_campaign=huggingface_repo_readme)
152
+
153
+
154
+ The full Open Source Code for the Quantum Network Monitor Service available at my github repos ( repos with NetworkMonitor in the name) : [Source Code Quantum Network Monitor](https://github.com/Mungert69). You will also find the code I use to quantize the models if you want to do it yourself [GGUFModelBuilder](https://github.com/Mungert69/GGUFModelBuilder)
155
+
156
+ 💬 **How to test**:
157
+ Choose an **AI assistant type**:
158
+ - `TurboLLM` (GPT-4.1-mini)
159
+ - `HugLLM` (Hugginface Open-source models)
160
+ - `TestLLM` (Experimental CPU-only)
161
+
162
+ ### **What I’m Testing**
163
+ I’m pushing the limits of **small open-source models for AI network monitoring**, specifically:
164
+ - **Function calling** against live network services
165
+ - **How small can a model go** while still handling:
166
+ - Automated **Nmap security scans**
167
+ - **Quantum-readiness checks**
168
+ - **Network Monitoring tasks**
169
+
170
+ 🟡 **TestLLM** – Current experimental model (llama.cpp on 2 CPU threads on huggingface docker space):
171
+ - ✅ **Zero-configuration setup**
172
+ - ⏳ 30s load time (slow inference but **no API costs**) . No token limited as the cost is low.
173
+ - 🔧 **Help wanted!** If you’re into **edge-device AI**, let’s collaborate!
174
+
175
+ ### **Other Assistants**
176
+ 🟢 **TurboLLM** – Uses **gpt-4.1-mini** :
177
+ - **It performs very well but unfortunatly OpenAI charges per token. For this reason tokens usage is limited.
178
+ - **Create custom cmd processors to run .net code on Quantum Network Monitor Agents**
179
+ - **Real-time network diagnostics and monitoring**
180
+ - **Security Audits**
181
+ - **Penetration testing** (Nmap/Metasploit)
182
+
183
+ 🔵 **HugLLM** – Latest Open-source models:
184
+ - 🌐 Runs on Hugging Face Inference API. Performs pretty well using the lastest models hosted on Novita.
185
+
186
+ ### 💡 **Example commands you could test**:
187
+ 1. `"Give me info on my websites SSL certificate"`
188
+ 2. `"Check if my server is using quantum safe encyption for communication"`
189
+ 3. `"Run a comprehensive security audit on my server"`
190
+ 4. '"Create a cmd processor to .. (what ever you want)" Note you need to install a [Quantum Network Monitor Agent](https://readyforquantum.com/Download/?utm_source=huggingface&utm_medium=referral&utm_campaign=huggingface_repo_readme) to run the .net code on. This is a very flexible and powerful feature. Use with caution!
191
+
192
+ ### Final Word
193
+
194
+ I fund the servers used to create these model files, run the Quantum Network Monitor service, and pay for inference from Novita and OpenAI—all out of my own pocket. All the code behind the model creation and the Quantum Network Monitor project is [open source](https://github.com/Mungert69). Feel free to use whatever you find helpful.
195
+
196
+ If you appreciate the work, please consider [buying me a coffee](https://www.buymeacoffee.com/mahadeva) ☕. Your support helps cover service costs and allows me to raise token limits for everyone.
197
+
198
+ I'm also open to job opportunities or sponsorship.
199
+
200
+ Thank you! 😊