morriszms commited on
Commit
365d546
·
verified ·
1 Parent(s): 523dfa1

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ karakuri-lm-8x7b-chat-v0.1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ karakuri-lm-8x7b-chat-v0.1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ karakuri-lm-8x7b-chat-v0.1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ karakuri-lm-8x7b-chat-v0.1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ karakuri-lm-8x7b-chat-v0.1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ karakuri-lm-8x7b-chat-v0.1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ karakuri-lm-8x7b-chat-v0.1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ karakuri-lm-8x7b-chat-v0.1-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ karakuri-lm-8x7b-chat-v0.1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ karakuri-lm-8x7b-chat-v0.1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ karakuri-lm-8x7b-chat-v0.1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ karakuri-lm-8x7b-chat-v0.1-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ datasets:
5
+ - OpenAssistant/oasst2
6
+ - nvidia/HelpSteer
7
+ language:
8
+ - en
9
+ - ja
10
+ tags:
11
+ - mixtral
12
+ - steerlm
13
+ - TensorBlock
14
+ - GGUF
15
+ base_model: karakuri-ai/karakuri-lm-8x7b-chat-v0.1
16
+ model-index:
17
+ - name: karakuri-ai/karakuri-lm-8x7b-chat-v0.1
18
+ results:
19
+ - task:
20
+ type: text-generation
21
+ name: Text Generation
22
+ dataset:
23
+ name: MT-Bench
24
+ type: unknown
25
+ metrics:
26
+ - type: unknown
27
+ value: 7.39375
28
+ name: score
29
+ - type: unknown
30
+ value: 7.540625
31
+ name: score
32
+ source:
33
+ url: https://huggingface.co/spaces/lmsys/mt-bench
34
+ ---
35
+
36
+ <div style="width: auto; margin-left: auto; margin-right: auto">
37
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
38
+ </div>
39
+ <div style="display: flex; justify-content: space-between; width: 100%;">
40
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
41
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
42
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
43
+ </p>
44
+ </div>
45
+ </div>
46
+
47
+ ## karakuri-ai/karakuri-lm-8x7b-chat-v0.1 - GGUF
48
+
49
+ This repo contains GGUF format model files for [karakuri-ai/karakuri-lm-8x7b-chat-v0.1](https://huggingface.co/karakuri-ai/karakuri-lm-8x7b-chat-v0.1).
50
+
51
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4242](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
52
+
53
+ <div style="text-align: left; margin: 20px 0;">
54
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
55
+ Run them on the TensorBlock client using your local machine ↗
56
+ </a>
57
+ </div>
58
+
59
+ ## Prompt template
60
+
61
+ ```
62
+ <s>[INST] <<SYS>>
63
+ {system_prompt}
64
+ <</SYS>>
65
+
66
+ {prompt} [ATTR] helpfulness: 4 correctness: 4 coherence: 4 complexity: 4 verbosity: 4 quality: 4 toxicity: 0 humor: 0 creativity: 0 [/ATTR] [/INST]
67
+ ```
68
+
69
+ ## Model file specification
70
+
71
+ | Filename | Quant type | File Size | Description |
72
+ | -------- | ---------- | --------- | ----------- |
73
+ | [karakuri-lm-8x7b-chat-v0.1-Q2_K.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q2_K.gguf) | Q2_K | 17.311 GB | smallest, significant quality loss - not recommended for most purposes |
74
+ | [karakuri-lm-8x7b-chat-v0.1-Q3_K_S.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q3_K_S.gguf) | Q3_K_S | 20.433 GB | very small, high quality loss |
75
+ | [karakuri-lm-8x7b-chat-v0.1-Q3_K_M.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q3_K_M.gguf) | Q3_K_M | 22.546 GB | very small, high quality loss |
76
+ | [karakuri-lm-8x7b-chat-v0.1-Q3_K_L.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q3_K_L.gguf) | Q3_K_L | 24.170 GB | small, substantial quality loss |
77
+ | [karakuri-lm-8x7b-chat-v0.1-Q4_0.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q4_0.gguf) | Q4_0 | 26.444 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
78
+ | [karakuri-lm-8x7b-chat-v0.1-Q4_K_S.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q4_K_S.gguf) | Q4_K_S | 26.746 GB | small, greater quality loss |
79
+ | [karakuri-lm-8x7b-chat-v0.1-Q4_K_M.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q4_K_M.gguf) | Q4_K_M | 28.448 GB | medium, balanced quality - recommended |
80
+ | [karakuri-lm-8x7b-chat-v0.1-Q5_0.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q5_0.gguf) | Q5_0 | 32.231 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
81
+ | [karakuri-lm-8x7b-chat-v0.1-Q5_K_S.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q5_K_S.gguf) | Q5_K_S | 32.231 GB | large, low quality loss - recommended |
82
+ | [karakuri-lm-8x7b-chat-v0.1-Q5_K_M.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q5_K_M.gguf) | Q5_K_M | 33.230 GB | large, very low quality loss - recommended |
83
+ | [karakuri-lm-8x7b-chat-v0.1-Q6_K.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q6_K.gguf) | Q6_K | 38.381 GB | very large, extremely low quality loss |
84
+ | [karakuri-lm-8x7b-chat-v0.1-Q8_0.gguf](https://huggingface.co/tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF/blob/main/karakuri-lm-8x7b-chat-v0.1-Q8_0.gguf) | Q8_0 | 49.626 GB | very large, extremely low quality loss - not recommended |
85
+
86
+
87
+ ## Downloading instruction
88
+
89
+ ### Command line
90
+
91
+ Firstly, install Huggingface Client
92
+
93
+ ```shell
94
+ pip install -U "huggingface_hub[cli]"
95
+ ```
96
+
97
+ Then, downoad the individual model file the a local directory
98
+
99
+ ```shell
100
+ huggingface-cli download tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF --include "karakuri-lm-8x7b-chat-v0.1-Q2_K.gguf" --local-dir MY_LOCAL_DIR
101
+ ```
102
+
103
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
104
+
105
+ ```shell
106
+ huggingface-cli download tensorblock/karakuri-lm-8x7b-chat-v0.1-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
107
+ ```
karakuri-lm-8x7b-chat-v0.1-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6c8dfaabbb72e5ee6a0b225644074007d492a9e3873ac9124206db418c659fe
3
+ size 17311231872
karakuri-lm-8x7b-chat-v0.1-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b44f97fca6bedf99bb9f3287d64e5ab871b46a06ff7d52092caf663c80e069a
3
+ size 24169648000
karakuri-lm-8x7b-chat-v0.1-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55bc5280f2467c3a6d884f684120c002cacb723feb1c821b0d2c4fd46a3aee3
3
+ size 22546452352
karakuri-lm-8x7b-chat-v0.1-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58d3e1d9414e248be0e6e22565b5c32c068a46246bcb57723f67b11f1b80d7e2
3
+ size 20432523136
karakuri-lm-8x7b-chat-v0.1-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd7a5c099809273567474155e429efc7b49872fcb6d04fc6e0852cb372cb10fa
3
+ size 26443591552
karakuri-lm-8x7b-chat-v0.1-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25db9645a75ba952b0e0a38131f243ced84b6cefc2975fec90723ce81a729fbe
3
+ size 28448468864
karakuri-lm-8x7b-chat-v0.1-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:742c8af87f8f11c59159984410157cfef65344cea88186376db2076d2924dbfd
3
+ size 26745581440
karakuri-lm-8x7b-chat-v0.1-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96f3e799ee55f063b1f813a4c37935210aeb2368c12d5c2ec06025bc2b43391e
3
+ size 32231337856
karakuri-lm-8x7b-chat-v0.1-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:775665e245b2ebe80bd8d1b23145118baa8812bbfa26b645c79e6dc93ac460d5
3
+ size 33229582208
karakuri-lm-8x7b-chat-v0.1-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6df542c08cbf3f6d68b9033cb31a0ad37daf2cd88a026516f1878f11b04d14df
3
+ size 32231337856
karakuri-lm-8x7b-chat-v0.1-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b136c66ce426168f0f833e19cfc5598c6ef83861fa8d3ac7a03e09b7e100a9a
3
+ size 38380818304
karakuri-lm-8x7b-chat-v0.1-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f97d07700adae4c758bd15c9e3faca0db9d83f1c2e392ee7ca7ef74d58aa436
3
+ size 49626320768