Triangle104 commited on
Commit
8592550
·
verified ·
1 Parent(s): 3e751c1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -31
README.md CHANGED
@@ -124,6 +124,7 @@ Refer to the [original model card](https://huggingface.co/prithivMLmods/QwQ-LCoT
124
 
125
  ---
126
  Model details:
 
127
  The QwQ-LCoT2-7B-Instruct is a fine-tuned language model
128
  designed for advanced reasoning and instruction-following tasks. It
129
  leverages the Qwen2.5-7B base model and has been fine-tuned on the chain
@@ -133,20 +134,10 @@ logical reasoning, detailed explanations, and multi-step
133
  problem-solving, making it ideal for applications such as
134
  instruction-following, text generation, and complex reasoning tasks.
135
 
136
-
137
-
138
-
139
-
140
-
141
-
142
  Quickstart with Transformers
143
 
144
-
145
-
146
-
147
  Here provides a code snippet with apply_chat_template to show you how to load the tokenizer and model and how to generate contents.
148
 
149
-
150
  from transformers import AutoModelForCausalLM, AutoTokenizer
151
 
152
  model_name = "prithivMLmods/QwQ-LCoT2-7B-Instruct"
@@ -180,23 +171,11 @@ generated_ids = [
180
 
181
  response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
182
 
183
-
184
-
185
-
186
-
187
-
188
-
189
-
190
-
191
  Intended Use
192
-
193
-
194
-
195
 
196
  The QwQ-LCoT2-7B-Instruct model is designed for advanced reasoning
197
  and instruction-following tasks, with specific applications including:
198
 
199
-
200
  Instruction Following: Providing detailed and step-by-step guidance for a wide range of user queries.
201
  Logical Reasoning: Solving problems requiring multi-step thought processes, such as math problems or complex logic-based scenarios.
202
  Text Generation: Crafting coherent, contextually relevant, and well-structured text in response to prompts.
@@ -205,17 +184,8 @@ that require chain-of-thought (CoT) reasoning, making it ideal for
205
  education, tutoring, and technical support.
206
  Knowledge Enhancement: Leveraging reasoning datasets to offer deeper insights and explanations for a wide variety of topics.
207
 
208
-
209
-
210
-
211
-
212
-
213
-
214
  Limitations
215
 
216
-
217
-
218
-
219
  Data Bias: As the model is fine-tuned on specific datasets, its outputs may reflect inherent biases from the training data.
220
  Context Limitation: Performance may degrade for
221
  tasks requiring knowledge or reasoning that significantly exceeds the
 
124
 
125
  ---
126
  Model details:
127
+ -
128
  The QwQ-LCoT2-7B-Instruct is a fine-tuned language model
129
  designed for advanced reasoning and instruction-following tasks. It
130
  leverages the Qwen2.5-7B base model and has been fine-tuned on the chain
 
134
  problem-solving, making it ideal for applications such as
135
  instruction-following, text generation, and complex reasoning tasks.
136
 
 
 
 
 
 
 
137
  Quickstart with Transformers
138
 
 
 
 
139
  Here provides a code snippet with apply_chat_template to show you how to load the tokenizer and model and how to generate contents.
140
 
 
141
  from transformers import AutoModelForCausalLM, AutoTokenizer
142
 
143
  model_name = "prithivMLmods/QwQ-LCoT2-7B-Instruct"
 
171
 
172
  response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
173
 
 
 
 
 
 
 
 
 
174
  Intended Use
 
 
 
175
 
176
  The QwQ-LCoT2-7B-Instruct model is designed for advanced reasoning
177
  and instruction-following tasks, with specific applications including:
178
 
 
179
  Instruction Following: Providing detailed and step-by-step guidance for a wide range of user queries.
180
  Logical Reasoning: Solving problems requiring multi-step thought processes, such as math problems or complex logic-based scenarios.
181
  Text Generation: Crafting coherent, contextually relevant, and well-structured text in response to prompts.
 
184
  education, tutoring, and technical support.
185
  Knowledge Enhancement: Leveraging reasoning datasets to offer deeper insights and explanations for a wide variety of topics.
186
 
 
 
 
 
 
 
187
  Limitations
188
 
 
 
 
189
  Data Bias: As the model is fine-tuned on specific datasets, its outputs may reflect inherent biases from the training data.
190
  Context Limitation: Performance may degrade for
191
  tasks requiring knowledge or reasoning that significantly exceeds the