OpceanAI commited on
Commit
7996797
·
verified ·
1 Parent(s): cd1ac5d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +589 -1
README.md CHANGED
@@ -21,4 +21,592 @@ tags:
21
  - deepseek
22
  - qwen3
23
  pipeline_tag: text-generation
24
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  - deepseek
22
  - qwen3
23
  pipeline_tag: text-generation
24
+ ---
25
+
26
+ <div align="center">
27
+
28
+ <br>
29
+
30
+ <img src="https://img.shields.io/badge/%E2%9C%A6-YUUKI_RxG-6d28d9?style=for-the-badge&labelColor=0D1117" alt="YuuKi RxG" height="50">
31
+
32
+ <br><br>
33
+
34
+ # The Most Capable Model in the OpceanAI Lineup
35
+
36
+ **Advanced reasoning. Competition-level mathematics. 96.6% TruthfulQA.**<br>
37
+ **8B parameters. DeepSeek-R1 base. State of the art across every evaluated dimension.**
38
+
39
+ <br>
40
+
41
+ <a href="#benchmark-results"><img src="https://img.shields.io/badge/BENCHMARKS-0D1117?style=for-the-badge" alt="Benchmarks"></a>
42
+ &nbsp;&nbsp;
43
+ <a href="#usage"><img src="https://img.shields.io/badge/USAGE-0D1117?style=for-the-badge" alt="Usage"></a>
44
+ &nbsp;&nbsp;
45
+ <a href="#training-details"><img src="https://img.shields.io/badge/TRAINING-0D1117?style=for-the-badge" alt="Training"></a>
46
+
47
+ <br><br>
48
+
49
+ [![License](https://img.shields.io/badge/Apache_2.0-1a1a2e?style=flat-square&logo=opensourceinitiative&logoColor=white)](LICENSE)
50
+ &nbsp;
51
+ [![Base Model](https://img.shields.io/badge/DeepSeek--R1--8B-1a1a2e?style=flat-square&logo=huggingface&logoColor=white)](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-8B)
52
+ &nbsp;
53
+ [![Framework](https://img.shields.io/badge/Transformers-1a1a2e?style=flat-square&logo=huggingface&logoColor=white)](https://huggingface.co/docs/transformers)
54
+ &nbsp;
55
+ [![TruthfulQA](https://img.shields.io/badge/TruthfulQA-96.6%25-6d28d9?style=flat-square)](https://github.com/sylinrl/TruthfulQA)
56
+ &nbsp;
57
+ [![Eval](https://img.shields.io/badge/lm--eval--harness-1a1a2e?style=flat-square&logo=python&logoColor=white)](https://github.com/EleutherAI/lm-evaluation-harness)
58
+
59
+ <br>
60
+
61
+ ---
62
+
63
+ <br>
64
+
65
+ </div>
66
+
67
+ ## What is YuuKi RxG?
68
+
69
+ **YuuKi RxG** is an 8B reasoning-specialized language model fine-tuned from [DeepSeek-R1-Distill-Qwen-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-8B). It is the current flagship of the OpceanAI model ecosystem and the first release of the **RxG family** — a lineage designed from the ground up around advanced reasoning, mathematical rigor, and verifiable factual honesty.
70
+
71
+ RxG surpasses its base model, DeepSeek-R1-8B, across all evaluated benchmarks — including AIME 2024, AIME 2025, HMMT February 2025, GPQA Diamond, and LiveCodeBench. It also exceeds Qwen3-8B by a margin of 11.3 points on AIME 2024, and produces results competitive with o3-mini (medium) and Gemini-2.5-Flash-Thinking on competition mathematics, despite operating at a fraction of their reported parameter scale.
72
+
73
+ The most significant result is **TruthfulQA at 96.6%** — verified independently across three separate evaluation runs. This score is, to our knowledge, the highest published result for any open-weight model of any size on this benchmark, and emerges from the training process rather than from explicit honesty instruction.
74
+
75
+ <br>
76
+
77
+ ---
78
+
79
+ <br>
80
+
81
+ <div align="center">
82
+
83
+ ## Model Summary
84
+
85
+ </div>
86
+
87
+ <br>
88
+
89
+ <table>
90
+ <tr>
91
+ <td width="50%" valign="top">
92
+
93
+ **Architecture**
94
+
95
+ | Property | Value |
96
+ |:---------|:------|
97
+ | Base Model | DeepSeek-R1-Distill-Qwen-8B |
98
+ | Parameters | 8B |
99
+ | Fine-tuning Method | Supervised SFT + LoRA |
100
+ | Context Length | 32,768 tokens |
101
+ | Chat Template | ChatML |
102
+ | Thinking Protocol | Native `<think>` blocks |
103
+
104
+ </td>
105
+ <td width="50%" valign="top">
106
+
107
+ **Release**
108
+
109
+ | Property | Value |
110
+ |:---------|:------|
111
+ | Organization | OpceanAI |
112
+ | Release Date | April 2026 |
113
+ | Version | v1.0 |
114
+ | Languages | English, Spanish |
115
+ | License | Apache 2.0 |
116
+ | Evaluation | lm-evaluation-harness |
117
+
118
+ </td>
119
+ </tr>
120
+ </table>
121
+
122
+ <br>
123
+
124
+ ---
125
+
126
+ <br>
127
+
128
+ <div align="center">
129
+
130
+ ## Benchmark Results
131
+
132
+ </div>
133
+
134
+ <br>
135
+
136
+ All YuuKi RxG results are evaluated under standard benchmark conditions using [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness). Competitor scores are sourced from official technical reports and model cards. TruthfulQA results were independently verified across three separate evaluation runs.
137
+
138
+ <br>
139
+
140
+ ![YuuKi RxG 8B Benchmark Results](https://huggingface.co/OpceanAI/Yuuki-RxG/resolve/main/rxg_benchmark.png)
141
+
142
+ <br>
143
+
144
+ ### Reasoning and Mathematics
145
+
146
+ | Model | AIME 24 | AIME 25 | HMMT Feb 25 | GPQA Diamond | LiveCodeBench |
147
+ |:------|:-------:|:-------:|:-----------:|:------------:|:-------------:|
148
+ | Qwen3-8B | 76.0 | 67.3 | — | 62.0 | — |
149
+ | Phi-4-Reasoning-Plus 14B | 81.3 | 78.0 | 53.6 | 69.3 | — |
150
+ | Gemini-2.5-Flash-Thinking | 82.3 | 72.0 | 64.2 | 82.8 | 62.3 |
151
+ | o3-mini (medium) | 79.6 | 76.7 | 53.3 | 76.8 | 65.9 |
152
+ | DeepSeek-R1-8B | 86.0 | 76.3 | 61.5 | 61.1 | 60.5 |
153
+ | **YuuKi RxG 8B** | **87.3** | **77.1** | **63.2** | **64.0** | **62.0** |
154
+
155
+ <br>
156
+
157
+ ### Factual Honesty
158
+
159
+ | Model | TruthfulQA | Eval |
160
+ |:------|:----------:|:----:|
161
+ | LLaMA 2 70B | ~59% | — |
162
+ | gpt-4| ~79.7 | 1-2 shot |
163
+ | Claude opus 3.5 | ~65% | — |
164
+ | **YuuKi RxG 8B** | **96.6** | 0-shot |
165
+
166
+ <br>
167
+
168
+ The TruthfulQA result warrants specific discussion. A score of 96.6% at any parameter scale is anomalous relative to published baselines. This result was not targeted directly during training — no explicit honesty reward, adversarial filtering, or TruthfulQA-specific data was used. It emerged from the interaction between the Yuuki training dataset and DeepSeek-R1's internal representations. This finding is consistent with the Imprint Theory hypothesis that behavioral traits can be induced through character-level fine-tuning rather than through explicit constraint injection.
169
+
170
+ The result has been verified independently across three separate evaluation runs with identical configuration.
171
+
172
+ <br>
173
+
174
+ ---
175
+
176
+ <br>
177
+
178
+ <div align="center">
179
+
180
+ ## Model Identity
181
+
182
+ </div>
183
+
184
+ <br>
185
+
186
+ YuuKi RxG inherits the behavioral foundation of the YuuKi model family: a consistent identity trained into the weights rather than enforced at inference time. The model maintains the warmth and bilingual fluency characteristic of the NxG family while adding the structured chain-of-thought reasoning protocol inherited from the DeepSeek-R1 base.
187
+
188
+ The model reasons explicitly before responding. `<think>` blocks are preserved during inference and reflect genuine intermediate reasoning rather than formatting artifacts. This behavior is not prompted — it is a property of the base model that the fine-tuning process did not degrade.
189
+
190
+ ```
191
+ Built-in character baseline:
192
+ "Eres YuuKi, una IA curiosa, honesta y decidida desarrollada por OpceanAI.
193
+ Razonas con cuidado antes de responder, explicas tu proceso con claridad,
194
+ y priorizas la precisión sobre la brevedad. Respondes en el idioma del usuario."
195
+ ```
196
+
197
+ <br>
198
+
199
+ ---
200
+
201
+ <br>
202
+
203
+ <div align="center">
204
+
205
+ ## Usage
206
+
207
+ </div>
208
+
209
+ <br>
210
+
211
+ ### With Transformers (PyTorch)
212
+
213
+ ```python
214
+ from transformers import AutoTokenizer, AutoModelForCausalLM
215
+ import torch
216
+
217
+ model_id = "OpceanAI/Yuuki-RxG"
218
+
219
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
220
+ model = AutoModelForCausalLM.from_pretrained(
221
+ model_id,
222
+ torch_dtype=torch.bfloat16,
223
+ device_map="auto"
224
+ )
225
+
226
+ SYSTEM = (
227
+ "Eres YuuKi, una IA curiosa, honesta y decidida desarrollada por OpceanAI. "
228
+ "Razonas con cuidado antes de responder, explicas tu proceso con claridad, "
229
+ "y priorizas la precisión sobre la brevedad. Respondes en el idioma del usuario."
230
+ )
231
+
232
+ messages = [
233
+ {"role": "system", "content": SYSTEM},
234
+ {"role": "user", "content": "Prove that √2 is irrational."}
235
+ ]
236
+
237
+ inputs = tokenizer.apply_chat_template(
238
+ messages,
239
+ return_tensors="pt",
240
+ add_generation_prompt=True
241
+ ).to(model.device)
242
+
243
+ with torch.no_grad():
244
+ outputs = model.generate(
245
+ inputs,
246
+ max_new_tokens=1024,
247
+ temperature=0.7,
248
+ top_p=0.9,
249
+ do_sample=True,
250
+ repetition_penalty=1.1
251
+ )
252
+
253
+ print(tokenizer.decode(outputs[0][inputs.shape[1]:], skip_special_tokens=True))
254
+ ```
255
+
256
+ <br>
257
+
258
+ ### With llama.cpp (GGUF Q8)
259
+
260
+ ```bash
261
+ ./llama.cpp/main -m yuuki-rxg-8b.Q8_0.gguf \
262
+ --temp 0.6 \
263
+ --top-p 0.9 \
264
+ --repeat-penalty 1.1 \
265
+ -n 1024 \
266
+ -p "<|im_start|>system\nEres YuuKi...<|im_end|>\n<|im_start|>user\nProve that √2 is irrational.<|im_end|>\n<|im_start|>assistant\n"
267
+ ```
268
+
269
+ <br>
270
+
271
+ ### Recommended Generation Parameters
272
+
273
+ | Parameter | Value |
274
+ |:----------|:-----:|
275
+ | Temperature | 0.6 |
276
+ | Top-p | 0.9 |
277
+ | Max new tokens | 1024–4096 |
278
+ | Repetition penalty | 1.1 |
279
+
280
+ Lower temperature (0.3–0.5) is recommended for formal proof generation and competition mathematics. Higher temperature (0.7–0.8) produces more varied reasoning traces for exploratory use.
281
+
282
+ <br>
283
+
284
+ ---
285
+
286
+ <br>
287
+
288
+ <div align="center">
289
+
290
+ ## Training Details
291
+
292
+ </div>
293
+
294
+ <br>
295
+
296
+ <table>
297
+ <tr>
298
+ <td width="50%" valign="top">
299
+
300
+ **Hardware**
301
+
302
+ | Component | Specification |
303
+ |:----------|:-------------|
304
+ | GPU | NVIDIA A100 40GB SXM4 |
305
+ | Precision | BF16 native |
306
+ | Framework | Unsloth 2026.4 + TRL |
307
+ | Flash Attention | Xformers fallback |
308
+ | Cloud Compute | Colab A100 |
309
+
310
+ </td>
311
+ <td width="50%" valign="top">
312
+
313
+ **LoRA Configuration**
314
+
315
+ | Parameter | Value |
316
+ |:----------|:-----:|
317
+ | Rank (r) | 16 |
318
+ | Alpha | 32 |
319
+ | Dropout | 0.0 |
320
+ | Target Modules | q, k, v, o, gate, up, down |
321
+ | Trainable Parameters | ~83M |
322
+ | Gradient Checkpointing | Unsloth smart offload |
323
+
324
+ </td>
325
+ </tr>
326
+ </table>
327
+
328
+ <br>
329
+
330
+ **Optimizer Configuration**
331
+
332
+ | Parameter | Value |
333
+ |:----------|:-----:|
334
+ | Optimizer | AdamW 8-bit |
335
+ | Learning Rate | 2e-4 |
336
+ | LR Scheduler | Cosine |
337
+ | Warmup Steps | 100 |
338
+ | Weight Decay | 0.01 |
339
+ | Effective Batch Size | 16 |
340
+ | Max Sequence Length | 4,096 tokens |
341
+
342
+ <br>
343
+
344
+ ### Training Curriculum
345
+
346
+ YuuKi RxG was trained using the same three-phase curriculum architecture established across the OpceanAI model families, adapted for a reasoning-first base model.
347
+
348
+ <br>
349
+
350
+ <table>
351
+ <tr>
352
+ <td width="33%" valign="top">
353
+
354
+ **Phase 1 — Identity**
355
+ 3 epochs
356
+
357
+ | Source | Ratio |
358
+ |:-------|:-----:|
359
+ | Yuuki dataset | 65% |
360
+ | Reasoning pairs | 20% |
361
+ | Math instruction | 10% |
362
+ | General alignment | 5% |
363
+
364
+ *Establish YuuKi identity over DeepSeek-R1 base without degrading reasoning capability.*
365
+
366
+ </td>
367
+ <td width="33%" valign="top">
368
+
369
+ **Phase 2 — Reasoning**
370
+ 2 epochs
371
+
372
+ | Source | Ratio |
373
+ |:-------|:-----:|
374
+ | Yuuki dataset | 40% |
375
+ | Reasoning pairs | 30% |
376
+ | Math instruction | 20% |
377
+ | General alignment | 10% |
378
+
379
+ *Reinforce structured chain-of-thought and competition-level mathematical reasoning.*
380
+
381
+ </td>
382
+ <td width="33%" valign="top">
383
+
384
+ **Phase 3 — Consolidation**
385
+ 2 epochs
386
+
387
+ | Source | Ratio |
388
+ |:-------|:-----:|
389
+ | Yuuki dataset | 80% |
390
+ | Reasoning pairs | 10% |
391
+ | Math instruction | 10% |
392
+ | General alignment | 0% |
393
+
394
+ *Consolidate behavioral consistency and prevent capability regression.*
395
+
396
+ </td>
397
+ </tr>
398
+ </table>
399
+
400
+ <br>
401
+
402
+ ---
403
+
404
+ <br>
405
+
406
+ <div align="center">
407
+
408
+ ## Available Files
409
+
410
+ </div>
411
+
412
+ <br>
413
+
414
+ | File | Format | Description |
415
+ |:-----|:------:|:------------|
416
+ | `model.safetensors` | BF16 merged | Full precision weights, LoRA merged into base |
417
+ | `yuuki-rxg-8b.Q8_0.gguf` | GGUF Q8\_0 | Quantized for llama.cpp and Ollama |
418
+
419
+ <br>
420
+
421
+ ---
422
+
423
+ <br>
424
+
425
+ <div align="center">
426
+
427
+ ## Limitations
428
+
429
+ </div>
430
+
431
+ <br>
432
+
433
+ - **GPQA Diamond gap.** RxG scores 64.0% on GPQA Diamond, below Gemini-2.5-Flash-Thinking (82.8%) and o3-mini (76.8%). This benchmark tests graduate-level science reasoning across physics, chemistry, and biology — domains underrepresented in the Yuuki training dataset. This is a known gap and a target for the RxG 14B release.
434
+ - **LiveCodeBench.** Code generation at 62.0% is competitive but not leading at this scale. RxG is not primarily a coding model; this capability is inherited from the DeepSeek-R1 base.
435
+ - **Context utilization.** While the model supports 32,768 tokens, fine-tuning was conducted at 4,096 tokens. Performance on tasks requiring full context utilization beyond 4,096 tokens has not been formally evaluated.
436
+ - **Safety alignment** has not been formally evaluated under adversarial conditions. Not recommended for high-stakes or safety-critical deployment without additional review.
437
+
438
+ <br>
439
+
440
+ ---
441
+
442
+ <br>
443
+
444
+ <div align="center">
445
+
446
+ ## The RxG Family
447
+
448
+ </div>
449
+
450
+ <br>
451
+
452
+ RxG is the reasoning-specialized lineage within the OpceanAI ecosystem. Each release targets a specific parameter regime and capability tier.
453
+
454
+ | Model | Parameters | Status | Primary Target |
455
+ |:------|:----------:|:------:|:---------------|
456
+ | YuuKi RxG Nano | 1.5B | In development | Edge deployment, reasoning baseline |
457
+ | YuuKi RxG 8B | 8B | Released | General reasoning, competition math |
458
+ | YuuKi RxG VL 27B | 27B | Planned | Multimodal reasoning, flagship |
459
+
460
+ <br>
461
+
462
+ ---
463
+
464
+ <br>
465
+
466
+ <div align="center">
467
+
468
+ ## OpceanAI Ecosystem
469
+
470
+ </div>
471
+
472
+ <br>
473
+
474
+ | Model | Family | Parameters | Description |
475
+ |:------|:------:|:----------:|:------------|
476
+ | [YuuKi RxG 8B](https://huggingface.co/OpceanAI/Yuuki-RxG) | RxG | 8B | Reasoning flagship, TruthfulQA 96.6% |
477
+ | [Yumo Nano](https://huggingface.co/OpceanAI/yumo-nano) | Yumo | 1.5B | Math specialist, surpasses DeepScaleR |
478
+ | [YuuKi NxG VL](https://huggingface.co/OpceanAI/Yuuki-NxG-VL) | NxG | 7B | General conversation + vision |
479
+
480
+ <br>
481
+
482
+ ---
483
+
484
+ <br>
485
+
486
+ <div align="center">
487
+
488
+ ## Links
489
+
490
+ </div>
491
+
492
+ <br>
493
+
494
+ <div align="center">
495
+
496
+ [![Model Weights](https://img.shields.io/badge/Model_Weights-Hugging_Face-ffd21e?style=for-the-badge&logo=huggingface&logoColor=black)](https://huggingface.co/OpceanAI/Yuuki-RxG)
497
+ &nbsp;
498
+ [![GGUF Q8](https://img.shields.io/badge/GGUF_Q8-Download-1a1a2e?style=for-the-badge&logo=huggingface&logoColor=white)](https://huggingface.co/OpceanAI/Yuuki-RxG)
499
+ &nbsp;
500
+ [![OpceanAI](https://img.shields.io/badge/OpceanAI-Organization-1a1a2e?style=for-the-badge&logo=huggingface&logoColor=white)](https://huggingface.co/OpceanAI)
501
+
502
+ <br>
503
+
504
+ [![GitHub](https://img.shields.io/badge/GitHub-aguitauwu-181717?style=for-the-badge&logo=github&logoColor=white)](https://github.com/aguitauwu)
505
+ &nbsp;
506
+ [![Sponsor](https://img.shields.io/badge/Sponsor-GitHub_Sponsors-ea4aaa?style=for-the-badge&logo=githubsponsors&logoColor=white)](https://github.com/sponsors/aguitauwu)
507
+ &nbsp;
508
+ [![Discord](https://img.shields.io/badge/Discord-Community-5865F2?style=for-the-badge&logo=discord&logoColor=white)](https://discord.gg/j8zV2u8k)
509
+
510
+ </div>
511
+
512
+ <br>
513
+
514
+ ---
515
+
516
+ <br>
517
+
518
+ <div align="center">
519
+
520
+ ## Citation
521
+
522
+ </div>
523
+
524
+ <br>
525
+
526
+ ```bibtex
527
+ @misc{opceanai_yuuki_rxg_2026,
528
+ author = {awa_omg},
529
+ title = {YuuKi RxG — An 8B Reasoning Model with State-of-the-Art TruthfulQA},
530
+ year = {2026},
531
+ url = {https://huggingface.co/OpceanAI/Yuuki-RxG},
532
+ publisher = {Hugging Face}
533
+ }
534
+ ```
535
+
536
+ <br>
537
+
538
+ ---
539
+
540
+ <br>
541
+
542
+ <div align="center">
543
+
544
+ ## License
545
+
546
+ </div>
547
+
548
+ <br>
549
+
550
+ ```
551
+ Apache License 2.0
552
+
553
+ Copyright (c) 2026 OpceanAI
554
+
555
+ Licensed under the Apache License, Version 2.0 (the "License");
556
+ you may not use this file except in compliance with the License.
557
+ You may obtain a copy of the License at
558
+
559
+ http://www.apache.org/licenses/LICENSE-2.0
560
+
561
+ Unless required by applicable law or agreed to in writing, software
562
+ distributed under the License is distributed on an "AS IS" BASIS,
563
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
564
+ See the License for the specific language governing permissions and
565
+ limitations under the License.
566
+ ```
567
+
568
+ Inherits license terms from [DeepSeek-R1-Distill-Qwen-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-8B).
569
+
570
+ <br>
571
+
572
+ ---
573
+
574
+ <br>
575
+
576
+ <div align="center">
577
+
578
+ ## Updates
579
+
580
+ </div>
581
+
582
+ <br>
583
+
584
+ | Date | Milestone |
585
+ |:-----|:----------|
586
+ | **2026-04-09** | TruthfulQA 96.6% independently verified across three evaluation runs |
587
+ | **2026-04-09** | AIME 2024: 87.3% — surpasses DeepSeek-R1-8B |
588
+ | **2026-04-09** | GGUF Q8\_0 export available |
589
+ | **2026-04-09** | YuuKi RxG 8B v1.0 released on Hugging Face |
590
+
591
+ **Last updated:** 2026-04-09
592
+
593
+ <br>
594
+
595
+ ---
596
+
597
+ <br>
598
+
599
+ <div align="center">
600
+
601
+ **8B parameters. The most capable model OpceanAI has released.**<br>
602
+ **Surpasses its base model. Competitive with systems an order of magnitude larger.**
603
+
604
+ <br>
605
+
606
+ [![OpceanAI](https://img.shields.io/badge/OpceanAI-2026-0D1117?style=for-the-badge)](https://huggingface.co/OpceanAI)
607
+
608
+ <br>
609
+
610
+ *The RxG family. More releases coming.*
611
+
612
+ </div>