| | --- |
| | language: |
| | - en |
| | license: openrail |
| | library_name: diffusers |
| | tags: |
| | - diffusion-llm |
| | - parallel-generation |
| | - custom-transformer |
| | - cropmark |
| | datasets: |
| | - OpenAssistant/oasst1 |
| | metrics: |
| | - cosine_similarity |
| | --- |
| | |
| | # πͺ DiffReaper-5 (Cropmark v2) |
| |
|
| | DiffReaper-5 is a **Conditioned Diffusion Large Language Model (DLLM)** designed for high-throughput, parallel conversational text generation. Unlike standard autoregressive models (GPT-style), DiffReaper-5 operates in the continuous latent embedding space, denoising an entire response sequence in parallel. |
| |
|
| | ## π¬ Model Details |
| |
|
| | - **Architecture:** Custom 12-layer Mercury-inspired Transformer. |
| | - **Task:** Conditioned Text Diffusion (Prompt-Response). |
| | - **Latent Space:** 1024-dimensional continuous embeddings. |
| | - **Training Objective:** Cosine Similarity Regression (Directional Loss). |
| | - **Sampling:** 10-step iterative parallel denoising. |
| |
|
| | ## π Autonomous Training State |
| |
|
| | This model is currently in **Autonomous Growth Mode**. It is training on an RTX 3090 cluster with the following parameters: |
| | - **Conditioning:** Hard-prompt conditioning (32 tokens). |
| | - **Generation Window:** 32 tokens (parallel). |
| | - **Optimizer:** AdamW with a learning rate of 1e-4. |
| | - **Sync:** Auto-checkpointing every 2,500 steps to this repository. |
| |
|
| | ## π οΈ Usage (Inference) |
| |
|
| | Unlike autoregressive models, DiffReaper-5 generates the entire response in parallel through iterative denoising. Use the following logic to run inference: |
| |
|
| | ```python |
| | import torch |
| | import torch.nn.functional as F |
| | # Assuming DiffReaperModel is defined as per train_autogrow.py |
| | |
| | def generate(model, tokenizer, prompt, steps=10): |
| | model.eval() |
| | with torch.no_grad(): |
| | p_tokens = tokenizer(prompt, return_tensors="pt").input_ids.to("cuda") |
| | p_emb = model.token_embedding(p_tokens[:, :32]) # Hard conditioning |
| | |
| | # Start from pure noise |
| | r_noise = torch.randn(1, 32, 1024).to("cuda") |
| | |
| | for i in range(steps): |
| | t = torch.tensor([1000 - (i * (1000//steps)) - 1], device="cuda").long() |
| | pred = model(torch.cat([p_emb, r_noise], dim=1), t) |
| | r_0_pred = pred[:, 32:, :] # Extract response |
| | r_noise = 0.4 * r_noise + 0.6 * r_0_pred # Iterative refinement |
| | |
| | # Map to vocab using Cosine Similarity |
| | norm_weights = F.normalize(model.token_embedding.weight, dim=-1) |
| | norm_r = F.normalize(r_noise, dim=-1) |
| | logits = torch.matmul(norm_r, norm_weights.T) |
| | return tokenizer.decode(torch.argmax(logits, dim=-1)[0]) |
| | |
| | # --- Loading Example --- |
| | # model = DiffReaperModel(vocab_size=50257, n_embd=1024, n_head=16, n_layer=12).to("cuda") |
| | # model.load_state_dict(torch.load("cropmark_latest.pt")) |
| | ``` |
| |
|
| | ## π― Fine-tuning |
| |
|
| | To fine-tune DiffReaper-5 on a custom dataset: |
| | 1. **Objective:** Use `1 - F.cosine_similarity` between predicted and target embeddings. |
| | 2. **Conditioning:** Ensure your data loader provides a fixed-length prompt prefix followed by the target response. |
| | 3. **Architecture:** Maintain the 1024-dimensional latent space to stay compatible with the weights. |
| |
|
| | ## π Diagnostic: Cropmark |
| |
|
| | The model's progress is monitored via the **Cropmark Diagnostic**. |
| | - **Cropmark** tests the model's ability to manifest a response (e.g., "I am good, how are you?") from pure Gaussian noise given a fixed prompt. |
| | - Results are logged in `checkpoint_log.txt` and uploaded periodically. |
| |
|
| | --- |
| | *Created by Darwin & Clawd.* |
| |
|