File size: 14,253 Bytes
1269259
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
// SPF Smart Gateway - Transformer Encoder
// Copyright 2026 Joseph Stone - All Rights Reserved
//
// N stacked encoder layers: self-attention → layer_norm → FFN → layer_norm
// Sinusoidal positional encoding. Bidirectional attention (no causal mask).
// Input: token embeddings → Output: contextualized representations.
//
// Depends on: tensor.rs, attention.rs, ffn.rs (Layers 0-1)

use crate::tensor::Tensor;
use crate::attention::{AttentionCache, AttentionConfig, MultiHeadAttention};
use crate::ffn::{FfnCache, FfnConfig, FeedForward};

// ============================================================================
// ACTIVATION CACHE (for backward pass — P2-C)
// ============================================================================

/// Cached activations from one encoder layer forward pass
pub struct EncoderLayerCache {
    /// Input before LN1 (for layer_norm_backward)
    pub ln1_input: Tensor,
    /// Attention cache (Q, K, V, attn_weights, input, scale)
    pub attn_cache: AttentionCache,
    /// Input before LN2 (after attention residual)
    pub ln2_input: Tensor,
    /// FFN cache (input, hidden_pre_gelu)
    pub ffn_cache: FfnCache,
    /// Original layer input (for residual backward)
    pub residual_input: Tensor,
}

// ============================================================================
// ENCODER CONFIGURATION
// ============================================================================

/// Configuration for the encoder stack
#[derive(Debug, Clone)]
pub struct EncoderConfig {
    /// Number of encoder layers
    pub n_layers: usize,
    /// Model dimension
    pub d_model: usize,
    /// Number of attention heads
    pub n_heads: usize,
    /// Feed-forward hidden dimension (default: 4× d_model)
    pub d_ff: usize,
    /// Maximum sequence length (for positional encoding)
    pub max_seq_len: usize,
    /// Layer norm epsilon
    pub ln_eps: f32,
}

impl EncoderConfig {
    /// SPF Writer default: 6 layers, 256 dim, 8 heads
    pub fn spf_writer() -> Self {
        Self {
            n_layers: 6,
            d_model: 256,
            n_heads: 8,
            d_ff: 1024,
            max_seq_len: 2048,
            ln_eps: 1e-5,
        }
    }

    /// Smaller config for testing
    pub fn small() -> Self {
        Self {
            n_layers: 2,
            d_model: 64,
            n_heads: 4,
            d_ff: 256,
            max_seq_len: 512,
            ln_eps: 1e-5,
        }
    }
}

// ============================================================================
// SINGLE ENCODER LAYER
// ============================================================================

/// One encoder layer: self-attention + FFN, each with residual + layer norm.
///
/// Pre-norm architecture (norm before sublayer, used by modern transformers):
///   x → LayerNorm → SelfAttention → + residual → LayerNorm → FFN → + residual
pub struct EncoderLayer {
    /// Multi-head self-attention (bidirectional)
    pub self_attn: MultiHeadAttention,
    /// Feed-forward network
    pub ffn: FeedForward,
    /// Layer norm before attention
    pub ln1_gamma: Tensor,
    pub ln1_beta: Tensor,
    /// Layer norm before FFN
    pub ln2_gamma: Tensor,
    pub ln2_beta: Tensor,
    /// Epsilon for layer norm
    ln_eps: f32,
}

impl EncoderLayer {
    /// Initialize a single encoder layer
    pub fn new(d_model: usize, n_heads: usize, d_ff: usize, ln_eps: f32, seed: u64) -> Self {
        let attn_config = AttentionConfig {
            d_model,
            n_heads,
            causal: false, // Encoder uses bidirectional attention
        };
        let ffn_config = FfnConfig { d_model, d_ff };

        Self {
            self_attn: MultiHeadAttention::new(attn_config, seed),
            ffn: FeedForward::new(ffn_config, seed + 100),
            ln1_gamma: Tensor::ones(&[d_model]),
            ln1_beta: Tensor::zeros(&[d_model]),
            ln2_gamma: Tensor::ones(&[d_model]),
            ln2_beta: Tensor::zeros(&[d_model]),
            ln_eps,
        }
    }

    /// Forward pass: [batch, seq, d_model] → [batch, seq, d_model]
    pub fn forward(&self, x: &Tensor) -> Result<Tensor, String> {
        // Pre-norm self-attention with residual
        let normed = x.layer_norm(&self.ln1_gamma, &self.ln1_beta, self.ln_eps)?;
        let attn_out = self.self_attn.forward(&normed)?;
        let x = x.add(&attn_out)?;

        // Pre-norm FFN with residual
        let normed = x.layer_norm(&self.ln2_gamma, &self.ln2_beta, self.ln_eps)?;
        let ffn_out = self.ffn.forward(&normed)?;
        x.add(&ffn_out)
    }

    /// Forward pass with cached activations for backward.
    /// Output is IDENTICAL to forward(). Cache is additional data only.
    pub fn forward_with_cache(&self, x: &Tensor) -> Result<(Tensor, EncoderLayerCache), String> {
        let residual_input = x.clone();

        // Pre-norm self-attention with residual
        let ln1_input = x.clone();
        let normed = x.layer_norm(&self.ln1_gamma, &self.ln1_beta, self.ln_eps)?;
        let (attn_out, attn_cache) = self.self_attn.forward_with_cache(&normed)?;
        let x = x.add(&attn_out)?;

        // Pre-norm FFN with residual
        let ln2_input = x.clone();
        let normed = x.layer_norm(&self.ln2_gamma, &self.ln2_beta, self.ln_eps)?;
        let (ffn_out, ffn_cache) = self.ffn.forward_with_cache(&normed)?;
        let output = x.add(&ffn_out)?;

        let cache = EncoderLayerCache {
            ln1_input,
            attn_cache,
            ln2_input,
            ffn_cache,
            residual_input,
        };

        Ok((output, cache))
    }

    /// Total parameters in this layer
    pub fn num_params(&self) -> usize {
        let d = self.ln1_gamma.numel();
        self.self_attn.num_params() + self.ffn.num_params() + 4 * d // 2 LN × (gamma + beta)
    }

    /// Collect all weight tensors for serialization
    pub fn weights(&self) -> Vec<&Tensor> {
        let mut w = self.self_attn.weights();
        w.extend(self.ffn.weights());
        w.extend([&self.ln1_gamma, &self.ln1_beta, &self.ln2_gamma, &self.ln2_beta]);
        w
    }

    /// Collect all weight tensors mutably for optimizer updates
    pub fn weights_mut(&mut self) -> Vec<&mut Tensor> {
        let mut w = self.self_attn.weights_mut();
        w.extend(self.ffn.weights_mut());
        w.extend([&mut self.ln1_gamma, &mut self.ln1_beta, &mut self.ln2_gamma, &mut self.ln2_beta]);
        w
    }
}

// ============================================================================
// POSITIONAL ENCODING
// ============================================================================

/// Generate sinusoidal positional encoding table.
/// PE(pos, 2i) = sin(pos / 10000^(2i/d_model))
/// PE(pos, 2i+1) = cos(pos / 10000^(2i/d_model))
///
/// Returns: [max_seq_len, d_model] tensor
pub fn sinusoidal_positional_encoding(max_seq_len: usize, d_model: usize) -> Tensor {
    let mut data = vec![0.0f32; max_seq_len * d_model];
    for pos in 0..max_seq_len {
        for i in 0..d_model / 2 {
            let angle = pos as f32 / (10000.0_f32).powf(2.0 * i as f32 / d_model as f32);
            data[pos * d_model + 2 * i] = angle.sin();
            data[pos * d_model + 2 * i + 1] = angle.cos();
        }
        // Handle odd d_model
        if d_model % 2 == 1 {
            let angle = pos as f32 / (10000.0_f32).powf((d_model - 1) as f32 / d_model as f32);
            data[pos * d_model + d_model - 1] = angle.sin();
        }
    }
    Tensor { data, shape: vec![max_seq_len, d_model] }
}

// ============================================================================
// ENCODER STACK
// ============================================================================

/// Full encoder: positional encoding + N encoder layers + final layer norm.
pub struct Encoder {
    pub config: EncoderConfig,
    /// Positional encoding table [max_seq_len, d_model]
    pub pos_encoding: Tensor,
    /// Stack of encoder layers
    pub layers: Vec<EncoderLayer>,
    /// Final layer norm (applied after all layers)
    pub final_ln_gamma: Tensor,
    pub final_ln_beta: Tensor,
}

impl Encoder {
    /// Initialize encoder with given config
    pub fn new(config: EncoderConfig, seed: u64) -> Self {
        let pos_encoding = sinusoidal_positional_encoding(config.max_seq_len, config.d_model);

        let layers: Vec<EncoderLayer> = (0..config.n_layers)
            .map(|i| {
                EncoderLayer::new(
                    config.d_model,
                    config.n_heads,
                    config.d_ff,
                    config.ln_eps,
                    seed + (i as u64) * 1000,
                )
            })
            .collect();

        Self {
            final_ln_gamma: Tensor::ones(&[config.d_model]),
            final_ln_beta: Tensor::zeros(&[config.d_model]),
            pos_encoding,
            layers,
            config,
        }
    }

    /// Forward pass: embeddings [batch, seq_len, d_model] → encoded [batch, seq_len, d_model]
    ///
    /// 1. Add positional encoding to input embeddings
    /// 2. Pass through N encoder layers
    /// 3. Apply final layer norm
    pub fn forward(&self, embeddings: &Tensor) -> Result<Tensor, String> {
        let seq_len = embeddings.shape[1];
        if seq_len > self.config.max_seq_len {
            return Err(format!(
                "Sequence length {} exceeds max {}",
                seq_len, self.config.max_seq_len
            ));
        }

        // Slice positional encoding to actual sequence length
        let pos_enc = self.pos_encoding.slice(0, seq_len)?;

        // Add positional encoding (broadcasts across batch dimension)
        let batch = embeddings.shape[0];
        let d_model = embeddings.shape[2];
        let mut x_data = embeddings.data.clone();
        for b in 0..batch {
            for s in 0..seq_len {
                for d in 0..d_model {
                    x_data[(b * seq_len + s) * d_model + d] += pos_enc.data[s * d_model + d];
                }
            }
        }
        let mut x = Tensor::from_data(x_data, embeddings.shape.clone())?;

        // Pass through encoder layers
        for layer in &self.layers {
            x = layer.forward(&x)?;
        }

        // Final layer norm
        x.layer_norm(&self.final_ln_gamma, &self.final_ln_beta, self.config.ln_eps)
    }

    /// Total parameters in the encoder
    pub fn num_params(&self) -> usize {
        let layer_params: usize = self.layers.iter().map(|l| l.num_params()).sum();
        layer_params + 2 * self.config.d_model // final LN gamma + beta
    }

    /// Collect all weight tensors
    pub fn weights(&self) -> Vec<&Tensor> {
        let mut w: Vec<&Tensor> = Vec::new();
        for layer in &self.layers {
            w.extend(layer.weights());
        }
        w.push(&self.final_ln_gamma);
        w.push(&self.final_ln_beta);
        w
    }

    /// Collect all weight tensors mutably
    pub fn weights_mut(&mut self) -> Vec<&mut Tensor> {
        let mut w: Vec<&mut Tensor> = Vec::new();
        for layer in &mut self.layers {
            w.extend(layer.weights_mut());
        }
        w.push(&mut self.final_ln_gamma);
        w.push(&mut self.final_ln_beta);
        w
    }
}

// ============================================================================
// TESTS
// ============================================================================

#[cfg(test)]
mod tests {
    use super::*;

    #[test]
    fn test_positional_encoding_shape() {
        let pe = sinusoidal_positional_encoding(100, 64);
        assert_eq!(pe.shape, vec![100, 64]);
    }

    #[test]
    fn test_positional_encoding_values() {
        let pe = sinusoidal_positional_encoding(10, 8);
        // Position 0 should have sin(0)=0, cos(0)=1 for first pair
        assert!((pe.data[0] - 0.0).abs() < 1e-5); // sin(0)
        assert!((pe.data[1] - 1.0).abs() < 1e-5); // cos(0)
        // All values should be in [-1, 1]
        assert!(pe.data.iter().all(|&v| v >= -1.0 && v <= 1.0));
    }

    #[test]
    fn test_encoder_layer_shape() {
        let layer = EncoderLayer::new(64, 4, 256, 1e-5, 42);
        let x = Tensor::randn(&[2, 8, 64], 99);
        let out = layer.forward(&x).unwrap();
        assert_eq!(out.shape, vec![2, 8, 64]);
    }

    #[test]
    fn test_encoder_layer_residual() {
        let layer = EncoderLayer::new(64, 4, 256, 1e-5, 42);
        let x = Tensor::randn(&[1, 4, 64], 99);
        let out = layer.forward(&x).unwrap();
        // With residual connections, output should not be identical to input
        // but should be close in magnitude (not exploding)
        let diff: f32 = x.data.iter().zip(&out.data)
            .map(|(a, b)| (a - b).abs())
            .sum::<f32>() / x.numel() as f32;
        assert!(diff < 10.0, "Output diverged too far from input: {}", diff);
    }

    #[test]
    fn test_encoder_full_forward() {
        let config = EncoderConfig::small();
        let encoder = Encoder::new(config, 42);
        let x = Tensor::randn(&[1, 8, 64], 99);
        let out = encoder.forward(&x).unwrap();
        assert_eq!(out.shape, vec![1, 8, 64]);
        assert!(out.data.iter().all(|v| v.is_finite()));
    }

    #[test]
    fn test_encoder_seq_exceeds_max() {
        let config = EncoderConfig { max_seq_len: 10, ..EncoderConfig::small() };
        let encoder = Encoder::new(config, 42);
        let x = Tensor::randn(&[1, 20, 64], 99); // seq=20 > max=10
        assert!(encoder.forward(&x).is_err());
    }

    #[test]
    fn test_encoder_num_params() {
        let config = EncoderConfig::small(); // 2 layers, d=64, ff=256
        let encoder = Encoder::new(config, 42);
        let params = encoder.num_params();
        // Each layer: attn(16640) + ffn(33088) + 4×64(LN) = 49984
        // 2 layers + final LN = 2×49984 + 128 = 100096
        assert_eq!(params, 100096);
    }

    #[test]
    fn test_encoder_weights_collection() {
        let config = EncoderConfig::small();
        let encoder = Encoder::new(config, 42);
        let weights = encoder.weights();
        // Each layer: 8(attn) + 4(ffn) + 4(LN) = 16. ×2 layers + 2 final LN = 34
        assert_eq!(weights.len(), 34);
    }
}