Skip to content

Batch 2: Transformer + Scheduler specs (CLARA-critical) #490

@gHashTag

Description

@gHashTag

Task

Create 8 Transformer component and scheduler .tri specifications for CLARA TA1.

Why These 8 Specs

  • Transformer components are core to HSLM architecture
  • Schedulers required for training loops (cosine decay)
  • LayerNorm critical for stable training

Deadline: April 6, 2026 (DARPA CLARA proposal)

Specs to Create

File Description Priority
specs/algo/multi_head_attn.tri Multi-head attention with concat/projection P0
specs/algo/positional_enc.tri Sinusoidal position embeddings P0
specs/algo/feed_forward.tri FFN: Linear → ReLU → Linear P0
specs/algo/layernorm.tri Layer normalization P0
specs/algo/adamw.tri Adam with decoupled weight decay P0
specs/algo/sgd_momentum.tri SGD with Nesterov momentum P0
specs/algo/scheduler.tri Cosine decay scheduler P0
specs/algo/residual.tri Residual connection (Add & Norm) P0

Template

name: {name}
version: "1.0.0"
module: algo.transformer.{name}
description: "{description}"

types:
  {Name}Config:
    fields: [...]

functions:
  forward:
    params: [...]
    returns: {return_type}
    formula: |
      # Mathematical formula

behaviors:
  - name: {behavior_name}
    description: "Implementation details"

Dependencies

Acceptance

  • 8 .tri files created
  • tri gen produces stubs
  • Sacred constants verified

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions