Status
scripts/forge_model.py v3 is committed and ready. Memory-tiered architecture:
- Tier A (4B fp16): ~11GB VRAM, 21GB headroom
- Tier B (9B fp16): ~21GB VRAM, 11GB headroom
- Tier C (27B 4-bit): ~18GB VRAM, 14GB headroom
What's Done
What's Left
Usage
python scripts/forge_model.py Qwen/Qwen3.5-4B --domain general
python scripts/forge_model.py Qwen/Qwen3.5-9B --domain general
python scripts/forge_model.py Qwen/Qwen3.5-27B --domain general # auto 4-bit
python publish_forged.py output/forged/qwen3.5-27b/ --domain general
This is temporary
Python is the fast path to published models. The real pipeline will be Rust/Candle — see #88.
Dependencies
Status
scripts/forge_model.pyv3 is committed and ready. Memory-tiered architecture:What's Done
use_reentrant=Falsefor LoRA + gradient checkpointingWhat's Left
continuum-ai/qwen3.5-4b-general-forgedcontinuum-ai/qwen3.5-9b-general-forgedcontinuum-ai/qwen3.5-27b-general-forged(the headline)Usage
python scripts/forge_model.py Qwen/Qwen3.5-4B --domain general python scripts/forge_model.py Qwen/Qwen3.5-9B --domain general python scripts/forge_model.py Qwen/Qwen3.5-27B --domain general # auto 4-bit python publish_forged.py output/forged/qwen3.5-27b/ --domain generalThis is temporary
Python is the fast path to published models. The real pipeline will be Rust/Candle — see #88.
Dependencies