"Chase the SOTA pipeline, not the MMLU slop."
i3 is a proof of concept for democratizing AI. Built solo by a 17-year-old researcher, demonstrating that efficient architectures matter more than massive compute clusters.
Parameters
Context
Architecture
Vision
The i3-200M (RWKV-Pro) is the result of "Chasing the Pipeline". By prioritizing training throughput on limited VRAM over raw benchmark chasing, we created a highly efficient hybrid stack.
Uses RWKV v4 Time-Mix and Channel-Mix blocks. This provides infinite-context capability (linear time) for the bulk of feature extraction.
4 Layers of Standard Multi-Head Attention (12 Heads) are placed at the top to capture complex global dependencies and reasoning capabilities.
A specialized Unconditional U-Net designed for high-fidelity citrus synthesis. Unlike massive text-to-image models that require H100s, Stable-Lime is optimized for a singular, perfect purpose: generating limes from the latent void.
Trainable on standard CPU threads without discrete GPU requirements. Democratizing generative art.
Pure unconditional latent diffusion process with 100 inference steps for maximum detail.