Published ResearchAI/ML

Supernova

A 650M parameter large language model built from scratch - architecture, tokenizer, and training - proving that efficiency can rival scale.

Supernova Logo

Role

Architecture & Training

Technology

PyTorch, Amazon SageMaker AI

Timeline

6 months

Status

Published on arXiv

The Challenge

The AI industry has a scaling obsession. Bigger models, more data, exponentially more compute. But we asked a different question: what if you could achieve comparable results with a fraction of the resources?

Supernova began as an academic research project at Ovidius University. The goal wasn't to compete with trillion-parameter giants - it was to prove that thoughtful engineering could close the gap. That architectural innovation could substitute for brute-force scaling.

We designed every component from scratch: a custom tokenizer optimized for compression efficiency, an architecture combining the latest advances (Rotary Position Embeddings, Grouped Query Attention, SwiGLU), and a training methodology focused on data quality over quantity.

The result? A model that achieves 90.29% of the performance of leading 1B parameter models - with 35% fewer parameters and trained on 100× less data.

Key Innovations

Custom Architecture

Decoder-only transformer with RoPE, Grouped Query Attention, RMSNorm, and SwiGLU - modern components working in synergy.

State-of-the-Art Tokenizer

128K vocabulary byte-level BPE tokenizer achieving 4.78 characters per token - outperforming GPT-4 and LLaMA tokenizers.

Exceptional Efficiency

Trained on 100B tokens - up to 360× less data than comparable models - proving that quality beats quantity.

Zero-Shot Performance

Benchmark Comparison

Results across ten standard zero-shot benchmarks, comparing Supernova with similarly sized open models.

BenchmarkSupernovaQwen3-0.6BLlama 3.2 1BGemma 3 1BOpenELM 1.1B
HellaSwag48.1847.3163.5662.0664.81
WinoGrande54.0655.4159.8359.0461.72
ARC-E60.9860.4065.3671.8962.37
ARC-C32.4234.0436.2638.1432.34
PIQA71.3867.6374.5974.6575.57
SuperGLUE56.1352.1455.5057.6057.30
MMLU26.7340.2436.9325.1125.52
MMLU-PRO10.3126.4910.908.999.48
SIQA43.4439.2542.7842.9442.84
BBH27.3340.4931.5927.3116.85
Average43.0946.3447.7346.7744.88

Supernova reaches 90.29% of Llama 3.2 1B average performance with 35% fewer parameters. Best score per benchmark is bolded.

The Architecture

Supernova combines modern transformer innovations that work synergistically. Rotary Position Embeddings for efficient position encoding. Grouped Query Attention with 3:1 compression to reduce memory bandwidth. RMSNorm for faster normalization. SwiGLU activations for improved gradient flow.

Each component was chosen not just for individual merit, but for how they amplify each other's benefits. The result is a model that punches well above its weight class.

Model Specifications

Transformer Blocks16 layers
Attention Heads12 heads
Embedding Dimension1,536
Context Length2,048 tokens
Vocabulary Size128,000 tokens
GQA Compression3:1 ratio

The Training

Training was conducted on a cluster of 8 NVIDIA A100 GPUs, sponsored by AWS for academic research. We achieved 54% Model FLOPs Utilization - indicating efficient use of available compute.

The training corpus was carefully curated from Nemotron-CC: 100B tokens combining high-quality web data, synthetic question-answer pairs, and distilled content from larger models. Every partition passed through rigorous filtering for quality, safety, and linguistic coherence.

Total training time: 350 hours. Total cost: under $10,000 - compared to the millions typically required for models in this performance range. This efficiency wasn't accidental; it was the entire point.

8× A100 40GB GPUs
100B tokens (Nemotron-CC)
54% MFU efficiency
By the numbers

The Results

650M
Parameters
100B
Training Tokens
90%
of 1B Model Performance
350h
Training Time

Why It Matters

Supernova isn't about building the biggest model. It's about proving that the race to scale isn't the only path forward.

When training costs drop from millions to thousands, AI becomes accessible. When inference is 40% cheaper, deployment becomes sustainable. When you need 100× less data, you can iterate faster.

The future of AI isn't just about what's possible - it's about what's practical. Supernova is our contribution to that future.

Building something with AI?

We bring research-grade expertise to production AI systems. Let's talk about your project.

Start a Conversation