Table of Contents
01Abstract
WWWIII is a decentralized initiative to fund, build, and release the first large language model that is publicly funded, publicly developed, and publicly governed. Through an ERC-20 token on Ethereum, WWWIII creates a global coordination mechanism that allows anyone to contribute capital, compute, or expertise toward training a frontier AI model.
Unlike corporate-backed open-source releases — where a single company controls architecture, data, and release schedules — WWWIII places every decision in the hands of token holders. From architecture selection to dataset curation to compute allocation, governance is on-chain and transparent.
The resulting model will be released under Apache 2.0 with fully open weights, open training code, and publicly documented training runs. No strings attached. No corporate gatekeepers. The people's model.
02The Problem
The Concentration of AI Power
As of early 2026, fewer than ten organizations on Earth have successfully trained a frontier large language model. The barriers are immense: $100M+ in compute costs, access to thousands of high-end GPUs, teams of hundreds of specialized researchers, and massive proprietary datasets. The result is an oligopoly over the most transformative technology since electricity.
OpenAI, Anthropic, Google DeepMind, Meta AI, xAI, and a handful of Chinese labs control the trajectory of artificial intelligence. Their models power search engines, write code, generate media, and increasingly make decisions that affect billions. Yet none of these organizations answer to the public.
The Open-Source Illusion
Models like LLaMA, Mistral, and DeepSeek have demonstrated that open-weight models can match or rival closed systems. This is a genuine breakthrough. But "open weights" is not the same as "open development."
- Meta releases LLaMA — but Meta alone decides the architecture, training data, safety tuning, and release timeline. The community receives a finished product, not a seat at the table.
- Mistral publishes weights — but Mistral is a VC-funded startup with investors to satisfy and a business model to protect.
- DeepSeek opens its models — but DeepSeek operates under Chinese government oversight with its own constraints.
In every case, a single entity makes unilateral decisions about the most powerful technology in existence. Open weights are a gift. WWWIII proposes something fundamentally different: open ownership.
The Funding Gap
The cost of training frontier models has grown exponentially. GPT-4 reportedly cost over $100 million to train. Next-generation models may cost $1 billion or more. No crowdfunding platform, university, or nonprofit can match this scale. Traditional fundraising mechanisms are too slow, too limited in reach, and offer no governance to contributors.
Cryptocurrency solves this: global, permissionless, programmable capital that can be raised from anyone, anywhere, with governance built into the protocol.
03The WWWIII Solution
WWWIII is a token-funded, community-governed project to train and release a frontier large language model. The core thesis is simple:
If millions of people each contribute a small amount, we can collectively fund what only billion-dollar corporations can fund alone — and we can do it transparently, with shared governance and shared ownership of the result.
How It Works
- Fund — Contributors purchase or earn
$WWWIIItokens. Capital from the Development Fund is converted to compute and researcher grants. - Govern — Token holders vote on every major decision: model architecture, training data sources, compute providers, safety frameworks, and release strategy.
- Build — A distributed team of researchers, engineers, and contributors execute the community's decisions. All work is done in public. All code is open source.
- Release — The trained model is released under Apache 2.0 with fully open weights. Anyone can use, fine-tune, or deploy it. No restrictions. No API keys. No corporate approval.
Why This Works Now
- Decentralized GPU markets (Akash, io.net, Render Network) have made compute accessible without corporate cloud contracts
- Open training frameworks (Megatron-LM, DeepSpeed, FSDP) have democratized distributed training
- DAO tooling (Snapshot, Tally, Safe) has matured to handle complex governance
- Proof of concept exists — open models have already proven they can rival closed ones at smaller scales
04Token Economics
The $WWWIII token is an ERC-20 on Ethereum with a fixed supply of 1,000,000,000 tokens. There is no mint function after deployment. Supply is immutable.
| Allocation | Percentage | Tokens | Purpose |
|---|---|---|---|
| Development Fund | 40% | 400,000,000 | Compute, training, infrastructure, researcher grants |
| Community | 30% | 300,000,000 | Airdrops, contributor rewards, governance incentives |
| Team & Advisors | 15% | 150,000,000 | Core team compensation (2-year vest, 6-month cliff) |
| Liquidity | 10% | 100,000,000 | DEX pools, market making, exchange listings |
| Reserve | 5% | 50,000,000 | Emergency fund, partnerships, unforeseen needs |
Token Utility
The $WWWIII token is not speculative — it is a coordination and governance mechanism:
- Governance voting — 1 token = 1 vote on all proposals (architecture, data, compute, releases)
- Contributor rewards — Developers, researchers, data labelers, and reviewers earn tokens for verified contributions
- API access — Token holders receive priority and subsidized access to the model's inference API
- Staking — Future staking mechanism for compute providers who contribute GPU resources
Deflationary Mechanism
The contract includes an optional burn() function. Any holder can permanently destroy their tokens, reducing total supply. The community may vote to implement periodic burns from API revenue or other mechanisms to create deflationary pressure over time.
Vesting Schedule
Team and advisor tokens (15%) are subject to a 2-year vesting period with a 6-month cliff. No team tokens can be accessed for the first 6 months. After the cliff, tokens vest linearly over the remaining 18 months. This ensures the team is aligned with long-term project success.
05Governance Model
WWWIII operates as a Decentralized Autonomous Organization (DAO) where token holders govern all major decisions. Governance is designed to be transparent, inclusive, and resistant to capture by any single entity.
Decision Categories
| Category | Examples | Quorum |
|---|---|---|
| Architecture | Model size, attention mechanism, context length | 10% of supply |
| Data | Training datasets, filtering criteria, language mix | 10% of supply |
| Compute | GPU provider selection, budget allocation | 15% of supply |
| Treasury | Grant disbursement, partnership funding | 20% of supply |
| Safety | Alignment approach, red-teaming, release gates | 15% of supply |
| Protocol | Token mechanics, governance rules changes | 25% of supply |
Proposal Process
- Discussion — Anyone can post a proposal to the governance forum for community feedback (7 days minimum)
- Formal Proposal — Proposals that gain sufficient support are submitted on-chain via Snapshot or Tally
- Voting Period — 5-day voting window. 1 token = 1 vote. Simple majority wins unless otherwise specified
- Execution — Approved proposals are executed by the core team or automatically via smart contract
Advisory Council
A rotating advisory council of 7 members — elected by token holders every 6 months — provides technical guidance on architecture and training decisions. Council members are compensated from the Community allocation and serve as subject-matter experts, not decision-makers. All final decisions rest with token holders.
06Technical Architecture
The initial model target is a decoder-only transformer with 70B+ parameters, trained on a curated multilingual dataset. Final architecture decisions will be voted on by token holders, but the following serves as the baseline proposal.
Baseline Model Specification
| Parameter | Target |
|---|---|
| Architecture | Decoder-only transformer (GPT-style) |
| Parameters | 70B (initial), 200B+ (stretch goal) |
| Context Length | 128K tokens |
| Vocabulary | 128K tokens (BPE, multilingual) |
| Attention | Grouped Query Attention (GQA) |
| Position Encoding | RoPE (Rotary Position Embeddings) |
| Normalization | RMSNorm (pre-normalization) |
| Activation | SwiGLU |
| Precision | BF16 training, INT8/INT4 quantized inference |
| Training Framework | Megatron-LM + DeepSpeed ZeRO Stage 3 |
Training Data
All training data will be sourced from publicly available, ethically curated datasets. The community will vote on inclusion criteria. Proposed sources:
- CommonCrawl — web text, filtered and deduplicated
- Wikipedia — 300+ languages
- ArXiv — scientific papers
- GitHub — permissively licensed code (MIT, Apache, BSD)
- Stack Exchange — Q&A across technical domains
- Project Gutenberg — public domain books
- Community-contributed datasets — curated and reviewed by DAO members
Total training corpus target: 15+ trillion tokens after deduplication and quality filtering.
Compute Requirements
Training a 70B parameter model on 15T tokens requires approximately 500,000 GPU-hours on H100s (or equivalent). Compute will be sourced through a hybrid approach:
- Cloud providers — AWS, GCP, CoreWeave for guaranteed availability
- Decentralized GPU networks — Akash, io.net for cost-efficient overflow capacity
- University partnerships — academic compute grants and collaborations
- Direct GPU donations — contributors can donate compute cycles for token rewards
07Training Plan
Training will be conducted in public with full transparency. Every metric, every decision, and every checkpoint will be accessible to the community.
Pre-Training
- Standard autoregressive next-token prediction on the full training corpus
- Cosine learning rate schedule with linear warmup
- Intermediate checkpoints released every 500B tokens processed
- Live dashboard displaying: loss curves, token throughput, gradient norms, compute costs
Post-Training Alignment
- SFT (Supervised Fine-Tuning) — instruction-following dataset curated by community contributors
- RLHF / DPO — community-driven preference data collection and reward model training
- Red-teaming — open bug bounty program for discovering safety issues, bias, and failure modes
- Evaluation — standard benchmarks (MMLU, HumanEval, GSM8K, etc.) plus community-designed evaluations
Transparency Commitments
- All training code open-sourced on GitHub from day one
- Real-time training dashboard accessible to all token holders
- Intermediate checkpoints released as open weights
- All compute invoices and treasury spending published on-chain
- Weekly public updates from the core research team
- Monthly community calls with live Q&A
08Security & Auditing
Smart Contract Security
- Built on OpenZeppelin audited contract libraries (ERC-20, Ownable)
- No mint function — supply is immutable after deployment
- Team tokens locked behind on-chain vesting with cliff enforcement
- Treasury managed through Gnosis Safe multisig wallets (3-of-5 minimum)
- Third-party audit to be completed before mainnet deployment
Operational Security
- All core infrastructure protected by hardware security keys
- Multi-signature required for treasury transactions above threshold
- Transparent on-chain accounting — every outflow is publicly traceable
- Regular security reviews by independent auditors
AI Safety
WWWIII takes AI safety seriously. The governance structure ensures that safety decisions are made collectively, not by a single company:
- Pre-release safety evaluation — model must pass community-defined safety benchmarks before public release
- Bug bounty program — ongoing rewards for discovering safety issues, jailbreaks, and harmful capabilities
- Staged release — model released first to researchers, then to token holders, then publicly
- Kill switch proposal — if critical safety issues are discovered, an emergency governance vote can pause API access
09Roadmap & Milestones
| Phase | Timeline | Key Milestones |
|---|---|---|
| 1 — Foundation | Q1-Q2 2026 | Token launch, community building, DAO framework, whitepaper publication, Uniswap listing |
| 2 — Architecture | Q3-Q4 2026 | Hire core research team, architecture vote, data pipeline, compute partnerships, contributor program |
| 3 — Training | Q1-Q3 2027 | Pre-training run, live dashboard, intermediate checkpoints, RLHF alignment, safety evaluation |
| 4 — Release | Q4 2027 | Full model release (Apache 2.0), open API, fine-tuning grants, next-gen planning |
10Risk Factors
WWWIII is an ambitious project and participants should understand the risks involved:
- Funding risk — insufficient capital raised may require scaling down model size or extending timelines
- Technical risk — training frontier models is complex; unexpected failures, hardware issues, or suboptimal architectures may require restarts
- Regulatory risk — cryptocurrency regulations vary by jurisdiction and may affect token trading or fundraising
- Governance risk — decentralized governance can be slow and may face challenges with voter apathy or whale dominance
- Competitive risk — corporate labs have significant head starts and may release models that reduce demand for a community alternative
- Market risk — token value may fluctuate significantly based on market conditions unrelated to project progress
Despite these risks, WWWIII represents a fundamentally new approach to AI development — one that prioritizes transparency, collective ownership, and public benefit over corporate profit. The risks are real, but so is the opportunity to change how the world's most powerful technology is built.
The best way to predict the future is to fund it, govern it, and build it together.