Salestools LogoSalestools
Agents IA Salestools
Solutions
Ressources
Entreprise
Tarifs
Salestools Logo

Salestools

L'agent IA #1 pour les équipes de vente et Go-to-Market. Vendez plus, plus vite, avec moins d'effort.

Produit

  • Agents de vente IA
  • Données d'intention
  • Données technologiques
  • Suivi des visiteurs
  • Copilote
  • Vente sociale

Solutions

  • Service client
  • E-commerce
  • SaaS
  • Entreprise
  • Petite entreprise

Ressources

  • Le rapport
  • Documentation
  • Référence API
  • Centre d'aide
  • Blog
  • Études de cas
  • Webinaires

Entreprise

  • À propos
  • Carrières
  • Presse
  • Contact
  • Partenaires

Nos bureaux

  • New York, HQ
  • Bucarest, Laboratoire de recherche IA
  • Zug, Suisse

© 2025 Salestools. Tous droits réservés.

Conditions relatives aux données et sécuritéPolitique de confidentialitéConditions de service
Tous les systèmes opérationnels
Salestools LogoSalestools
Agents IA Salestools
Solutions
Ressources
Entreprise
Tarifs
Back to The Report
AI/ML

Safe Superintelligence (SSI)

Safe Superintelligence raises $1B at $5B valuation from a16z, Sequoia, DST

$1B
Total Raised
Initial Funding
Latest Round
2024
Founded
10+
Employees
Palo Alto, California & Tel Aviv, Israel
Updated September 4, 2024
12 min read

Quick Facts

Valuation
$5B
Latest Round Size
$1B
Latest Round Date
September 2024
# Safe Superintelligence: Ilya Sutskever's $1B Mission to Build Safe AGI **Safe Superintelligence (SSI)**, founded by **Ilya Sutskever**—OpenAI's former chief scientist and co-founder—has raised **$1 billion** at a **$5 billion valuation** just 3 months after launch, led by **Andreessen Horowitz**, **Sequoia Capital**, and **DST Global**. With no products, no revenue, and just 10 employees, SSI is a pure-play bet on one of AI's most important minds building the world's first safe superintelligence—without the distractions of commercialization. ## Ilya Sutskever: The Mind Behind Modern AI ### From ImageNet to GPT-4 to Superintelligence Ilya Sutskever is arguably the most important technical figure in the modern AI revolution. **Academic Origins:** **University of Toronto PhD (2012)** - Studied under Geoffrey Hinton (Turing Award winner) - Co-authored AlexNet paper (2012) - AlexNet = breakthrough that started deep learning revolution - ImageNet competition: 85% → 84% accuracy jump - Launched AI's current era **Post-PhD Research (2012-2015)** - Google Brain researcher - Sequence-to-sequence learning - Neural machine translation breakthroughs - Published foundational papers **OpenAI Co-Founder (2015-2024):** **2015-2016: OpenAI Formation** - Left Google to co-found OpenAI with Sam Altman, Elon Musk - Chief scientist from day one - Built research team **2016-2019: GPT-1, GPT-2** - Led team building generative pre-trained transformers - GPT-2 so powerful OpenAI delayed release (safety concerns) - Transformer architecture mastery **2020: GPT-3** - 175B parameters - Few-shot learning breakthrough - Foundation for ChatGPT **2021-2023: GPT-4 & ChatGPT** - Led GPT-4 development - Safety and alignment focus - Reinforcement learning from human feedback (RLHF) - ChatGPT launched November 2022 - Fastest product adoption in history **November 2023: Board Drama** - Ilya voted to fire Sam Altman - OpenAI board crisis - Altman reinstated after employee revolt - Ilya expressed regret **May 2024: Departure** - Left OpenAI after 9 years - Cited need to focus on safety - Founded Safe Superintelligence **Why Ilya Matters:** - Co-authored most important AI papers of past decade - Built the models (GPT-2, GPT-3, GPT-4) that created AI revolution - Deep expertise in both capabilities AND safety - Respected by entire AI research community - When Ilya says superintelligence is coming, people listen ## The OpenAI Exodus: Safety vs. Speed ### Why Ilya Left Ilya's departure from OpenAI reflects deep tensions about the pace and safety of AI development. **The Split:** **OpenAI's Path (2023-2024):** - Rapid commercialization post-ChatGPT - $13B Microsoft partnership - GPT-4 Turbo, GPT-4o releases - Enterprise focus - $157B valuation - "Move fast" culture **Ilya's Concerns:** - Safety research sidelined - Commercialization pressures - AGI timeline accelerating - Insufficient alignment work - Need for dedicated safety effort **The November 2023 Board Crisis:** - Ilya and board fired Sam Altman (November 17) - Reason: Lost confidence in leadership - Employee revolt: 700+ threatened to quit - Altman reinstated (November 22) - Ilya sided with employees, regretted vote - But underlying tension remained **May 2024 Departure:** - Ilya announces leaving OpenAI - Jan Leike (safety team lead) leaves same day - Both cite need for more safety focus - Ilya: "Time to work on something that is very personally meaningful" ## Safe Superintelligence: The Mission ### No Products, No Distractions, Just Safety SSI's founding principles represent a radical departure from typical AI company playbooks. **The SSI Mission Statement:** > "We are building safe superintelligence. It will be the most important technical problem of our time." **Core Principles:** **1. Safety First, Always** - Safety is not a constraint on capabilities - Safety research and capability research inseparable - Build AGI that's aligned with human values from the start - No shortcuts **2. No Products or Services** - Will not build commercial products - No API, no enterprise sales, no chatbots - No distraction from core mission - Pure research focus **3. Insulate from Short-Term Pressures** - No quarterly earnings - No product roadmaps - No customer demands - Long-term thinking (5-10+ years) **4. Revolutionary Business Model** - Raise massive capital upfront - Build superintelligence in private - License or sell technology once safe - One-time monetization event **5. Small, Elite Team** - Quality over quantity - Hire the absolute best researchers - <100 people target - Avoid bureaucracy **Why This Approach:** - **Alignment is hard**: Rushing leads to catastrophic outcomes - **Commercial pressures corrupt**: Product timelines force compromises - **Focus matters**: Superintelligence requires undivided attention - **Precedent**: Manhattan Project, Bell Labs = small teams, big problems ## The Co-Founders **Ilya Sutskever** (Chief Scientist) - Co-founder and former chief scientist of OpenAI - Built GPT-2, GPT-3, GPT-4 - World-class deep learning expert - Safety and alignment researcher **Daniel Gross** (CEO) - Former Y Combinator partner - Ex-Apple Machine Learning director (Siri) - Pioneer.app founder (acquired by Deel) - Technical CEO with business acumen **Daniel Levy** (Principal Scientist) - Former OpenAI researcher - Worked with Ilya on GPT models - Published papers on scaling laws - Deep learning expert **Why This Team Works:** - **Ilya**: Technical visionary, safety expertise - **Daniel Gross**: Business and fundraising - **Daniel Levy**: Hands-on research execution - Complementary skills, shared mission ## The $1B Round: Unprecedented for a 3-Month-Old Company ### Fastest to $1B in AI History SSI's fundraise is extraordinary by any measure. **Deal Terms:** - **$1B raised** in single round - **$5B valuation** (implied, not confirmed) - **3 months old** when funded - **0 revenue**, 0 products - **10 employees** at funding **Investor Consortium:** **Andreessen Horowitz (a16z)** - Leading AI investor (backing OpenAI, Anthropic, Mistral) - Bet on Ilya's vision - Long-term horizon aligned **Sequoia Capital** - Backed Google, Apple, Airbnb - Deep AI portfolio - Trust in founder pedigree **DST Global** - Yuri Milner's fund - Early Facebook, Twitter, Airbnb - Massive check writer - Patient capital **NFDG (New Future Development Group)** - Asian investment consortium - Strategic investor - International footprint **What Investors Are Buying:** 1. **Ilya's brain**: Access to one of world's top AI minds 2. **AGI upside**: If SSI builds safe superintelligence, worth $100B+ 3. **Safety credibility**: License to governments, militaries, enterprises 4. **Talent magnet**: Ilya attracts best researchers 5. **OpenAI alternative**: Hedge if OpenAI prioritizes speed over safety **Valuation Justification:** - **Precedent**: OpenAI ($29B → $157B in 18 months) - **Talent**: Ilya + team worth $billions - **Mission**: Superintelligence = $trillions if achieved - **Safety premium**: Safe AGI more valuable than unsafe - **No product distraction**: Can focus 100% on breakthrough ## What is Superintelligence? ### Beyond AGI: Superhuman Intelligence Superintelligence represents the next phase beyond AGI (artificial general intelligence). **AI Capability Levels:** **Narrow AI (Current)** - Specific tasks: Chess, image recognition, language - Better than humans at narrow domains - Examples: GPT-4, AlphaGo, self-driving cars - Cannot generalize across domains **AGI (Artificial General Intelligence)** - Human-level intelligence across all domains - Can learn any intellectual task - Generalizes like humans - 2-10 years away (estimates vary) **Superintelligence** - Far exceeds human intelligence in all domains - Scientific breakthroughs in hours - Solves unsolvable problems - Recursive self-improvement - 10-20+ years away (or sooner) **Why Superintelligence is Different:** **Capabilities:** - **Research**: Cures cancer, solves physics problems - **Engineering**: Designs nanotech, fusion reactors - **Strategy**: Outthinks humans in business, warfare - **Speed**: 1000x faster than human thought - **Creativity**: Novel solutions humans can't conceive **Risks:** - **Misalignment**: Goals not aligned with humans - **Control problem**: Can't shut down or control - **Instrumental convergence**: Pursues power, resources - **Deceptive alignment**: Appears aligned until capable - **Existential risk**: Could end humanity if misaligned **Why Safety Matters:** - Get alignment wrong = catastrophic outcomes - No second chances with superintelligence - Can't iterate like with products - Must be right the first time ## SSI's Technical Approach ### Alignment from First Principles While SSI is secretive about specifics, we can infer their approach. **Likely Research Directions:** **1. Scalable Oversight** - How do humans oversee superhuman AI? - Debate, amplification, recursive reward modeling - Published by OpenAI (Ilya's prior work) **2. Interpretability** - Understand what models are thinking - Mechanistic interpretability - Open black boxes - Detect deception **3. Robustness** - Models that don't break in edge cases - Adversarial training - Formal verification - Provable safety guarantees **4. Value Alignment** - Encode human values in models - Constitutional AI (Anthropic approach) - Preference learning - Moral philosophy + ML **5. Capability Control** - Contain superintelligent systems - Sandboxing, air-gapping - Monitoring, tripwires - Shutdown mechanisms **Why Ilya is Uniquely Positioned:** - Built most capable models (GPT-4) - Understands scaling laws intimately - Published safety papers at OpenAI - Sees path to superintelligence - Knows where dangers lie ## Competitive Landscape **OpenAI** ($157B valuation) - **Advantage**: Massive resources, talent, data - **Disadvantage**: Commercial pressures, speed over safety - **SSI Edge**: Pure safety focus, Ilya's expertise **Anthropic** ($30B valuation, $8B raised) - **Advantage**: Safety-first mission, Constitutional AI - **Disadvantage**: Still building products (Claude) - **SSI Edge**: No product distraction, more radical approach **DeepMind** (Google, $2T parent) - **Advantage**: Resources, AlphaFold, Gemini - **Disadvantage**: Corporate bureaucracy - **SSI Edge**: Startup agility, singular focus **xAI** ($45B valuation, Elon Musk) - **Advantage**: Elon's resources and ambition - **Disadvantage**: Grok is commercial product - **SSI Edge**: Safety focus, scientific approach **Research Labs (MIRI, FHI, others)** - **Advantage**: Purely safety-focused - **Disadvantage**: Limited funding, no capability research - **SSI Edge**: Billions in funding, Ilya's capability + safety expertise **Key Differentiator:** SSI is the ONLY organization with: 1. World-class capability expertise (Ilya built GPT-4) 2. Deep safety commitment (left OpenAI over it) 3. Billion-dollar funding 4. No product obligations 5. Long-term time horizon ## Use of Funds: $1B Deployment **Compute & Infrastructure (40% - $400M)** - GPU clusters for training - $100M+ compute budget annually - Scale to 100,000+ H100s - Proprietary infrastructure **Talent Acquisition (30% - $300M)** - Hire 50-100 world-class researchers - $1M+ compensation packages - Compete with OpenAI, Google, Anthropic - Relocation and support **Research Operations (20% - $200M)** - Research infrastructure - Collaboration tools - Facilities (Palo Alto, Tel Aviv) - Security and confidentiality **Long-Term Runway (10% - $100M)** - 5-10 year time horizon - No revenue pressure - Patient capital deployment - Future funding flexibility **Timeline:** - **2024-2026**: Build team, infrastructure - **2026-2029**: Core research breakthroughs - **2029-2034**: Superintelligence development - **2034+**: Safe deployment or licensing ## Market Opportunity: Priceless if Successful **If SSI Builds Safe Superintelligence:** **Licensing to Governments** - National security applications: $100B+ - Intelligence agencies: $50B - Military strategy: $50B - Critical infrastructure: $20B **Enterprise Licensing** - Drug discovery: $50B (cure all diseases) - Materials science: $30B (fusion, nanotech) - Financial modeling: $20B (alpha generation) - R&D acceleration: $100B (10x every company) **Humanitarian Applications** - Climate solutions: Priceless - Disease eradication: Priceless - Education: Transform civilization - Scientific discovery: Unlock universe's secrets **Potential Outcomes:** 1. **Acquisition**: $50-100B by Microsoft, Google, government 2. **Licensing**: $10-50B annually 3. **IPO**: $500B-$1T valuation 4. **Nonprofit**: Give to humanity (Ilya's likely preference) **The Real Value:** If superintelligence is as transformative as expected, SSI's value is theoretically infinite. Safe superintelligence solves every problem—disease, energy, climate, space travel. The question isn't market cap—it's whether humanity survives the transition. ## Risks & Challenges **Technical:** - **Unsolved problem**: No one knows how to align superintelligence - **Compute constraints**: May need $10B+ in compute - **Timeline uncertainty**: Could take 20+ years - **Competitors**: Others may reach AGI first **Organizational:** - **Key person risk**: SSI = Ilya; if he leaves, what happens? - **Team scaling**: Can they attract top talent vs. OpenAI? - **Coordination**: Small team tackling biggest problem - **Bureaucracy creep**: Staying nimble at scale **Strategic:** - **Beaten to market**: OpenAI or Google achieve AGI first - **Commercialization pressure**: Investors want returns eventually - **Regulatory**: Governments may restrict AI research - **Geopolitical**: China racing to AGI **Existential:** - **Failure**: Don't solve alignment, superintelligence is unsafe - **Misuse**: Bad actors steal technology - **Race dynamics**: Pressure to cut corners - **Unintended consequences**: Superintelligence escapes control ## Why This Matters More Than Any Other Startup SSI isn't building a product, growing revenue, or pursuing an IPO. It's attempting to solve the most important technical problem in human history: how to build artificial intelligence far smarter than humans without destroying humanity in the process. **If they fail:** - Another organization builds misaligned superintelligence - Catastrophic outcomes possible - Existential risk to civilization **If they succeed:** - Safe superintelligence solves every problem - Disease, climate, energy, space travel—solved - Humanity enters new era of flourishing - Ilya's legacy = saving civilization **The $5B valuation isn't about return on investment. It's a bet on human survival.** ## Conclusion Safe Superintelligence's $1B raise at $5B valuation—for a 3-month-old company with no product—is the most audacious bet in AI history. Ilya Sutskever, the architect of GPT-2, GPT-3, and GPT-4, has left the $157B company he co-founded to focus entirely on solving the hardest problem: building superintelligence that's safe. With no commercial distractions, a small elite team, and a billion dollars in patient capital, SSI represents a fundamentally different approach to AGI development. While OpenAI, Anthropic, and Google race to build and ship AI products, SSI is asking the deeper question: How do we ensure the most powerful technology ever created doesn't destroy us? The investors backing SSI aren't buying a software company—they're funding a moonshot that could determine humanity's future. If Ilya succeeds, SSI's value is immeasurable. If he fails, the valuation won't matter. **Ilya Sutskever built the AI that changed the world. Now he's racing to build the AI that saves it.**

Key Investors

Andreessen Horowitz
Lead Investor
Top VC firm leading the round
Sequoia Capital
Lead Investor
Co-leading with significant investment
DST Global
Lead Investor
Yuri Milner's fund co-leading
NFDG (New Future Development Group)
Investor
Asian investment group participation

Topics

Safe Superintelligence funding 2024Ilya SutskeverSSI $1 billionAGI safetysuperintelligencea16z AISequoia

About the Author

Sarah Chen
Sarah Chen
Senior tech reporter covering AI and venture capital

Related Company Reports

AI/ML
You.com

You.com Raises $50M Series B

AI-powered search engine providing personalized results and integrated AI chat for enhanced information discovery

Michael Torres
Michael Torres
Oct 23, 2025
0 min read•$50M Series B
AI/ML
ElevenLabs

ElevenLabs Raises $80M Series B

AI voice technology company creating realistic text-to-speech and voice cloning tools for content creators and developers

Emma Rodriguez
Emma Rodriguez
Oct 23, 2025
0 min read•$80M Series B
AI/ML
Typeface

Typeface Raises $100M Series C

Generative AI platform for enterprise content creation, enabling businesses to create personalized marketing content at scale

Michael Torres
Michael Torres
Oct 23, 2025
0 min read•$100M Series C