Agentic AI Cloud - Investor Summary_vDraft (2) (2)
Agentic AI Cloud - Investor Summary_vDraft (2) (2)
Agentic AI Cloud - Investor Summary_vDraft (2) (2)
Investment Summary
December 2024
Note: TAM represents AI Infrastructure including compute, inference services, tooling and infrastructure software, networking & storage.
Sources: IDC, Gartner, Morgan Stanley, Bain & Company
● Autonomous Systems
Decision Support
Total Latency ● Algorithmic Trading
Systems
● Emergency Response
Do not
NVIDIA HW support 405b
1 rack
API Services
haimaker's compute backbone,
federated globally
Lepton for AI Cloud Architecture
Integrated compute to runtime,
ready for agentic flows
● Hardware sales: Sell SambaNova AI Compute ● Rack Sales: $350k margin per node/rack sold
HW to data centers where a portion of all HW to data centers, with a target of 200 racks sold
sold gets contributed to our managed cloud for in year 1.
a revenue share. ● Inference: 10% revenue share on inference
● Agentic AI Cloud: Provide managed cloud services from data center-contributed capacity.
inference service, in a federated model, ● SaaS: $25k SaaS revenue per rack per year for
enabling data centers to quickly enter the AI
Lepton AI orchestration platform.
market and optimize per-rack profitability.
We have line of sight into more than $70M of contributed margin from hardware sales
in year 1
Investment Summary
An inference cloud purpose-built for Agentic AI, powered by SambaNova, the world’s
most capable AI hardware
Note: TAM represents AI Infrastructure including compute, inference services, tooling and infrastructure software, networking & storage.
Sources: IDC, Gartner, Morgan Stanley, Bain & Company
● Autonomous Systems
Decision Support
Total Latency ● Algorithmic Trading
Systems
● Emergency Response
1 rack
Note: TAM represents AI Infrastructure including compute, inference services, tooling and infrastructure software, networking & storage.
Sources: IDC, Gartner, Morgan Stanley, Bain & Company
An inference cloud purpose-built for Agentic AI, powered by SambaNova, the world’s
most capable AI hardware
We manage SambaNova’s
Cloud and cater to the Agentic
AI market
● HW Sales via Revenue Share: Sell ● Rack Sales: $350k margin per node/rack sold
SambaNova AI Compute HW to data centers to data centers, with a target of 200 racks sold
where a portion of all HW sold gets contributed in year 1.
to the AI Agentic Cloud inference platform. ● Inference: 10% revenue share on inference
● Agentic AI Cloud: Provide managed cloud services from data center-contributed capacity.
inference service, in a federated model, ● SaaS: $25k SaaS revenue per rack per year for
enabling data centers to quickly enter the AI
orchestration platform.
market and optimize per-rack profitability.
Notes:
An itemized breakdown of all uses of funds is currently under development.
Investment Summary
An inference cloud purpose-built for Agentic AI, powered by the world’s most capable
AI hardware
$120M Delivery of
Cash 100 racks
Principal,
Interest,
Colocation Fees Distributions
Asset Lender +
Data Center SPV
Colo Services
80/20 debt Equity Investors
equity $$
Asset Lender +
Data Center SPV
Colo Services
80/20 debt Equity Investors
equity $$
Notes:
1) An itemized breakdown of all uses of funds is currently under development.
Performance
Token Speeds
Serving Profiles
Built for Agentic AI
No Data Center
Buildout/Upgrade
● Operates at 11 kW for inference on a single Required
rack
● Host 100s of models per rack
● 5x more tokens per MW than NVIDIA 100
● Works in traditional CPU datacenters
Copyright © 2024 SambaNova Systems Inc. | Confidential & Proprietary
SN40L: SambaNova’s new CoE-optimized RDU
520 MB
102B Transistors
On-Chip Memory
12.8
TB/s
RDU High Bandwidth Memory [512 GB]
Super Low Latency Model Switching (Eg. <0.02sec for llama V2
7B)
800
GB/s
RDU High Capacity DDR Memory [12 TB]
4
Up to 5 Trillion Parameters! 3
10 Nodes Memory Required [DGX H100] 1 Node Memory Reduce Data Center
Required Footprint
Space, HW, Power
Simplified Hardware
Capacity Planning
Throughput + Memory
capacity → Throughput
Auto-Optimize for
Various Traffic Patterns
With less predictable
model traffic patterns,
system can self-optimize
Notes:
1) TAM represents AI Infrastructure Enablers including compute tooling and infrastructure software, networking & storage.
2) Sources: IDC, Gartner, Morgan Stanley, Bain & Company