AI Chip Startups — Probabilities, Anthropic's Bet, and How to Get In

Companion analysis to the candidate list. Focus: which silicon startups have a real shot, what Anthropic actually signed, and what it would take for you to land a role at one.
Honesty note up front. The "Anthropic signed with a startup" item — I do not have a verified, single-source-of-truth answer for which startup you're referring to. The big publicly confirmed Anthropic compute deals are AWS Trainium (Project Rainier, ~400K Trainium2 chips, multi-billion $ commitment) and Google TPU (the long-running multi-billion deal, expanded in 2025). Beyond those two, claims about Anthropic + a specific startup should be verified before you act on them. If you can tell me which startup you're thinking of (Groq? Cerebras? MatX? Etched? Positron?), I'll dig in specifically.

1. The landscape: who's actually trying to build AI chips

The AI silicon market splits into four buckets. Most "we'll beat NVIDIA" startups die in bucket 4. Probabilities below are rough subjective estimates of "this company is a meaningful business in 5 years (2031) generating >$1B revenue or being acquired for >$10B." Not "will they survive at all." Survival rates are higher; real success rates are what matter for equity outcomes.

CompanyBucketWhat they actually makeStage / ValP(meaningful in 5y)Why
NVIDIAIncumbentGPU + CUDA + NVLink + networkingPublic, ~$4T95%Software moat is the real moat. Question is whether they keep 80%+ share or drift to 50%.
Google TPUHyperscaler internalSystolic-array ASIC, tightly coupled to JAX/XLAInternal + Anthropic90%Already at scale. Anthropic deal proves external viability. Most credible NVIDIA alternative today.
AWS Trainium / InferentiaHyperscaler internalAnnapurna Labs ASIC familyInternal + Anthropic85%Project Rainier is real. AWS will burn money to make this work because they can't afford NVIDIA dependency.
CerebrasWafer-scale trainingWSE-3, single wafer = one chipIPO filed35%Engineering wins, business uncertain. G42 customer concentration is the big risk. Inference pivot is interesting but late.
GroqInference ASICLPU, deterministic dataflow~$6B35%Real revenue (GroqCloud), real latency wins on small models. Memory architecture limits big-model inference economics.
SambaNovaInference (enterprise)RDU + integrated stack~$5B (peak), declining15%Lost momentum. Enterprise-only story is small.
TenstorrentTraining+inference, openRISC-V + open ecosystem, Wormhole/Blackhole~$2.6B40%Best non-NVIDIA training story IMO. Jim Keller. Open stack is a real differentiator. Chinese demand creates real revenue path.
EtchedTransformer-only ASICSohu — transformer baked into silicon~$1.5B20%Binary outcome. If transformers stay dominant + Sohu actually ships at promised perf → 50x+. If architectures shift OR they miss yield → zero.
MatXLLM-only ASICDesigned for ≥70B param modelsSeries A (~$200M raised)25%Ex-Google TPU founders (Reiner Pope, Mike Gunter). Strong tech bench. Hardware is brutal at this stage.
Positron AIInference ASICAtlas — transformer inferenceSeries A15%Tiny, early. Genuine lottery ticket. Claims good perf/$/W.
Rain AIAnalog/in-memoryNeuromorphic computeSeries A8%Tech risk extreme. Sam Altman backed. Most analog AI plays die.
LightmatterPhotonic interconnectPassage interconnect, Envise compute (de-emphasized)~$4.4B35%Story shifted from compute to interconnect — interconnect is the right bet. Real customer interest.
Ayar LabsPhotonic I/OOptical chip-to-chip linksSeries D40%Less sexy, more important. Intel + NVIDIA-adjacent. Picks-and-shovels for the photonic-interconnect era.
d-MatrixInference chipletsCorsair, in-memory compute~$2B30%Real customers. Inference-focus is the right side of the market.
GraphcoreWas: training; now: SoftBankIPUAcquired by SoftBank 2024n/aCautionary tale. World-class team, lost to NVIDIA's CUDA moat. The default outcome for chip startups.
The base-rate truth: Of all "NVIDIA killer" startups since 2016 (Wave Computing, Nervana, Habana, Graphcore, Mythic, Esperanto, Cerebras early days, etc.), zero have become independent multi-hundred-billion-dollar companies. The realistic outcomes are: (a) acquihire, (b) acquired by hyperscaler / Intel / AMD for $1-5B, (c) IPO at modest val. Your "100x from current val" math only works if you join very early (Series A or earlier) at a company that ends up in the top 2-3 of survivors.

2. Why most AI chip startups fail

  1. CUDA moat — every model, every framework, every researcher's habits are in CUDA. To be picked over NVIDIA, you need to be 5-10x better on $/perf, not 2x.
  2. Architecture lock-in risk — Etched bets transformers stay. If the next architecture is something with non-attention mixing primitives, their chip is a paperweight. Specialized > general only when general is too slow; this can flip.
  3. Capital intensity — a tape-out at TSMC N3 is $500M+. You burn cash for 2-3 years before first silicon. Dilution is severe.
  4. Software is the product — chip companies that don't ship a credible compiler, kernels, distributed runtime, debugger, and profiler will lose. The hardware engineers founders hire are obvious; the compiler engineers they need are scarce.
  5. Customer concentration — most chip startups end up with 1-2 customers (Cerebras/G42, etc). One customer cancels = death.
  6. Hyperscalers eat the middle — AWS/Google/Meta will keep building internal chips. They take the easy wins. Independent startups have to find a niche neither hyperscalers nor NVIDIA serve.

3. What Anthropic actually uses (publicly known)

ProviderChipStatusSource / Notes
AWSTrainium2 (and Trainium3 plans)Project Rainier — ~400K+ Trainium2 chips for Anthropic trainingAnnounced Nov 2024 with the AWS investment expansion. Anthropic is the anchor customer making Trainium credible to the rest of the market.
GoogleTPU v5p / Trillium / next-genMulti-billion-dollar long-term deal, expanded 2025Anthropic has been a TPU heavy user from early days. Google has invested in Anthropic; deal sizes have grown publicly.
NVIDIAH100 / H200 / Blackwell (via cloud providers)Used, but not the strategic compute baseAnthropic's strategic bet is on TPU + Trainium because of supply, price, and not being captive to NVIDIA.

If the "startup" you remember is one of the big two above, the answer is they aren't startups — they're hyperscaler internal chip groups (AWS Annapurna, Google TPU). If it's a true startup, the most-rumored candidates worth investigating are MatX (founded by ex-Google TPU folks, has raised seriously, plausibly an Anthropic contender) or Etched. I'd want to verify before stating either as fact.

4. To work at one of these companies — what it takes

Roles that exist (and what each requires)

Role familyWho fitsWhat you'd need to add
ML Compiler / Kernel Engineer
XLA, Triton, MLIR, custom kernels
You — closest to your current ML eng skill setTriton kernels in CUDA, study MLIR/XLA, write 1-2 fused-attention kernels you can talk through, contribute to an OSS compiler stack.
ML Performance / Distributed Training
Megatron, FSDP, sequence parallelism, comms
You, with focused prepBuild & profile a multi-node training run. Understand all-reduce, ring vs tree, NVLink/RDMA. Read DeepSpeed, Megatron-LM, Pathways papers cold.
Inference Systems Engineer
vLLM, SGLang, batching, KV-cache, speculative decoding
You — strong fit if you've shipped inferenceLand a contribution to vLLM or similar. Understand PagedAttention, continuous batching, speculative/medusa. This is the hottest hiring area.
HW/SW co-design / Architecture
RTL, microarchitecture, perf modeling
Not you (without 2-3 yrs of pivot)EE-heavy. Skip unless you want to retrain.
Applied research (model-side)
Quantization, sparsity, MoE, sub-quadratic attn
You, with publication or strong demoOne solid paper or open-source release. Quantization is the most accessible angle (smoothquant, GPTQ, AWQ are tractable to extend).

Concrete prep plan (3-6 months) if this is the lane

  1. Pick a target stack — TPU/XLA, AWS Neuron, Triton/CUDA, or open (Tenstorrent's tt-metal). I'd pick Triton + vLLM as the most leveraged: applies to NVIDIA + many startups will support it.
  2. Ship a public artifact — a fused kernel, a vLLM patch, a quantization extension, a benchmark study. A blog post + GitHub repo. This is the single highest-leverage thing you can do.
  3. Read the canon — the FlashAttention papers (1, 2, 3), PagedAttention/vLLM, Megatron-LM, Pathways, the GPU MODE / GPU Glossary materials, Horace He's posts, Tri Dao's recent work.
  4. Map the people — for each top-3 startup, identify the 5 people most likely to be hiring managers (LinkedIn + papers + GitHub). Reach out with the artifact, not a generic intro.
  5. Pick a focused interview prep — these companies test (a) systems coding (C++/CUDA), (b) ML systems design (KV cache, batching, sharding), (c) sometimes architecture trivia (memory hierarchies). Less leetcode than FAANG, more depth.

How a $1M+ equity offer happens at a chip startup

5. Honest take on you doing this

Concentration risk warning. You already have Anthropic exposure via the SPV. Anthropic is heavily exposed to TPU + Trainium + (maybe) a startup like MatX. If you also work at one of those chip companies, your liquid net worth, career income, and lottery equity are all correlated to "Anthropic's compute provider succeeds." That's a bad portfolio. The hardware bet is most diversifying when it's a chip co whose customers are not Anthropic — e.g., Tenstorrent (Chinese + open ecosystem), Groq (consumer inference), d-Matrix (enterprise inference).

If you want my best single pick for "chip startup with non-trivial 100x odds + a role you could realistically land + uncorrelated to your Anthropic exposure": Tenstorrent. Open ecosystem means software contributions are a real entry path, Jim Keller is a credible bet, demand is real, val is still ~$2.6B (room to 10-30x at minimum if they win). Downside: Toronto/Santa Clara, no Seattle.

If you want maximum 100x optionality and don't mind 70% probability of zero: MatX or Etched. Smaller, earlier, ex-TPU pedigree (MatX), Cognition-like ASIC bet (Etched).

If you want highest probability of a real outcome with chip exposure but lower 100x: AWS Annapurna (technically a hyperscaler internal team) or Google TPU team. Comp is excellent, you'd actually do real work, but it's RSU not startup equity.

6. Open questions for you