On April 15, Elon Musk announced that Tesla's AI5 chip — formerly codenamed Hardware 5 — has successfully taped out. Built on 3nm, it's dual-sourced from TSMC's Arizona facility and Samsung's Taylor, Texas plant. The specs are aggressive: 8-10x the compute of HW4, 192GB of LPDDR5X (a ninefold increase), and five times the memory bandwidth. Musk claims a single AI5 SoC is roughly equivalent to an Nvidia H100 for Tesla's specific inference workloads.
The "radical simplicity" framing is the interesting engineering philosophy here. Tesla is explicitly stripping traditional GPU components and image signal processors from the die, replacing them with silicon optimized solely for its System 2 neural networks. This is purpose-built inference hardware at the SoC level — not a general-purpose accelerator with Tesla's software on top. The tradeoff is inflexibility: AI5 will be useless for anything outside Tesla's training and inference stack, but that's the point.
The dual-sourcing decision is the supply chain signal worth watching. Splitting production between TSMC Arizona and Samsung Taylor is a direct hedge against Taiwan geopolitical risk — and a quiet acknowledgment that Tesla is planning volume at a scale where single-source risk is unacceptable. Both fabs are physically in the U.S., which aligns with the broader "sovereign silicon" thesis Tesla has been quietly executing.
What's not clear: Samsung's 3nm yield rates at Taylor have lagged TSMC's at every comparable node. If AI5 production ramps unevenly between the two sources, it could create quality and firmware consistency headaches at vehicle integration scale. Tesla has managed that problem before with HW3/HW4 variants — but AI5's higher compute footprint makes it a harder fallback scenario.