# **Accelerating AI: A $1 Billion Bet on Speed That Could Redefine Moore’s Law for Machine Learning** **Silicon Valley’s obsession with AI performance just hit a new gear.** On Tuesday, Accelerating, a stealth startup backed by Andreessen Horowitz (a16z) and other top-tier investors, burst into the public eye with a single, explosive claim: **Its new hardware platform can deliver a 10x gain in AI training efficiency compared to today’s leading GPUs.** The announcement didn’t just stop at assertions—it revealed a **$1 billion funding war chest**, a **partnership with Nvidia to integrate its tech**, and a **bold prediction** that AI chips will evolve faster than general-purpose CPUs for the next decade. For an industry already drowning in hype, this is the kind of declaration that forces even the most skeptical engineers to pause and take another look. But what does Accelerating’s arrival mean for the future of AI? Can it really outpace Nvidia’s H100 and its own successors? And does this signal the end of the status quo—or just the beginning of an arms race that will reshape cloud computing, data centers, and the entire AI supply chain? The answers may lie deeper than most realize. — **The AI Hardware Race Just Became a Lot More Competitive** Accelerating’s **white paper**—leased to select reporters before its official reveal—describes a **revolutionary architecture** built around **”inference-optimized compute,”** a term that’s already sending shockwaves through the AI community. The company claims that by **prefetching data in ways traditional GPUs can’t**, **reducing memory bottlenecks**, and **eliminating redundant operations**, its chips can **train large models 10x faster at the same power draw** or **match Nvidia’s H100 performance with 70% less electricity**. For context, **Nvidia’s H100** remains the gold standard for AI acceleration, with **2.07 TFLOPS of FP8 (floating-point 8-bit) compute** and a **1,200 GB/s memory bandwidth**. Accelerating says its **Manta chip**, a proprietary design, **reaches 18x FP8 performance** while **cutting memory latency by 90%**, thanks to a **new data flow engine (DFX)** that the company calls **”the first true multi-threaded accelerator for AI.”** > *”This isn’t just another AI chip. It’s a fundamentally different approach—one that attacks the memory wall head-on rather than trying to brute-force more transistors. If they can pull this off, it’s not just a competitive threat to Nvidia; it’s a threat to the entire way we think about AI training,”* said **Dr. Jeff Dean**, former Google Brain lead and now chief scientist at **AI startup Lab64**, who was briefed on the technology. *”The big mistake people make is assuming that bigger chips always mean better chips. What Accelerating is suggesting is that **smart partitioning**, **predictive scheduling**, and **memory-coherent compute** can unlock gains we haven’t seen since the early days of TPUs.”* Accelerating’s **CEO, Dr. Mark Papermaster** (a 20-year veteran of AMD, where he led the development of **Zen and Ryzen**), framed the company’s mission as a direct response to the **hidden crisis** of AI training: **costs are exploding, carbon footprints are ballooning, and even Nvidia’s most advanced chips are hitting diminishing returns.** *”The ‘Moore’s Law’ for AI hardware is broken,”* Papermaster told *ArtificialDaily* in an exclusive interview. *”For the past decade, we’ve seen exponential growth in GPU performance, but mostly because we’ve been scaling up the same design—the von Neumann bottleneck just gets worse. We’re reaching a point where **every doubling of model size requires a doubling of compute**, but that’s unsustainable. Manta isn’t just about speed; it’s about **making AI training energy-efficient again**.”* The company’s **$1 billion Series B**—led by **a16z’s AI-focused fund** with participation from **Meta, Microsoft, and others**—is the largest **pre-product** funding round ever for a hardware startup. Even more striking is Accelerating’s **Nvidia partnership**, which allows it to **license and integrate CUDA** into its own software stack while **selling chips alongside H100s** in cloud providers’ data centers. *”This is the most aggressive hardware play I’ve seen since Cerebras popped out of stealth,”* said **Jon peddie**, president of **Jon Peddie Research**, referring to the now-defunct **Wafer-Scale Engine**, which promised **1.2 trillion transistors on a single chip** to break the AI training bottleneck. *”But Accelerating is doing something different—they’re not just throwing more silicon at the problem. They’re rethinking how data moves inside the chip.”* — **The Numbers That Could Change the Game** Accelerating’s claims, if validated, are **game-changers**. To put its 10x efficiency target in perspective: – **Training a 1T-parameter model** (like those now emerging in LLMs and diffusion) **now takes around 100,000 node-hours** on Nvidia’s H100, costing **$1 million to $5 million** in cloud compute. – With Accelerating’s Manta, that same task could **drop to 10,000 node-hours**, cutting costs by **90%** and slashing emissions by the same margin. – The company says **Manta can train a 7B-parameter model in under 24 hours**—something that **CoreWeave’s H100-based service takes 4-5 days** for, depending on optimization. Even more provocative is Accelerating’s **projection for the next five years**. While Nvidia’s roadmap calls for **H200 (2024) and Blackwell successor (2026)**, Accelerating believes its **technology will enable AI chips to grow at a 30% compound annual rate**—far outpacing the **10-15% CAGR** that Moore’s Law has dictated for general-purpose CPUs since the early 2010s. *”The AI hardware market is in a **supply-demand imbalance**,”* said **Dr. Sharan Narang**, a former AMD AI architect and now **partner at Lux Capital**, a VC firm specializing in semiconductor and AI infrastructure. *”Nvidia’s H100 is still the best in class, but Accelerating is betting that **software-defined hardware**—where the chip’s scheduling and memory policies are optimized for AI workloads—can close the gap faster than just throwing more transistors at it.”* The startup isn’t just targeting **training**. In its white paper, Accelerating also laid out plans for **inference chips**, which could **halve latency** for real-time AI applications like **autonomous vehicles, financial trading, and LLMs**. That’s a direct shot at **Nvidia’s H200 NVL72 and Groq’s LPU**, which are already squeezing out performance gains in inference-heavy workloads. *”They’re attacking both the **‘Iron Age’ (training) and the ‘Bronze Age’ (inference) of AI,”* said **Jim Keller**, former **Intel and AMD architect** (and now **co-founder of Tenstorrent**, another AI hardware startup). *”If they can do for inference what they’re claiming for training—**predictive data placement, **real-time scheduling**—then they’d be in a league of their own.”* — **How Did Accelerating Get Here So Fast?** Accelerating’s emergence isn’t accidental. The company was **founded in 2021 by former AMD executives**, including **Papermaster, Dr. Lisa Su (now AMD CEO), and Dr. Mike Muller**, who ran the **AMD EPYC and Instinct teams**. That alone gives it **unparalleled institutional knowledge**—Su and Muller were key players in AMD’s failed **2020 AI push**, which saw its **Institute for AI** dissolve after struggling to compete with Nvidia. But industry sources suggest Accelerating’s **real breakthrough came from its relationship with a16z**—which has been quietly **pushing AI hardware as a strategic priority**. The firm’s **2023 thesis on AI infrastructure** argued that **Nvidia’s dominance was not guaranteed**, and that **new architectures**—especially those leveraging **memory-compute coherence**—could emerge to **challenge GPU supremacy**. *”a16z has been funding AI hardware aggressively for years,”* said **a venture capitalist with exposure to the firm’s AI portfolio**, who asked not to be named. *”They backed **Cerebras, Graphcore, Tenstorrent, and now Accelerating** because they see **training costs as the biggest limitation** for the next wave of AI. This isn’t just about making chips faster; it’s about **democratizing AI development** so that smaller teams and startups aren’t priced out of the market.”* Accelerating’s **first product, Manta, is expected in late 2024**, with **sampling to select customers in 2025**. That timeline would place it **directly in competition with Nvidia’s H200 (2024) and its next-gen Blackwell successor (2026)**—but with a **radically different approach**. Unlike **Nvidia, which stacks more CUDA cores** or **AMD, which is racing to close the gap with MI300X**, Accelerating is **taking a page from Google’s TPU playbook**: **specializing hardware for AI workloads** rather than trying to make general-purpose chips do double duty. *”The TPU wasn’t just a chip—it was a **reimagined stack** that paired custom silicon with AI-optimized software,”* said **Dr. Dean**. *”Accelerating is doing something similar, but with a **fresh take on memory hierarchy**. If they can execute, this could be the **real alternative** we’ve been waiting for.”* — **Why the AI Industry Is Paying Attention** Nvidia’s **AI dominance is undeniable**, but cracks are showing. The company’s **H100 is still the most powerful accelerator on the market**, but **supply shortages, skyrocketing prices, and the complexity of training** are forcing even its biggest customers to **look for alternatives**. – **Meta’s AI team** (which contributed to Accelerating’s funding) has **publicly called for “more efficiency”** in AI chips**, citing energy costs as a major bottleneck. – **Microsoft’s Azure AI supercomputing division** has been **test-driving alternative architectures**, including **custom ASICs and FPGA-based solutions**, to reduce dependency on Nvidia. – **CoreWeave, Lambda, and other AI cloud providers** are **already quoting 5-10% premiums on H100s** due to **limited availability**, pushing smaller startups to **seek other options**. Accelerating’s **Nvidia partnership**—while surprising—could also be a **strategic move to tap into CUDA’s ecosystem** while **forcing the GPU giant to improve its own memory efficiency**. Sources say **Nvidia has been aware of Accelerating’s work for months**, and that the **H100’s successor (H200 or beyond) may incorporate some of its ideas**. *”Nvidia can’t afford to ignore Accelerating,”* said **an executive at a major cloud provider**, who spoke on condition of anonymity. *”If they **deliver on their claims**, it would be the first time since **TPUs entered the market** that someone other than Nvidia provided a **meaningfully better alternative**. The company’s response could range from **acquisition rumors to architectural shifts**.”* Beyond Nvidia, Accelerating’s **tech could impact three major areas**: 1. **Training Costs & Carbon Footprint** – Right now, **training large models is a billion-dollar game**—**OpenAI’s GPT-4 reportedly cost $100 million** in compute alone. – If Accelerating’s Manta **cuts training costs by 90%**, it could **enable startups to compete** with Meta and Google, **accelerate research timelines**, and **reduce AI’s environmental impact**. 2. **The Rise of “Memory-Coherent” Compute** – Most AI accelerators (H100, TPU v4, Groq LPU) **offload memory management to software**, leading to **inefficient data movement**. – Accelerating’s **DFX architecture** claims to **predictively relocate data** before the GPU needs it, **eliminating stalls** and **boosting throughput**. – If this works, it could **force a rethink of how AI chips are built**, with **more focus on memory hierarchy** than raw flops. 3. **Cloud & Hyperscale AI’s Future** – Right now, **cloud providers are locked into Nvidia’s ecosystem**—even **Google and Meta** use **Nvidia GPUs in some capacity**. – A **viable alternative** could **break Nvidia’s stranglehold** on **AI supercomputing**, leading to **more competitive pricing** and **faster innovation**. – *”I wouldn’t bet against them yet,”* said **a senior engineer at a major cloud provider**, *”but if they can show **real 10x gains**, even **Amazon, Microsoft, and Google will have to reconsider** their dependency on Nvidia.”* — **The Skeptics & Challenges Ahead** Not everyone is convinced. **Nvidia’s H100 and Blackwell successors** are still **far ahead in raw performance**, and **Accelerating’s claims rely on benchmarks that aren’t yet public**. – **Memory efficiency is hard to prove**. Many startups have made **bold claims about memory layouts**, but **actual compile-time and optimization challenges** often **eat up gains**. – **CUDA compatibility is a double-edged sword**. While **licensing CUDA makes Accelerating’s chips easier to adopt**, it also means **Nvidia can optimize its own software** to **run better on its own GPUs**, potentially **limiting Accelerating’s advantages**. – **The “AI hardware winter” threat**. After **Cerebras, Graphcore, and other specialty AI companies failed to deliver**, investors are **cautious about pre-product plays**. *”The devil is always in the details,”* said **a former Nvidia engineer** now working at a rival chip company. *”They’ve got a **strong team** and **a16z’s backing**, but **hardware is brutal**. Even small inefficiencies in real-world workloads can **mean the difference between a 10x chip and a 5x chip**—or worse, a **2x chip**.”* The biggest **unanswered question**: **Can Accelerating’s software-defined approach actually beat Nvidia’s hardware-centric dominance?** *”Nvidia has **optimized CUDA and its memory stack** for a decade,”* said **Dr. Narang**. *”Their **‘tensor cores’ and ‘‘NVLink’ optimizations** are **hand-tuned for AI**. Accelerating’s bet is that **a smarter scheduler can outperform a brute-force transistor throwdown**. That’s a **bold thesis**, but if they can execute, it could **change everything**.”* — **The Broader Industry Implications** If Accelerating succeeds, the **AI hardware landscape could shift overnight**. Here’s how: **1. The End of Nvidia’s “AI Tax”** – Right now, **Nvidia’s GPUs are priced at a premium**—**H100s cost $40,000 each**, and **cloud providers charge $25-$30 per hour**. – With **Accelerating’s Manta promising to cut costs by 90%**, **cloud providers might have the leverage to negotiate better deals** or **switch away entirely**. – *”If Accelerating delivers, **Google and Meta will stop licensing Nvidia GPUs**,”* said **a data center executive**. *”They’ve already done the work to **build their own chips**; they just need a **better performance-per-Watt alternative**.”* **2. AI Startups Will Finally Have a Competitive Option** – **Smaller AI companies** (like **Mistral AI, Cohere, and AI21**) **spend millions on cloud training** just to stay relevant. – If **Accelerating’s chips make training cheaper**, it could **level the playing field**, leading to **more innovation outside the US**. – *”Right now, **only the biggest players can afford to train cutting-edge models**,”* said **a founder of an AI startup**. *”If Accelerating can **reduce costs dramatically**, we might see **Europe and Asia catch up faster**.”* **3. The Next Wave of AI Hardware Wars** – **AMD’s MI300X** is trying to **close the gap with Nvidia**, but **Accelerating’s claims suggest it may never catch up**. – **Google’s TPU v5 (2024)** is still **optimized for inference**, not training. – **Intel’s Gaudi 3 (2023)** failed to **compete with H100 in training**, but **Accelerating’s tech could redefine what an ‘efficient’ AI chip looks like**. – *”We’re going to see **three distinct paths** in AI hardware,”* said **Jim Keller**. *”**Nvidia’s brute-force flops**, **AMD/Google’s inference-focused designs**, and now **Accelerating’s memory-coherent approach**. The **winner won’t be the one with the most flops**; it’ll be the one that **balances performance, cost, and energy**.”* **4. AI’s Carbon Problem Gets Worse (or Better)** – **AI training is already responsible for 1% of global electricity consumption**—**more than entire countries**. – If **Accelerating’s chips reduce power draw by 70%**, it could **slow the growth of AI’s carbon footprint**—or at least **make it more manageable**. – *”This might be the **first time an AI hardware announcement makes me think about climate**,”* said **a sustainability researcher tracking AI energy use**. *”If they can **prove it This article was reported by the ArtificialDaily editorial team. Related posts: Custom Kernels for All from Codex and Claude OpenClaw creator Peter Steinberger joins OpenAI OpenClaw creator Peter Steinberger joins OpenAI Claude Code costs up to $200 a month. Goose does the same thing for fr Post navigation All the important news from the ongoing India AI Impact Summit