Microsoft Maya 200

, the Maia 200 packs 140+ billion transistors, 216 GB of HBM3E, and a massive 272 MB of on-chip SRAM to tackle the efficiency crisis in real-time inference. Hyperscalers prioritize inference efficiency and cost (40-50% reductions). By 2028, custom ASICs could capture 20-30% market from Nvidia’s ~90%, with total AI chip sales ~$975B in 2026. …

Read more

Insanely Profitable AI or Just Insane Bubble?

Amazon shows how capex-heavy bets can terrify investors yet compound into fortunes. From 1997–2003, Amazon had over $5 billion in cumulative losses amid relentless infrastructure spends—warehouses, servers, logistics—totaling ~$10B adjusted, or 5–6x annual revenues at peaks. In 1999, Amazon’s market cap was approximately $25.7 billion, fell as low as $3.6 billion in 2001 and got …

Read more

Nvidia Maximizes the Revenue from a Data Center

Nvidia with dynamo, Nvlink and other optimizations increases the revenue AI data centers can generate. AI Data Centers are power limited. They need to have energy and compute optimization to generate more tokens per second and more valuable tokens. Brian WangBrian Wang is a Futurist Thought Leader and a popular Science blogger with 1 million …

Read more

Groq 30 Days to Starting With Large Customers

Groq said that they will start operating an AI inference cluster with large business in 30 days. Groq made a presentation at the GenAI summit 2024 in San Francisco. They are processing 30,000 inference input inference tokens and will put together about 1500 chips into an inference data center that will process 25 million inference …

Read more