d-Matrix
  • Technology
  • Product
  • Ecosystem
  • Blog
  • About
  • Careers

d-Matrix Blog

Featured

Scaling AI the Right Way: Introducing Our Rack-Level Inference Solution  

Scaling AI the Right Way: Introducing Our Rack-Level Inference Solution  

October 14, 2025
What is AI Inference and why it matters in the age of Generative AI

What is AI Inference and why it matters in the age of Generative AI

June 4, 2025
The Complete Recipe to Unlock AI Reasoning at Enterprise Scale

The Complete Recipe to Unlock AI Reasoning at Enterprise Scale

February 13, 2025
Finding the middle ground: how smaller models will unlock the next wave of AI

Finding the middle ground: how smaller models will unlock the next wave of AI

Delivering high-quality AI-powered applications historically relied on massive models. That came with significant scaling limitations, as deploying models with more than 100B parameters and maximizing toke generation doesn’t scale up without losing latency… Read More
January 27, 2026
The power of the middle lane: why a hybridized approach to memory gives the best of both worlds 

The power of the middle lane: why a hybridized approach to memory gives the best of both worlds 

Using what’s on hand: spare data center space is an untapped gold mine 

Using what’s on hand: spare data center space is an untapped gold mine 

Batching just right: how interactive apps serve as a new battleground

Batching just right: how interactive apps serve as a new battleground

Why modern AI workloads demand a disaggregated approach 

Why modern AI workloads demand a disaggregated approach 

d-Matrix and Alchip Announce Collaboration on World’s First 3D DRAM Solution to Supercharge AI Inference

d-Matrix and Alchip Announce Collaboration on World’s First 3D DRAM Solution to Supercharge AI Inference

d-Matrix and Andes Team on World’s Highest Performing, Most Efficient Accelerator for AI Inference at Scale

d-Matrix and Andes Team on World’s Highest Performing, Most Efficient Accelerator for AI Inference at Scale

View All Posts >

Trending

How to Bridge Speed and Scale: Redefining AI Inference with Ultra-Low Latency Batched Throughput

Blazing the Trail Toward More Scalable, Affordable AI with 3DIMC 

Transforming AI: d-Matrix’s Pivotal Moments in Pursuit of Gen AI Inference At Scale

What is AI Inference and why it matters in the age of Generative AI

Why Datacenters are struggling to keep up with Generative AI

Featured Video

The DeepSeek Moment

In this short talk, d-Matrix CTO Sudeep Bhoja discusses the release of the Deep Seek R1 model, highlighting its impact on inference compute. He discusses the evolution of reasoning models and the significance of inference time compute in enhancing model performance.

Learn more about d-Matrix

From the Media

Alchip logo

d-Matrix and Alchip Announce Collaboration on World’s First 3D DRAM Solution to Supercharge AI Inference

Andes logo

d-Matrix and Andes Team on World’s Highest Performing, Most Efficient Accelerator for AI Inference at Scale

Datacenter

d-Matrix Raises $275 Million to Power the Age of AI Inference

SquadRack d-Matrix

d-Matrix Announces SquadRack, Industry’s First Rack-Scale Solution Purpose-Built for AI Inference at Datacenter Scale

d-Matrix Announces JetStream I/O Accelerators Enabling Ultra-Low Latency for AI Inference at Scale

HPC wire logo

GigaIO Partners with d-Matrix to Deliver Ultra-Efficient Scale-Up AI Inference Platform

View all media articles > For all press inquiries, please email pr@d-matrix.ai>
Transforming AI from
unsustainable to attainable.
  • Technology
  • Product
  • Ecosystem
  • About
  • Careers
  • Blog
  • Newsletter
  • Media Kit
  • Contact
  • Privacy Policy
  • Terms of Use
© d-Matrix, Inc. 2026
X Twitter Logo Streamline Icon: https://streamlinehq.com