d-Matrix
  • Technology
  • Product
  • Ecosystem
  • Blog
  • About
  • Careers

d-Matrix Blog

Featured

Scaling AI the Right Way: Introducing Our Rack-Level Inference Solution  

Scaling AI the Right Way: Introducing Our Rack-Level Inference Solution  

October 14, 2025
What is AI Inference and why it matters in the age of Generative AI

What is AI Inference and why it matters in the age of Generative AI

June 4, 2025
The Complete Recipe to Unlock AI Reasoning at Enterprise Scale

The Complete Recipe to Unlock AI Reasoning at Enterprise Scale

February 13, 2025
Scaling AI the Right Way: Introducing Our Rack-Level Inference Solution  

Scaling AI the Right Way: Introducing Our Rack-Level Inference Solution  

In today’s rapidly evolving AI landscape, it’s clear that inference — not just training — is becoming the new scaling challenge. As models grow in size and capability, the infrastructure… Read More
October 14, 2025
Open standards are the path to the next AI breakthrough 

Open standards are the path to the next AI breakthrough 

d-Matrix Announces SquadRack, Industry’s First Rack-Scale Solution Purpose-Built for AI Inference at Datacenter Scale

d-Matrix Announces SquadRack, Industry’s First Rack-Scale Solution Purpose-Built for AI Inference at Datacenter Scale

Why we needed a new Transparent NIC solution 

Why we needed a new Transparent NIC solution 

d-Matrix Announces JetStream I/O Accelerators Enabling Ultra-Low Latency for AI Inference at Scale

d-Matrix Announces JetStream I/O Accelerators Enabling Ultra-Low Latency for AI Inference at Scale

Blazing the Trail Toward More Scalable, Affordable AI with 3DIMC 

Blazing the Trail Toward More Scalable, Affordable AI with 3DIMC 

Why optimizing every layer of AI workloads—from software to infrastructure—is now critical as apps take off

Why optimizing every layer of AI workloads—from software to infrastructure—is now critical as apps take off

View All Posts >

Trending

How to Bridge Speed and Scale: Redefining AI Inference with Ultra-Low Latency Batched Throughput

Transforming AI: d-Matrix’s Pivotal Moments in Pursuit of Gen AI Inference At Scale

Impact of the DeepSeek Moment on Inference Compute 

The Complete Recipe to Unlock AI Reasoning at Enterprise Scale

Blazing the Trail Toward More Scalable, Affordable AI with 3DIMC 

Featured Video

The DeepSeek Moment

In this short talk, d-Matrix CTO Sudeep Bhoja discusses the release of the Deep Seek R1 model, highlighting its impact on inference compute. He discusses the evolution of reasoning models and the significance of inference time compute in enhancing model performance.

Learn more about d-Matrix

From the Media

SquadRack d-Matrix

d-Matrix Announces SquadRack, Industry’s First Rack-Scale Solution Purpose-Built for AI Inference at Datacenter Scale

d-Matrix Announces JetStream I/O Accelerators Enabling Ultra-Low Latency for AI Inference at Scale

HPC wire logo

GigaIO Partners with d-Matrix to Deliver Ultra-Efficient Scale-Up AI Inference Platform

embedded logo

Breaking the Memory Wall: How d-Matrix Is Redefining AI Inference with Chiplets

D-Matrix Targets Fast LLM Inference for ‘Real World Scenarios’

Cerebral Valley logo

Forget GPUs, d-Matrix uses DIMC to win AI Inference

View all media articles > For all press inquiries, please email pr@d-matrix.ai>
Transforming AI from
unsustainable to attainable.
  • Technology
  • Product
  • Ecosystem
  • About
  • Careers
  • Blog
  • Newsletter
  • Media Kit
  • Contact
  • Privacy Policy
  • Terms of Use
© d-Matrix, Inc. 2025
X Twitter Logo Streamline Icon: https://streamlinehq.com