d-Matrix Blog - Archives

Think more vs. Train more

January 29, 2025

With the arrival of Reasoning and Inference-Time compute, we are at an inflection point in the AI computing journey. Finally, revenue generation from AI models is aligning with the cost…Read More

Introducing dmx.compressor

October 16, 2024

Quantization plays a key role in reducing memory usage, speeding up inference, and lowering energy consumption at inference time. As large language models (LLMs) continue to grow exponentially in size —…Read More

< Back to the d-Matrix Blog