Featured Insights
Strategic perspectives from Hume Consulting's expertise in data center optimization and growth strategies

NVIDIA's NVLink Fusion: True Innovation or Strategic Lock-in?
NVIDIA's NVLink Fusion True Innovation or Strategic Lock-in? Earlier this week at Computex, Jensen Huang introduced NVLink Fusion, positioning it as a means to "democratize scale-up" by allowing customers to mix and match compute architectures. On the surface, this suggests flexibility: integrating CPUs, GPUs, and specialized silicon, all interconnected via NVIDIA's high-performance NVLink. However, upon closer examination, this appears to be more of an illusion of choice.

From Silicon to Token
The world of large‑language‑model inference moves fast. Meta’s Llama 4 and DeepSeek's range of models turns yesterday’s “good enough” hardware into today’s bottleneck, so picking the right platform is more strategic than ever. I compared eight options that keep popping up in various engineering and sales conversations, including consumer RTX GPUs, Apple Silicon, NVIDIA’s H‑series, Groq’s purpose‑built LPU, Cerebras’ wafer‑scale engine, and turnkey DGX workstations. Each proves valuable in th
Latest Articles
Fresh perspectives on data center optimization and strategic technology consulting

OCP 2024 Regional Summit wrap
The Open Compute Project (OCP) Regional Summit was hosted in Lisbon, Portugal last month, the 5th (and largest) regional summit the group has hosted. Whilst I wasn't able to make it in person, I’d be remiss if I didn't write a (very) quick summary about the conference, and pertinent updates to scaling digital infrastructure in a sustainable way. The hot topic continues to be GenAI, such that OCP has created a new track for Artificial Intelligence, and a strategic initiative for Open AI Systems

Here come the Inferencing ASIC's
The tidal wave of Generative AI (GenAI) has mostly consisted of training large language models (LLM's), like GPT-4, and the huge amount of compute needed to process these enormous datasets, e.g. GPT-4 has 1.76 trillion parameters. This compute has mainly looked like NVIDIA's GPUs, but you also need... 1. power 2. networking 3. capital, AND 4. a nice cool place to host them (data center) The looooooong tail of AI Inferencing will dictate that compute is installed closer to where it's neede

GTC 2024 post-conference
Upon returning from GTC24, I've been able to reflect on all the new updates across NVIDIA's platforms and below is a summary of the various announcements. Blackwell was the star of the show, with the B100, B200 and GB200 chips announced. Note that there were no consumer facing graphics cards (RTX) were named, nor was there a successor to the L40S or BlueField3 DPU (though there was a new ConnectX8 NIC). As always though, a little bit of devil is in the details - as well as marketing waxing lyr
