Featured Insights
Strategic perspectives from Nick Hume's 20+ years of experience in technology leadership
Latest Articles
Fresh perspectives on technology leadership and organizational transformation

Here come the Inferencing ASIC's
The tidal wave of Generative AI (GenAI) has mostly consisted of training large language models (LLM's), like GPT-4, and the huge amount of compute needed to process these enormous datasets, e.g. GPT-4 has 1.76 trillion parameters. This compute has mainly looked like NVIDIA's GPUs, but you also need... 1. power 2. networking 3. capital, AND 4. a nice cool place to host them (data center) The looooooong tail of AI Inferencing will dictate that compute is installed closer to where it's neede

GTC 2024 post-conference
Upon returning from GTC24, I've been able to reflect on all the new updates across NVIDIA's platforms and below is a summary of the various announcements. Blackwell was the star of the show, with the B100, B200 and GB200 chips announced. Note that there were no consumer facing graphics cards (RTX) were named, nor was there a successor to the L40S or BlueField3 DPU (though there was a new ConnectX8 NIC). As always though, a little bit of devil is in the details - as well as marketing waxing lyr

GTC 2024 preview
