Introducing Our Benchmarking Tool: Powered by dstack

1. Introduction Benchmarking is an essential part of optimizing AI models and software applications. Whether you're testing AI model inference speeds, profiling different hardware configurations, or ensuring system performance over time, having a reliable benchmarking tool is crucial. However, many existing tools suffer from issues like inconsistent environments, difficult configuration setups, and lack of automation. vLLM's benchmarking capabilities are excellent,…

Optimizing QwQ-32B (by Qwen): AMD MI300X vs. NVIDIA H200

1. Introduction In the world of large language models (LLMs), most benchmarks center on Llama or DeepSeek derivatives. We decided to diversify by adding the Qwen2 architecture, using our Paiton framework. This 32-billion-parameter model pushes GPU resources to the limit, perfect for comparing NVIDIA’s new H200 to our AMD MI300X, which leverages Paiton for advanced concurrency and custom kernel compilation.…

Eliovp Featured on AMD “Tech Talk” Podcast

We’re excited to share that Eliovp was recently featured on AMD’s “Tech Talk” podcast! In this episode, our CEO, Elio Van Puyvelde sits down with Jim greene to talk about the origins of Eliovp, the passion and expertise that brought the company to life, and the innovative full end-to-end solutions we offer today. From our humble beginnings to our current…

News & Updates