1 comments

  • Palmik1 hour ago
    Similar article for vLLM: <a href="https:&#x2F;&#x2F;vllm-website-pdzeaspbm-inferact-inc.vercel.app&#x2F;blog&#x2F;deepseek-v4" rel="nofollow">https:&#x2F;&#x2F;vllm-website-pdzeaspbm-inferact-inc.vercel.app&#x2F;blog&#x2F;...</a><p>Bechmarks from InferenceX (they do not have apples-to-apples setups to compare the different engines for whatever reason): <a href="https:&#x2F;&#x2F;inferencex.semianalysis.com&#x2F;inference?i_hc=1&amp;g_model=DeepSeek-V4-Pro&amp;g_rundate=2026-04-25&amp;g_runid=24943464864&amp;i_prec=fp4%2Cfp8" rel="nofollow">https:&#x2F;&#x2F;inferencex.semianalysis.com&#x2F;inference?i_hc=1&amp;g_model...</a><p>I find it odd that sglang, vLLM, TRTLLM don&#x27;t seem to want to publish benchmarks comparing each other. They used to, but now there seems to be some unspoken rule against it.<p>At least we get comparison against &quot;other OSS engine&quot; this time, but that could be HF&#x27;s Transformers as well :)
    • imjonse39 minutes ago
      They&#x27;re OSS projects in a friendly competition, both working towards the goal of having alternatives to big closed source players. No need for jabs.
      • Palmik13 minutes ago
        I don&#x27;t think &quot;friendly&quot; and &quot;publishing benchmarks&quot; are at odds with each other.<p>Model makers (both open and closed weight) typically publish benchmarks against other models and when they do not, people rightfully call them out.<p>Including comparison against &quot;other OSS engine&quot; is just not helpful (what if it&#x27;s a sandbagged baseline like HF Transformers?)