This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the Meta’s Llama 4 Scout 17B…
Tag: inference
llama-bench the Phi-4 14B and AMD EPYC 9554 CPU
This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the Microsoft’s Phi-4 14B with different…
llama-bench the Qwen3 Coder 30B A3B and AMD EPYC 9554 CPU
This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the Alibaba’s Qwen3 Coder 30B A3B…
llama-bench the Qwen3 32B and AMD EPYC 9554 CPU
This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the Alibaba’s Qwen3 32B with different…
llama-bench the Gemma 3 27B and AMD EPYC 9554 CPU
This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the Google’s Gemma 3 27B Instruct…
llama-bench the Mistral Large 123B and AMD EPYC 9554 CPU
This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the Mistral Large Instruct 123B 2411…
llama-bench the Qwen 2.5 Coder 32B and AMD EPYC 9554 CPU
This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the Qwen 2.5 Coder 32B with…
llama-bench the DeepSeek R1 Distill Llama 70B and dual AMD EPYC 7282
This article shows the CPU-only inference with a relatively old server processor – AMD Epyc 7282. For the LLM model the DeepSeek R1 Distill Llama…
llama-bench the DeepSeek R1 Distill Llama 70B and AMD EPYC 9554 CPU
This article shows the CPU-only inference with a modern server processor – AMD Epyc 9554. For the LLM model the DeepSeek R1 Distill Llama 70B…
LLM inference benchmarks with llamacpp and dual Xeon Gold 5317 cpus
LLMs or large language models are really popular these days and many people and organization begin to rely on them. This article continues in the…