ARKONE

What does on-premise LLM deployment actually cost and what hardware is sufficient?

All questions
deploymentinfrastructureopen-source

What does on-premise LLM deployment actually cost and what hardware is sufficient?

CTO · Government contractor·Asked Mar 21, 2026·172 views

We have a client who cannot send data to external APIs under any circumstances. We need to run a capable model entirely on-premise. The workload is moderate — maybe 50 concurrent users, document Q&A and summarization. What GPU/CPU configuration is realistic for running a 70B-class model at acceptable latency, and what are the operational costs teams have seen vs. the API alternative?

5 Answers