Community
Join LMSpeed QQ Group
Join us for speed-test updates and support. Group ID:
llama.cpp is an open-source inference engine for running large language models locally, optimized for CPU and GPU performance.