Back to list
toolC++CPU reasoninglightweightefficient
llama.cpp
LLM inference engine implemented in pure C/C++, supports CPU and a variety of hardware, is lightweight and efficient
23 views0 stars3/5/2026
LLM inference engine implemented in pure C/C++, supports CPU and a variety of hardware, is lightweight and efficient