Llama.cpp

Llama.cpp

Llama.cpp is an open-source C++ library designed to facilitate the inference of large language models (LLMs) like LLaMA on local devices without the need for specialized hardware. Developed by Georgi Gerganov and the community, it enables users to run LLMs efficiently on CPUs, making advanced AI models more accessible.