llama.cpp
productllm_inference
Overview
Developed byGeorgi Gerganov
Founded2023
LicenseMIT License
Open source✓ Open Source
Primary languageC++
Use caseRunning LLM inference on CPU
Technical
Protocols
Integrates with
Also see
Based onC++
Knowledge graph stats
Claims43
Avg confidence93%
Avg freshness99%
Last updatedUpdated yesterday
WikidataQ125998452
Trust distribution
100% unverified
Governance

llama.cpp

product

MIT C/C++ library for LLM inference with GGUF support, runs on CPU and GPU

Compare with...

based on

ValueTrustConfidenceFreshnessSources
C++UnverifiedHighFresh1
C++ programming languageUnverifiedHighFresh1
GGML tensor libraryUnverifiedHighFresh1
GGML libraryUnverifiedModerateFresh1

supports model

ValueTrustConfidenceFreshnessSources
LLaMAUnverifiedHighFresh1
LLaMA modelsUnverifiedHighFresh1
Alpaca modelsUnverifiedHighFresh1
Vicuna modelsUnverifiedHighFresh1
AlpacaUnverifiedHighFresh1
VicunaUnverifiedModerateFresh1
Code LlamaUnverifiedModerateFresh1
GPT4AllUnverifiedModerateFresh1

programming language

ValueTrustConfidenceFreshnessSources
C++UnverifiedHighFresh1

open source

ValueTrustConfidenceFreshnessSources
trueUnverifiedHighFresh1

pricing model

ValueTrustConfidenceFreshnessSources
freeUnverifiedHighFresh1

integrates with

ValueTrustConfidenceFreshnessSources
CUDAUnverifiedHighFresh1

primary use case

ValueTrustConfidenceFreshnessSources
Running LLM inference on CPUUnverifiedHighFresh1
local inference of large language modelsUnverifiedHighFresh1
CPU inference for LLaMA modelsUnverifiedHighFresh1
LLM inference on CPUUnverifiedHighFresh1
CPU-based LLM inferenceUnverifiedHighFresh1
Local LLM inference with minimal dependenciesUnverifiedHighFresh1

supports quantization

ValueTrustConfidenceFreshnessSources
trueUnverifiedHighFresh1
4-bit quantizationUnverifiedHighFresh1

supports platform

ValueTrustConfidenceFreshnessSources
LinuxUnverifiedHighFresh1
macOSUnverifiedHighFresh1
WindowsUnverifiedHighFresh1

quantization support

ValueTrustConfidenceFreshnessSources
GGML formatUnverifiedHighFresh1

requires

ValueTrustConfidenceFreshnessSources
CPUUnverifiedHighFresh1
C++ compilerUnverifiedHighFresh1
quantized model weightsUnverifiedHighFresh1
no GPUUnverifiedHighFresh1

platform support

ValueTrustConfidenceFreshnessSources
cross-platformUnverifiedHighFresh1

maintained by

ValueTrustConfidenceFreshnessSources
Georgi GerganovUnverifiedHighFresh1

developed by

ValueTrustConfidenceFreshnessSources
Georgi GerganovUnverifiedHighFresh1

license type

ValueTrustConfidenceFreshnessSources
MIT LicenseUnverifiedHighFresh1

supports protocol

ValueTrustConfidenceFreshnessSources
GGML formatUnverifiedHighFresh1

supports hardware

ValueTrustConfidenceFreshnessSources
CUDA GPUsUnverifiedHighFresh1
Apple SiliconUnverifiedHighFresh1

uses quantization

ValueTrustConfidenceFreshnessSources
4-bit and 8-bitUnverifiedHighFresh1

founded year

ValueTrustConfidenceFreshnessSources
2023UnverifiedModerateFresh1

alternative to

ValueTrustConfidenceFreshnessSources
Hugging Face TransformersUnverifiedModerateFresh1
PyTorch inferenceUnverifiedModerateFresh1

Alternatives & Similar Tools

Commonly Used With

Related entities

Claim count: 43Last updated: 4/9/2026Edit history