ExLlamaV2
productInference Engine
Try in Playground →RSS
Overview
Developed byturboderp
LicenseMIT License
Open source✓ Open Source
Use caseLLM inference engine with quantization support
Technical
Integrates with
Also see
Alternative to
Based onExLlama
Knowledge graph stats
Claims11
Avg confidence93%
Avg freshness100%
Last updatedUpdated 1h ago
Trust distribution
100% unverified
Governance
EU Risknot classified

ExLlamaV2

product

Fast inference library for Llama models with support for various quantization methods and GPU acceleration.

Compare with...

open source

ValueTrustConfidenceFreshnessSources
trueUnverifiedHighFresh1

pricing model

ValueTrustConfidenceFreshnessSources
freeUnverifiedHighFresh1

supports model

ValueTrustConfidenceFreshnessSources
Llama modelsUnverifiedHighFresh1

primary use case

ValueTrustConfidenceFreshnessSources
LLM inference engine with quantization supportUnverifiedHighFresh1

developed by

ValueTrustConfidenceFreshnessSources
turboderpUnverifiedHighFresh1

supports protocol

ValueTrustConfidenceFreshnessSources
EXL2 quantization formatUnverifiedHighFresh1

license type

ValueTrustConfidenceFreshnessSources
MIT LicenseUnverifiedHighFresh1

requires

ValueTrustConfidenceFreshnessSources
CUDA GPUUnverifiedHighFresh1

based on

ValueTrustConfidenceFreshnessSources
ExLlamaUnverifiedModerateFresh1

integrates with

ValueTrustConfidenceFreshnessSources
text-generation-webuiUnverifiedModerateFresh1

alternative to

ValueTrustConfidenceFreshnessSources
llama.cppUnverifiedModerateFresh1

Alternatives & Similar Tools

alternative to
Compare →

Commonly Used With

Related entities

Claim count: 11Last updated: 4/11/2026Edit history