Production ready LLM model compression/quantization toolkit with hw accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.
[email protected] is safe to use (health: 71/100)
Get this data programmatically — free, no authentication.
curl https://depscope.dev/api/check/pypi/gptqmodelLast updated · 2026-04-28T20:40:49.731846Z