Skip to content

llama-cpp-python

The main goal of llama.cpp is to enable LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware - locally and in the cloud. Simple Python bindings.

homepage: https://github.com/abetlen/llama-cpp-python/

version versionsuffix toolchain
0.3.2 -CUDA-12.1.1 gfbf/2023a
0.3.2 gfbf/2023a

(quick links: (all) - 0 - a - b - c - d - e - f - g - h - i - j - k - l - m - n - o - p - q - r - s - t - u - v - w - x - y - z)