llama-cpp-python¶
The main goal of llama.cpp is to enable LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware - locally and in the cloud. Simple Python bindings.
homepage: https://github.com/abetlen/llama-cpp-python/
version | versionsuffix | toolchain |
---|---|---|
0.3.2 |
-CUDA-12.1.1 |
gfbf/2023a |
0.3.2 |
gfbf/2023a |
(quick links: (all) - 0 - a - b - c - d - e - f - g - h - i - j - k - l - m - n - o - p - q - r - s - t - u - v - w - x - y - z)