Hey folks,
I’m excited to share the initial release of llama-cpp-hs — low-level Haskell FFI bindings to llama.cpp, the blazing-fast inference library for running LLaMA and other local LLMs.
What it is:
- Thin, direct bindings to the
llama.cpp
C API - Early stage and still evolving
- Most FFI is "vibe-coded"
— I’m gradually refining, testing, and wrapping things properly
- That said, basic inference examples are already working!
Contributions, testing, and feedback welcome!