This is to announce the version 0.2 release of guile_llama_cpp, available at
(link to mirrored copy) http://download.savannah.nongnu.org/releases/guile-llama-cpp/guile_llama_cpp-0.2.tar.gz (direct link to file) https://download-mirror.savannah.gnu.org/releases/guile-llama-cpp/guile_llama_cpp-0.2.tar.gz This is the first release from the savannah.nongnu.org site, major changes: version 0.2 20240629 * introduction of chat.scm, with more control over char parameters * chat scripts now output long replies from LLM Guile_llama_cpp wraps around llama.cpp APIs so llama.cpp can be accessed from Guile scripts and programs, in a manner similar to llama-cpp-ython allowing the use of llama.cpp in Python programs. Currently simple Guile scripts are provided to allow simple "chat" with a LLM in gguf format. ## Project Information guile_llama_cpp releases are available from http://download.savannah.nongnu.org/releases/guile-llama-cpp/ sources in git repo at https://git.savannah.nongnu.org/git/guile-llama-cpp.git Bug tracker at https://savannah.nongnu.org/bugs/?group=guile-llama-cpp Mailing lists can be accessed at https://savannah.nongnu.org/mail/?group=guile-llama-cpp Currently, we have three mailing lists: bug-guile-llama-...@nongnu.org for bug reports and general development discussions guile-llama-...@nongnu.org for general user discussions info-guile-llama-...@nongnu.org for release and other announcements For more information, see README.md in the distribution. Enjoy!