diff options
author | Sisiutl <sisiutl@egregore.fun> | 2025-02-05 01:32:52 +0100 |
---|---|---|
committer | Sisiutl <sisiutl@egregore.fun> | 2025-02-05 01:32:52 +0100 |
commit | 5d9133598de6de36be2bdc808f6b85002258f9fb (patch) | |
tree | 87a116c22a0819302e6392fbe158ffc2b9532c20 | |
parent | 0761ffb99bf34a68a832cf2e0d1e78d85a0f10d4 (diff) |
new file: llm
yahoo
-rw-r--r-- | sigils/packages/llm.scm | 113 |
1 files changed, 113 insertions, 0 deletions
diff --git a/sigils/packages/llm.scm b/sigils/packages/llm.scm new file mode 100644 index 0000000..e6d91cf --- /dev/null +++ b/sigils/packages/llm.scm @@ -0,0 +1,113 @@ +(define-module (sigils packages llm) + #:use-module (guix packages) + #:use-module (guix git-download) + #:use-module (guix build-system cmake) + #:use-module (guix gexp) + #:use-module ((guix licenses) #:prefix license:) + #:use-module (gnu packages) + #:use-module (gnu packages bash) + #:use-module (gnu packages machine-learning) + #:use-module (gnu packages maths) + #:use-module (gnu packages pkg-config) + #:use-module (gnu packages python) + #:use-module (gnu packages python-xyz) + #:use-module (gnu packages vulkan) + #:use-module (gnu packages version-control)) + +(define-public llama-cpp + (let ((tag "b4640")) + (package + (name "llama-cpp") + (version (string-append "0.0.0-" tag)) + (source + (origin + (method git-fetch) + (uri (git-reference + (url "https://github.com/ggerganov/llama.cpp") + (commit tag))) + (file-name (git-file-name name tag)) + (sha256 + (base32 "13mc5cv3iwgc41k6mdpxh6cxb5a4zm3g1g9r3jr03fpw6426pvxr")) + (patches + (search-patches "llama-cpp-vulkan-optional.patch")))) + (build-system cmake-build-system) + (arguments + (list + #:configure-flags + #~(list "-DBUILD_SHARED_LIBS=ON" + "-DGGML_VULKAN=ON" + "-DGGML_BLAS=ON" + "-DGGML_BLAS_VENDOR=OpenBLAS" + (string-append "-DBLAS_INCLUDE_DIRS=" + #$(this-package-input "openblas") + "/include") + (string-append "-DBLAS_LIBRARIES=" + #$(this-package-input "openblas") + "/lib/libopenblas.so") + + "-DGGML_NATIVE=OFF" ;no '-march=native' + "-DGGML_FMA=OFF" ;and no '-mfma', etc. + "-DGGML_AVX2=OFF" + "-DGGML_AVX512=OFF" + "-DGGML_AVX512_VBMI=OFF" + "-DGGML_AVX512_VNNI=OFF") + + #:modules '((ice-9 textual-ports) + (guix build utils) + ((guix build python-build-system) #:prefix python:) + (guix build cmake-build-system)) + #:imported-modules `(,@%cmake-build-system-modules + (guix build python-build-system)) + #:phases + #~(modify-phases %standard-phases + (add-after 'unpack 'patch-paths + (lambda* (#:key inputs #:allow-other-keys) + (substitute* "ggml/src/ggml-vulkan/vulkan-shaders/vulkan-shaders-gen.cpp" + (("\"/bin/sh\"") + (string-append "\"" (search-input-file inputs "/bin/sh") "\""))))) + (add-after 'unpack 'disable-unrunable-tests + (lambda _ + ;; test-eval-callback downloads ML model from network, cannot + ;; run in Guix build environment + (substitute* '("examples/eval-callback/CMakeLists.txt") + (("COMMAND llama-eval-callback") + "COMMAND true llama-eval-callback")) + (substitute* "tests/CMakeLists.txt" + (("^ llama_target_and_test\\(test-chat\\.cpp\\)\n$") "")))) + (add-before 'install 'install-python-scripts + (lambda _ + (let ((bin (string-append #$output "/bin/"))) + (define (make-script script) + (let ((suffix (if (string-suffix? ".py" script) "" ".py"))) + (call-with-input-file + (string-append "../source/" script suffix) + (lambda (input) + (call-with-output-file (string-append bin script) + (lambda (output) + (format output "#!~a/bin/python3\n~a" + #$(this-package-input "python") + (get-string-all input)))))) + (chmod (string-append bin script) #o555))) + (mkdir-p bin) + (make-script "convert_hf_to_gguf") + (make-script "convert_llama_ggml_to_gguf") + (make-script "convert_hf_to_gguf_update.py")))) + (add-after 'install-python-scripts 'wrap-python-scripts + (assoc-ref python:%standard-phases 'wrap)) + (add-after 'install 'remove-tests + (lambda* (#:key outputs #:allow-other-keys) + (for-each delete-file (find-files + (string-append (assoc-ref outputs "out") + "/bin") + "^test-"))))))) + (inputs (list python vulkan-headers vulkan-loader)) + (native-inputs (list pkg-config shaderc bash git)) + (propagated-inputs + (list python-numpy python-pytorch python-sentencepiece openblas)) + (properties '((tunable? . #true))) ;use AVX512, FMA, etc. when available + (home-page "https://github.com/ggerganov/llama.cpp") + (synopsis "Port of Facebook's LLaMA model in C/C++") + (description "This package provides a port to Facebook's LLaMA collection +of foundation language models. It requires models parameters to be downloaded +independently to be able to run a LLaMA model.") + (license license:expat)))) |